var/home/core/zuul-output/0000755000175000017500000000000015140063762014531 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015140077047015476 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log.gz0000644000175000017500000336654015140076716020276 0ustar corecore}ikubelet.log_o[;r)Br'o-n(!9t%Cs7}g/غIs,r.k9GfB… "mv?_eGbuuțx{w7ݭ7֫d% oo/q3m^]/o?8.7oW}ʋghewx/mX,ojŻ ^Tb3b#׳:}=p7뼝ca㑔`e0I1Q!&ѱ[/o^{W-{t3_U|6 x)K#/5ΌR"ggóisR)N %emOQ/Ϋ[oa0vs68/Jʢ ܚʂ9ss3+aô٥J}{37FEbп3 FKX1QRQlrTvb)E,s)Wɀ;$#LcdHMeBmFR5]!PI6f٘"y/(":[#;`1}+7 s'ϨF&%8'# $9b"r>B)GF%\bi/ Ff/Bp 4YH~BŊ6EZ|^߸3%L[EC 7gg/碓@e=Vn)h\\lwCzDiQJxTsL] ,=M`nͷ~Vܯ5n|X&pNz7l9HGAr Mme)M,O!Xa~YB ɻ!@J$ty#&i 5ܘ=ЂK]IIɻ]rwbXh)g''H_`!GKF5/O]Zڢ>:O񨡺ePӋ&56zGnL!?lJJYq=Wo/"IyQ4\:y|6h6dQX0>HTG5QOuxMe 1׶/5άRIo- MR9z_Z;57xh|_/CWuU%v[_((G yMi@'3Pmz8~Y >hl%}Р`sMC77Aztԝp ,}Nptt%q6& ND lM;ָPZGa(X(2*91n,50/mx'})')SĔv}S%xhRe)a@r AF' ]J)ӨbqMWNjʵ2PK-guZZg !M)a(!H/?R?Q~}% ;]/ľv%T&hoP~(*טj=dߛ_SRzSa™:']*}EXɧM<@:jʨΨrPE%NT&1H>g":ͨ ҄v`tYoTq&OzcP_k(PJ'ήYXFgGہwħkIM*򸆔l=q VJީ#b8&RgX2qBMoN w1ђZGd m 2P/Ɛ!" aGd;0RZ+ 9O5KiPc7CDG.b~?|ђP? -8%JNIt"`HP!]ZrͰ4j8!*(jPcǷ!)'xmv>!0[r_G{j 6JYǹ>zs;tc.mctie:x&"bR4S uV8/0%X8Ua0NET݃jYAT` &AD]Ax95mvXYs"(A+/_+*{b }@UP*5ì"M|܊W7|}N{mL=d]' =MS2[3(/hoj$=Zm Mlh>P>Qwf8*c4˥Ęk(+,«.c%_~&^%80=1Jgͤ39(&ʤdH0Ζ@.!)CGt?}=ˢ>f>\bN<Ⱦtë{{b2hKNh`0=/9Gɺɔ+'Х[)9^iX,N&+1Id0ֶ|}!oѶvhu|8Qz:^S-7;k>U~H><~5i ˿7^0*]h,*aklVIKS7d'qAWEݰLkS :}%J6TIsbFʶ褢sFUC)(k-C"TQ[;4j39_WiZSس:$3w}o$[4x:bl=pd9YfAMpIrv̡}XI{B%ZԎuHvhd`Η|ʣ)-iaE';_j{(8xPA*1bv^JLj&DY3#-1*I+g8a@(*%kX{ Z;#es=oi_)qb㼃{buU?zT u]68 QeC Hl @R SFZuU&uRz[2(A1ZK(O5dc}QQufCdX($0j(HX_$GZaPo|P5q @3ǟ6 mR!c/24مQNֆ^n,hU֝cfT :):[gCa?\&IpW$8!+Uph*/ o/{")qq҈78݇hA sTB*F$6 2C` |ɧJ~iM cO;m#NV?d?TCg5otޔC1s`u.EkB6ga׬9J2&vV,./ӐoQJ*Dw*^sCeyWtɖ9F.[-cʚmD (QMW`zP~n"U'8%kEq*Lr;TY *BCCpJhxUpܺDoGdlaQ&8#v| (~~yZ-VW"T- 0@4V{g6R/wD_tՄ.F+HP'AYM; R j"b~PO"dwEр%}5zW]O )-1 8/%\hC(:=4< ,RmDRWfRo>UJy -71ieQ*-=hiך_%הrΒ]rύW -e]hx&gs7,6BxzxօoFMA[D,{'$ь'gISzp; AQvDIyHc<槔w w?38v?Lsb s "NDr3\{J K纂P//Q۲6+ktwMzG,87^ 9H\yqū!)\(v8pHA"ΈGVp"c ?X)hm.2;sl$瓴ӘIe~H|.Y#C^SJĽHǀeTwvy"v܅ ]?22R.lQPa ˆSܛ1z.Tvg.0Mn}#~sz1}z}V|:f7.PWo˷ *ŮޢoArO6r ~pu2qG&ڊcN8xlDZ3ᕡg:3n%x$dޒywUnj|_$cݢaȤ금~{죤z[=_d rtc\MU&IzhBYň(&&WV#@..$ 91XMBQϷmQm{mO!_ysW1!x`~to|\n5 lW#@UTkտ,Fպ̥ 9MGb&0ۺ*'qpZX7c&͌ӒҶW r@/m@6P!{`ͱ)m`6*-5F 6=X#leU d6xTV6 gn&i"@*"mrf IEVpq 0÷bp៚U|,}S@t!:X _ .xM_7ve Z@7IX/C7@u BGڔE7Mk $qZhڗ}Ow4ǛJ5H 7Bbj:E]`C 8蟫n'Ą6[#P-@i'/!×6?oK`8 )v3v֠:b(v6& `-K;~:|B6vXpw*t]IJG 5 ƻ7{۱ַ P񷍋 ߂ gSPC=]U҅yY) &K|Zn$N(lYiV =?_e^0)¿]{ @lW GpC ӫByݝ0ߒ5bZւ 6{Sf觋-V=OߖmtC\/`m(9B< f`mPіpJЦXn6'5m 7aTcTA,} q:|CBp_uBȆf]չoݴmqR".lCqDsn:6GkM\7"Ǻzfbx[ۮC=1ÓOU/zIoeXX۱oK)滲L RnM'L]O-xRb2Hӕ$~؏ .m}xQ~f> War1|VϟT;Cq%KK-*i ѩQٰW`DݎGu( 꿡\cXn~w3n:U/.P珸XECh;l bcWa?\34 P샃 U!7Kx5 ¶]fXE }}5/1҃TXskwqRtIڢLhw/:?\3Թ-hrxE݂\Ej7EKvا, w悇_[uWhP nW*7GKi*7F ѡZ afrjcl ^2BQG7'Dff^f!8:/p6>TV*P,rq<-m/K[[ߢm۽ȑt^, tJbظ&'JE%"2.*""]wn` 扪MV)m8[6-Te@`E|=U D(C{+0_t[%XU͍ &dtO:odtRWon%*44JٵK+WoC.F3 %N%FF"HH"\$ۤ_5UWd̡bh塘ZRI&{3TUFp:4TƓ5[۲yyz+ 4F.Aw`'*7R(#ғ [K&#MpGXާ/5܉z,4kA=:z7y1^}<ϲVrn~F8I !6WB3C%X)ybLFB%X2Ug6^vw8uUF+X|YukX ^pQVΖj?+ױV}F#N9Ǥx;Dk |Cbm3ù//k Ɠ=x?ZI,y,X+Q,S*6 !ȴ ,!NB ^wKyⱳ]x99?u0軡*uG, 势Ąu3=Az4 u5'3qSO0|d1f_Fw }yqR\,%B~`$ C^waB4s=Oi$LXI zPzMD{]4ü Q̦ Q^Ղu ;` .Тr yFBQ#C`Jyn,m93B%Z~O/_LKCQϰԨN\uRT{;eqQntq"Occ°NRjg#qSn02DŔw:ؽ 5l)Fa/TT8mCԤ{"9b{?-TxF8'4sѓOJV@ڝy?ʕ+9M+CWp q)-" >,uˮ◶>waPcPw3``m- ԧAeE7մ{b\No*{:Mzw =lQυo,\*wٗYS%g,0\ Rk k8P>x?myڈۥLFTk"5l9O'ϓl5x|_®&&n]#r̥j'=K)lsXg\{Md-% >~Ӈ/( [ycy`ðSmn_O;3=Av3LA׊onxlM?~n Θ5 ӂxzPMcVQ@ӤomY42nrQ\'"P؝J7g+#!k{paqTԫ?o?VU}aK q;T0zqaj0"2p؋9~bޏt>$AZLk;3qUlWU Ry==ck vz(vb$^Nyo$p[DtUCE9sBz%lOONRѦmDVmxюݏX}K6"Qi32\-V_kR(I-wtSJR^m{d a|y,F9$^@mdH֙toN1 < ҷBq/ ۓ,j|z6OSu;BKŨʐPqO K\{jDiy@}b|Z79ߜih(+PKO;!o\戔-QB EM;oH$$]?4~YrXY%Ο@oHwlXiW\ΡbN}l4VX|"0]! YcVi)@kF;'ta%*xU㔸,A|@WJfVP6`ڼ3qY.[U BTR0u$$hG$0NpF]\ݗe$?# #:001w<{{B\rhGg JGIެE.:zYrY{*2lVǻXEB6;5NE#eb3aīNLd&@yz\?))H;h\ߍ5S&(w9Z,K44|<#EkqTkOtW]﮶f=.*LD6%#-tңx%>MZ'0-bB$ !)6@I<#`L8턻r\Kuz*]}%b<$$^LJ<\HGbIqܢcZW {jfѐ6 QڣPt[:GfCN ILhbB.*IH7xʹǙMVA*J'W)@9 Ѷ6jىY* 85{pMX+]o$h{KrҎl 5sÁbNW\: "HK<bdYL_Dd)VpA@A i"j<鮗 qwc&dXV0e[g#B4x╙✑3'-i{SEȢbK6}{Ⱥi!ma0o xI0&" 9cT)0ߢ5ڦ==!LgdJΆmΉO]T"DĊKٙ@qP,i Nl:6'5R.j,&tK*iOFsk6[E__0pw=͠qj@o5iX0v\fk= ;H J/,t%Rwó^;n1z"8 P޿[V!ye]VZRԾ|“qNpѓVZD2"VN-m2do9 'H*IM}J ZaG%qn*WE^k1v3ڣjm7>ƽl' ,Τ9)%@ wl42iG.y3bBA{pR A ?IEY ?|-nz#}~f ‰dŷ=ɀ,m7VyIwGHέ 2tޞߛM{FL\#a s.3\}*=#uL#]  GE|FKi3&,ۓxmF͉lG$mN$!;ߑl5O$}D~5| 01 S?tq6cl]M[I5'ոfiҞ:Z YՑ"jyKWk^dd@U_a4/vvV qHMI{+']1m]<$*YP7g# s!8!ߐ>'4k7/KwΦθW'?~>x0_>9Hhs%y{#iUI[Gzďx7OnuKRv'm;/~n-KI`5-'YݦD-!+Y򼤙&m^YAKC˴vҢ]+X`iDfKkBYx-qCfqsn[?_r=V:х@mfVg,w}QJUtesYyt7Yr+"*DtO/o۷~|hw^5wE of7cꃱ.)7.u/}tPTGc 5tW> l/`I~>|灹mQ$>N |gZ ͜IH[RNOMTq~g d0/0Љ!yB.hH׽;}VLGp3I#8'xal&Ȑc$ d7?K6xAH1H#:f _tŒ^ hgiNas*@K{7tH*t쬆Ny497ͩ KVsVokwW&4*H'\ d$]Vmr달v9dB.bq:__xW|1=6 R3y^ E#LB ZaZd1,]ןkznxtK|v+`VZ3JϧC^|/{ś}r3 >6׳oƄ%VDSWn 0,qh! E-Z%ܹpU:&&fX+EǬ.ťqpNZܗÅxjsD|[,_4EqgMƒK6f/FXJRF>i XʽAQGwG%mgo 恤hˍJ_SgskwI\t`ﶘ080ƱQŀllKX@116fqo>NrU Ѣ9*|ãeeH7.z!<7zG4p9tV|̢T`˖E ;;,tTaIUle*$!>*mBA2,gJIn_kSz)JC]?X(OPJS3.}clݨ{e!MB,cB߮4af祋,1/_xq=fBRO0P'֫-kbM6Apw,GO2}MGK'#+սE^dˋf6Y bQEz}eҏnr_ ^O^W zw~Ȳ=sXअy{E|!4ӥ2 ]8â6 U`V%`!c%؎ʨTzrKh! c.}.D>)d_ 8rcu,wf2?Ǡ*_lDn}rauyFp*ɨ:UiM2r:9ct X1lmĪ o玓,R%!`hGT LYF#g<cm${|Xdu4tmtїUJ\~dc0KcMlf2?mμQ ߉J4WrSHTdp"ӹ'cJq2zPlX̯.0H!ND@UapVoGڧD5>H]f@!=߸2V%Z 0"G4ȇʩ@]>Y$ًF_Mm_Tt)ib+q&EXFu򾬳ǝ/RS>r,C2NfOjpcm{Ll9vQOT>9U;])>6JdbXԠ `Z#_+D[7IIjJɟUh ҙ"`"a ߒ"G̾H`6yiCk(OA/$ ^%K^+(Vr[RR1"u4A.1X0=7f/"(o9/L1X{]q`Ȝ/; 9a>E)XOS K9mUxBa"'4T[Jl /K/9,rlCAj_TiǘP,:4F%_0E5IE'rX-|_W8ʐ/=ӹjhO%>| :S Px„*3_y.g9| ;b`w NtZtc> ײ1KĴ{3Gl& KT1ZWX8?C]~We$9; -.D087?1a@P5B,c}jcGȱ WW/ @a#LA4.ٹ^XڋXٝ:^Izq. ٽƎDn6ٹBc5Lt;3#i3RAٽ9| cbpcTfp> 6L/_x 'ۙz7~w~);qU9GDT! 6]c_:VlnEUdn6UˇKU;V`JUݵޙEO[)ܶCy*8¢/[cչjx&? ՃJȚ9!j[~[' "ssTV2i sLq>z@JM->=@NỲ\쀜*/) ̞r21.y? bO]3?C!yw3ޯL_Su>o>&lrw&i"< :]_<<7U_~z5є/rfn͝MLmc 6&)e+n7cyy{_~궼07R7wPuqpqo{ߟ+[w_uOq?u-|?WS_tOq?Eu-L_p?Cz .e ϿO*3 `Ђ6a-`kIf-s,RL-R`1eL~dپ&+IhYRczr?㐟,v~,b6)up)3K,RLW"Qd9JgT\1f3@Kh% a4x,kA k ^d kYj5Ah𚄓vXZhX1xҖ51Y +Id ZZ\C| fD>hB֡#-$+Jpሟ,Cg:6 3 xH "}C[`ӨOAFn5ʬLHϰ:N@VcyBI#Dr. "h hg ۃm-qu>V&൘ G7qi#^tҒ[JI!{q*lrD܇Gk@;oI<5xZ4xM"؇'k!>V|lk'{d+ :sXӄc)?W`*|\v aVT0"tMًcΒVz]T.C$cEp._0M`AlF̤@U' u,—rw=3}resLV&ԙy=Ejl1#XX۾;R;+[$4pjfљ lݍ3)`xvcZRT\%fNV Q)nsX }plMa~;Wi+f{v%Ζ/K 8WPll{f_WJ|8(A ä>nl"jF;/-R9~ {^'##AA:s`uih F% [U۴"qkjXS~+(f?TT)*qy+QR"tJ8۷)'3J1>pnVGITq3J&J0CQ v&P_񾅶X/)T/ϧ+GJzApU]<:Yn\~%&58IS)`0効<9ViCbw!bX%E+o*ƾtNU*v-zߞϢ +4 {e6J697@28MZXc Ub+A_Aܲ'SoO1ۀS`*f'r[8ݝYvjҩJ;}]|Bޙǖߔ 3\ a-`slԵ怕e7ːزoW|A\Qu&'9~ l|`pΕ [Q =r#vQu0 M.1%]vRat'IIc(Irw~Z"+A<sX4*X FVGA<^^7 vq&EwQű:؁6y\QbR9GuB/S5^fa;N(hz)}_vq@nu@$_DVH|08W12e_ʿd{xlzUܝlNDU j>zƖݗ&!jC`@ qэ-V Rt2m%K6dX)"]lj齔{oY:8VmS!:Wh#O0} :OVGL.xllT_oqqqLec2p;Ndck[ Rh6T#0H Q}ppS@ώ@#gƖ8sѹ e^ CZLu+."T#yrHhlكʼE-X'I^=bKߙԘ1"+< gb`[c1髰?(o$[eR6uOœ-m~)-&>883\6y 8V -qrG]~.3jsqY~ sjZ+9[rAJsT=~#02ݬf¸9Xe>sY~ ae9} x* zjC.5Wg󵸊y!1U:pU!ƔCm-7^w]斻~[hW$k sE0ڊSq:+EKٕ|dvvjjy6 æ/ML-yz,ZlQ^oAn-})xǺǍ--qcl:WLg ӁvJ[ǧc~Of+8qpçco#rCtKӫce0!Y-+cxMK-H_2:Uu*corD~@N`#m~R:ߙ歼!IZ5>H;0ޤ:\Tq]_\_>e˲\oUQ\Wߋ47WwߋKpwSSۘF,nC.\UߋoVEuY]^VW0R=<ު˜9x}[ێ'|;c^ M7 >5\-> m-8NJ\ALd!>_:h/NAC;?_ξqĎ6xMY(=ͯl~l8V0٨T zL{Ac:&$ ^CpH*DW\r2aR|=(L X1|wrO_g ux1^^V2޲jMi^b``Q#dBxV#NBk1;DAV$"*1]Y~ d->'I`{6_Ȗ4zC`7K@lJQ#idWdc`[ n&~@X(%!H:*,yD0E,$u2ܓ[懪9#B"JE^Dؼ64BCUR\l1X/(}1ug<*_jD07-U( x{Q U):qHi ~ȇ{! zƽE%ΣarAP%+d|OOIي='n0 7Ɨik喣k:Z{>325;d_W7<9x+fjPp&ڰtKsmg偡,; w }m1nxT5= +JQ-7^74]9SIc썯,_jpZbĠ&h=uO Z8LxH"q.s|+ l + 00}/0^~oͿYB{MzB3 kׇ PCZ)긊X\U ξ釾oyw,PT}D'疣]98Ǻ~u3Us זK4f/A [XP^ƀq\oWn̶lfxx x}e0>޼1<T6mS%oDKkP8[ d1[^k1k 7D4(kXRqfKZΣX;-L ;΋2ݫYDD y5OTb\f0{ ŶКwR+yRE3chaa04_=iỡ Ug0@34u4F ;t\=5af~23IV(p.id(.W-8O"%heg>'^6 y#,ņ:>/ŭe76mؖ~l(+֔7J" "P9""x%7̣r'XŨgu #uG\ kTo_Gn1ORoXƣyU= 2T*'*Zk]rj9ϲ}<͒ 2ᔸD"Ft,`"q-'U hwj抻Pp8ȘeTx21o2ͫSѠ ws)N&cwg 's'DŠ/$# 㬎8_Р;CRu+I\Ǎ-q@,m^O-i0a ֪+Y^k?q5bQvm;^5񇌃9>\K,K#a{P+BjnKDGCXװN?$[ \Bt'OO[(cc_Ճ@ ΧEʓpo adop[k:(4|]5dGMB4{g`h`v늧mKEEhO>xR(7?:5 *` i.D@A\G5Nv*6M]m UKҜIN6)kDBMVG1^&R*M_y^Mჳ5MLlg%~pHQx=2+z7\67ORԶS~SfZspwx(涔?9/f_gv _>QR2>ȕO([l 3˛Q:U2׊WG?uC5A( ()i5&'C"\JS.iV7tk{#>ص5$VFxcռ« %Gfh[AWf9>E qk+iCg܎>Bldyp ز \{uQ+!;9[j_suRXă ](!JzUdTc);W'g+kIcՔ9wԻv`c%DMu[sׅPeԾNL馶-طʳ<ؽeQelFD^9|} .ylwVZfiVTzÌgiz:s1=O$,9-B]ZU؂byseK>MCa[KRD0MSrF`xW ):\XS Lgu)ѳPy N\rls@&@D_c+`F\a[Q?] O(oA#2qFyح c~'d2 q-՗8sߴ$tQ1*2=[2'q}dq"6g2fsþg18-h斄f.ͩ e#۶|$"gV}XKqh<+" {K?TLuzeWW4\%>etlJF0|ΛҴЫU~U4p &jmʳ#qt> sGB,64ٝd^A@Ȓ)"Rqʖ9H6r %#'u=솗 C8Nز 9+%Gn^-r%ny-)3n66W (%&JQ ,q>|=>۶D4")Y]H5V6Hމpsmk@CJG!bMZ3e7n}x`QS\ Dpu:uoJ8՚55hFQM/"^Ǿp=~>^MO2 eo#òy>1JC1ӭWbݵfYeRXr]HPr`Џ9瘦΃ݣ|RєIʵ1V׊9M-cẆ%tam愁kx3U]]|'t[m=|恣ZKsk vWD!Gahض4+!F빇uzwd*W$\wQ#pdhW܈rdefe_&Wl:diX[EZx$)iS"(fmI"쉘U<~wl$36]咉-)[1Ki^S3̶=T̿)+u=~ݕKѕI.ыweiӎn{۱oLUeliZ'wWI;e\BX*mHՃw/L%g|)|?Kcыe4–p :{H1UI-+x˜]- !ÃeH~aqxԃoձhdg=9/˛ o%vEĦKXx7D=]G.1G-`X 1iXܙ4cw,Z.R#GB(^^Lff C]U7"DwUCfgћMk^i<=lTDU\Oy+L (e5Y I¾RTeU$+EUgq6[Dw"ce@z&*-I#oHg)Hv#WIBo>mҰ Plvo0118 l @̽k6 tս }oqɘ\Bn4( [W]eQCTGkqY\I侊 Xeh0F۬,Ɉ71~NJdξ !d0(㻍5T֋O$e03InܻݫpcF礗w 1Y{0ތHe=^l/xսkf=TC}.4y 3''cFW>ț{28/? I}OF _hL 3{ tþ0f?`G.3 L7iXcz"'y5'.4(xէ MwՂpS0X_6H?{8 !`w-Rctg9`:dIN4mK^WEJ28i'ttq,WW*$hJ#DcMYx̏BJb49}}Eat{GQ&WQJK&_M8kl0ڋ{OuzXYR?tGPHeU{:cUϞA}lpR)>G>O2Yl#Gt-RߛNlU=\NFqzNIk/ ԮqM{6D2WUiT % X.Ა+~ý=:"yog9ϻ$w`㱺d1E@n|R5`aF}s&u53[]iD@ΏN X#ECu߰`xW\{Q9KlD ?!]c޶})|A]۩DlG`Isz65#ޅ)l@h"fA޶M.İ"}6#8# K-vĶX- 3w-6whmʻQ("BBq#jAs+>̈%膥Fpׁ,.mU!@a+NwC" :i 4tg$A8^|y7rv- ʍo0f[pZw#BF@dJ2(Ց0F 0BF uٔ4@aWCP\rǺ%sg?isVн_wʉS[!FCN5`7bv3fmt.p NĩOBNJr" kSM@Snu 8 P~B8c(0^Gl9iWuOtȈǩɐJ»dL#sFhbo;0t>F3 sn^,amL & L#v킝pl= pX& ˙kw":@XJϙK†=r݉'ue3 ͺi{{uoۡۜp:-P&-څ$T$Mg?O{nO31swŕsqܶYV"ޑDn;y91qK XdආYe+0/o+>5䡁eS,*R;SB(0 `\̜'Lm&AoG8hfepXCH,U*CB** кaJq">ޑIdXcb?:l%k6cq3&i֪ <'RQ 'Wci+,EX] TZBŒ&pfPHWā^$x9u񦷓lCܴbB'9;ߙQ2}o^V@(BٖrףJi.}6ƅcAL=#57D7qE02Fh#Xn)X7`m&ғ#- UF3"s  ~:sZ35Tc~ګڞӅڪe̠.g,MBޖe`Ӳ55hށop/型, %-XbqQaPH( X`X VW /q5N`f}Me!woT}u kpY_W_}6BD~Z=F=Qka/A ػ;gˇO=_z:l-5%7QW=`fs̢J`|Y%& -&I:0=N3< i⪹`ԅWQ݁;Ntj⨤ ;<Փ7 n ,'QdƌGf숛dOu9.ӲLۅ򁪴癁v^c97d@Є K |Ӳ׍=>n됾&Y } N~jޏQ~0G"AU[K:Or-V* Y6;˘Uj 6;Pc 4S4sG @Si-2;.=t['Ey&ܛ Gw|1Z&ϣ 6D҉P/dN9h̷D[(0p!  Ƴ9A<"ΚsÃ|X*&7,|חG eKؓO]Gx<qcY<8z60Qߢ),?v!u>7Q19.a{Alo uə4iّgqJ=Ӂ "C8 5oR<%a=6oͺ2)۷Oa_ºDkTCCxy_X=LR2l"b;0&_dr_ 0/rqq6lK (41vR`5ԵphFM`n`IaG:jBQoeQ+ a8.cZ8:!l0/.+K,LX8*VN`}J+K;A!(= TQplvrquI"꾌Pk k'PLBzzkO2Bg tNbOu.X/yq-^:?sRi yqbd8' 9:8tq$7MI|Y1dRIpӪ Iz-sxx̉GڭV@Q!OGwG(VQpn}~/b}j?%H̪|٭>]z3s-j  W#A WtzYhr1f#qBIѪ'."n,>>T]ODy*p'N<?ʬ@wC-Sb  E`Wm5]/(U]RMX~f-̷9?K`0{hPG S2GP+޽7 FhNcǨ~IU}/5^5]C*4%yMfOOpmK4B3yԀu Ӊ!t"[n X$7Z=g`n +8N$i|xnsyux2Nj \1)f0Fax2etr2nٮuP &x ;nU17Db<,KU*jP3Of";aMCŘ(X2S@cfql, ;^@2m}!L+Sm\Q&f^P&RI0I9ZIW 氣~XegcajQShDQmKXQBzzb<*,>d V6 x\4q]ibYzVKȳ:wZ7zLPW7vsFA x2htmV1%\LDҕ8,l<-K&j+DiJK6*sWP^fw(t'Yho}ӈD~GJl)b*y^dfAgӞk:***:e6MUzmdZ}Mƃe~ӿ? [e:H:7sRE?C1ԏ?#adkC'$)Y PGI*aI@}7P IeTMYއ/Fr%*y]Z٨0I߲jyWPXFQ`6*@oy~ór/|勏 tV}zZG- lqM*NA֯J2Co)~ҵ 1K f`m(}юpS,L? e;|=HJzFmH e:)dUQMM!y@RSڐanǪ4.tA6_YW{}7MT&djK0âXd 8ARDe4фmjFAKwzS-pW-gQ&7 Ixk;DjN<Ų]5VeM]jYd0++Ɏߵ*}p[Az'?_ pjk)C1Z*gDX˾]cg"1":bW(Uv!4} mjeuqFO3Ers χɁfOf`96=C1T PSmk;Imp`,`kŕ#BE"V /j[BDfϨ ipc0`\de\H@1aԚFhRuHqjQN"W{+ R漑Ovel.g.9OC^'߀kRf ]+tyK }@JtLwK]g'XlK銸}>cn %'g]^}Ս /t {)6?2߷|*G\5<\P0?m 8F= УxJ84xNdՠ+fHfBjKJE)%P]AKNMyF[[?)2e2Jb6Ұ9n61hXyԺŸ"sO( sVz8LZ|ñ^4UΉ$%fݔʰ"?ѝeH4(M'?]{vDۤnyaE Ƽ%2s_l(?a 2rIg ;m P^>E%y6hz|Qs.Bkb9)b<)r\4v0y@Ӭpk1ߧp6c<·hcMHwDG9юkmFJ0vrW^uՖm{˒e돵pA'Ag_~ho;蠔{{\ǠmK aᤆVwgo~9ؘݧ Lۡ]huѨo rHGݯÿjM,* $ &-9Vyk E <Zʇco@ 2&GBQL=lhCN^-;NM,ZOrr"$߃GؾP$OCqY"h4xQc b$1M+ G3ݍ۽p\]w-a2h>AE:'~{ E'Q5{ǁ_XgVU+hQ VRļU a,؅oOjb]RYfm*.0EG@0$?:;cwuj%bLe!HtM~] 8J'htFhM.`]iKԏ_ R$18NN/w\j|:Wk:y&8xI}2\ gu;*cGǶuiS$MrydJdQM`{G!1fB%Q|<0E -U2H cI 70K*{93F$x_  *L /4 t&Ig m뀅LYj!H<̈RRMdem1at.LBhlM4 CߨtL4K[]μQ9=> zH]s3BiR` d,Ta]4VUb on\ - ?wb&zmX"H:s- _vjQ0o<4X':5sHϼ߬foqn|s2ƌҙ>HjVj4 )%[kS]M JZ\G v aZ`SZ&Ay XpTscg40H#5rAϙAaĖAyR~}0hX%Ӂ,HF{UMSyTt87r hܢV΅& MHl^4BGΣ`yG;K ]'H*t}m8m2%:P!`h.r D?1r G3grlXVj#Dj)P r?Aa_c &C͋(ƊJGy_k1 kZaJi`|orѮ5Oqr$ B ~0`|0)u$(t=N({U:||$-:)BAdL.)Yw=&_(jL.qJ^Ѣ.CA~rcq8sm:**>6@sVF(bSdڛ҄RV "@/!)^vuAhb Ɖ՞g"+jn1Eٮ;֐jjf1L>sP#|9IݱkNy =XN`-$jɤt'-G e+N?yLnru|4.ul͂: FR1HJ+(EYuO8Z>pg\d쒖XTǗh$2WR> %>su`ϼ sxUK$5A?_woXpscl]rY+h"50{ȱ5f%Io _Wİ2&JVd,k>R= >{h-**O&H[Uda! ;!;cM 5R-Ey+>wv29mт0Ƞ*b*]i#EO{-8mj-lx aN}zL1Nn RPـC Vt=vJjh䊥f)b( WWvl^G;7G9:ˑ)!jC 6#Gmxֵ2AeFR 7 رnro C<֌/X2g^]ߛ`Sku[:E <>=| 4Ct9''pZ2&S%T)LCtV@P$&ew֤&xՇx{^ј7V.92i#k|aq'ָBx[=+=D&w}hrV3Xj>{2d֔P8H3{\T+ShIpmC9;dvv3;zX'Ϣ^@/@J Y1snrp2q# f%Cc5: SL潲 "pIc|e>sV,cqӏ3]ip=/&/~c@O=J1o̍,=1ۃ}|y7oɬ0)(,ԗɡpz?uepeD/XO_A>7RWT$鉁-WudJ0KͶΘ`@lDo{t)UZحnFw~c.Ln܋h bDCYEH-.\L"KO)b;>fP7.ep[=P>ZzQ z 򰂼zHAaQd'U!?D_nYpו /(Wuc_8t?}o[.Z䙐KN&4x:1KYcÖGE~COɓ^>=?ӏoBQO^xL|Ur)Eb:!8Ɍ:e԰#c™QX#H:TsGjP-&N)cef E L=@ iLGcK:bPeG )Ff\ZY(.^^Ǥ:x7kx_09|~?τ]zIYwøp]7;HI;ٹmZi|!c-L!t^FqC54hg Ed)-~q Ȃ21f{QO9~{~'nEb?׻=G(VWUW]|Щ~կ{,M^#:dihQ tm ^A"ݤa:fs_,v~kILcaP[B|k$/XL_R: C/ν1sZU<-]aFEbb2D *\Uj1F)Y͠=mXpͬ1G/jv3Y比5rG3~Z{J^ E\Hb|@ KA &޳U]z?uM8uu*4Is`93q) G٘cA>VuW59'_:4xzƾd'G$If?쥹Pr]&,o& }A0O!H rWu6b f<=O^ÿ>!H<0986rػɏ{aAa2t'[HڧTF-תW(>@sDmffT 5f,:^5 | N5P$Mh5J鶼K#5 96&ѓف/owV*4V&7F:-i 8G-%lbrv2{0RZ$1-]EG Kyx]'Od??n|mnVlKve^?>W;Wu7r͓n2ǥSË25.nwc*/#}A2_01,yN02фTB@,.O]"\ݧn{a+u1.k (9gWUp]L1P 8tRU< &c6BvǼ]vklk$MCSC0&KJIA=CRȏH2% -39 g[fΤ9ASAb*qLZi [eV٧ks1Ɯ@v)1uS'Ү6JK;Zt:Mbo=o@t `\eMpNq" vZΡ ƣ(FYqlu \{&z~}Q+"D(Ҡkɂ39-79KG}!޸`(#ҋFZ͑'|<i#3:_15&WSVct#=Bq[R^}Kz4MoOo]W㓬,;y#3Ќ{vz@Z .~n C7V?Jgo[O[,tGY7G,Л3g@1*uOK8OG$3?@#|Xsu-72b78xVptMKpQ>|<'Qk"m\"D"FBk5Jgt~8M.dC&`m N) C D8r_ ,̵\>6txGJܑG߼3y賡\$0ǐBR}rǗ?@ޙS? ՅC2Mߛ,ip:i4hJQяJ0!@>.m ߣ?T{f#`ˋ+?1@(5@Mփ罢+(Ub2,>u$= [N_0FšYq?0q5m^?0,IrQu9PIx#x«_|O=-m`>D56z65$ܕm5<xhWV# K2Wu 7't;,=Io^7OT+,cnoYyUط!L:,yݳq9y AV\GS3LˎfSDž߼$ F$86V*C`}{,P?_QGꟐ{sV^F#ֽWA'/N{8{E |k pf{?Us={S/A~| . x4W XJɋ0܆}0©i\sDBãс"U-ޏf4~t|h{s_](#:d^;?@_W`\}wO-U4ss,,I.z_žVL܏2ENt7|fWF%?C)D89?RB~Gº"˝="qzp ,D/Ӂ߷(6Ѿ^4.6Th3˟&extm&Oa2.lá㬢+Jqy?4j„d浐{ FdUȗ zq[o| \drw.Oz{7Ҋhyu6]}x_)\vx7]ޘmU&w#&+[_D ѥj5LڋYW meJ+O'٠ $QY _%nZ/LzrVsSL}]MuMdB*r1[< 9UAX4_bܠ[ i-kSiANP ܢX3mDjy@uX[h_jBhg؂":IRvM9;8]2_-;PW~FBzĊFO":APLv޷9;T)fh%=@5 %]ԄRݝ1He$Jl5\['AN7'r k5&8Sb/ :zu:q%%k ?<_p-¢1^Mo-:ԐSDCuy20u`K_\_lfqC{+i]5zѷoGٟu>\B`W9ZcRnUNh(L+b}cS'S?~uD0Ֆ:4O!u"1K`m[5+ΣCn\0fZ9 hj , $a Z>H&UIeZqbq6ڠn{T;,_7]R1K._p34Ё&lK:x]Z/O-D[vOm}4xKy\"ZE@c˶#Oq**o\Wn<Ź$PZ'%=p?J-Wpw*w M^VG8\/"$h3?B,@gH``z?MHiM<\爫`f4e = x٥ zP^Y N xHpPvP 4K*Vpͣ wr HG=<'pg$Wsu|?OE,,qWӼ(r -B%1%NXELFb4EJ:0i( 8Rfڢ5-0^zfh 3x?>F`6C+O% ԏ+X?oEȳz#q0GC # f#7)_@(5f偔w*2@PVs'YENs]B̠3͡7H9ʤ/YR,Jˀ):xFuV'7rՍ8,P JXadT Xs$`&<)^wtcWs<ˬ">"E9CF=(z^@ ;#!ѧ!١b>D(5e\mxTu>&K$Hg#ȵSn8 ezxhC*עzS1SO2Ǘ?|ׯCȺy9:܃xtIEop,}SXߜ_o|{_]،1|ӣP-[Z͠lV,/4 M)$\nȵ\V] ͆EFhS(ql*6Xk+@tkrSk>"j'_oY )w9- .%bJg;>Z6>`&Q 5l$ l۲+v9ﶌFf|ٯ[YGL{ٙ}z23I+ggVg4%+~3+FH΃a^16TD_ُ^T&[,3! cۥH/?@iDEޯ|&(A,@sM ՟[Q%Ew|VWe4-vrzr 7_#)\\$4X;LX $0na0lG3jR3O[S<4AQQpY1(P3wˊ&Foi|l g$_Zcŕdh!g(KBjc3"5K(?i)Ռt {t3fŎ8 _0$Y`i20R8ch$RF,׏ TH@$2&q+dV0 M"B4Z;nЯŊIP)0RH$uLiHJGd8%(qcDFu&?"4FM41FpƖ4j'pIO4׆"%逛VBN5Uv N8 0 q@\:@L4ụSu[3lʩ(B cX;bw)_&Ij3|eV1[r+ Q.P D;m&;DSF+%2Z)îP hb0o,/o`*N2~~;g'0{Do6uSvZR 2Db c .6N> Qŵ w-6YcB,HS_vw+y/5o]LT' =eDӭ~M쀱dcbF=l(;fY/#F3ufxJ2=Ĝ ViQgD9"Z}⿆R?RU|*(R%ŠFNg"3dR)"da:)s%t$N84w"Lm\AxNxDdRcy#'AexhKRqUya(T 3I%yN4Ef%,h&4O)0!gZ) .K K}P{(V`H8]ŚTvjt\!\Nø Z0*EXm ~d`ƂPŋٵg ٮ}?~m'^E8 W9XAm OXvSIقtUǭ!LBz\hgLzjFhV#fAv %:nDتhX3'ּ7XC tfW"(gی'Mj+X1D7'Z>T0MP*Y=8'ǂWw^`Ƃnb㫍`>P_ix,<5M!x Of4dpbSTXSR =؆F3ZdoljŞňK9p516 BlS !:Kifp=5@+lj=+[!"h/P"t!DE-JkE+9?CT /b&\ΰܖP'7&^-S-18B4=Zb0}YbS >q"[rOVyMkVt@X1%#ʔ5FXi Bl\)%rPTB_±M9[]a\Ӊ=?\@u1o*ߚfx /ʗXԲUg|D)=vgUh7<\xώe x7OIz1^ӫvj,Scؼ֯AW!΋gI%in^#]};;`v496Ջᛵk}{y_b``Ab֟c$&uxbkStf6tH3OK%.e/L}57=*KVt`1r͘&X%aae?nK0]Ew08f k3㨵Qĭ|,g6 vkS}f;jHF\S1b;Ag=2ی;^OJW>hf0&ܰrgtR,(e˳|SiY;{f3D1d,b+BsЖ4F?rF&X]Om+3Wh5Dn X'|dBP,|ϫe8q.a9c$ 6Df8P n6@5#ԜUqː0xm$T[YS7WԞnH kPB Q|ϳKQS[{xvc aq>0:2Z0{’@Jr$TJXnwrĞG%8 *5=`VYmw7칩׷o_X|mnXf<}i^7_!b'7<V oYmZ#<g?=Mc7g7gӗ zAUxl:B[`eY7_>G9y>OnnTGnk{rᔃkogAD7i؛r& NOxM,c^L`Uk }5 ShI]-?\!wBW0Xx//KцL[*>`(:ګ"&Si_n"{/g8FxM?Y%yXc2]zB}8&w:4+x26c20=MpТi@Zg+ߨG辑#),}ۙXs\0Df]m -̶#)ݕ"V6g2ͽ&f^Ox/vUĭ|h# bQ,Gk;Lys4w/*y@M߾R~y$h?¼6^F@eY&lrF7лC=qq.ݠrLXȑd;\⸭pP`erXvEcJki=ٺfsĄv=S޹c'V`JRH#;=S|ƴYR|Ȕa_GC5LۓL{r3PF +X~p`SP ^X7+9y>-;"7IVN 8S[T5\VVZť 2`>+k:/7 hr%ˌkiWzqꪈYXp@=Ŗ kX-nbFh10S^48[bqɸ05S ~mXo | m9Hg(½,$ 嵣Hwvc6@U˳V]§Niw?|c) )_C0wMXV,!J`}.Ԉ{Ũ<꠿^U'-٫; d㭄%{ܒfo~q>'c.&"s8Za`\{%=꠿\`a.MLX s?c.&<4'+M %[.K_4"~D#/vbggt^ W1<ܚk˻6ueӔILZ2F& p'VK5Zs7Y+MP^7d]ͅ"2n8R 2^ú+\r{AUƫNnɸ|֦wM.cNџ w>oHuPE7@a!HiְE9 ?s\L%3Wpח`žb_(7_, 03-dphn1^/9``lO',|/<(^AiRf- *h'DW8ϫh`\WW?q2^|Q>>uPchX2W?L*-9gۘMأp6 uyl, #M"Y"a9d1m'"uR*a#$Q)j[_N/Nb B)ٝ !TmՇl̘jj+_i*P|f9 5ೃeL IKjɨ9eo+4^pc0{5 NHлUt:6/,7mU/_isV3zlp׏zS9ZbvfE f@nY+ao]៿Y1V&杍_QT۔YKal\H/Fyҥ?qІ;6wCc$~YT}[qagJň$ _wqTHSc:4j_>6(b^̨`eJu̥&݌ӏ8p| 5AtWA]<& Sle~(-weqI~ZRއ^׋3'h5T7X"Xr  uVdyqz"m\]Mp^]g35DNWi׽o/,ݯ[;[[&O7ZWv'|8统/+_'Cw =/O-Flw^6?hj*Z&Ovv맋ucYwUsWWˏ36Oױz>@vqe7j?W ?63f<><2ן:%8ehUϲjbwB >M/?'܃Rpٳt-nn>P>9#G75^M>;$-m2Sy_oZTOog(/ӄ[+4: ݻ}ǰϞn>{'y.Q8-JF)".h)U46&Z}6>{цg6q}]y{11ְ>lIJLGv >1!+\vU=nCiwni.73m6!~h^:A\}7z !%—rQRZ%jKd=ьrJڿy'fd:$M.4Ը(r\2Fpe\+}`dNQ%=45oa.e;m?ǕO^Z/l?W{w|>KS {M闢K *fDX [;Ie ;LIVWWm|3 d ¦2ͮbKvwm~qx ?z)ζ;1hx(#X[!i!%IEJO$6&O?BR=i&#kҀ^S-G/DUEg!җL8"f1j =ȘqcA|2*wxFqM DrTxvGr 84  K8oKmhx vt0Q{H\8F(@ Ij"g-IRIJWi`w53|,ȆNA]'Ng #%4E !&g<(N/ 9G.|,VM4F-)/K?D$AfEUIioobu@őι%x^q 9 RY;.'ýo~|y NX}dI1)#V)3mhvBệ-|,L'Õ`zv3GnF[jT6J~L"%E,8[DDOBy+rY(NEEV.Ljԁ% DxU'p ;Z;RE]E>T%f$0 UCllPL<@Q?k%h9S,cw2pG7\" Znqy^n6T~l}wu~(`ۛJ_WOv8*Gㇳ˧s|8 y}-H[ uZY EJ|<`H !Z:0 |¥M?].W4 AGFՐOԺLC 颀E#2DNnLX yQyY0ck @k掘O>@_ jCWT9! 7c83[_:&0̙=<9 TN[WR&c0jE> ŧhM*1K'‹4yC{1k;BQ"'$SƐ%<&* ĺ@VO$.}s-jUR~4LWqwy] νϪO;WO 1yj"3mPw70r@ ?C0}A.x}V:Ѣ0x/pKS4[/'MŪp7^5|?Kt=mzCo*{gϔ[x {=b΃ -yYnm528! *:Vf"x)Z+ar<:1 0cAvyToQlw.c$2\~"g%6E1 ,IT4Tgc9`wDZPҩk.e]sA,$ ϯZ Jɲvx,|A1'|,NהtI*ţ!1,yj:K) c"H'Yڍ/G $Cj7f6ݱx(A 5=2[4d O-]Qc!ݍ =k&E$MdH ѦHoD%O9h|UxeSGMIK.)]N,R;GnpXȝ˾y۱;[kqp^|c  _|Yb= 0gA /d lp0ƽ9dZEPc!粮("/锬FT?U#RcJMW/T_stҡSMpu -_5} Ձ[ y߾c@ȨU'×_o Q@–cF]푱ã9|,bB-~q\vC߰";`[qgdwbkbvé Փ 6sq:I)4r^nc.I͋"O #wAFyIJݿvԹnƣW5*PC{+*o.^^ݽU<+0;fDcdwȨװ O|8#)N 85Y3fAndW7fSJ!!4Z¿S+2C0V~#^3dTgK%WxBpSoM}Ľ#&ɽ~h=`^sWtD&5}ǽ ]W)sU"*~}7n%5Bɓ:Wq8wc_@?$sȿp󹫡0wBE{d+tLIx81MSr͐6Q-]g# xt2(OXzvmo/Ẁ[aI4N6%gϑF#Ei7n E>pQOҘ 90k\t*K9܇|E>T5]͚ *ܮ9&ԉc # NaBaU!S+gIJzXLqbRkw~hLTƣGF%FoPƂ+i n`W+19dGQB\êju{'zOesVeh\*C`UFQ©YI"*"xt Qc!qz1zdT SLG"]O6IoHV]s:m5M0UsUԖH S{/opoq4dbညȨ9 a1Ԧ',Xώ28g@VRJ43{k ,jƂQ J,N6a}Ðw 6e۫rۚ㓆goW3$̍Q Yc|$ۘAm D+|\EH" k\ Z;P?m'kyʃZN]c]Rq@Ȩθ4v_ӷBQ>6o9Xn1^{eXcdQ#!uii]=2*X1_ו0FS'҃}*K ~b2Nc\eqA X唳fհrPG'W|@_(o|oŦ.0\,}auUkZ7_[QNMJK0TRMhS0|$ 'R녏Sm7]=$&AS&<nćiS$ؼ<yl X" l3IrH- )UX뺡'Xh-s,Mnj;kkBUl{_Rm Eܗ]Z܄)^L:h#gMV35^+`+JUת > }ϷAj8.zMLWqS޿6GyW8> b'RlJ˨2$25 '6{dVl[6㔹a9TvGFM& /q"BVTcL[roi| r}Xu+ڿba"eje4o.7ը,e* eYj/Bq +>-ynaCE"|,|_2ZQ#yvq0*z}+LyZ꩐@͹=R |8k{ t wr3_^,!oU-#)+sޢ֖FK)vQ 6:">~.>[DԹʡוb5'&0C#9zLlԜءc?TMrTe契y"g9)iV/pzEZK\I#c9GrXMUuj ޺D!< O8n#W=aE].yMmrJvێE%[?d%%1mXTY*ևq: ɌY;Y05Tu"SʬtS.eEO+Z@s#Y `a#E);rEM#hk([ .9~{|-=(RԔ0s P! #D*1 ݷ8j1F[@J]^_,JN N@ V\b$ (1ǎ1Z$ǁJr/aySgkwlQ- SeSR6wd1QzJ%PA'KL1X{1 ,!f$2b ܦ4R[ubv VPoϜ>S:-WLVl=>`xépd C1N$O!rƳ:R;;M0_ $M TҀʨXĒA4;y]~+]pu-=K#=a"DvpK^a?6'ۤzOIRnEiXH.]> \K-c) >#fG{0ි|@c{Y]'5 .ɧ UcbM_8KpbYXehbj[([:`@ b̝ST!0M#¸҇C ' ySY/םN}2& @k UUf6+uʅ`$I .g)  "-^$Ɉ: BcQ_Y#h_Yo%,P'Iޕ9ק.4rmr.]WkMq2 3d8va_Wd*̤N`m2_Im7W aE1L0|!>mE-̐(L¿)9Ae,G{W2 қ8/9|]t]+oG`0Uј&ffG &Xv:xh1 N=W}6Iul"Hd̂ӑh)8I6`bq[)y#WϛĂx䜲"۳[/<˩tn' mTtj޵K K#_.hkwdy"g Lsw@=,K Pu[AR*" _7%68. q"hJ0礬mX,AƉ`P(*ţR Z񩠷9*>4P|x<3^ z"5,jp{tA9ZA{ 8= ՙ`iNꚽjկv_{ϊ%iԮ|~yo Kui.eW/ F, lt 9ibU4F980Le4} ǛYf m5۸6 ˂9_̀4ToCJPD3{OeJj€5}Hha?O9fy׵ Y,) H/gfGe75.v*V{[̾Y;nn/v3[raQPq6_Ί;[bfU2nj c̻pyiTzzÌx1Ӵ8))Q7wJ-#*XU_l#QF{?x¡Zw |~KMPӧ2K+UCk'O Y]a{@ݖ{G`=^`R\%1#7Gx"QmFhp6 .A1g1UFfAS_ .$1nbWzZzmcbx65;wY%W 꼴uccc6NKIzTt QFzG霭I8dRc K:Q$Kc8K"|x^op*‰(` N5q8"_`Ք\ncb܆m.5KN}\.uā $,-SB]qJ#3{y~Co zvAk.`v1v-ðsmGW$E y2Kk<՘IyW_e<,BZm!Mai}2ėAT?T̸{4"*$Oh- Moe$˹O1t+Dw&[c`b;բSR_(C], ~@*.V\2r =x H=9*FHT4eBg$ca,g"q?/fNEO-&Ip2LqPhlUdSҌ<%Ivz 5 j4FT6 I"P "$2BRm;̈ͭ!xNk_A40q@G~Ĥac;6b bfaC9$aNƩ?T+`Ʈمd_+$.˨0HJ}NJv"ƺpvv{ 1~0珰1N2#R9V*F9H+7pmU&Ɣ_|^k` U O̿UgaqnØbYXeȤgaĈ/<=#!~?:^| ipQzl͉S_Tǫgz:#Uhpez|YkaWK(1~;6ר&^XYSSwO 82rVf Qpz]Y;C,>e _U~zcpQ*sX撙R:{::,{>za?̳_WGPr=KlH&DGBm^uAT}6EC˛/N10 yJ$qL2f4%Z NV3n_r[|OvYIi鼬^*>mU5 )Y- 3_€f1d.r g`^?ϯ|ɋ_ϟaws1?lXx3I 4{ ]f֒]|Y֨._-OPo&ڟ/Z"xVZ?ڨ~̶F?N,5{~c0WKRr"^[[6?D1RAPJSmi`, ZUi&<@"djU\O9■S8&o-]JK*m *el6_s0_$7?YM)"- vX%VݹjY )\@R?=/]^EWq 'jboTo:k»̭wyS/ ۈ}1ջT_t7r;J+C^>)A^n*ݡJ[& {cz`b*?,a.&؟?osTmvO U~˛Q gc^s{_Ŭ4ff],@ٶ謜er~N*_ !5bѢaA`}-WI"j&-Rg;칎HD=#"Jc\)< i?#&YݕC_ObQ>?8D啭Tᴎ*9FO9rB3Ĥ RWQ_ i_ !ė). w23.>,(Q CsF[9,eyHdirqL[LO0* !`wHbQkM29L!1nIOA#YtF]V͠8_cQtnW9.m$..((LAO[4х@ģ=!*B.<[kZ|d_s@0⮩eY{F.(o@SUv-#9j܁GV(9Ow Ԥ_))!۪64@m7Z&chAH*tܹ|]rw%b(g Gr-.PrGDѨ'~>PWcv h9ZL"E79!F*4zk@sz1ThT-iudYQ&XC #DPbCʘRmw:?_?bCm 7\Bc sXL .َh[qQI,tMiR%'>x$c'#@|i^FTtCJ/osu 뙧县F=*xʬ $WtUQP&+a>Yq)0p羴ݾCmC۹Wb(D&WV#UТڪ1![.+C5ڊ1F56g` *Ľs`3<4 kB mEi`;і8h?:JMC6^V뷹_BDY i|JMT߂maX !yAO0}F w ď{O4Έ kdgrTJ+aye$ K덣 ;nQ.Y\XIjkz=Ej1"vA4Hө;2 PbQ|8Xd >} i( ٳ.(Jńl' uha1F<. 9q8# Vti#Aj+/Hgx8BKݹ~~!kc j.,:h iwB.!]kGBך-R<#E}Z\wD"&B#F80q"u𛁖T/!YcY!.)Ac""F=UQDɂW99?ΐUq^!Wgj+8LFw)n-u1g A&؊1 rZLHOG)Qg]XdqBv('-5b־ͨ'ס|G^|lֻK6ibruWGjεWzjڌcTc 'A l:4w'y5Aن׫ tɟc4' _gdk٤/oyRpEN&7E+5w(|} O: %›K|גܻ:Suf{t - k^^B>)5A#7QE5=LoqR l n'_u JKi,cygU/_8+-y1@!0VYǼCE(<0*]\Q'Xp*gclC.3F;9dٜYطic#a)W^tn^^J>ۧ+~V/ԕֻyŇn@ T`iP y˅\"FGA'M/R ]U\D#͸*su{41W3VVL j}6_φ1F$s$vQ%EY-vD2"~RtUia ʖO@|rSxI--U/ 1.y,%* ¨it?;f z!Ș0 ) lYaei1tƓ ŐO JQ!nwaxam+M~ ߀ cL&8 Yqumo9[>4H\ouPi01«߀fǞt^)X>z+ Nk*L`iIƑPY'6]8d ڈ1ž`XND8R_Ca4Ђ [밵鷏hTR{G4FĀ9'"s>Vst;j|3]DBO@*RJdH/g%J>{#^DWg9Źri!J=uƮ$:$\/4T$㕊TH:((iD8@vn]xHi>\G$OH6Ĺ`4ƕrZXF-uW> /!6bBK3c%ۛIJӇ1GȋWw?"GyCzQ^P0s\C5.|aAr7#TP;A[&Yk Tb]'ϯyDŢF}UyK/} ^X g}qɆ_K]5-nI;< ~֒&9SlPG*QZ_5P~sv {$ip\ %LxNT )SgxVvJU8#Sw *A*P&{dҚ o[mK4q$f uLsZ|9o|:}{zbډ/y_Nd [<>,᥄7bfD< (.@FseAH:LV*>9X-uׇ@mCKMjj \~ܲ-(i0f`pRjg4Q4S̡vy)jaUO㴲 a.O6chQ wڹ2x7}ԋ2†A:#a*n-m]t|^M^M7˗Do϶徥[Y +VÓe׹!kCUZ-`?vUvn>yjΦ?F4x˞I$׏)w w̮Vvniz7ӻ?uݗa{6;imDi}Unb3?my'b>eZ^ͣ~MA~cX$ ɶ_<# $k@>|xU?B Y˴&ym X [k|ū7N4w[q_deCou1G,|A[*TlΧd2w٠: mOf٣0-$'7a>]^Rzs<|:ur3ӥgnQշ<-Umvdvg]fHx|n5WbaiycoO3 [[}Ɨ7F^)ɳ}|v9]j=~:tTڝeVwUnXtOjzn2'S!}oy:]۫an@zPFnk4ZRC7u\yP QNw] ܐOkohK>l[ %&`5dPҵ`-ua1oxF)~4"%A:#g̑ ki>)v{A1$m V*~EB y&<+HF/xK"_PgB}A!I#|p('N/s_|/YɟN݌>zC*'\K4(6F_QYw= iד< B$ T"b+C'ߩ#0z/% 6r#DPb;␩RoKo6ѻJv{B|Iz!O ͪ  8SY #pD #m.l` H;A1KH 3F ^Ŗ:SVz(eہ6m^N2F5/rĮGRvH_m!4]ƲD "ytn(a!#d$̰{YKT/އQ2nŋljYڠ-.M~\Xja@1V9jLsă qy>o1!>ĘK]3pٲ->a YZ& ClZeG?T83S%&B- G]U_;,-ua1FpX(aJHGͫū50]2 %h/x=1S1]r6wPk }qH>B򿪐ÁWGҷA\~0TLT %3Ϲ5$sMHknpb1Z h}vi9x$_8p?'W_;"f= yn>Z G"7ݶ`Xic/\f)z*h" E nd@.mUpI8 O!A0-.qPiq=0"!%esI.OQX(dOSFYneOiq?@|qYIxfsloL{#PVWeTX˟WĨުq~;*COZ1F81b #啑jRbuiCO'+AGxt~tYg'y,gal.nv\X0ch9#bq/9r^FDAz#Ho6#|,ۈ1qEbZV TJp *>Fz ޜ$0z nъ.CRLRS4.`yFҜE!*Ro '@kmƲE0, "&dN_e+%EU%Yֳ%tıK%)>νԩ|KLMòh$] 6?L_f6|u52v?Zzy6m];,nAxz FK#q ĥQ*)-VF `k&p~H}ϧ 2^NSn>Jw@(1,<ͦ8$.lj>b7h5YvOiowrт|A47o'o6%b>4qW+}hM<)݆{} LiC?{҄XwOj4vQ 6DOx~:ǍhSq\v%UŻ_񳻿WU0$j~ Ay1 w0(z6 anHr<="08uc+YRm[Iƹ:6I2e#5G3y1s\j0&!ITcktDFip%GLRGF!q@k;7&}DTúktL3,y,*͉u:O!pj *' J InᨪK, Aq*gcD) ց! RRP0i0 ƺtfB)6-ίGtα`«k +&eVJa'^HQCY U~f B9^]WSO ;"},8f!TEv25WB}¨Q9tɾqO]ډD@U V(M\j:D:PNÅ(`|0kRd JY j^B7άd2f$3ސ"GI&%o^FO8t.q=L;l@wsI=|&`\EW`3 Ч$Ɗg̏o{<0g5M@V-dSV+N] GdtZ\տ5}\+D|I/`E7aCGd:4nÇTe"&n5[ժ|s& pǢ.Ҹ[MRLnЃY%)u_LL}N9[?qXN?$"Ƴ M:Aa8h&jza@cY x{~Ӄ0SHӼѽkk9 舌 f:UNLfP{J2< i-)̤ XS( MJTe8u2 vHG"ОBhiSoxQL{) ̴(c#{T?u^UL.RłL2_Ak5eZ/2+lMmLIp|ܳL<JTad[yN\`AkLCCVjJ\6JFap ]9 ,( ]{GdG_~W:3JW*',^1ROz̴$ꈌPq9[zz5-]j#$#PG2 n2NmlabYtr<\0Z*!Sk 'dG]$D[*b_㱤vk+`"ACdK)n@ s jTt6rDFap F7B"]c@Qq##a r JFap Ft{ڹ489]y&>13eq[1)ׁ #"R&xxf;>:,!1㩉dz[@G=l[칪= FًheNaU~敌ΐbq\{QgDC0 E_ɚa6}\Na2Nɘ(oƩ*2}-0Zsԥ3} |i0$bɚdi蜬Wƕ$0$.\̍5KL).ɝ.I\a4\ H F "'T;0-q`L}dzE=es}D'iGdE_ `_<$it`:(v:QNTQQ|uyI[FA{'e8",8U536ZY9ʐ|TTHh>xх.*c7m/rRf #!ں ( N.YS4ܱX!MS"뮔I He-1}zl{+fP z<1&;LQT8h=8"08'(8,58#A}ljq0!mީiB &Iߚon{):PwiMtA9lZ n>[OZ/ Hw?i.٘AjSQu~}l=TBGd's)֓!-rɧ92QEgb/h~ͱUGbR9GdK|H2M=:Ω^ #Ps=ҁZ5{4wKGdGϿw j]ysDFap2 \UlݗITa}E]sRvm`$CZf +>Qk2cCA0* :W uMv„r+eδf2v' 9Z/ H ݸu d3agjNsi6|u5"W\bfyUd\6mnu EmnAyzE72FWʍXCI7ա VGX=^0l`fQ;Ͻif{#q]oow/ ,wBtW#Xsׂu7r~ߟěwLyQ*:npÊ Dx??-/Mh4 w_x&R8Mj?Z38~ؚEGx&YJwShLOk؇/>z<׏q`&iZi}(yO _g3i(*#ni8%~ڭlnZ6̦ YFO~*.I84֣cɏ]castC6~72e& bO Fkn ջ@@GcJ0^i&)JVy]*>Pc]{]?ÇI x ˵~ /"Lv߭qS~29: 5R/hG6&3ot}q3y4 Ѓ3m@V:_隣l'_mz;s/ڭgra;6-G»:յf9Ya>vnnA40gY^8֛?wxg9;ֺuo_vug|3kx1~?,.{!uqcdk><fvq?] <.z?y'GieQt L"+geъ 83M3i;n?Y?~&ލ~ ݴy- ;-9{ k &6a/n>ھ]/S Ur# Jq2DGޭ+"iMn>N[8D3֠UyS1:Ihto [(!( aTP82WEqٓ<]ظ Bp *ysn4 j}9]Dyk,b %Y_x6?r{呫aQc˟ocrv=KP^ɤvWJ}i,+QAg(yk(pwPw*^ht>a^ ( {|#eДJľ "Ka1,̟/БW+$g_kl(6e@VDƣ`1-O L'Ԃ3XR(*e2jPE0[¿2/mEǾ~_^!ͱ^oO-G܏w_:Rv뿪bGT1y_nPX > p+DUW[d\O[9l~U lw} X| PƄqLq-Nv椪y]q.ŎXRJJbVsY@|ӫCM# w3غ^0F-`n_Oޕqd/︩&kg2ZJ"6XWMI5ElI>JHlV{B0=)1fXx[dc_V!'CQdMx<6n D@ PO \ rS.ըRR qJ!N))8BHqYRK))8yR qJ!N)_")RI))8BRkBDEJ!~\S zߪ~xg#R^Bz՝=9goy 9kn`8R}\{rS}ʽO) I)>{rS@MʽO)B q?qUfbfGL6qgӉ#)ٺ@ h fR0βo_:g@qՋ_Z_2A`G0&mUvn:~=Hv8:_śU( 2kY }-|#?߅~1DYtNG t5 rWH82P]{ޅ@w@f7 3q!(PY/ZkMHs,Bi{O 2{VSd!s}%%ZYaVp/d _Έ/qdO~4Xt H[|QggYET:ko[_?{ιPJ2|STg%Jꬤ:+YR:+ :+Jꬤ:+JH ,NuVKuVR9KuVRTgFᥝTH9'JBRGePOJZ`ceM7ʉ \.%H A2 QdQFr'/q-1{e;Og [KҴtfA0-2_ïF w7~a&2DD|5h1|hv;e5(Րb'weBIKb Z .pb`A(s-*6J *vρ*h*mޒ䵧D GԗL*"㍟04㥆c  H&^$&w_~{&aF c|:'f`C_X1/KN!zpaE.\qZꕑH:dRIT!#0$%.,B^QA}ACb8[OE `U .GJ"%Fꗑv:Â$抆/%LG;.8><`ƻ[84o`c+jG8um<T @W*xYF*rcW{GoHaofiY |Q0|*xmPlW%P7癙Ww3WH*^JVl$PvM$ͦ-Ȣ ,яnXim!\dד ޗKȦy7w o|+Lkx(jfJl?+VCdP0Fρ?շ $Ze8q;b˥]A& P+0`(F"2bZLUb|3nJ g_1_\aҢDR j6rTkH]SIVXSc꬚X.PDUb[d;x1TCU6F$otgv5 w5O-#x wWf$.k{#rHTZ݀:_:w4(xs'{ڣ[6ٍâ91eJ_Z#0p ˢI{sbHh͢%#)Fw8H.k2B 0Wڸf _}@pWj"H)R,^rư!V;i(CH"D܋rىQo4aMڭ;Y rrn;Xts/T"uDrθE["wsJ}VRc곒>+J곒$N}VR$*R4ŤDvK"R."H"R.qE YjXjE]DjE|@ *!5`^G8f@דHyM׆7>if\P P!F?aF>]{˞5uH8_ ~Yvn8$Ff4bb<;S?~z>͋8d?bg E|lT"hE+ ]A0 ^2zBD#pH&ճA`mPgg@i,E>$6{.eWr]d\%yYnBkR}! I8G #hr"|azhƸ iVUl>]Fg]?R-:&J|T&.j۝tvZ=ٵJ=;C큷N頂:]GuDA裕B+"g^ bYv7EO!'$e"Mgg<g-\NZ(S}/ti]#p5h nf1,L ܗ.S0~2g<.N9 V!a3$JT)~@R vnǎ.ށdw"b9[,Tw-=^yhn%zrYⱰg% ^O'T$`&zEݻ-?G)2-p( ]n~ݞ7}@ǁJHv>>4nK{ӭ:b37nOVHuUmۥX8M1zT`ڿOT]ԏZgx4qwDQ!CZh->eU oꈡLcuE*NPԉho"-Kԫ`~?LWβ1 /-P!x&T%?>zvq]k@UfbǕ+$D;gsÎfSDo:q̲]>~֡B>nY~ j>ߜhSW7/e'M1~[w;a ;%S(JPT%fIYRiT%fIJ,IT, #˃3vy|3 e5_Tbշ2`1'̔!Sd*"2L*"Ȥ"2L*"Q,Q*"X*"Ȥ"2L*"|@H!ّOygrOWTe9wG|! q!K*[}]G#=c 5)+d)sVly,h)@z |YI{P '3uވ(ࢇ]+ۂeom`b} "ꓶU@5 7݁SG]JEq vMf.=T%ƙ /rȍ4,qm;vԇDq  =:x.MLGS;,ok'U!,Ry@HQo-*TJ*c`c I-۽4j \C0QJh#dˏWMcOi|cs&`XCւ \2%X5-&`NE%Wkls-Z?:d{ni%䵧`9"%dRJ0fy0},#9fpdObrHP@n$vG_|:'f`C_X1/KN!zpaE.\qZꕑH:dR+o?j:p&ąE+*/hptaTQi VRy$BJaQb~i%]gX^$\Q +a:qAp 8o+CX`6\6n ]7lDv{ߴ< 2+ʫ*F*8rcWI/鿝|=^;-adEF!cCQѺMGPp|Ll WƱiTRT{=NH~~/;n/ Bo,.mZ$n6UK cq7#rjz4[Hr]OsY{;wow) TL ag q 9|VhҖ(Tp-_r28Rˮ͠]lI +051"a`RT*7jjj:o>[r%!- .K$랮Zko`#G廆9&ue>,.VX`bEL =3QU'9X&N*^XjS 7#o'\H Lڢ0EC-}M,Xߗ5O).#x ^W !>s5=u͔#i8Riuj]ŵ:w4(xs'KX"iG@Rm$t'{9h~h`aY A%t #\h²b9\-{lhv=|#sHRAJOA"e>f1w3 I@B'H^ϕ[\N$ߝQ z lonX\b}3>: XtT'Ohv'l-:*t)R_!4?\Q2rG* 6J ֕N3ŝ/+EiIOR)+IiH) 4U.BOz6H#Dx{vWvYq+D{@jP ]C  tvPپ$v>{ϒ>g`kJKk,Hl-eO{9*EZImqFJ{=, @36`;7`LH ~5}v=8/m𫇅$Y]{A@bOFU١JمxDf& %H{=˦|a{^ȰZsș ,9&AJñum˼pe5:dΛwx݆{=-^0ɝr2gv^͞0BYw(yaЅ7X?@(q4A[. i6W[ެwxv%]#_r ؉!g 9Cbp9+7~\BPoX )m \"Q!1 x7oٓ*VJ+pd[.SW3xЎDi dL7%Yjn?EB6%7hS`?ABȫIoB,փpdeڌ =QftUWkH"4L6:O8|p>YaQZXd WSFE( 1km):T=!hTjeGъRo^ 7y͵V\dKP1"KLj xtT3Q+Qdy "D&f$5EBz1JHc :J,H5"xvʦZV8:l%+q243EB^9iJp* `D -T&1$to7xLV9cXi]1n?EB]a0<]._u2= s1Sm‘z-Bk y$ʙ"#6YOOٳtrttr<ܚNl젾2yh- zN MTr|U_Nj7֡ ~1xkr&+UA: .1:l!л7ӓuS6Y7l>fzS&A[ppǽy{4cqzv6ǎewxG&Un+'G&Y[q$pt UkQpHe6#[K 0B>"5dG6HJ;I1 XXi%h csE|4EBޑRfT*% vNJRx [ѐcS$}9&#ORQ [{AE+jXNO7 /bV:D1}u۲)UQe2"x V I 4}eKf5QD73EB|D[+^{;zK5 N26 śǯgo2I˥huv7|,wWBٴ^!_N;F2To7ͷ]C5lKmfԮPwMRBD&A%2#)zġZvOr ETxJc)퇋- (~j4F Ke9~"x;Vj[vj BG&&I))z7cԗ$ŊO+$SWedzIyIR`@$"b$NVPX)zfuݫQqt'sqw0`]~eSn) GO?3yKŕ3E`&`Sn89!!*!+)Ɉ~~9J{ rPZ%jLl=0}EMR2AKUhb!uM\W/kj4L~DלөqNrmHgℝ/EdkW11E׌:r]sѣj%Wy=IqX^R2k/"qgc0)#866(J NO1)6#tf]]s_=N?4ŐɁNQ@\A @o9"c]JZ 9R_Cڵ{Fe"{BonY?/ ;vU'B%w\UlnFjw{p t>]lYeKkorXy;kt]~/IL.^tYq}CZW,k ao8/~+oGWnX1r*J^j;nBS{#$TɐZ].@ҕo$D^_J*zxur-/N~o;g{M]bH75< 1՟V?lYDŽ٦_姖;CC)$l$1<ߚ>zm0ħוdϰlq{,HT{r%A\xvoy^9b}GMgCIZW?nk0-AzVgO㡆^N'!AΦѧaX*w O*Z|~l>]k_o9z?W)+I뮁9uf"ԯyE˿W?PCXSVv䬼|4C`4?,RÃnؿ~yóeC>=yV׏mvwa F[fsqٜXF>>&` #Vf.Rg6aC=l>%SWRռJyf0E+r:堲/60QdxF}9%,N奄;5"W""99Iqlt8z>/rckڡOڹ t;%vެ?k!z0y;$~G?=ylΈU\ӈX\Р3jP}߽4WԶhʙ+xp5;T)6eLKRϘ2  GN鈵)-ҪTS.3| 6RVS-GӝU9}@&cF]l$S9VDGd[bRNVLI+{ $/ |tsH42(ɾ59RѰ(#; dݺdJ>Cj޸dpnx//kZ 羆Ə7V^LsVEX fgiMQ4üh3WN}|y]ግ~2p[+lj #66;.P<rbodSn( $U^&bZA8}R'=j>T. 2XHŐ|Ly+(;3.T*Y lqP\lD !"$L[b$!YJ[. ]˜$Č0I3 6$${Z0#Bc+,nP3yIx9o2i4toʣH{ґ^ Sd3qIFuaI]ݣ;cM]wtf!oޱw1 z3 CDtH9~` $J1CW[Z bfWԪͯU7݄>u3 ( 5caڢw/zfo#35ے;E;cLy8/#g=,'+]HqMsS:_tQk t0y֩USc P dviMwRs=5链ĨU<󔥕r98C*ƬIH:қ= \ 4J+ZryR7N`Ci ל>gS^.!\Cm! !\mo{*]YdYIIkϽ r3_K yq1߆iǷQwE?AN?į_f(ؚ}߸!Ro'ΧW׏ȗ4_,Ͼϧ&hg=۩$i!r6?PXҕg3*p?H0w3,Ò}io7:Y,/oƟGh{mhh/w[4Zt0)Qs?rm& L=]]3\S1R#r9ou3X x@.vWL #bQ>}ц[+R:9~QevKc^{Z$bgh&ϴƇ 9.gB.  %b`BEs-C;(/AxZ)1K zMW()$H'Ha;꼮NۙM[rGdҥ`^>_slgG|͗ϡz$ $2m)|9~CR= ˝iG_L S% - p%c(PP²>@{S"{ ޱґ8q,5EoLeC%їa C,zRщRџkkcȑJl>6W!/R^#_b+bI{S˨7FoCòd Z<3_3j儐4XI4XI4XI4XI4xHXŬJJJJJJ*i:R% V`% V`% V`% V`% V`% V`% V`% V{>wD% V+tݫjE,B%W2؀ IJ2|4RŢ$dt杢}xch#:o'λ$wt[r(o?8zLf'ۏ.l:QӍV6Yhʐ_q,P*yjDǫTL*(gЎ\`tNV52`wDd=CdwcC`x6B|XMQ(X|<]i6RF1i׉[3N{zb LקfK ΀B͢;{2h0.FJs>lv}NQNo%yDrdsȸ/|4 K$SNNn lXpK,v`ܭ.S{Zxf?o/@@(Ofhw64?Γ''Wat  1AzpsNwRj Xl}l3bVM{l1#%}{֚fLhy.[;m/z fТ2?|gu/GWQ'}y6*ÍۇZkf &p<760n/n?V#6v;x-F.Fzhڢ0{I 1_6 z[ z#[?m3 |^Mo3'zdf1zrǡ.] LY䘷Bx3sXQ@hB}% G' Atf'׳,"M´.F2-&sf/RdoQKS3"fDhlEPFO:37:\h3{@01xQkc,"{]R)łpָ$#ں$ѝ{܃P3 ywIЛx{'Cf~)h@$=%޽eCcb崋ʒCȍ6L5 |4( dNdǪ_E=ȉICA;%KI;9CmNDL^g/dPpɰrng,=eh\8e:تо@b 壛?/^=i?+::h5:01 1{iNj h"(\O2ANP "sMP5Y(h ڊNJaD7S:!""L֖i?q,OdIѕfL9kdNU!w3A黾=+<,7?e[DxD 3=(~ͳЪ<t|="QaP4 YjQDkWxUi 'u0*UQ݋n?D#ZJ+ QgSY͞Y͞Y͞)B=%3U/»_kBW_H Z=n85uA,%©ݨ`$8Wîvn1֎.c#je6XS.C>L3˜; d"2j+@2PCToG+~;m;fmk4+~?z0Jlv2ūVEg{΂!S#]Os$қw M]Yo#G+?1.)Kxvk<{<՜H^!b%Ht5j2q}ui\_Z OgW SyWTptϣv;&]]fnVG3nm>}6ݓ:4yiv~E=Mx~8Hne/dh6W![n/XsMoov~i1S ؇/Ю{{Q',deg|?<yBsoՂc.HPx!%c9;=Ϫ<9vTݨEG(Xb/vQ rF_H7ef?p"P4шIؾ']]H: 5:j+@aHG%k2>aW4eYZ1.@)EokYoO/zoZ5U !9Grb:$NVkB2N AR-lB)lE.8tPV"쓇RSD|eàu~{Kv:\ m`;8zK%8 A1TBHI L#:nC9IXRYﴠ0)G"ED[BX#6$YN.HƌHPNM򠀥,H.8!(Q{pVQf2V$N)ig-aS"yfGxK]#_n pA>ܟVE};rd8{9.>WLl>X{cgIi__<Dѹ)R$g{&ifLB/YRqF7^EǣZTXl((PL.-s.TD*3. 擞͎su_Qf>۷if ;ޮߏ=qrk3Hi %nJ#'gqʥI+ ,{pK4h2fNw屯y<"%ȑ==lh)֣}]@qf*?*] S ]}ct:وT'y\v:PcgC}DvCi&fiv5d4Ri5ϸTfqUڳz-r#քi{4Iƚ\;-c[/Gvӡh*yƹB1=jN&T &IG>>ȏR'U.q60ǗՒz<q.ԐBms=/Ğc"*,Vh󧜋)fG iD~R>>*{9^(vƨ7|l^"#Zm&4r)t8a~zhOzP}Lquaq\}) ΂ =8c.(ʡHdޏi$і ,!FYZeTRYiJB}CfG,Zﳢk-ڢڏjY|tmʏzD DGe:|L)y Rk|r`@DSez%1"B3jH#hYm\DPEtA!؉ӰBntiŴص{i.VlQ6+gJzD;?[Sscuzvn,+34:)w&\MTx,c/gT2(b@!D{@N ;dM1>|ƹ*՚ٞ5Qޙ)%x4SUR bP˧ %"Z\Y\d P N |{hB}8- ~~*ǽAyt;/֧ոOӎ>OyfI?fFOwVi8UQ qz7c++'w. 4W Ow_ ƮGBcx2~~I*WL~\ f _z0 0\ G_n~}G1QD]|fE󠬀w7ѓōCa7_L]?okCDgor3Asg+f'c _g\n`g+˵r®խSx\xf*M9n_-I#y39DP(7/WiJB7үYOb */CPl3GTT#*:yeœ8zQ3\eWcH"Ubz'U۩L<\uWf"Y6'fr3~̂'W\: ?%ҌH/KW2s]I4Sݬ2ڊV%Eh?8sԺr[Bx)l%QgRB.  1D"Csvh[t@u䈴"z$)%J} \Բ`Q HQce}n%'{F=o_3m&QP@Wi9z>XP:1W7d$wNT1@#ƀfg1&Уb% y-ä֖&N <\cMc9/ްqPr^kZYc_D /Q9I و޼IؤIJ8nU =Y -M"3v ;׳P5mpouʠҵF[5[mߜƿm߿A YxCKAJ'q•@(?GU4].YK~Y0A$ J3hXEg10afQڨ W!Q#v)/R4 :rB cE0V*Nzv4賊:-VN.ѾG MqHY )/ 9d%(ل^*Sv2`Ю$g|)$QFTu @DNDYMeӟ6-r \NpgBFǔ@2 DEu V&/GиsUhkM'6YjsɯO$с[`*Dfv Z[nh.$3 lFUqYX]qk'qr}ͫW; 2ql[k;tt"tĄ6`(Y~%2fȮ-x('~ %P2J$!Bf +:k_?Z/@SoxTk7?ʭw8,6e|Eb?KKƭĄ(WB4uj!ŸR.9e Ȅ_\obEY7H?σAM*}Sw@!od{TX2{ڴ}])|T Z JM K-8ՔvDK.uHO@5kӁr(u^6JR%Y_ޏ>,2;ݮsߖV %p) bTmAL&ʌdXGhk1/uUe5̖cxhG㟿rdxY PY3\rǭiŽ"oeRi6r:᭶&9ےLhR /E!H6X|9sF5|4l'=|W֖Sv YIqi/qTB >)aY2y YsY s.#DnMMBPƹIB|iP IQ$¿LIExb]\quF;vl}rGvXؖ̉12ڇ#M$w܆&z'zRQ̮:mhbm})"+%OeKc1xK&NYc*2E9@MQ?l7[')ѓh%J&,3QqnvWQ.b-®=i_`>gy sբrfr39&*|dz/A>N pWW_*:|SJT}@7zFu^dXQffq7⸸N$2Wd¦L+ş)~bb B_V{Ľ@5}.UQuߺ)VOюD]cCZ72sVo6]qS<[bhSNP[EPL۵a8wTu+vtr{! y⶗;D3Iǯu麙C512f?;ڵ)I/jq}(a &?u#ɿ"/{ >  3`fL6<=w-߷dJ"%Jb˦$$ث˫݂{{𾀛PweYqN0'q?܂v&c@Vguy,& L15m8`R^/K1K 9 jtrހ400wܖ AD]d$dQ&# W}qxx*?+`F5okHgz__?p}8}-y8m|D.!Ds\-Rj ; JSi<:j'x]Jn䉽7\^+s'"9z#*u :q*뱾0Rk1J46L 2$7I&٧x7&E)dG@@.d֚y{xBOp|N *&}_v_U:bG7ϾJi%UD* [inxF 6TJSvyOkQvaI.WWttEr@ I)7PP9*=]t\PnUuGcvkn ^Xppr n qt2K~iE2@~ 1ikހ_*9E_16o/.]$̻HwU\`Y&8`cP8>0!2͍XB\`J)PD~ht)ȏ_i;uD7{D@ -Z YsUҟ'-OwЎ63!r햻hEjTK9\@5W2!7*"9 aCLTsIذ hMNt;8 >$[~S^|6qK<~LG>zw+N'ed$k_zGUgieE4sFڼEt ;F f#`MM$r*V]C&T 2wH!/D![SU~hБzbno-v),\T˚7|Z!LB5'#X3¥XgDI}! Bx03,,2h쳕h3^ٱ3%6sd?McIZN:зCfS!PpV$"f sc)/lڌRF G lj ͜'"a@rp΄8.hue}ʼ^Povg~VJy\;,;.K1FۮN]  `I%x:2xi_ٷ!`BG}uyQ 7+9.cMs1o'!գ+SMѯ]Qw, ؽ.¦^k6CtE GAi! h7ZvGRuJMPo-ߦF$Åvn.!$RHn3DG"!ɒ$Y"\췭H>")&m{CyV I7X$Ŝts9IsXC .!D CƹX#C;ZÞ;YN~zRR.a Mu6{?E. >N+ǂQ;_nXh)8/XzO(0F ,D^6|a)YUG\wijfY0wD9^yUxu.n7 il瓗L$gíSb>Ε/J "*مIߟz #A$5ea]-q]mqwA/),V ͈S8AxIj_5%NtRP ;S )\_SBJ \h Di )C-Xe8\+ͤ`!fnl:zc.BCeU[QN(F*<АkH#H\c^UA&g*VZ?Uj4nZâEŢcUM?֗0/ժZYRkἼZ;.]+GvV뻃 8Z `'mԊ!(s cJ+gߎ@O*+L{6.+f]n C쫏gAtUr\[֒kor/SSq*oZ4ܴʾ3Hkk i)lsEyW*:B@hVH  z'(ea5x@)-zsʩZ:w8-t?umcTBk7LJTF#6@.`xm/qtiZE0kZOv*ƯVqVTy7"υU)[퀱UQVݕzK at/{l^ar{ѽaY[DA4.Co lsse6^Hx10]wƃæFg+llgxAX&vjh#_#~]ۃ+ ]+R-dL}FCNR%'p#L^XDЅL2x"酱RoU[LV0 ;/x)'Za^Hp?Vl*8 \%fxETx[8 z\x*kٓ> ٓٓ=iN \g2H@ lmV+::˲ҦT~3|RrQZ@bmZBBD>H5=B嵒XI׈OGC` 'k˟ 2}q6nʎiJK+=|ʆ20,#/aucƔѥ=0VkCŘ Z/r슸`.CQߋ*,#k|L~mQǧ×C*Y3Ç_eO~zeI[jܒt߂WE[ [mb;Inx:y4W_=Ypag.y'l i6{2=Jt]Sq;;{ηй,\%zCK$sQ||oW61|py㩙}x͞Aw?*`DWQG:BۣU{!^l CU};s~/ B㛎FL ~X"5'Excpd@^Go _)dqDS,LS9tAǍj}୻!+:5&k b&[#(^jd{ 2 D[Qw䮍)hU2(b7| Sd/P,p,XkQ qÈfr$-ZBzD#y<0\t!A Z q+$AƆ`3Y C  LPzjcZR]pNeుIcoO TG 3Hj#cA0&p8$&6 6pZJN5%`SzdN]쯈?2_o~:82|02W$`qi>Oy Oi*r֗<6,/܋pOTHDžfGbD8u0vkl+$B{=@$q=)S(A 8MQ+Ƙ:f--[ ,'Egiyڼ5+;4@l@^.u~=ivdv݀^@OA鬇l&8Y2h# !H NbBDz x.0y<&[Z<+>k38p$x! 52,N9iG *w,?MWE ЁY+چxnd& 8 ĤʍV3 4q{8X/@/BR #E~ 8j @,^c!C@@ ;|xD+4i2ހJg_e??G=jzKR82lQEy ㌧y wF,镰I$ߦVH?;őZ*0|lEশT,[[$gE+nRy;*q~ɗ|܇+7N ; Lr1Փrpe&rȱ[]n(52:S6Q)QxOUKŋIiSK_n6ͭ!ۢ!trP/yZ=-bvM'aۣu@̌!f˙QYdi,R$<eɯ瀍H/ \ܔ6E"V-&Geƭ9o5p- ZA- ўjKa4 Vܥݢ5F"H׀)D *B` È-%*d*!8bR!I9w3ɷ1X3@kzs$jQĠ\hmYf,3 63j|!e 8ҊPԖV,( 8ؘTۢ(j'BD6oSPs5lGڳ"h0'q?KuvP>!D`Z!fI\=`w@r~2|;xiЅH Cb \hhкhT~ЧHT?V0z^Д 4ntBP٪vQޱYG3Bdvl1$md@{ҡ\M3&5P7A@*l,xѵ5v?p>@{ ~ `qsTiCphdgj{ɑ_ 1_X&99 _چ5ג3{b%KZjK%y8xR7]dU=$1ö\`qBSw6 z cSww}i뒪9|Kn3]sWL# A!z)KǛc>ܘs!j4y1q\" Z"9?-%6hy2.,$'_a0. ц(Dbx'.59e:+o5W=~Vqz;ի2L(ER(|g&n[j0̵*N@s>Zt$Nq4ܫL5岕\'IO;Zt'ʿg-ݟ`gLLOӛ1 t4ߵ6ohn-s#SH4vSӔiMm̚? >ipDH,~]UV 6[7qɃ04~ `y.`Cvnh02dϑջslo]C?FwӨ]n0d( /|1ݗİ"Yqк-=Pbg3Nm^s7O/msZv-2м4XS-6gDO&{o[ԧ&C%g;QT/j%|)u=I=q88)7)ߋMXhP?)Y{1Q^[jsZ~?E!xb!ka@БQAJ8'A 'ă̈́*(cPC+hC)٣G cfHHѾ{;ԜRͩNn /ZFcP빖I"+"˰D cI@Uֹ(mԆaxrK?HAڜwԑh\`l(U4rC DG'nLm(v"ŵk@9oj:fkBgo@ 6A;'sGo=2ߡ۾l[jyVMׇ<"Wo4rsC).oCl\l9,QeH6=[U(ƃ"0&ǞR5n ~p zv%^y'hM+A97<Īē: )1gB2MTPk2Eȑh ĩ^Y̤4 ͡W:FtakXa=+֝z@~[&jJx]0pRfjC gD0@/~D0`@e <"s@I^k{8Ic̒Α`y޴y8t^ XO,+*tXIuP m3Ru1$VSj9+31v`:j-"{bIWQpttlttm&sbTfjNZӶp7&̝(YMOL ;pp|󒩺#RpG0ГsQbSdO壸lp^zeRFX{,aBy0x3%F+.8P%zSC~|\ W;Ha &$;VoWmTˆ\|^qZȓǥ( a'lr`Adpq2 q=-Q1'dֈs,x2OkCJ>/-G yxljk7x䀙`΄ )qI5&*SМ`eH8wv.vG|#fhPhK>RqkKUy1-po>b!;k9/r!NXщБ`<6;7"\F/_> .Xqjy郴ĕDxxDIT]~W/ug |㈥^9JN) Ukԙ謠$hVŇ$4.QhƂ3{o#:H MZ1C8 +QXbdr{8P9"הgjAp,$pÒyGp,8jT|1H`:eDo"ay-\ n"astZiE>qxii\3S)',hT_a*}D(1e!\q_f U&dĜaM%XkCqF&a"Sl?B]4ޑDJ(:%9M+έ1;DA"TTc9 {z% -sەc\KBiA3x7Y!2| l9"̅{_rrМ4bqHqvvnӄ '$1ApV3qnԤuZ RD`i&_ZxFb3>d I:'> 8iV-{=)W" !Kldhښ"qI{SD)Ĵ$m)\D+*}NwU[7-.ʶ-eRKI{?J2Avr8M{#՚!nC:^".zZmqO}uy}ڡě +p%͏0)o\_o`!)04ANM nh0?|sbq0XX:w=ekgcn4~u$c\hÎܰcl}3!܉7 N6=0p0/p7oCT<lw Y*sme]5CNXU)X)7N˫cJ^%Nv:*Z6ߺ _B޶:^9n2wqZ9Y-j6mi_x^'&9v$ϮA ɑtX>_ң֝HeɎ6W9ޛ'V"ܡ#wɏnsc%YP\AD¼R[PN.Pр艻{W,4ZgTku^Fe['(yk''.gg.u@Nֳ/G':(1?[.=~R\3[d f] |TS ?_o ?1^Dm&ip4K"5<m$'E$P9@Smt&=Zrw>sXYΒ겛$Zd7mb}6I*OME )++f8危HXB`lޏ4`l  x{hqWLJNiwU|R +:X |:u}WN}y=_&AvYhP0uFe'ܒڠk"TFVm`tI!ţQ(uL^1$Rq,MGTBI$-*ǸђM9PGB(_'=AN[.TDi/j۶X᫨LqD?OK4%[wS@3Ș-qU:q٢<3xVE;xHIkFt05_"$j &(vJ|$YDՐw;\T. bs&pCH#:[Ik^~KEU !Vqt ơ`5=qbp~NˮSO(S.]^tsgn.w-RY 9+lcmʒ $6Z ) dsV0n\+ \񭰿]()('yL=;U훿|h4`neV4" $ AHQ4 F%.A&yYIALXp!h ҁeem`5W$3(h  ZBBC:Hr,C`gFc93H@E$w B *4h8H18+H\1e+ULvHL{vQvY#Q;Wa}᪝ۃ]C'tAs.:#ϕ٤ϓoki]/edi݊TI f@-(OR/ifW1 IK rX޽ ϗ&HG"fY4,Aig*ڝ6},T<֍rr3bҬ#@6bS,֗3\4?J 3ΐw󷼲? _F/?YIcQJNΏ7o͏{CD8iȏEص{aюI)i6sْ9.Hh?^n,>:O{cN_#/?ezRo'|[Sa9Ksm@3ƇIP+I,qpÜ= c\_7]"n%jh,;Xkr|_$-<* o g7t: ~EM^5V\i `f!4>$Gc w*"[L"8,t͟!8_Uo}𧧎ڜa,<`YbY$m>x# Alt bdAHlyVࣩ;򮤷xZMg8~麳`wgkRv)۪<~cLI#6-FYf391D45 \f2;:r[) *7ou@ڼP$Z#S!Y2K ([ %l)INP`C-;L{~Nm<>X"܃+n1k̶JZ$i ١N"GAhaZD6Z@T؀8UV|/_^DB0cX<9rmթ9Epj3g.inU7lֺͷ&KmNX*fkZ2zj>?4?lX>qnh9cN8;QȕzvFZ%7^KRZ6u9?L~IYǭ&[?Yr߾;y@{iU6@'D{ݿIcmEХ;;6ǔ9_>IhL0,x%5Gr ("G̃=b#z](v@#w m2 ^~2R1ڍ`ZOKjWt4,ktipVSV*c@҂vCIftE`HN0-Byl_lRb8X>bg{G0_o@^9};Gc@h>ͰncY B`Ձl]- KbύGֵyo5I潵yom[潵yomjL&U &VjZMP j5A&'oB|kZUnjZMP j5A&*iBҲ\VjZMP j5A&VjZMP j5AtsNgu VH*{Uw%FMqY38-Hnh7kzQLU]ůnsu{:< qo7ЖC3 zȫۓsv0+FKĆ~K}п^ [MЈIrn)2}D^"k^~KEU !Vqt Ɓ Mg7rb۴;gwNRecmʒ $6Z ) dsV0n\+ \񭰿]{b5⢐x=my\|h4ʒKE@I0i J \L0.0ʙ1vpCtr7aY1X L"5JB,FCKuC:d<Xű7cr (PQ+]6e%rAB m$ Wz̲*UL;$j ;R *,F.v ]x. ΎX 'ݽU&V!փW-a^G,ҟ?h~xV\2ԙ`'A:`I̢*9vdNJ&dw"JHŸprPg3!Jm *\8uQo Rj4 v@\.$p<'ǒY7to J e:IbK,v~>žx"oQ~xzvs,"] OMH^J]SRp#d(mS*֑Nrq8G>ty`(Kw9*T +uJGG@fFG"&w_|+5 財g"5d]Veb$2ƭ0.(BĔ,(L2[ lmg$Vh*ִ"|ьZ!1VdZ9n7]*%8=Z .4Jr.цL @nZ0dXSEVHqu6Flʙૅ@N-QgP7;F+;cJ{Hu"H%9:^k"џc I"ƎS5^7(ljz5,{p":&E\Ci"Aèk)ch&b"@o:% 埥u%[v˨5H@PR0h3N md:0syK n6Lݕ?/Rw3nDó^ f p fz,h/7k剽,>s<-p{YU:բ+/mQY|r?Ͳa[e{|F3{tYE.OXRRlu姎iHjJME.6l\m>J޶+:!kom{K:+G;?;?wu*h;%)=c}}mӫ+DN0.TySZt^۲@]Ѓw5]Sň$ɂ#>ȨŲS*92G_E_[<>PT8Oy+%N*F?^G O7I]G|֤س\IWr\cp!ŢA q^xTWf_\{Y{"bXDBAd41saz羰Փ&hk O!j,玲6IO;ƜF$pБ@O1+C+%<M֪$4]*pT1g׻L|!:>M[t!,@E CKΙyXY-MFR@=&"T>*Ug *LZ'= c`yb^8k\., {+0քhG5 yv?cι.U?D4%Rv,HMNTlbɬH h]ҴX(C1 !8ke|exp}_ 01;Oav )3ueyxrMɔ̽V*Jq! ȲR4P Y XB̜j0LfQU~WU;YkSj.Q^8ELs$(H&HVx1JN2-bgAu@ }N0> VPَm7x߼z? 89I ?bbJ椨]pK.fGm~8AL[\ c۸bJiMXNsWͯo ?EVeyb sg (<`W>?v'΋;Yl%5xl}nrvkpHK6/m56e!*\Ipɚe>k&'@J3֛$(dl|U G\l8& hS)ȥYqm6'HJ]Rm:5Yxz?%V[㾖 [b.W"Kɘ'`&II{tF5Y# nܔc*?Me_=wO⽗{P21M։۱=f pJc֍7&btTsgE0^7\+'L\YkUsJx<*P9RR$SU3|&%LI,,rKz8@"]q&s#\*Ԑ`.}5*gM6x6}ANWGvWϻ˧pgÜ#Q{Dž3 >d֍bc;6<*k&Xv_l{AӠ!vdfQnWW29^cjĢwmׇݵ?|>ˑ׮-%;}OrߏaG W:|,}#c+\Cn_~(Kv,n~-vI֡YDzG1CFub2zMT5%q^6c C>d߷-1E߅;k:kۘ%dR*"b6g@Z" 059& рE|BIs# #pc[R̽k15]kbٻHP]{}ʼn-,MW/X Q4|k%zk7n %ݠAh E 窯6sջdב%#RA[c9W6ƅiBCV q!X`hh2hz)U1 y99:6MX>-dD| "N(sZi1k!KH+(Q6rT:I\.ۺqΑ+/_uz ioeUD%0: cr@ց iDBYR 8jٳK(ܛ4.3gѨW'z;-d\y3d[]A]5!b=(ELb1[ۋIĿ}^5HE_p+5Q2{YnsyJf9jPTYØ<*+R]Ez=P,SVr% t\RZ˥?K82yf R.b{E?2 _۵X?|bI'>o[8q[]/6ro%#mRDlki ;}Ț/. ?N]bw#co_5nL%wKz .gwybeF8|%HVv"{]@}uY7?7GTjD0yZGq=xbɾܤ[w3M9MYEcR.*`H~E =6Ju%{%W~T{5?DƮ}&IAh0 ]]nҤglZec˫3-dp)$%ז[ Jڍ θdb.npm񢝑H#,[%9)TT trz5H 9 /Db'x ׫{F"\EwEm]dvJlY?અ/v7~ 4S0";\\:3u- 5mRpB"[6M3R]@5҇g+eNF5 ~:_m+ zN֝HF9ќ/gPd}ԊϜQ`Ľ˙ wm$_iKiClƋd"3· &A2!)XW&բxȢDe@ȮWϚ#&Q!Xļ-h͢کrщh0<8Mq*E}c([Æx !)(YNav3>l4gKMZw<ݾg_}I'5Dq0x&]Ivv0 X0>KlhJPo wU> 9#mЭHNֆ:xG=wo~^$r㦃ynIͼ 0{B(j֒ךiTk ˆJ'`4Ce毪Gl_lw5 R@ 6F1SVY׽Y#_Z}֣J|ٴPl2jT,CUͪPai'P}x')KMq/ſX_M_W vYݣ_}.XCvfI-k |Za\?:b/Hfy2I]T&Ⱦtw]_޽KqnfL7}Hf.IWCm5M-!^6ϗ^hE/WM 1`gvXM hzmsגy@Ne8EȲ7oNovI 0*=k]sW"/5-v>w:s#tAe:mx!yMWeAߏyWtN&m]bya&V:d X#"@i6(҄Hƀ8I1LWӣ;SF$@20 ɝ=jj8 l*7դPÀ:hQ1f6\RR/@` f-_9>*'D$'illddLȅ˛ L&Fg3P"4@+&-ȈNZYjDs$r`T*"Z"`>6|vzMa j!9tq'(5S/48ш{ Tx̂HA* dۗB!Sj1'[VcHǠ`%hoVP+P+rb';i% X; >r52Xt@b@DGUGqR/ )tp5vh'F¸Fa"a xTĈ[_Hޜd<j[6zz*̗Dd>]qE]up`{h$f!шtTjc1 Dc j7>F!r[<Nu.ZAitSu: N 䙜MTrK;ZEѭ/jlSkKw#0 ZjA  s$z,7q=hdp6yz/ 12As:HW8oDv$r;YB1T)b+GY@St+5X@ .pzd}YH/$ c65|b2GH}J]'tIdwQ9tLHce^` YlQ\H:`9!][nu ';+/tаu7аVfr/ujyZSlku핤ګhَKIԆ1ߪw__Ovx2-8낂M!uD5hFu5;˜7.MxfZKjSQfxw=2ȑ1X=cg$t2 n\'KSjL}-`/C-"GuDZ(ATWecPjQ?ªk绽j]O'SB:DsrYq|ZKIL@1$]' ZLn;(v+<(Z0~'52dI'4k6}=pMM06p2D|8M8]Ū~j wr:v^r)s̷'=j6>]uUn]CG25_ན1y Wg񇪇#: .H#Q*=&蜖@GlO|L&5=&M.1 VwP-W ۬6Tc8ޏWaX͠]RCʹq>w"7r]$޴y;|?}nNcc =.K-͗v9l<ݑo!pIPFZ{K;^p^iߵqW0m; FÚ #`l:j_3Nuˏ%5)lSɧM!{Q)5'uP>Ce:* ؘ32#%aObYLfN^x2!CK [+t+_xЯ[fvՖ@S=#?F4d}L̊(V:B% (FA*(7h yQڬ4 1E  62(aTF刊`) NPLm89x-z|*X嵯Áw`S46XdY{AÃ,д6װ.aR!aMJnӢ&Jњdkk,,^S{t&8g"~H3o2(4b8># >j) d - 0h4+࢜cB9M(rnS Ȣ`;X+\$J!RHDcBنpJSZ{&M8o|#/>Spz%X@( ِ_>^ƀQJ11fwY3,HF,|{Y)O{pN% Fy9@G`x(J 0u e,($t}OBʶ䊙tSOe8y[9G1%atm^G } :?$*8ʝ7\ 3f4)m!x:FY^1Tk%i?@؍۔6s ¸!BKI}5M=(lpHT]ruIUK5Rww=oEx7'lpLiTMѧ%;3Lr7Ϧ0FUhN!KN8PM 9E&8%~Nj~0çi\Pj#/;h"*T:~׍rF,@Iժj. Rd8˂Kx@O?7T ڮwAZsq^gXPǓ-"ۿ كut: -%ޣ&\|@!R7ce/ȷWX!˔|vٻgɇ$+: EޏW_iT#bg[$ h\}7Nz^Uskf|֚σ7i'0[uT`,^IT8ͨ%eѝ#;?y#ot334㞾GCRXNSbq_,}/oQBŲ/Xbq_,}/btI;UgI;UgI;UNS%TI;UNS%TI;U,˽-KN%0%0ƕWJ`\ {PO܋._C'ep&q5`XZakP@3D($(/ N%ņ8i ABY hMJ)Qkt pCzGz(C0!B* aj3jX$\2b@"g4 21fNϗs^ gMuX T?Iϻݒ6.vm}p<,|``ңtڪ]Zףv꒶zrЦ#>I@h{שjzÀJkh?筷w9tfn_ͦChYC˺n[=5yi֣N7@s i7c϶w/lvf ]WW)뻅66OfH)^̥CVclsT_NE2WUMdn6 5j(;wg?9<#LP⽩FfƊj|ȝ} к{D,ybc)/Ra&awL S\S s``T8WtҘ?ZB&%DF8!R"[dELe̵ю3X6CJNi8:p XBGUrT [{GWb>% TǫkkػAIFz(fSc}^u7)%(Q60nVz8FdBHPj% ^ l$䬕p CL 5>iSɚmy1'񞴅JMDw/ ,B5.vR"HFT &.R QXRYﴠ 6 S {SKe-!~`]H":yT֖Gz{C4f4%D*@5 $( Xʄ*2P* >V-.zJ)휔) DQ0U8_Y⮘{X2l@7^ &ؒs, ^LeeJ!T)= ϙ@eg Ћk}T 3E͏e{:L3SԖ웣&*d :BCeq/m1+& Ҕ$B);ɄHL(\ RrC`ଶq. Pʹ M907`M/DCp%a6C2q1Yw͏Q/*&7[rCZQY#Ψd=n|Sy[捲3HSן[VPIxA`O?47Ed_lh?7ۼwu$ɞ:WTnQB9aWSru$+[֍\ssWqd⧖$Z@ޑ<30(:sP(7+KpȡC3㦾Ray؀dԎ$xx1zjQќ㷋VQX<C.SDӫuw\h.޴4{go+<Ƅ)l8ޡQ}Q!~t$7XFϫx_V0!'"LfShكa;5$mW_6r r[Bx)*jY'J0Τ $\dLb~tsgml21:4 DB}65d16@D ǭJ%A '"ioov{t=J;:e~7~>F lMڛ1nK#ϸ`XljT AD*ιYRM37ojCDZ~O+^]YZ5  94*`wItOZUȬӮWf{ :x+XhZ.RfR(T{R'PATN?Gh)}YZ_ì-6D+"ϖ*< sQڨ W!Q#lbrK>H+$䤎G cE0*'SޖN (0^RZ +#e-[+]'2kD/bOJc#r[@jTA( *F,GtCpV΃> zY= YV~tv-^y'RK( DZ ik"̚tfI\1Τ!9R m",3Y*+C%1*-*8rv *^i<8ŘWI8% ,QD0@e'HKr_m BO%I[Q_sw'MOJ8GMN3磇<{( J(I&(\UEV'B *5C0WFZc[TE ioN_wISoWVҴE* T \TWVPd2|}Qo 3(|9ov)0*q_u7ޏAmM~_⤷̠rc{P #OAgp(N>GE)YM1RuLNJQO#pI|deŕ9At?,g~T&77Y45sW/ =0 s? yu3U| fǃ@$SB2DVHu mm"LM/g4t;A~6SO? d'u5v#7Zhԟ.o% /馧b|JdЭ̍ #Ζ,M̧F=VC; I0zx۸޳{ Ά;,3 siۢXnuwGo1Uس#nŜShf,Gˎ-qf'NYUiEXg Jh'u: s[O(oF㤀o+n$1+89H\ґ'qJ8ֆ\lrlw\@КVOp~P H Mϫ^MT8Y qu_2 QJSg1O,{.5nԔ=# A lr;_[€w!{?gsRjz@ c 1?N'$y6n2\5QT>:>{2E e+У4H˳v2nKxrqI[>wِK'wy~F, :qּn+d׻#ʆV, $kIH,P9o\Gje?w AMszyg`֥/KѾӘ&9c84<\);4.=v풯mߙa|/_?W ԷIRVIWH"AE3շ o#\#ᎂ|.G-1Fv TThE19m˜s"T6F)ji (w[ %+F%S69k5<8ֵvoAi7hvzqj[j)nœ{{6kfP'f,I×L͚[;Ž5 wf%ePʾOB0OͲ;1`owW  oGq{}%2n+ܹ`z:?Ѳ==񒆺,޵#Ev( pa~ p7F@HGEHr2+IeʒXn6Yo'JV 9Zy%x}K3oY qMǂsP䫞[?Mx\&@78sUpҸ[5zk|dLӒ8rU "GĘg޹& !݀;0Wdo$3u3N0nZ uPa6._k:{u5ߋVNE3AU֋>=*UʏwMr{m /c:-Ӻ-7Bd\X&#U, ѓV. mMB͵+TmXm:KVida-8TʲPt ˰D0gg|yc͋?`uПL߀Į!Lh5( AHQ$&j&'Ƭd$p:T:_>ZU;`xK6Q JU"3LQ&C̄ERFMV1!f,58ec!s3 iH6\ # BIs# qFz!m98㴎ӎiÎр8^v":8i0H=˻\'.H%=e86tYpE֗+& g 2kihWimorWbs-5ʋs)TuBIˇ64 QkPs|:Rҍ7^E咱A# QwL9#2Ȩz&+ə)壍Œ$?܎c\^(^+*pryp,# ˑ[ܤioV#q"cG?AEMn}LvE =:<[`S_[˭!v A z4_׿ S._UO-n^^eٴ:I"\pK^9b]-%O&s|LI}H?[%$;r${d6SɸQ&2cjr?ICi8mfmKӷ_PoUE;_JWW;x:ZcԜ 4S뢁lh콇}xM?W+<-DDD ݑ|,;A\qY9>JMyqdz~jB+͹BʠX@&裑4%%%AЙwΪ:#8:BCk_bkh76sJ&wrP!ː0N$\ZhllUɩgAFGly&MuǼUwn<ޝӴ-0V\/w- ֫aviKŬqh ͐'Kfv,B2&Mщ`v2 )wtCbn0sUY'HHoR"dQsҷ掠Gc.1 7*P3#l5:E& ON'l b/S7SoRnWK׏i؛Ҹ0F(E̺HAzypK}өn{Z YLLϨor0۷GEQv֩,-gWMv̫z?0}D$!fr4cYJ'Xkwݻoǣ$$ފ߿_$mszB˅v1-lqʹ dqD|(G#?j"idfuFM)Ad* [4_\굘,L"yLݸ :gY:D0[9ByCH'̅MK#L"j4MR& 1' &(൑'ڀZ٣isˉ*S/L3߂]>rjgU[E)> ,{i$q"-R l:Fyjd~:ųtÕaYw*}|.qFh($WRd ghIDc|.B,wF 5eA CatciUgjr(wT { MƳ/А,֐( CQEeq4-sb40LqodfK=mB.iddҐҎA)gBtԙL ,R !W5?t7IS6[mؖxN6n~6kd2~/̷$UpVYGd$O7wMTqm B3jI"),h@2lC&Z+|"kЫHz}k3_o&_RԽGS0s&^YUڙT'2Jؾ~v˰Y.{$@KVxc /ljz7,vqrwN]Ng\Dʟ>Ėv<[f?Y4wI z߂Aor\Hia",@Z!emD$pZx[E#gl9o+_O Y= D_WVd?.gkpM;]}z_Zr\.B}ۃ[q˄,o21L>խVfEu|Yъ`>*4Ld -0rUv˄qtWkYF̼ѧ/SZ˗8sgv1sͳ8t _]:xaDKW{P;E>6'fs?<>AX mkȸAS&)M;)wNRn-IM|[4H $4f[gc;ZɑTL ?y szĽcFsu<: N1{܎_,YgyJq/Wl- 48yh Ǽ+TdF&v^ǁF u@r(rh6:i&my }#ڟjoaGE'Q{5PAc.t(넾*oO<.r<лK]ȬS%2A#K|2N[y+رϑN(ev__l|&m>4Of &JgULn6 ^nyRLF$SkM@2*HjsL˻h~ufF0c2g9e̙;GLxG>e8H#.I8^KdLZx,cHb!CզUw hЏvjNU?7\Jv6Xm@oʶamB23& ƝY,qrJ4nhTV(cO P]%IJ4&h^Dh7I]̰74 0vik,K J$.T)9 Ȓ*&3__4΢٣GMgBYS)bSU5 -W+Ť+gꨠ2&?u )~KB傺1 BбTAfT 4,ƤJM.i,!l+c ,wD9lUiroZ Z|>/MI4 #$DHZiG4JT-1`9"5 C4؄tc-dkO{-֡ܐ ӁGJ1|Iޚ휠 NWԊM_n^g4oF?iw˟O>^\]e1\o2F{br?I4e| #;po ,w'ûON}QYlk »cG.;q1_l3ʟ?"ֻ9UeV>rFUO~r?&k~3|Z.r4Mբs΀hߒ1FV{`%"78ߏ<!೓o$H&JfcۜOWVlG/i^cϔJΈ& QM-BXuQ9JK+vZUلF@cD==WV|-o (M ̊el$L|VoW衷hFBZ\(!:c1(*:ݚE$LJ zzxz.?uHfuut^-;x +[wg}pyɿ''aV۟~yf'W~G2MD^E,DWk 8[˩+\mF$tOϾ{`rHq/H1gL|֫Wh,"SNP\[KSt%J>5 :-蹒1QlqK_KK]0PZ_Ah&D_ C)PM4qd"yKh5%â B 4i>i+&*WƸ\PrtEFنKN@e\ſ5WS5̥%ԯNn`~7暍Ŗl.})D̠,cO֏Zç3)ϖ`erߨߊ/wKz_ )Q' C1H2RpP0bdf|FmԶVA7@E/\Ot2nH {Ԩ+rjo~;XW|V8##&ڃ#>ŷ"_^BC[}9VyYIUf20_R=ˋSyǕʻW'g/X rwFW:C#At!.g v98CыӋ)Ӌ~7I˙:x{K5bU@/;t(9,{yպ?ky}/e_x#>g0<pݪfK&֛Td`~ps>/UanٲYofޘ䛋@M~ڃ&sJƀVL"|lCoz DfMzOׅnְuՒߦ+ɫjn5s=?r]z9\Y3Y1gA;&9ێq]W>(ldgKޮ:7q8Uk_QdES )|w)- 3Zux˰Ŗ;pwz# xJpNoNsYը-j[Lnʻz^X,ߢ3noݲ3*r}#ɌE_͎g.s噡żp88rD!]7j4X.E~0a2`٭04E4L6$|Py ltO{Bz.?N14^>MڍTN>5U];FFEq0n6OKj]qt;v#{h7b(8C# hz/3]i\)EÝWy,z-9]`KwK|6Ϸq:'`YBekaF٧uyqvZ>o8c/k⸡Ԅ@P9SVLct! >c"@"dVOԼ#9jjKS6(hQٱ[2kt4E"xJjU 8y1.ʙ:*(#1LO-m6oI\P7&tB]X:1ȌŘTs䒦rb˶2rGV,֫`2ؔtLcp% BJ>iUbbmͱ4jΤY&ck"?_~C[= !4 :sx˫^(ߍ?;FtN}QYrblUn.QSWaDŽ{oO ?%TCn`٠Z8x[2hj/D$9䑧 /|vʍG<6q٠}h*ջO.GLǣFmЍ}[ϔJΈ& QM-BXuQ9C+vZշلF@cD Ⱥ=$ZD1+GaمTzi9~ 0u?wyfn9pδ(~nzjkm+W`?/7b] ] ѵ"y%yίŲcٲud$%"g deXal0V> co%x0xU2˝[1Y:BK|0j 9ry=YUt*[[̵AfD A[`M(X?)>8"+ɉI&8$Z,8Y#sL;2)o7zk\{RBxYoo,\΋Ƕ_ށb~*LOUeaկ=Փ&iѷnQxh6^{m2z8L=)?+#zOѻg囨'ؠz:bH$2Ȩ)&s&l:e%W" ǡҊ _h n3h ל>gSPjBla}- !\MJ8i F)s;J&lRDwo\_ƥg?.u?͏"ŋGr< ߸!RƏRs4欻ĮIKJ#`X>w|=5qB>mv\$?\W$W}UǿQ<1\ iFoF%X/8kƟKy'Hֻ/Jb>?у5g%=,g@NOWn$Z~Y~egg}Y 0V?*/q20$wDG8ܽvc3.8d'FKe<)~ 錶O'8tV3@#G3 Ldp5=ANh W_cH|'EOD_nZtu=;{Pgi-Z-|qoOMkӽ_~klAy\>[!M~$d_h?(5ЇKvAQgcc;]}(&G0o:iI䡵LA j΄ЩTSHwI!u/TtgB%W(33(>VjUA >hY}μ2@J6q2>y\VIܒ+|S b3MO/}h*.Vm C%fEBf*` Fdq$gyHm:2rajp_r^D`Srɑ+2A*{m)O֣9DT$SF G* l'2CpdLvř˝ 'iH\o2fnBlrem @ ~U90JaSQyc[h:\gm]/g]a.GxoA>8iZ֥O.^ \0f>jRnu;op8GG:]y <4eK-ۅnCo?y nZˇ塔P݇@o漍̏Gyҕ-:E;U0Y)+\Nt n# ҁ/Q{nk[+=>]F?6y%i.UDkVhz$\vu9])ge.cQ\ޜLzvkBL򴼼Q۴_z}VvwG/jxϚ7BkoQ'6<L^D’ ɧYSL|$!Ր/zl L sEx3BD`ƉrkRL_J/?%"HoJZ \1ƞC qIn˝ %䙗{xEgcȭh:yuLܐl=xKh23Bd\hF۫erԇ֦, Ӄ]c97XJJϹHm{ar[ؗd<mp֡D072>ޠp,5{}B] 4$ AHQ,&j.L.Ap:G,ۇOv'j&Cò`rp )J4L YQب %`{|HYrӢ *|Yu1g9cAkT p \)-F͍D#qFz!6>tlÁ@մiUiZ=fWPVDM9P:bʪTY*ReUJURWTVZ?*UV*U*UVʪTY*ReUJU*UVʪT\eUJU*UVʪT J0NH^YjʪTY*ReUJU*UVʪTY^%*Run*ReUJU*UVYYjʪTY*ReUJU*UVʪTY*ReUJ TeU.{eUU(UʢTY*ReQ,JE(˕E(UʢTmaeQ::m ¡EG^aD1FvS>0 x|62K|$b V7xՖ=[mYvE2TLQU )xLDYWMs[B5ripճS&^90 hܛ 2 r``{,Y)ci \ٕcR(@jP2rg=m8[~@Pc^3BDL:@r%E&pR1'P8"k#Y*Ч VaRiTN2&Q!s"DPpz#cL6ZtC@[e*VԊatXG.] LY䘷Bx9d,phChB}%qT-o V HB\z1IAa:glH*b)eI45"cDVJlZɽMD||7Df[ECkWF=䳈u  {g Y㒌 uaI]գ?Xכ(Fp{wI8Nw6ǑG-EJ՜|Э[^Ez܅9 ?f>';,E6v8>?/ӭI𱽜ˇgfd8&C ʹTЛ/XXC>)@i#*fmHLI:s [&#gzu^R|Xޢ4Nes4//.O7{|fӋw f˪T9JᦐIĔ`frJ.gZ{8_)b#quA'0SeJ!>N 3N(W*iN)=ZL+nW_< IM՟Τ;78.r,lª|lcѪ].5kXMZYn12@V675i#Z,$,Ox>u9^piV&#@'E_[>;)EUqxhdo .O1 G2h%Gv>4aJ"J6 BJPz.lyƄOdqywZܜwD# a窓>S ʧ[Y2R 5 RڡEm}_-.>n6:caIJ$NөRھ&y ͫǐAy YZVEVnvխ aDmh|aw}o[mn|@(? Dni(nWZ]pK\fX[w_iݬvㆃj3gǠh|樏xbdȡ&0-: Ψhۄ0/ v}gI6uZ%'I3?%`%]IJ~8hu: G;&z{Pd5Wì|vKM~|7q=O(*]쨩GTF"*m?0k28HyOKnMvfNID/LmH?~*ǖ =7o]Uer4NdNg I"6*O"_ۿ|\{G.yS:ODϖ㾃5ɩd"j4-at4ժ.[Fڨ%tfUNд1{!h-Rځ+KeWUV$SjUEu0w2 J!0wZkbB㦴ַZ B2V1K:ׂɈ`M*x3hrՑGtH1+!d٩S^{&@_4IUxPIɜ; F L.\&QeuK>cHE`{ d p`҈"#tv{}J{"Tj `-iC]G68U& 'u$jɭZ@$5T^Ue^i C.k].w؊PD4cҀ'7(3`Q'] +U 2PhQLj0R܌"2!0xPuJc-0'`ҕȪTy19Mk.,8R7kҪTg $uBͺyha3\Bv Z:-Q] v%eD/l z;k* Y[0m@IVTFZ6Z4`ԋ@+Bbx)Q32`p('Stm!X7ϦrUa4D2_%T,<Ҭ .#!:.P,=uQՒ-0 Uyۈ'3vEC@L ##B 0:N W[h~+nݒi匘& Q΀̌v)lI}'m^љ13ꘜ9EK⍖jؙMr@*_DTkt@07 SJY-OޅOYd=[|]*@eati3A1KYIje?N9RRGX\gcW/9gm^.ɵxRi0Ju@bX: u@bX: u@bX: u@bX: u@bX: u@bXꀢBAT [ w@{ꃧ=ya_zC0W-*z؅ KK{M-"6{ݼnܮz@Lg%yS! 'UQgگ-M?\↘_[=oǗ2s9ob; O#;0У-;l׈oFsσnN7K. Zh U_JP)zej}= 9+/l=J벌x*gy~w W[}Xi߉g&.uw(U?tƿ/puq~J>-׷[WK6}`ADe[?z ~>3٢vSO~F楥R֓U3\Mi xij옅~ϊ}ph}lS@eiz$0׸W#Bް[@2!l,熱z)?|~r3Xs"!L3+3=Y޵,#_Qx}aYt,:bnozxMjQlOO)HŒvU,fU!y2"#lypE!4GZR2)ER 9=ެ?X1/󿈇c1JccF܍'crsZщ%Hfw>K@9L49}79}79}79}79}79}79}79}79}79}79}79}79}79}79}79}79}79}79}79}79}79}79}ZN){VbOT5MZj"Yeggk;Y":9V߰$p8$pGzn1V/VY`̦ ȹBqK+x¸W!`"4d-"VdӐVBɷBzD9CkTO#Ķ`9-iQiTdHb;LsmRgh1Cp =pK ʬʌR~>jR~[:9iqOZlY\tG8h-ú |0DC̒cCk0Đ9~(sȏ'DNZ֨7dK\v@8-I%%_c ~~Ly} ~.FWZlU80E7V>1*s  8)> ?klZT?;t|,?~ˆ,v#jж'qyns:˵o{=XeY.LdMҐ'DžF"36đORaE6Vj8tz{ñ7_=D{+<d(N3n*&4>F&y=@!uR1$he4%Prk%D'bR BQH1 ixާ`Dēu%Yf`F$ sM._\V3g?]C8yw}kqwGƯ%.`}׆<_ckXn%ix\x"{C`gLn k}+ZX~@&6kdw;ӗſqp/?Z'V)sʑK/+o,Hb2#AG2/* ɃI6n>)V]܏.M:] \oW_}R¸Zt[:ǡk[4 isLW! HK"8RL,2DYaKofwz4K}1/^)Le)\-j<~\!ș.8_W=NPUƇ}Ǥ?&_'_#lc}#n v-1{?`};Xl]ma ۹kLi:x?1s=fdx^nWCz·ʞ.m6'GLL'Y#p2'&/ξ)qHoOJ?ZXI;:Ӟrcao,p6'X8ʖ~]kLT?jlpDY$uFA)U#O/:a Rrt+uT3g^GC9rk+_F95H01WmL})=/neˡ.`t< ]B%(.tHGnK*" S$P2BB m#4r`sy.ATs;Ͼh6WJNNyef5BiaI$"U+Y5sί~7r@ïK,;Y2(Q jUZnxRI+=}h2ՕĹ˥s1 ਷{) 3sO-ݫ>F怫lQe☥o7/6zm&ʘر+d?y6%{w]Ex?>$L[bc'g^ҏeFCȘT Ι* YxZ"RqEmv_l(Wm|u>!|xU&jZ]x] dyr H](p{s#qzs(QmwmVk/a-gBx N!S2CXlft@,ZyMxYoHQ 6 VZ( X dF̝L*\lRHQ=YҷAf⼵DJWOJZ&xam\WO{#* rTIlYjnet0[^gZ\e r2+2j 4! hcḘ__BĩU*Dku=^F39C'CG0|8W vJ+)I)$w7 ɁQ8yD %I#|r\z@ЈRt*C+[Ѓ 5=azG@낤ԘkatN RR ƹoh} Lb.Woe}X+>^Tm}#  ]Hu51V`z\j<ьR"}sg?MRh;Q)^6Rl\;n1}c2^^xDfO>,_Z?\t{w%@?#堯Ц=9IMGxi=‹Sjv9.x x;[;Mk|ܙ߻(C[O}3|s+g_붻y65i~3HaFfKF21^ԛk9zI`&0%DG^oJoA;=";_xuhpzY.SE2-£d;nj~&HgϚ_/;|Ĭ՗3;VsWvOf`҇Fa}F=+ֆͣ@<X&3sXbBT8\lNEad=b E,4)Τ NKY=zY\K畚%0d$fP,Xm Sӏj vt>fnG!eFAI0KA:khqPR絈93dQAZQUM9)btt7Nc[BbR+8-$2f~0ftvbYz(] p>UVw,3-ZJÜ2殬bψ_&[9 K`i­DzHFB)2V׋XQ2^R,ꗣR&x/dQ6iR(Ar̂VRPqdȲG ˒,REN# Ykc!mȵpN=䑍zoi/RZz Wg@=6s#`mLboF"Zzb%F(qTUgj0}GYCNF |w#>#6e}P> << 56JSqlT%#IF"icϣku'4de2Ły(Q c2J,Buuo'+{fs]R5wg%y-2A۞fZ5>D_uw3B`s359$KC6qΔi| + ulN'ǘFkOBzI.ZB\r *jѩrX9=x+z~|;後ձ7_=ӘpQYY*[VgThvws=^O*PHT  Z"E $aWahDR BQH1 )A! Z8xN eF f4Q;餴9+12:|]*Dq 0]d.x@z U^z ȕh5sC8yw}kYkn0,ƅy"uD7^^ŋqyVK_s{hTbl.]ye+ӹ)x}cW7x^_l>,hfT5Ω9* Ή< G/. iEJíh+DdT, &*" RZQYϝDx K7OҺkE:jm_$H̀e) 6ϐKU{&WwI yjǧg**`dLjd[{O\c"J+Ldn-Jc23]eYvԣG}MTS*~.&.Α|i^6/uxzN&ԫODJ lԀ J7`;Xjqg4Yث댜^ةW,N>UUX>ӏƕX+[/sZfRsN@W~G YI%rlR ǽ'YgJzem!qe,ۉ.ͯAv[g&{El6`tN򻃉|:c6^fCsuG2O}e;u3ܺj[?=N͞{x6e-%CEϷwL{9獖CbwoN[!G5U!3qPNJ 2XPle>%DʨQ&pS N5*#nuH (E*$h1蔸U1F3 M֌_3>l.3NՅc]H.|R]u); n/QE +p4ςMEIףvf"/fJsѬ6T^r'yP#$"72Wc95N†sy.Z38Uk6+ZX) g]Z;+N"P5A"jᤎ l09$$'#0H$ 8zhG/ڳPׄHOT^_(UsFo?6&cB^E`Y-/}~o36U^COCo>]M2=.p\>#*9=C*E/\3O|c2! *¤3,Sd(v/LN(/;SX/^,xo͒w7ԦһgAU[<ج>meQVVY^)n*Ewg=ey;':c|NewfWkN'%̘ht 19K3ǼZ(7R[H^B)(%qh㹳jǘ2D*]e CB.dgӶ֦gT?{/F#A mZ|;6q`T=BeW$Z` #;-.I/ ϐOPF7L#{uAnFhFJ2m]+7 qlw$rɣ=KQ1W~r0rMnKVzD40bOm@1 g>dfhZ4MŸμ~@%/O7L=4$j@bm.4ХLŻ]=`1{ kvegȲt?\QY,Q Wy[e9.mOՃF ަ̼Tkt!9"PI}bm%}< \DePDPT+Ks׳sIW'NLWXP:c.0Ir=HH0A)ܞ 7cKm"=<h<߱9$ԉy3sj.ժ@D՝yk|r;WϝYi?{Wȑ 1/`UyD^ks;lv\2Y<ãL6%@ Y̌I وlY$jlRE>J*L+jK(eoEwޞb4uj%{[3(½|3 jj)+襴Z*VKYFiD"Ǵ}lM7>r*δUi/=fp6Ajt]/jۚ8+t _f͇GhP`i2{c`4nWslPԴp:-8)}v BC8QA@qJGhƍ'>hme?}SM@w@181 H貰=֘B>"mu2L'4OG C6 }T W r2zoZFoBVYZ&l W"D*< sQڨ W!Q#LR+@")<3`(:T"tob쇙هv +RaBɷ(ex1m*ռ)>"p h*(EAq?#ACpV΃>"/`[rw_vv-^y'ZK( DZ !NBJdBrMUUYS,I+ƙ4"GA9&a ɻ2KYP*Rȉ]ztu^);FE9$ *i6vaO "t2P*i@鬒A]M |z.p-?"z҄s44s>zfpS-,TjZ~6yRϓ w%^d$.XC<b5c 33 qPuZD&&Gѩwt♔ 6u'%"8)S8>p)˔.8ꛗ\-7"Ӯ_19X72Zi ++l2"0D"ga* 8⌵F !u!)ٛG9X}uӐCyE7KU3hQvLHMΓl%ZƲЮne|%ĻY*FTu@Dcg}6pG$JrgHNAʄy$ 7wq.v:k֔z.hCRK}wIt#2 Y݅]1qZT}YsHϤf_[Nl|Wkv~,ݲnO< o_]<ͱGGgBGTPz&dahߍqQWiۆĄ(WB4}O-xj 1Ax0Վ_\ ۘg?]ol[ڼ~o@e752߹|Y^j0?COϞQ<:9B5Rڃ8s+!ՙ ߣ,ǝ>6܎+(q*ۥ"I&T ZEp1(nLxk">u,qQ[!-2I, ΊX@0FH $E)%40FGιGNV<@bqwuwšOZ68{hЦOy܇i@L BJ81R)VEM\Q4K.Fi"AyR{CEB"fRP9XV[RQg:EKbLx-UtD1q[F xr{;K2V]}>%I dl5+zfy&fɥyҎvv1uWK騛eu}GݗQ1fʅa5rDtn8M^[՝͸@KF}rR e*M.qƤ༏x(;❋xEx'"^icLB\OZ!BBʨ$^0:%AP і&Yfk"gpPeqql>ቖ,;[g7ּK\+Sp̻}AXOT-W :k5HK f2Y)C2J YFܸ@xBN ^4 5,c$!\R`P IQ9_">h.˿ <]/NR%JYÃ(RWx3RɽXD1WZc>F$4BhSgYH?~h;fe ʞvef,U,x|"Reʭ^J`O4er($b\^,-t>}٪PIXwb D_ruwݻgT˥WW7a0"gwן էjazBdKT x!%JmTDܳ B`r/09B, c@[ ᯷wI 7ޕ#Eȧ] }h`{ ;`>LAY=%%' aeV -QTU73ˆLO8y)x:l |3Z7$O'6Zw_Rnog 4a5ٔ ojb20rV&352 >F(x&*ֱS  uʤ(Db z\rDS5ZɐǛv P1RP,mb ("Qi!b6`l$.Z%(;]:nd91/Ї/^Vso^K1GU{^*e ^K>P>k1}me'#N>ZeCF74=pK_rhH'j d+09A:nU&<{g9wW3s DZM A5BDp:"c$#͂*2c33)^l:Q&ʔ%dž*ZdǸ3rvtyS4E:2~=w(0j= !PI ,J]sՉ%l㤄NeE8L&"x~jFcVecIkKHt1feɑcY錜ݲrFBR5_JT+hw qQy\ܣ^ƎI͚n$%cBl;%`UtsMࡒedVDGe͍Ѝ~Mh:8zw+>WHr_ teqT#̭[0u;OkoSˡ= bkT1 A@c(> P4|Zp/zR+>$Q* ,ٹ\6C'jw0|,4\$:;\jkBب*֢/li#]1cI!T;/tQ{.;#gP S<۸ o&̙"|&̈́ۈvr,aeg ؚkR{%K;]''CMjܗK D lct. fN'Cg1>.DQ8cQEDSfdFJI CR'L. v"z d 鯨lMmxio  t(` >zJ3.s>b/ t56fL6fʶ%Xl e!YQ?WiuR6q,1+0z7 ,EhtӲ~7#C+z3'ToO+#Coovn<[H]37莉n3+r}ɚ/{~\C7ss]E^]ﬣ q5nC"Gex7UUS;Zѹ ] .e陷 ٔNhqsJ1))l jF|ɹ BEKŪRms-r(ŀ 5PXIu;#g[Q®8P]:օׅϪ 7xٛ{'4~g}xM/c3#8T^!"sr:y ,C:cXbA6TQiUզv>9db=ba E&Cץ"g0Ǣ;CXk^kv+N" ,`$FFJ׬ ] /!B$:ZNaCby-dFҰ4y5DbBr YG ,|[F}-~}Q[ /vjDӱF4F5ֻdE2ل\P$mKdI4iV $,j,Zh-yS%6y7T=S0I)Ė2^&ջj]7튜9F֋E1:;CX/^/zqkӻ\9|&zSNP:a0:NN䂉Ѓ^|x,38T;L> 0cy۲tWJ4G<Й,Z/yI h+1D΃ y2y+=}P1g;y$eؐc 2L% UBp9OMe)N!U9/ie^n5pܛnyTl\&Au2!V)]s\)iudɆeT,5y,?gݱ l,vVd|P$_|L |Y"2rV ` pw= 8k|MhHq}ºGb}F0g( ńI^rFgu/<\BUz_XF׺9,24ѳUE/D2[ RtEք-2ɉdeW)J@!N'$7{;|o+%O?rru'0>~7ˤq. ]fraXN26' 뎒ԫ@*I@u6%7"'h8mBwˇ_#,FchvDZʲgg[xz9Ǩ9o:KS^8atߋ[lm3㫃7l/yNW0fuy9bm[yY$t}SFOyS=&^-݊KģМ|^4p4ǐ@:dV"Fc)\ ٴ.Xi"&c^,Z/h\P8r*G,-ieSlr$ Y rp7cI'LXuL!aS|G_aiژi`5`-'NsY^{(k_m>] ;H8 FZ$6A jBy}٥tXѦFxW}9`kUP(HXkQ`AYK#ヱ^ʼn]$gSYf%b4AE|<{c5ƠdFZC9iUT;I&NWx^C3?m(a9,bYri؞K_tD}b_.u\xAdgTGP[&xӛY~tӛ?+ )5~t5}Z=;1S-vsMO#׈?ᗧQkR4*J?MY.& f?{Hl|0C_v`3@2Ȳ\Dz]TJeaR6.%*xLB {9r_g~2s~u腾(P$p!>C$ҧ.px,3Aod z0h2Cſ W_2Fߝ Y3wWәB_0o$g^rL( qH^ 8[OҚc"p`H?ЉquRЁwssW0i5y)#->uG+)pN}N\D$dSDK:) ǘˆsHk7FCH) OTIi%r5Kod=܋Л5(*y(gIA B,`^*#W(%j6(7ɭ&~?79+-l41($901so2dlVZ3JKML\=^!v#'n̒2LdXe1,U6%cI2rVl|m)Km"l% ɕKGFFe,L/4)?> z=}d/L2pBO1 S㍌9tCV}Vy\Axw'ԋJ, 2X@b@f%1o3s%#j* cʶ)*cTUbxTx!QH B[. =ˆbNJ3 6 boQK""#"D[/.yѣWE> ceFOZ'=䳈u bd3q(#r] x$jtTg Oogc7%]=('󄄖lrd;orBl軧).H%er缵SY -Y̚,22ihǠdpٔ3ha!dhAD:x!?EC]a!y#ڭ*RS^F%Voq-7{lig->|W`ܵƱ##aȥ:B@HrhieDaG"tZA 𪁤Q҇t *HpjN[+7%MS'VҴƔ6Hu$H \uK1yϗ~<2[: R }X> :^M 5 : k`t5]8=?}.~- cws{^̞X>48}n  ig._rXPwǁ Qk602, |h3N "y+tdܫ+:!HR}]~EYmo/n6|(>桾ڽ~q,-LQfI5ir3J9Xk]9,M@ϬIC/]eL ݆^Z/Zbe.k(w ާaj7|A2`6<`xMBّ ; X&z^~jְ֫)tQ2aEN -,DV5t\a%p-5-7&-o%oH]]q@RQI!8^Sg?[r̔$ c{E9&o:-n59^De9,<0I^ͫ/YbR<=X,vv S_߈e6g8︙pE/&f8aŧo @i -Pkdd *tQuA?yc[" ĸw11b՚jُ4+e\cQdIys ƇOIaeݔ=) ;mN_1yQ遤ѠqIe!JQ02++SLzRa#Ċ3ַ;q nrvQhbPHR%r`F )$cdȖ٬f@$0ɤȅwU:rVg)\ly"]D6Mu܁JLrʨБ!̴gux@MQP;#? 'd٨81E(CM7dGnUlPԊwOj~Bx]p, w1 蘷Bx♹XAZ1rBeG1Bj~L~r賫u[?ۛ\}~Z*Pg,A^ݾ냤C\0RPjQ[Ԓ}6`pJ`NթjP ~"Tk(]\rOOw3ց¦NpWHbS |%0  >vًٳˇd6yRO@rAT8}"H )Kn&%'A̅ƒH-2>3WcK|8#l~2좰vo_\>',v2fI5ir&ʭ;io<jLlOL.{$l=&kȂՄ❅10dⰃ}ײ5t\a-5-7&-o%6丯8L ]5mAfk{zTnÃ|,qoh0'څWeW-|pw:gۋl3&)^]Ry4ދT-5W,ib/ _47x5ԅJKƺQ4/܆拯g lQ |(5m'J O nsL-n%h]R,ch]d 2k`b݆S%.C v d.B JPD1(3:EkcFs$ernR Sb6K,${&{HBV{UUl26&c^v( pPHELcОDS"jp:fO\ci)'8cHnS,"?>{NC[% !G  Ƅ9|gꋛŞv6([R5!: ke.@Wqyޭ:-.78h;꽺eY;ߎڒ@9_ɓ3F٠fT<9CM@&SJdPT rPJV*6.PQwژE~ц\ 9Ԫf='s7>>9FA#AgR#h*}N & Q7sVc#?kǻ;y.t6(1'1p$P4QҙڨD$.cz2#<-u}s21CYcmuUWjDKR;PuNF焟{O,de2C%{td9g`@ДGBn!zMWw ೴X=ku't]ߝ}Bw)1ns6uC7d,鷜I#+?%Gwu n^ָl".яj1M*$D^~3$E*DQCvZmy=5=Uտ r@B6G'!Cx%P=ɤBѷlΠ5ܣTޢѓKTm6 BG#= U彑߉>9N&-bu?P(+eBvL-Vn#2~܃ iPǘdh(qQ&L@l&LHy&<:R )„aDAƄ3y &R6g͝PYPp[& % kAH IyS Sxӂa22 @{#gҭ^>&)K`>ufOo8Riօ{_)熱"Q}Xy?07!KS Ui͚MF,xG_oy 6VrT5靹ҩ=oܨ\tpMZD~M޷2֧Wy4 4(uL^1$Rq% hZerMSm[93h5zpTvcm_[@]t9wqB~(uU6i zY Bwrcng Hح^m:|{9i^%v?P^.kGZ91\ Gs烫qߊyH{E^ۂEE.}/h烟xGe(%OIN?]ӭ!Rv;D Џx9߳3~ގV\c/ O~jppdHo3Iʈ[2ۍ θd䒝\xg3|F'9-sOpŹ0tGg< 3ACd`nrvfEnƷ Vz[-SpsM|"1-><-}ۥ_-βggr|4+U _l4W ӷUaw-dolq]{I,[6KgFCDٓCm׺f;}/kIy^ʄ=kب]s/۔G9~X4'h.(L:>DN@32xp9rQ!TbrD LPsEKӴM#or`kNYo5Jc I+ŽHP$atI2܋Zj{ZI룥־<޼x-Q ^*t`ALn`2d؅W_%_yi* LB^/{(|3PU7`g]3g= fv:Vq2VOOčkV^N'H/O?^ށ`kGӏ>a(2cEL7t:ՈDg*[tԓ߬$EB'W:؀Yi|6>$ɗ(|0YdymKd҉qE}} wo*l@s]w{+C"!d~#g.x# qlt\bdAؖg>*d`mZ<๪omu鎱;71K`M1M0VMH:ɉ>?*p5R(L|MQ>(j|V鼛>R7Kk9Ldsȸ/qJo18 Ib&9M)Ob {EQhztYk b[}Hi _iO3뼏Zi8(nL CF&l@Nr\܋Afl:, G:4\ׂ'y"jY^=q2,礠 )/"^qyIa4K 9;us(?O-aaV/nP&W98À-% ``n;feCO.a)]ۙ{I"mWs 2;;;!2Omo*JW7tuꦻi:dz]]@0?teCW6 n}#\F^~~(߷<{;q-&voWpi_ ebks`,~j(i>Vm%bw; ͋TDzL._߾l(|j8/zM SFX,(LFpEGQIׯ\Gg!K(q!h3 g`ZplD3Ju!G?=b Up`OGwSHưinq9])z=kZ3LHaAd?n+EjZlD~:h)BOo.I .H!9^J mDRLѥ|ȚzyDHn@HMRqt ƞEIOjd?{=?7x9*=ȵe`dqob@h蕗;@f Bf DEWh,rt0֦, ՃU]d&hk3)5co<G)/ }uY  7Wř%ܾ;v=wO~wAGozfm )C D1 i"x|ե1]$UcҊ> &CR) DQ KceP3`{ؽqZ3.Ekjmٳ֖UkW{/N ^V0Bb8OxM3$'W\R rs0k>le\QGc}5Y#FNhTRTId*G '(I1/M Y2cmUPEⴒp\&P"=RVAA,i@z*F}iFԁl+vL޸d_{֋ŪxdS&TV2EmL/0C^CՋЋǢ{}YTX6':z_T\(}R"JZKL.[.*:ĵY}P/ 1^&x.ɂ}Jhu-5fٚ9[3gkl͜5sfٚ9[3gE 55}ִϚY>kgMi5}PY>kgMi5}ִϚYvE5}ִϪFi5}VXbMi5}ִϚCZPJd"(8=oCi!7d(DxxyA;Fp/#(EվS[c or6,r"ɜ U5)&s!(WyF*ƅg& i -\8"ƺHzh VK}1qM(P.sZ(Ɔj,Ot#)vϲ1Eg5.dRQ$/sFcJ n,1ґmJ<v#k<9ZkN8ȹ[6]4teeI.2GVrUN|$24R P<(GAce(b ]B|:n.*Ctkm!c b*DL)͂Bd^[ lmea8VnikȨc%Z SV2˽YUFd2f-OUAWqiBZH8FAe >c>p=Øl1 wv =o5ς'pp>׬qYqw G /s6)a-Ͱl?|Xti16Lqj_fݣ@-P};&a x7I,>z_MD&cɷU^kw>hp=V]!vp5έ#@ŧX/8L~#S8Eރ/||W.* hv#~yF=gI+?O|? a3 ؽrwAPd7dgWl=,k$ٖ(tN7YUU0,3<%:˫&[C&D il@2JqV! ZyvǐL}Xgٛ;7ܑWnLp%o؛OHe|I7)-so7sΰtGgJn J*Y6k7Kr>y8gEOQ[sa9q Hb-><-}Rюh,{vIG3<G~Մeu@{6@?d7lڃY "%& f`FG &BSgOntp|4AQeS{ 3o۳_#*ډKē?.?8 %L:>DN@32x9rQ!뀪u1̒L)0Gsm-NJϷ.FOF,i)e >\')tĀZ)}@@ C3*I^_u%OLV#b~Aw|kkK D`Bz]c~cOÇScYm鹇rIۓY@[4&) [FaG'//KBk3~r+BAT R#,MI.&3'S)^2)Vm].yvGlp|v@YrMCVҹM:z?``rD:sGk뀐cfګ߶K`vm$Zx4)3zwmṆUokf]!7-YcKRƗ7ǃJ_nrhs3{01ELp&,ƾJ܆} z訃߬~rB@vu!eX@& D26@0KWd^ժʦ̐`+'J:{xq A ! ~Nƕ+%N _]b<k]Mlnketueо)c7=xWzn?ӉiW>-ow_+{gy{l,֟iҒ WiLJYh Y& Q7H+Y;,>iqL^@TdI$s  !$yi1sn 6gR hQE" HbWmD"F^rInCf\䀘^J,m IDjW)$k8جc.kcHmR6ǁeVHa<ȉ;UzIY&3ibY;ET3&$`rVGYa!;Nw%& mv+ɵ[e5FO W-I:z)7z =9d:!9X#rZh-; =ԪtI/,$)>0k*O.-.90#(tv⨢8:%V'^Iv$ | ;9 1'(UÎ /%[46"{ `u.]iF 1l ,")B,,G Ӆzc\CʝWQדrB4i4z %dPr\.ʨh m?Ms^RIxpQVhQZ2'/>oeiEF&--1h<.嬌p*d蔈&x!;EK@Bh1/Ln]!;MZY}'܏kf{ͼs )M_YJӖT/S~ͨY.(stHb`.5+:4,89]w;Ku.KC2VTE_GNF}#M>X\ t:`r0ώNsG 3h]7L($JXJYr0AhTd.D'wLd 3_XGyv R,y]hBx5]۽z49]x hCy8ʶc5|BʡdyF,=4p)^Zib 8!i khUÏx1rv,D9 F:qc4AEi,i#!@lI>(eNo&ǵbm[hmkMNliy/#.Jw֟?8iGk!ri*L?]}w6|h >Nxw,I{pA 1`{'N]0G 5W/ŗf Nw/*lBE3O:Rx\NӮhd32+Ͳ8d$F2PY`0xr9k*z T>ȝN

"o2hA}3V bzRx}ī ׏c~gZ=1H7y:q;傓߷'6eۓwtW:ӧ8CK$u)~9~9ZK")*KJ4%h)3Le)@ RvGϿWNf9mr:(`H*A:IBCcg:G^~it{nk:j~:ϼ^:-U{^G= ~~`v=FLjo@c+C/O %WzmkQ;[Wj$*5I(} ,4^EtǵIElR7l&gQ12$w wN(Ehi#cxȠ1nȲGPEfgAy^:eL9y,1Hj!zfr5r6zBKvsu9<݌|exTKY3\vxai:M8P̬Mqoy,l9d@;P!̄(Z)q!JKdp>\R'Ih= xN,ydu?C^R#) I$ ʄvh]PR5 }$C 6 2&Mz ҁfB!EE:LIDQ.Cܦ lLXHHl^$(xuʽefr;uIY%[:ZIrJ9F 5r38Q'ثz\Vٯ w/^fB0/WGku~5-;}h⋲UÑ\V49(` ULpFwxs@SGܵWCJl}%pi3,Yc0{k &@I͓dd20`tݪTRk.Ȳ&+u(|VdR: y9_ae2Fpj] u)d&[_iI:A{f][/Ԭk"C>{QC&b #5F-^"4\nm PH1ݫ5x\։cAxilD9@ 22RFJ/I\2#<( $ ] *Q dҔ[5>@mw $ >zBSx{. ~%f^ɶv!]\5 UF̛I6>{l޵-EWrٖvo6⿐,a.zw0PE.H ;CW*e%}RJ"ةw - I-eSBE 1䆡 N,jTHHhZk8g@-T/2着>Lg'h"]ЭZdǗ3anY$no9lS㊆[fRrNNL&ՒmhL9lmTi,U5BT$BTZ14v8k8؛կ^'ʖbQ+"K06؃m^+~R{W/MnhB>˗y@z| (-Ig8p /%|F)?a>Wsm EVr)zHwXE(D@VY@jT)^jk|pFOw9twu~%T6VYr󸮎u]w?'Hvtt"ol]_ܛ~uwYʶ8jT?9OR-F?GЎ#d^`2kOWY\|C-bӭcxQ/E)i~zz3NeP/H싙 Nٖ"_ŷysg<R5ɁOCw[ K R}Ɛ)V\(sT1$:cp>$N%bJ"5s>2;\w`*y8m蜚ܐ+/'n-(8u=8c/ً T\IR$Mg4]C LAEtSEsiIY? ^S%Z@5%"KAC-hѦ Zl5d伋XȵⰋɸ.wJ&7`=&Ξzѫ?߁] 5ٱQ̙KΕЋl3xk9I4dZNiܢ#U/ veR`BZOgK7=*Nr@aP.q' sf^UrWiK6@r57P""b;Np 1u5lOEкNؚad ˈ4D{pE(߰-Ds]d\ے(PjM4!bOJ1l8 )/4t1R\cz\}5Cvv؏'>22WL%smD/G*HXM^$L烻2|_WcXڔ q0aRo@G\[ꉟSO_nX=ʳl|Wzd1}\r,ܧ"Č5,6ApE(6/GԹ;I v< f%u'p[%/EA7FɁmHa`h7L=%OևS/Liu[K}/?5z8u\xt :z%t0t$'yX5@R]LI-ο;V檛4B :O$TWDmp"o`hliگ>{mUcAٵ7pNW@WT٢oh|r873>ӞO8*e^'ɾ 7vR<^ DDh֘ D,|4ehghx*>ZoWԩ/ 8,)j  l8LR Y6d%B.NSRbL˃pXn՜X}'||b|պݥ"߾l1flQ߻LIa815!@I ISw!?c)ƚG5 jbb^_kb>a@ʘ#4g c'DLAAhK(l=56AS#6#V:K`$HRм6@]#TFO[& <8? 327[zzMFԗ=z>BF֍6]ݷKGqR߷~"u-3Zy6b:8Ao+JtO*:J~_F5ӏɝ2qaֺ,¥Y]CK9nl~=o9ܙ^m@ ,RY.-r75^c/[_l\oޢSヮn{g,G#ׂ///ܕ+m'?vS$h3Q#Qb7XiV`05sUoxHgiJ}TL qxםt>h0T}x'B &2z|7U!or$Y>`RCE#J}i\ Ƒ>dءbZyEDc$R)3d'_4[p!'j- z6T&PXo_|4o;kD:nlQ:7r9"V 36'ۜi}fDk֮x)Y:; 5F.\A%e$f%Ξ; GS."puUcY4/T ( Ԥd깐2d`S?+ ^1^X &JҠctg$K9`jD-*Ԃq*м|^m6ƴ*I >E$t"V5v*Ε(K#5K™،t{-=1?Sֵ̻?Lxܑ Zҁ/ʡ\:8X|3u]\ G]6عsu)c\8> G0?y$HTdžp XH %! ĺ;la S38oҽtp˓xC+SaـsXMpHlrh[sX41꒜ϤAU -Ll3"易~mtA0qZ9=qSyyH8gWr\"׉\pye!=riF_!b^h^ĩ ZF} 9)O4Hy_ *?y .IEx5b4HT3\RzZ_D$U(8nY9gd;*\ztAh-SbČHV_+vul5JˆL[v-)GS[J1[Ob C3Xը5ВXq΀BS[2_!NNIWB=%>m/rICTr&1VL+n#j )^ 4D-flCcdʱfg]ꁰPH\r\(ѓ*U+F7 a g=~vv޶U]{LJzLrG\DYd123عf̺͌B72.eK,3 &aǮo?_W2^R1 ){*/e3^~v/?E+9"Jvre_9zC4dL3dl-Rx97rY,bA`Tl(XPc|,@H@1gDg.:_]ű76û\ҙ/F]D/u<]Qu0=1Kx:ut.-B}^)ݻzi@hB>˗y 0#6A1yyj:rǩf)O{Uzm)3܇5\ܼ@{JK?/RPH1A м dc5&o/nCn4o]Zף! BNecu/ G?\uS} oGGZ/nŽj fm;qL~.ӟZtn;聫0n' 鎊WY\|ُO [8*|\^ 4_+A=15z?N - #h;c4TzIr rc?UEDfv4o~Ĥ8D%"Q)ֱuPr"BoQ湑q|a91&K8_CHMD ٻ޶r$W<`," 4ۘ"%%oK,ɖe)>Eůů$d#o_ Ɂl B‚8+Gۺ9f! bbZt DƎjg^r^fX o&ͦDgzo 8*z2dJnOvByFY!1[zF߿*LZݹ1LML@tZTegY@pQб[c,JF(lzr睔*m9RSN}a#յeLGwU]iƾPvl eo _sE-lÀJg׳_4hp8sxus[l HHy$ci2Tg]߮LuWURa<ە~kcE-:3*2&+D B=dɨ0krpP(t얶YkibT"ڙBM6@*ٓH` n]Ow%nS.Κ:XggZ]E.n-Dje9cϔF\քU%;(WK*!@^So_.=L;c{h/`V7ܽv * U`Zb;G:@)F麍^rkkW=LUݭ~p|yaZ? %Q&sF(* F/\Q۔TuTf¦gžٶWvErd$ҪD$+ #ucxEdϒet]fwqQ]~/(/P^_w,& zov롍B%0}$h~ѪY?ѩu0-:Q= ї<lbif=KY?F&>^~[c~dXO. :5uןo=$Z} ]`'m56qXJj衇$ bV{3S!Al".(*, ٜOgy{#ێTb|4EA6?,άM KUډU]D/q~ݡIF1I*Tկ\.t-OQϣb/n)𐇽~Vl3|f| QϯTpuí.E7;_y6>S7@>>;u"+Ih^ͩ7BOF/=`+ C6* 0-Z]=ߎ5:rT0^褑(Jq&YISHL.g ѵϝDgyyr,aĞaOҺ ?ɾ1Oڞuo5A_χz|(cDbĿw&W/o2#-!4V &(BN=)s~oYӽr#nu|R3%rAч B &&Ii >B&U*ZȘr Ρ\"P+ތEh@cNQ1LL0mt|\{ HhVJYpek|Sf35tM !X!1oa-e1F=H _,%zC^HO|u U2Q`H  %sT5iR>Wa J1uLBp5 K"hxocJ$ AEcE#%/ @g`L eEJ=z;GTHE^F R caI%T:ep9x-!Y5,%t_zGThj<5Vm?.]/wwy׍[Ef{5_kT`U[7ܺm,Zohuv-n"||w=42,oܮxKwߌ1: w'?-L> |긷F%Cᯚ>&Cyl9yH~Sfk=7%˰o }AWڼ]eK{,/] ~5vs{5/jo/? iu FEUp`o>@tX!N氺 TRauRoJ&y~5\\n}0_nzR3~Pcyz<y0UKKbX"j\sn]xفm>_iڠAĈ|:wϛ0zq5x`v0Z$XPobqzPnm6? f&k_-%Y.twjW?~KZɀ(\tpM@>kt?]-z]" 1,T"546Zuv~$K~_vxf1Z8N \> '~[U=VJ{4?G9L;ѫ | uuՓ5Ҽzh\'+כФLAaO.-޹g?s?sOw6{.E+fA,Uax2K]tĸ}q_FN3oFg NQp@#K$8U}/S冓YvxʵDH%&Zǰ&&$/JSEy^?^6SQ˵<ZLC cV`EBgDdFwިX 8NBzH۬4'4GQ"Lt>ON &`3BC Ff!' v&5d{#-ט%bA5WW>V P<2t&Zc)0~Xy2`tV@[k凃A}uĬ' :<9<YUZJlxA ŨVHj2f}bCN,iљFZ"45b(c{E{8l1N6 -K|IiL0"xT#O1Fw[]i R)ružVA0AQ\-=Zd$tMݙ8V`?o㐶fNqe}E?/|C6]{x=WP C4Rc(Z|ZJPx|U*MH F2Q=y{_{RN qbP;)utw2v E0t(cylfY[QxtpP.rZ/Xf=|% ExYo$Q]DXy 5 4CN|-Ϟ~G ?hWKԉe`И>B I))DCJt ?%ʠ1lDI }QBi HgGUl& }cٙ8LAtsOW_B|?v 7{\>CO.eLł u_{Yv]ek^^9F$Ө &YǘTh yٸ bW`J. ]Hut.:"R Is&'`TLYf c02D!R4Trf-uKh58`tmk%w9^(Q*b[\Pf&/uNۉ$Nz y,CR)v9f-OdcqWΨr+O9ίܤY}yȍH"`uVsXSl=` co! f«/WC>wi]h4{}S? ֫g6sIԘqCe&2E]l \JA *->imԸ6|7\'xӑggya Yf{}(_ݶU{{ghO̩L| -h6]|{^QV-na5١7̹" ./Q2E)h(km+Pm$6OE_ 6S9YDܛjyާ&o0px=B\QQ-EW8ΣS^dte5QGM1pRFV;وPhj'.d( l k8!;+jlWr|1VG ,-k=gηZ,|׸{NޏtUޞg@c>hϤg( A,fGo`UhL)f}!XkJE}\6bzJnȭ{$SP+[U1R-cܣ3p7]ml d .WԮm!㣻tomξN-vlX6*=" Djft%Ϙ*ZN,۷7+9{Fٔ5lZCh"cNXC82bS^q1jw]ld' DIpW\ 5%HHל*b\&-mujHYIxȂ 5mA5\TdBF(qu8wJlSߺ_LӓcW-"Lq (D X 0>ٺR2:X1 2PKN6U}ƠM\eSIrhKY\)>:Iɮvun]č@YAC`% dW*הKBr8ŧbIǮw~7{&lh|Mͷ|ȭWq +4˂IpS e?j~Em^ G[n6=k=(gie'o5iN\k*4~R@H0G*qϹYN(y\`HU(jr)i+Mw9s 3&Rյz J7z//DS<4њrcF,΃Xf|@a6FΠX-; ̺6q3"TeYfl۷i[c}1  BƊAA \|de[4o {>^5_rZ.Q){Gc*؞&(z S6bW1JPqPaQ\CE8=rqD_B+HlMQih]5W UyJ٩I*. ɸ$0`xYN _MєJ$P{-Dd^/cI1c%` \c'ngޢ\'/9Y'ǧ-&#D=m;cڋo狳P~WLZcB\q(<%/?5tk@:像Y6&kMl,Xr7g|iP|$얗v.E1L.R`KGq{F.gLOf|1h|>m$-r\g_ǸX墈>zu<(lS.NnXk?2k'K<=:$w|BE>?wqw-vZbC [SoÈw m!i6# ^;%PF+m9ٷBOe׮ ؎ϴ%Qiu}P^w/F߯Df'b /(MDӸU#o׿wx17YǨ+ܼ4+w4{uRurkjך_tΣ߭L.*Ed[ggÇ@)5k )Ύ21-tG]m~}ou!>'t|*$ϗe\^lqY r4=ƪkm"qܼ)2|׫g4x]֤DdC:+HEijO9^!9:H\ &Z>&k ptc6]xy !-ļ"X bD-)4`QE /E D="Q{A 4C–gfB[][ΤX99Pb,8YV@7>1Mzަ0 ?^m[څ?un8?qiF1=O Ԁ۬P!_6zу.Go^n|ˤ1[O9N(#y9n?cɯjc救l|ׯnq{\Wo|,B\?UoHy]_ojSEnyS;Ld*֓vAPE;zsvGYS6<lR,Xe`^D$2+;b@=khbYT |y+ea_X[MZSPVt)iIdh6gֲCG,k' jl&?VqD( XǜR,GEhY,J%ml񔳎l\ b8%B7x#-CU|/zg9jq -l!$K1W+)"1""kktZљ<(_7Q|?zݣ'UFyQT-TY%"QD# F\c Ɔ}&!jW ~_\I\&&Cq2? DK8NHM^&Ī`}랁G0V1/j[1Bd7!yp \DB/SHvdVe} V&@ud&_2nUJ+{z^vhw $rzTgf^r&l~MӍ/|HbZvs #WD@!傂B@VkԾЎ /H@k /?*Z$2 Y.2:}i!!`ۋM>ht#Np!SYfIIe~Hs_g;Қt_tQCg_NWJ[g)t2a{(L( -`tzIz&}nf`~Z =^ǍFg%&DZBpI%_kJĥ1yYH1;x]t5`e U(ꐸPтU )lс*U;t[=F]$ٟi:G}=\|q=nsǓFyt}F9y< |>2T+,fրtrJ{[EhqI[/fX$,“X&WP;_0!zpj ?MgƠ`g]^'mj8ҀOOlyk~ Qquvܦ{xȈ}ׯ!yǨm)H6PNUZXU[uh1 {,OsvO yLYl&O#7sH[lLu05˱(6:7sQMix񭣾pxZ$qpP0D޵+8+e(Тh}-CwEd{$Gʖ -QZ3EQ P'Yۍuȿ7A7(cjAg-C%C"Rڲ," ^x1~J9ZrYR"SFc6BCғr_)B8^SyϔJkqT/ˍ0~>jLюVG7r+] B՞gG^Ңw2 ګ&1i?]Xs1Yũ`"ŲӖ БklϏ9o;O KHm-]/Ɉ.2$blg6uVMFXofwȹ.i9$ Vm2E ]1 S@_B5*\u'UlX 3F™*A >[>3Q #[61qRmP񬡽ܘy__iM՗빣$0m *[0 SQGThk!ڐ!ξ.;V>e=:74^>*b 웷o@${F y lZq. j Nx!R: AH yWy- ֡l|*]儾/}T,|V>u _Ppgv}hrf W70|#__FG7Ǭ>+V ]6tE)A6tHU ʼnW1\j-O*bPJX (q KL 9zLƹaizp9)=)@>'Tqa$KN{yT 郃fٳrO)lrW}ql'M;̣˧=&ܑ1OOs+vvqz ?scJ~sg".mYWJv ^rH$:n_ƻ$f]v}2-3.--ݝ=df\k÷9莉_oGs\-fDsyˡ6GoW{v>~! t6oc3֤Ж ж.W͡m]-S?4Cǹ_ЖCqdWOsϴs&8V(ews%.(SJ0yf+1'+RA526[ndlUaa+8 Uc,T2 u7EDh0  [yrd&dH@,)(QBHCFΚBDNA6m@GRP5΀/(uBPLH!dD)F(!1"6[S>g~ԁqq <%_g3)9Mc\4.Mm%kgL"% BUNDc(k:ZAb x \<C};^6׷/&EۛU/RKsg>f wף4w(T.X/Wyvq %O2?-41Gѳg6稺j?Z}:uM^>|jz?`7}Iڏ.f>~?Dq_j:1/V|t> IMA{&R V|BN006Cn'Y=0v|fsM1R~~8t7:+N i?=<7}?Y*?nhZe}Q _m/M¬CsO TůVv봿}+<^d,$q4\#XFA}W@w.ozQeE'yӉ%WV&X#^qٜ\I(z! c|ZlJaw1"9Ha`sm-Nk_hxm*AXortʪX4oeI)$ J.g u&jL& X=Ikj!9O>@HC8xz,7:C]쌹}qtut!d2j2KCQKBc6K:ʥ0}% J*@m`NXP6h ix9y: ~|w_9|% rM[ft$}Z^ح6ٸ9mؑU`ΨV./̳[kΦVn]+'|\^f ʵZ 1,DfQ %4LBWTVh:Ț֞S`ӽ^kZܴ Ln;>ۗNU>:|# I ۜR|>?{_I*%Y)D>+ʝfVā:<p-G&q9+S={dDx-f7/7E"DPK% oK㼋WVX@E3eV :Y "brђBʾNE*g=,) y_k|2l9{FFM~S: ]:l0ey-m%r]#j/6o=B>8+JI(ad+3J@v2!Fic^I6DCҖaI!B0yG΂A ֺ$ lLH'X-ߙm2FK)HOjZ!?H2)r%d|2ZY]VŃQ5f#"ĪB~2+_=?[uwh4rc,HeE {!}v6]Rn'<+X <bnbn9a2SŌ5o:'`|ђp,ŲӖ Бkŝ[nzSǃ'n 33 1E_ AeHĬk[6uVMFX鉝kwȹӮf9$ Vm%2E ]1a*v@ U ; bdfmq(tސXyJ<]\xleg,/Kmcٲu=ԒX]$밊bڄ"l Ta.){nUMb -ÇASZ!FأՂt% !5_6t2帊׺|4dlqTnmQ|Vu ZFHUΘ!Y `C 9fRd=Yno;,xƛXQ3%DfXJ^;oX$mz3V-k<)-\tƣ2 UIY.HJ},V}ӽ7q:?-Ih 2qhLl6ζYS.YoW$g x@$0@&kOXuXbԙkl%Aیf)e- L:*9fzZd8MP1T!@֔bXru6` ȌRQ[% j(!b<)W8zq:M'^ 9f쎐(N\A)[H+ZU 1(K\[wDDF+rIdC1TdG3=߲mU(mBv\/{ʾm椢LT4M`|IfOџy܀.9MCvݳu跸 z5XR [\r @\'$L&J61A0"&4q=>F5 >Iay1yy.˖9Dx}#kC"ߙlb9vT"|/JS}^66סzR`L(ʙZg] G"u2֘-@:zo 96:NdUx*h_zJ:j !jbD };&}/4zõCztgq~>Nځ"5 ,`MucڭP|0am.}=3>8A(7]]./͢h2eii5SΓY}&tC|:?Z(b:_>#iîk1a)DYOHm԰K?v*nf6.Iѯѐ, JɪF*+E,]cU) ~Cww<5 >ұh޵]o3/эG?^r*~ >&h}h-:1( 8C&Wc9i" IIz:y  3OrWnB=Yi3q RVvT)w,uAwbk!ƭ-hy'hF Tr%jpUegf]LhJgcH$(Q_ʍnKGlНrWV4t?y5fELDT>Åֱg]#{D"/i4  āIұdB-q[Ilܛ8O[8eݔ{hMov=hnCL{$6({dpn?//\&ӿ>BR6v;t w1Qkiv !_(!BH;"D-:Z;ͱDr1% Co|ju"$ TP!iAw/%-OhL1D <{rFm&6 (k/?_õܫ3_{17/YZc'ƚL0kw5a*iT'!غmx%#`͡+6}.fV5;e8LCe3s,3yfj0{ƜT̩D Mj:oȤ*eBLWH)ֱRBPN& JOg_<;ռ?IbkQ'ت,ܡ5_iC(٩QPWK%:a|+uHmPf l r`ZWNHfR˄>;%Ζ|_CRc)b F#N\ CvΆd\lL,W2V:P ؔ/_Pyv|*HtM yS%VJd.ˈ"' hXRlW Ɔdc}밗l&.u$5UϺ1;v|,ZR;OO6-m좤a8Pgԫ7TMhm>4mh;ЂzDuM`FS4jRzHT8] ^\!2I9Z=~[ޥ.8 GkG!k;CuW4.H]5&b S.lظo0cVt o.}94:]J>Q3G^pmoaR COBI֥,QAl3.6R\JK͡u5o R7ŅziC->ԈntvuJ,:WhS9]u ˢG&/|~isog_N74=nCI74yW1ݾ+Z7 yʝ(ZpyMS?7XAzw~o^ qDm[lOiSA='C-:2}>OܠA#gJ 3\0ԩ#Sdp-NAmD+Y[F[ŜݝwEm&0hg7O6N?N O:1&ףqWM\iI&[tW ?"wv~4a,I~I`rWo]x nD O08wդ54twդ87讜SV&Ҟ).ٱ+V& zuKrcڻAѸ&1cqWMZPCwWMu]wΛFD`kp4kGYѤu߻jR ]Iw(g|ayo&g;vW'|7O[n=:EAY&?c/Ӳ!ӳqK~*N>?2nп`v(k`>KW?G˜kJ,]W4_Z;`qQD\s2B̾n2țINj:R]!d8MKU,oSܷVcxjDrM0&-#9ҨX-Fr䕲cB㱸&.鱸+s},'[|v,ESE_N_̝JI}&hg Xn)u,šXDb+6 ٿۿ 06#x7[U;T(\אiU VnA*j֘}BNZ,LbMbb> sK󛛰tFܝ룯U;o߽yCnj<{ݻ>[WV,M r#u ޠŃ Gh$SٍڭQm^'tttU>O}z iJ0Eԥrώe/m싡wv;{z fIs um]J'Hu He\Xkhڦw_M^ k`p5]\_wxYBU[o5*cB; .ksQBF=L[|? wn0ؽcŗ,ir}%E-rVQS"YUUU`F{`(WJl7` xPJpB&ƅm`*Qf" Dyb-[#g梉_'tFFqȅ| n;z\4E~Oo l޷6}-D}G+\%JR' Ur&\O=Fu+T  vF8Fȼ|~qib~k'T> IQ>hC %@r+厵{jP tL$АC܅aڠQ? ΔrvTYS.!|.I}kRXAXסE? #Imˮ ?ȖRJ< 0jofx[j|iwN _;8@R8{a0B-*uH/~I{}HR;_ayyUr?w換: l[A2 =:GԣL¼M]ʞeӅnSׅ>'_y]Q`-Ns2}PM]lH5kSۧ)T{nNL[KwP4ݽca6Ɠ9n$?d 4o!tma;1j&ȶvha_Xwdibf"0;WƱ-Կxf> ʹFrQ>r| ȸHm5Y:D+.̮ȷws¿sYSp ӛ8M$ø-2H{!9O:׊uyVS[#Sԉ8 r|^q3K\† C)GJڅR0jy[X^-OD3@LԖ.IARg$Nռj{T)ӆڲ)x-EĤQ4ǩ!AXcH QP/a@Cnq(O 'SBP- Ks}b4H3q%Y@gCU½aqz}Ɉ*IF SU I"ZGF41xRJu Ei$rK#e,k\5^80`2DA4v90NkIH E#\dmYє<4A'P"j:yu,oI}lW8?(* .H.8Hr̂D<5 <:: #a~D\ 8@,H2{輨;1e pQ&;A;܁Ha/]O*>+[1ʛ\AKުKĕN^'hywy5)IU9ٔG~9V H$K -YBZKSāFѸq:%M -k, JIR'&AUI !=#:j1>14eZiS'Nmjsrd!=A %W*Q}+A}° T/SU'X d)xH^*UuN:s*b_p?+9$!@xrUz`~B 봷*N.2A+.A4AIv,VEKBj(D.D+=HuxfR^q@J@m_7 Ǔhn-}];Ř6l]9^*N-ZWC9w7lTQz+Xiˁ]Zrޙ'2!LH~F&$? R3xI*,C>DUHb++&%IZxk" @mf5(" Tz)B>42uA@O-9.R@ax WƳ:kn1$/ȃ]_Ʋ :a:y<]^eDŢUqWoQ9;/*tB$/-ín.T Ol餳eb9/thKSӛ~Xosv,XWɼ].2Xs(m/BTBKġK@j Q!ȝ "ĘA+"`%mRk)4&֔nvr-uX<#Z˻X Fyŧ8,.pON+wQIu~wsoZex;}6˴}lm⸧g'(EPiX?E' +7I/Pm34wAWzqaN/AkoߖPI"FUQjA)xwԞy 8*D޻h.٨S(P9#-|0j¸DSR\lqyJ62L擽yeGMIPFSYzA*Q(&21dd)}%!zܶ73s$"˅iô:iCdͱ (HYT\f B폡: * `xqvKS]XaI{WSzPÔ3ԫĉx2$$c:?xo<)~A" @H2<'\9DCCAQa5ZUJjфC$ +HTpL:ԄVN%ɱQ*/EbT[hETQ чz5Y`0g߅W_윊?U_mR}UL?v;yOG,zU|Sj@0Qg ݌3AP0BZugU6nzX~>> .Fn9{em5gQ!!SfuQ ˗= ̫jvU0{_}bڻ]hKiy=ڃ|}ls޼Vk8/,Bԏ]^ҫF?]Ȑ'>R[m28WLR̋(4ۍ0J('bL.rUᢚ6ww?[4D;G<_HrR{MNV- 8}aJiT]-FṖcE[e˝\<Th5gq˯_D]])_>t!^_w0F͇%=ŒG樺^v]HLkf~Yuj̭fkv"uxEs;GsNH3M㝧(AÈ:"tA.&H1:"KE SM3# d)ǭPD2r&}pA A5 PDL|:s+糪OYp$m$~2-]0x <|:]ҵEm8k"-+ QJhe讐> }dۭ]7xڏd$ixS>@8#Sv[):q2:y~l+/J.c*^:b_ZKx@"MsF<њwBM:S@kdȶ4 ^uۊ[ݩGS;*6Ke_Cξy%F*T;'oB8}_-_uuemܦFH(M䜻RUʬd!τНqio>Hq} r9o>pc(D`,PA ڠ :MA4!@b!L\P."B ȑkkƒh}r^;RdJ^@J>%RR  }E Wd^R}.=YT8FΑAklJOWmoGt܄/~p?0- o􎒄mvS=Y1{`)|Գ3ygGe)F{5s_SY?;"\L4zi9ގ9u=WF64p[G3 >ft7S 1X]ڤEت]+ZbyL () DQ!dut")2bW#0bwL̾v58eeԖj^=>em:l#M,)Y"Dܣ+`&qsU<ڀ$O\L!WdEG-  )c>`Xr:_I{5rakoG 0F"j@ !)TB(uZ^.5$QQ7Ȑ%:m,g̀*/7%LJ&2ɒ biXkp-r(Aphr'_g5.9ue\. 1'@E4d99aJ0^Fì1x .ոT<4Мgq ]9 _'QS]P"5 _L"5똽Dj Dj =$RDji{Ap@3W\. `Wʝ#vW9vAp]W\. Z fЮ"\B+!. d r21{*TAzp "^"\r  f7 W +4s?[m@j]ԆMwkWB~{&GϿ/lڜgtyà3'ڞ[ Ӆ`}9>BїӅZCo_ W`}9>B.\ϮPk{o*껁+q֋w(WuvіĊ~,lI[(%np*x>Nt߶ݶn e@lF`/ljWۤaWٳl2I+_ +]O_|ӿFtѫ/@8&ve-@5F6ӓ+OOJ1ÓާC4cJ֩(d6(c!I4h!K;3y^J(- \+wZȏh'?)sh1p|:)#ETyWn*kFIAcg&x"|x$`?ɪ1Pk2rIL2lp9!@$8UPހ92ah4 qzΗq&;BDZ32;3|NU|%_zyt\N)ߞOOTۉjW$cSKbW RaDc#FArInCZ͢]$xi#ĠJ@ J!x=d-YÔFZIAuMyQn'FNTTJr*C ڒmJ` I2de9FΑro߁m]{ Y:Fh@Jq$Rd uF WLZ)83dt}V"F'HƈtS}V{\Aժjd;Vx]w? Ჲ }HDǼ+U,3 \2:PmV'a|ܡQ8iiޗ:+x:+uV6H>UػAwrJc(i0T1'J3 61,ߢ0xD*zDOD?>N)^EN> ؖ9BͶ=;p)Ia!+L&r $@#9butBπ<Di2ϸBSK(D*0C](g6܂cbd SyyW/Z E ȌQ&AZ#*rw)z]o~jRXCX2{QCq(!ն043/"AO.gCZ}0?OISTe%>EN8 ZX DA)[rx:MkOq<:S3˹iSy3v0E8K8OwN0#ۊw,Rmsl>/_sŸK1y;(/fGͧ.>qՃhn뫸N6_}W$˫˫u!ײk\e_hڮb]+,`H:g-o<Hܸ-=ם[>ҶA[vP:MJ=Wwnzje3T}Qa^ԧNLnEup[wҰ"/mDqD&$y֎iGM hfe׌NH:uk>]SZ)j~*aI"W?~5:DZ@:-ž$#sZ̃*}ώ[1ͧAv}Y%M7g&ac*{(ژvӰǜM 0_?oDiu3$bVwXͨ}Xd4͌EY.):_|4|a;ųybp.D S=!y|}Y[AP%SJb9Ռˀ,&irLR&!VK L$*e"pSU~] [)ҮTG lbJ2+,"`, Vʦ!^^_'VS?*vzYzaﱪen##Tgǯylzb/ mQM@2j}qln3uOwL]џ'&/fQ!-yO8 "by#cܹxȄX-YvHj4"Lt'xio1%kǬ2Z Bgj{H_!R& vq $ꮖS+_E,45(y ؖ8=ꧪ`pvC4ް<=2^1^ux˾ϕݣ`lDΑ28Y(Xku"1Jml/9R5[>$G LX@ k򮆳g_)dzRX/z@RĶtl'M97>;*U;K9`$)rJ@UbîrU 6$?h*tB]P.%akʃ%KLR:B^?H]E9yFt\tpbm(, P =Z ӇpoY @)O¨Re3-Z\tDH[:mTZtlTcV -P,IJ$2xSQ*gХ뀔Ά)"TSwOHxe[*fu{v>M)6_LHK&Ak\.&$ .d6Z&#MH}t&>"R 锷sP PL|ZD(U6bd0A b4B'oIFwYd*mr !Ȧm0JPt @;>eZ *..pm>Y7s}fG|0Vy.`n%3Upg{t<%=)G<JўufM@7Q6$4Dh`h2 (t{4cq@} +3|:}\-u! d^.G0?0P^dơIqza1H2d Z $ۤ!xKފf[Y _?&A0Zi0] >}`L-`y7?Vwm~n68mᩫc,}gޟSUeM{l ÷IOPAׯO o0fz7{+5p8Aԋ6Gs{={&@8xBZp%׎VF b Mtg]ѺBN#;"{wX @Σҳ h 4d9ec.9,jRnqJ6Od'ײcM竈2XI*GͶ5V&Vi^zRldY*nî)nL@J,ZGќAry `c $EYm p`w'%v=&tF#4Ew/}?P\ m`P bיu G"{?:WNʵ{!bNbx] WuΒ\h%V5lD]HT@e5a0^K pkX$EIXe椕|QPjʘa(݇5ς'p?Hҭ~w8ʃtrκ.6zy>s_sǗSS'l#f"{xMrY)DfmݣW=0/ʰ:~p5}9`ZBb=aLkѲoNAY9ϪG{XO-k]ӷ[Cw8_p:q8}iu(V?T&~x -$EH^u %B\3g:.px}26wqB,-VP%1_#`Ѹ$kMk~駅(TfPH~20Ql*ї*{|m]dOOpZe}V _sSSNXant󥯩_=9BPէk" =ndq4$,1Iju/Ra$1кBP@=YD;Mz>itusAaUhJqdiH+2f ʁ\d )kjϝx>`g2q`}. ݾk{Ӗ )/bعC]KV:v:Jbix{;?m~|(ʆ&7LZJD>!ȶ;z̺OR4FyL\s1Eն1Ѧk&-x4[h쫿wRud|\DКe_I9NXIܽR75_e㋧];#pܣE$Ehe-. S %:btN9\`\L_,ޅ.þlZۻ~w| =Ů܅o{wc ҆Goy<#>Oӕu3?FdqFT "56mB̥J6덛Wλ)H^|Noz 7?jL0  k:9PS> os0 2n{+EeٮmPͶGZτ0kG_ (g]pER,l#HkD#AMX$K={#GUr] V$i&4@G ф@DAkN=q(YSRcN1 )bia9؎=q gϠn~ T ::oa.wl{mm\ߣJ6XYbb޿htn?ŵyЦdj.cBw]{'< Jxۋr[󯭛y9tfG94e-}o|Gk-o漏̏Geh؍;:^?ڬ_r˦fl9fSԇ3>a\;k[o?m~d<16.N@5# lM( ISqQ&7j3(LT(b1KcuL%&_4XS-+xp(ٔOddߢɒw8c_x4m1W2*I!&~9J3 + oĤ ӅJ \P;@Z^ T2U?e*DHF{Ңpv/T:2̈́a)Mn!3990dɈ6mGt?VރT Ck9hSexĨE"a-"REbHJ 0ֻsA΃>-Jʏ%ʑ/Jaba%յfl85cwX;]ؕd UǺPQui"-{:Ӵ(c|ß7hp84<jSO=c3E R(ZTNaLEdur(Y2iWl9~ ؋Y jS[,TȮm'|("NTDrcpvkly,Z38Tk뎵v`) Fd=FCx%H"5#5AɘC?U-I-Kb$-dF؊NL"1ZR>YT)D\LɰL:Շc}hӇ: ~* >yAHm /WBYG{u {JoEl@I~?ǗÛPû7Iw?P_Fk +JE>ޟٻFn$6ߏE6wX_nwX,9䯿baY#Sv(6&mqO)t2r)7SRp#dIpkEKF1U.-؃ $],Џǫy =a==pPS1rqW ! ҫ6&D0JDNlbEe{nHmͬ2VBe霋gY:+o)2ں̜e9TWk4Nr-T&ŗK,"dMf K lb{{p<\&8?M)4 s3;[C$27}7lIݒ\a9obh-ӜWϛ?_8j~,/k_vzr%⫼9~ZoNsoJ .3xb3gq9 k@1̒L B轹NJϟ] :|it%ӻO~V#ӛ[{Dgnw':ol$y' :.!)@OR)S0$L"nv޻d>v͸*$\]*!:Ҡ$-#ϗi4Ѹb0?E(S`p%hAM6ihyJlyw?_&W{n䶽@,Gvv튿nktl{Cc{Z/iM։4[!  D@qcM,1 a!ԻS1u1u1u1xт8%Y]vi,k>AЂ }JI%3k#fv 3hu_wysq {{%SkzȝKNZm~as#{2'7=eߘe nht2==tϤ1{`l' \?|]wOh'L!]ϫOy}c;jBVr衵'[n!f?9/FOzT6rc=c֜86߻FnX]5J<: 4G1s!P:cp\*z ȵYA!,!RM1\22.90!FQzQEqt O.|P]N1,F(!,ˆRŜN(M boH۟T<hBLIe59R+|wH%FtH$=<3L9*Ȁ[@xg=ʨ. 5x3֩ 8X|a1;;p{ i!fo4u$J2!ύ{8\%?>x<9Ehbk`$2LJn@QZgN_{ؗ%LZ2hK9+#>Y2:%"y8EK]eu+pONm*7&syمKzdbHKs6Lu~>ۭm5wRޑ֮#xis_z<*91 ؟[]f1b7o]vNla0;7Ap-_6",9:EG%Bl6'nɉwMfFʣ5F /]ZVq\@zZ$R nNDu)1 >Lm,|iRZrskr|һn]:nZg1|7)\Qp34w@}+rFs-3HFO7m<) {UIT #E} c`wY)ڈjfaOz5G&$%6jDD t6缔Fv♾^H 1{ |փo>S })J_TvP%N[+]>[v ՟#k_W(#bh gQLK~4,yDZ;]c== -0L:ɧ |M~ &,2tJ#q\V J216躭; ltQY,L!K^*a2ވD N&\IՖs BlGKkvxdR>Wbf;9N>7}-uDlu!g ʡȕ%gf.rKK99s@iH-(GA*i+\c&% ZcLD#?AoNC2%ۂ[M!lL99EJ9FAr*a"CC ]Cd>%O>^PA>(=>[;hio_},cS߷^V<(SAraA&N2!bN$˥Nj+-+:g<]bz^5cI0qWK(D H1Gw彍 O#}XzglE%1H(r(RڱZ-S]U:rl/AmCV ZGiEdbB)k/R2CF#jnв>YS.!!I}kR8XAXB2z'-{@l DPYF'DUr5N(MƘkKsljqB)AR fJ&HSOW$&eWDadhDAy\A\ōu놊mt1r 0 8JGz}7c]n[ e!:BN-#jߢz(~/2U/|w"-hx]\&\pr1\Kn.(ŪGz:> 8SO5\q A$|HKƈ{O[ChT=ys rϰ_nܻ^JͼnFYqM穟+;m<3qC3sr%Lmzv?λ0,aiFTK Ȫ7udFM]N=6ɭ]}R=M k5tVtԀ.P&j\f`Z{4OAoѾ}LaIMeڴ+iY;6nt7Z_{&6>˳0AuʻΰMLx3OLgS_ּhf/x2{ld8g"8)NB= gn)lSgwQZN䂈*RFk& cМGƥFrl WНhr0=>2]GU52׸h9 'l:%l!_1rH.T@"^n=xT;S[ P$IH&UMF겖CYsVP5Dp+D!qjH֘ QP/a@C.n_O 'SBP- Ksb4H3q%Y@êP5r;ƽav{>dԾ2Uƍ$#M5*cb$##jI<)xr:]'KҠ)HˑWMpg ǹ*D&.dР-M]H-r6 3A:I9IAh,-K5&hTU Ac K>V8&|CrB c$*rdd,g|GGyղ˄\] mS5Aqa5 L\lE /[8 ZV^ tpp&8A˻KGC~{-{HBwqWs۩>$ϧYp*>[@_Z-u^1շT.WŴY4d7`PM3 S{/(C1پf/ۮ N v4'f4zy4+'nTK~0UG>Csj(,tRxQŽ,=%?r?[Ov3Zyp 7Emem@{A;N&r\-ɗupܮ(@>:@1'!5Kb ŝ>bEȼ mi ܖV ZFr 䐒~~s+:e4 DdX`HHRJh`Ctn^thM1-I6'yhwC ˺& G]3fM!G P…u[ @F' qI ZR  ɜ:inBT!5"`EK3ϛfLz*ym E zr<'#lWm=>c`ST,dY D,_^TEo<}cx(o^Z0!Ur`] ^.l-9L=MH~t&$?" Q<$ !rq"*N@$`Z1VVj-56DG*!kq @TS @YH hklRV˯f s7uc9q_ȑ>7eI`l0w>db:z * =kݢ#\?;nn%ϗXr"J˻Xyŧ8(&.pOJ+wZw\>.ߴӵep5~4˴?,㞖3o-C٦aO06oޓxF||@f~^yN޽9\8#1M;|}R`2 b օeRMfThLBRh'K# ;ߡYJ/thLѓ!{1:G-{:A}s87gVf@Frl1㲛^6Ϩe3\*k[ot~祄 ;.qԉڟV!¹+**o?Q!%:zߙ~2,r }ew%/+\o[u4tDAB◟~WiS&8ߑ%sϥ/g|j!~X>5;ްd& fJ ॖ`q&Di+y) H8g4D#*i9󼨊Ll/e3BQsNcOաJ/rI}gaЃ"AeW%vp(s̟4&>:CJ ^M?6.Ai\[xl>wb`HчzkPT`s,b|իJ׷3YZɿҝ=~=oNK+j)Jo - wZddm,6\U+kUt_QOW;%M䜻RU|L #zR}An&xL˱wZA>{a}5.p)ivpu$99ȥVWOèQt|?A]N]m1uO]G?O6߼ۻ|{^3gހ;qvgS>[γO n" qD'fI!h *AےL6Pq4G `(Z#;M)_ǴzXӴ ʿ?r[Ňxd7S0_ϚpD}T ĒQFKҏ3z;ev)8X6_D伞y2`.. E>Ϩd&X~sBYhuB>:X8ER͔ LT*(iuZ臍~v?w#ǩ3 ]]?9+ EF!zV1[a$JfQ:(Bΐ5%.QdoN{T-0ӄQDOtH094>Pp5<5xNA֠;;»lAovӕȘMˆ@w\ I` @lH˗~hL&G+c'7;l#%|qz0iB) W)g\ !g9BOIp#A`$v.g[vdo!S"y|,S=3H8r=5=U_uW݊Hm1 E*[/z˦"GAy,0$[' 7AF%T#Rw"ђ#KS.):g+0V-֑MC:׸ՙ]B4)&E䡲q{i2Z%H ]Ŭ]+f=!ل"x1V_\Yd_c 00 m*C3@IW)9|kQ3/'Oa3GT¤,O^hbd(tXEqވH$LKdm P0vY4xuKFZ{Vknݙ]6q0rH-Z:jfKo…{; / e_7un1O쥔;9vLqvT)7::гA!E:eUF~2z2Li8/g}'Y9j*29evf S^:2f*Ô05p! p Lwt$!rΥuq h.p^ŨPQy X6,sBy%yNFxfqؕ޶yM9^ ywsuꮁ>hRA"Z7xm/x{4To#"6Qb9)TZb/O3o|z]Oy4BTQ+ c+43zYpL:=w:e#6XIAwAty,'uIﯷ3qyfꭞs#HcQAcID+Z0ei -᳤lVV3ǍuZ4#pH`hPOE#/ 4L!3N<pc_nI΍?ݍsL?>wsiƌ&jCP1'|gB]ne4լ7)$t ,v_c0 2.7TwBw=,>hVxbMe(;L.?K0`b߆2Z%>皥|h (0HimRzU{Ce/̓6.K@w6t*]2ikR6Ja'~Ɨ*U_0(CWڹy u-ԃ5bW妎:Hdm |P~GԣGHz+6ŤGKeSZMICLGΠcGv/>QlX94ϑëfҿ5ɟX^ח# ;xjY,aZ}G/eWq7p1KIm.^իL\]\ &},.`B7 ;^ GG]kJZTГx>vaa?wz"]AVV^k A1 X^ M?4w3h&ES^Q7 Co4ߥzwX |^bKUԗhcFtK1Ԁ?abxÝߦ4ֆ?m39>$ꥥ<#ip*4c~Ic ȁVQOFw}bg[it=U RZr?˫ߗΊla [?P!QHB:Vz]۴R>5ώ3-NVf`ڷldզX[ǎ˔ 9Jz Cze<۾wh6stS _~X$p`в8QխiOgfu*5Q:4;n\٭oXvdɻ꾶ȱ-_2 y{uz-szD흉Q*0Jr肗B9DnA>˾e K ͔E!+p<ΥoLdL^VHHr^r+RJ~ICx]vx68@+\㫒eąU-ʺz 5TKB ]3ΐ3Bg*)V^6|<&WE.b]MaAsl7!^BGўiyQ|QD$N A@8L+v!.Fu`q>Ft(j09XHML}J!])&8>zlMP[!2/A40qhDYP*x"2đHF5\&0{F (W/IlB-ViǬQFYJy>rzq <7F.q9,=~Q1HXK ŕ2pT5ZgQ[Sz"M҅6!e*D:A"U4c,ȴ""|l JFc&=<|KʣTbvf:k!#8dZNRJ]dLN'D  hlE66 .-̷E7^1%SJ;3-B0鉛yZ^|rl879- -ABpwWMD%iOtTp 4>Y Qq#xdQKɕV k8h  <!PJ&Ew9lVJH?B"sfg?BmJq92v>x 4ߢ=A`}}Tf>~4,_H JBkd5zS`bxN)  DuNDܣ._8(~P h*2IP5A⥣(7IJY'Î6"ڋfAࡋ &vV林[ד╷kx><$H2FC[n`O=,m}M 6,J(og[}8Յ tǤ~~(:1xP yN<68 0cμXc=!4k-xNDdzEQ}Am;_9m~tY|ӱ<~eJ1XdA! M@J-S!x2U(;՘G"hHM̨3TF刊)D*:drgg?;ؔbj뫎j)ʈrG\Eg,`U ˬe^f}L1q܎2"p<\LRh|aI j)B!$m]I[BV(AC\( ,,u>ȰL "L3bXs)*TGCJ9a@ws%{gqo!T9A 3lBU/gk~;Yõ٬E~ywg )wW%$߆cn `C ëg«&XպlU̫j>KUOrUtޔ}SHM޾I%bf?." E*=( ,a ]2牫 }xpK<)]] DA켅 l)Y9 oo/ٛ~qşS " m0kYht86jy5,Zik+ȗJaYw\^͏QmUH꼿d. /jy7)/ohV9>ܠ\]K =h}\gU$ػ/+Q''^ Wxrvc0,˓r/*Q*EǸ.Gq!:7uaE$XLjAu3NJPF~|MYn0-FoTż]. 8:#zs&&9͎r B($<3B1gz4‚S7O%qEM=L/3 }9h0͡]w}(]zA.}v;(izջyq45ԩK?ҨqO|3`Ewc}*-s6E 5C礍ߜkE bL)Uz[[Qf ɴΘ+7b~eȸ(0ɴ28:w-NNO/Xv텄[yB=j#>-@<ц!fQ ce2͊f^Othu'hycyѻdؙ_K3$/$/_5(p8^ {l،HtИ K$E읠NrnXy%VtɁD ޅaHQk %mR!`V0/2Blm#GEȧ]\K@p7ٻ`p{aRL%Gg_[ezܶ[qHbZbTкL c3ƶDז8=vJݚve^[^Gq| em meFxk!yr2+U6>=q!DC&!#(C*HA56JkluPX -38-#Bz{čTS:  l Ql,2%Jr "=Ȑ%*{b1mX*h8¡-9);u1x,eX_\^EsUz_fRp}}"3=<.}멢wVUdφӢ;5W9į5afk Lw:O-rME ƦQ_%k\}C?_ ~"Q},IQY\Xõ6 M3Fї -#y݊Ο {jHh#mBsiќ|h.M&Cʜ93!Yue10HGs-NU[ϗ.Fzr`+NhY A P1ɡ Ъs+)ή.RjњNYjFw7Jn6Wt8' ^qQ5(^~|Z(8@BZy,ӶUW3a't*,K9;aśUYTBJ +J;cfP |,$&*ɠcHe&ƲQAЙ 'NZnߺH+\K|ÚY霊/ ˙!3`"EkIcILeʾddqY|8]F7M^|UmTS$2t|Uq>(L|SVtu] Kk9 B9dܗBP4 K$SNN@J(|?"#֨7pN\> mo{ym+-d#BZZßͧY8;uAB؀{8JYYY'"PZY)!:] S+,T :߬jABlgȱY;BXdJLs:egA͂8:M.tW)UOIR(IJRd\3d ` Y`00iM;rz.mmYۥ;l%P?N[/9<b6aGC֡O+RxׇUs P1 OHP\M\ˆD:o_t˖)'1I&1 R6>A.cz22# wf]!E0#JnYg Ɯ2̝CQ%Fne;gG-)Z1崷.%$aJDE#)I@{b(D+t̞ƲD+mC66XI(웶Wk~VInHJiv&}i+ 3 =Lڗbua0=C l.0hI8р6O1^z;[_N?fdFV㓅t6eGU>]Ob'n&HmBSyS&rZ("h$E/B㵐s[ƩYVm4m ޏGLT5.*UҹŠ*iVykmܸ_?7AC{ MPbbR,9n~ϐLɢ$#[p69gΜ;g#!]Znu '1bePc QyUKO6grQZ4*⾪ތ.wǶ.ٴ5];KFf sA/UP6*il1W6G&9 ʔqQ2iyi8)!8Eep(0m ~oѳ!J ndS&>,@ ?G.AiAÌ?o*^O&U0/ߡE8xGϓi-`kLLnz&wz4}V<ȍCu6rluFMaIKh]j]67wYMthg8gZвQn'=?6`[-7C* oޅGU-c-ςW 99y]s?uN]57+r_RJ-ͭ7 Yr*Łus kфNJ=>8L;Vhi{AUُ/ݾ:"]bœ(142q*R ^2$3/$=/<(/\:͙'d=%ǩgZ\iO4,F6Q{C.>q6TWT |m aI-ȩ7}~4tTSK&RI Zˣ[lϋӽ+D[qVXw?};*ϹPoU1w1\pn'ߞ7?pߖn˵f{mVﴝ6tiݒM6Oҋ#vc}aʚc/kF͐oy++N٩K_*j6(%}jgNvte˼VR& 1 &p 5nDh:vFt899>nn^tC v& ή>P,K%&(J(4BaDⅬ[9$5l%j,ݺf5TjW(RY ,ū4'yZK'-) QPFuD:R)0q0Fe)A@ ,zF>L3,' hE4(D#3?߹Zۻ"[ iҭJReYwBg.g %1wJ\`x) s=r\y+Mb 4[bxy¢nW߉HitL*PRafXTqaXŽFiH&W ;=w?{فD;vYó_[XNd!BXut{+8,rzHR= =h) GJl=B)K *HE U .UDJ")Fnj(9\p`|1UKMB41Xa$Ơ|G8|gGcBOٲ,|XXrTd 0dpU%#ϣ$/*n> ,<}-MEP:TcҼXuJ]a*_܄ϵvXR٨襨$lqܥg駑sБ@->md4vڛ?7sּY?W{ϩ{o:D}̉BFY..3{[*%4 Y4LK=u X<.ηbK=5_pYS7.}y?_&O*OO/Z U"IIꑝc^2!bdnd b< u97 L )mIWp-a"ӌ>%>xFl豩EEݢ^w${yIZ|vO# WY`3`nkJjv]5{OոECɦ^w{uO&wҙ2m=i&bEV#KGBb:]|Ap mDr;Yh8)#x0$#FaclֳĤ# 4={mhvz|9Fqh0<8Mq(E}c([ÆxB+Yϓ0Ke63}#r^ ڠ>.+\ 8(ŏ[ O3M9OSF6dۑ!rsuz`KiQ^g,(U2xi23ŃcQc՝@^e= Z*:b V+F7"J OI%5KdmLsMuΐH :McCyWvn 7;2Gq_]S8 &LɃ#?0WL1 5֙sc̱rz7IK\ԖR2|%gS|6I/IP}Z1aDeSC;{ S ]^]Cĥ5CHgj):T%?MS\|t:J|]5Du-}6?V7갸;~ε!h~AK- jYڂZbt Nm=zHE0eT_V z\W Љ0A+JCT( fp50*Ӂ*#*zGLpRsԈpBݔXt+n1Z4x\8( !VO#ɱyvfZ(S-=Kp`;rLySAhkJ" y&7 S^:2fa!0%LqM%aC0(Sl)O*!GH-g =cxzF+-w9a>% Krd{ɇ(H[LSdXO0%v*TEXhs'h ԉq|Tm M!GWNx>i2N)Av \Q C8{J >h,ݼrB( O@!cL[:伊2H2Fqb'(C=?9J0FKB^xz,)/pzE ғB _P},So5qep ~ #?~I(sh;m,}|#μ`F\=f)R\ptZn_Jѹo &WÛp'Adm.=:3Lc˜)Q}MWݮS\2&x8ZN*/~n bvoAI^_"ˍW56Jn#0k*O.M.90#(tv=pTD4f`ޓ53FRҝqai#JsR:49@C̒-i+"$[)^RAOȁ'*Ո߃m:ZzтdL3C1`R!FXjw֣JU/ͣy&NN:iȡ+[Aoeb0:_z5i&CE|H,eBBBq1eVl>塴mVӥ4NrT7W\UL8 I\Ǜ}tMZ`MB-Y+} ;ɁQV{A[c3(Q)H56K`,07,#+@'dpXz^IFpꋮ~=K7Q8Q )%8@GIՁy"kA6Qǒ* y]b^1ECUM6'-}2d܊gb?܎yG D.Y4 O4CmBcKfL#D-$Q#Ӄ$:Rdz,#AD`ˍȵs{*Z_J: yާ$NM R2: \pAv!m9^GVgȴS";͉Op7٧Uv&j|j{p)mG۲Lni,.bb޿sc ]I@ȭut#0?mU^M׺"Y~GMα]v>q ek-l2nrvB}tGdyKs[oɍ[/o_X\߁lo%j_bsGFѥ?X1CC)dQ5>$9"l!Wۓ $B%,*z 623De@ǼɠJ˓.G91H\7@)EӐu}xgu %euY+1.g6FZA&ٍ|zXXK3BQ EG/TZ-񆬩 Ox3oAp~ۑox>y.X$<bV!0EB#b"t0`P046B+<*9ɇJʗrE*#v5qv#vq-<Zq(jʨ-{=ح~U:@ ˹:#O,)-Y"D(5倾?#x0KaB${"d (:d k0HޒpѐHEo@E ڬ5ٍmPb.x2~l-8UeDT="!8/tRRP9t xdhd FE% 2dɤ*"z3VJ͍LJ&ner1R$-`82"Vg7"rQ2s:iɡ+qr8H}Ve%ɳ Sҁ9e!c1 x \<<⡩0<<UL"_U$> ce?.͒` i A,'SFYiGБ"Qq :~>t+rHfCۇ0/R&zb#pW`E!AL6J4`A5X0` 1{OTAջtN5$X3Ԇ>Y 2TI)`e O8g0"9h/lbFd&D $.C,  |T Nȫ=+3fX)d_m1rGIIPgjoQ1r7B{" GC乨 t 9kNo$Lh7j*KwaMjhu\+=-YD,XevSnpOVkmH] ~0sv$`l2m+%EqSMJd[mC3&YdWUU]dܓ8GNXojzjD!!0e +I8qH ӱu5 ?t\IШ!9'ύ1D!qÉNQa5ZUJjфCAN>2QZi8L$F$*QmYL.EO'MSE 62)mmiɳ"\a#CpٝnʞEQA*0;y&? Z9,NQ"lek@ᬌC˪Vq&h}/ 7,^!e|xrT6nzX|\ ssGch#bxoXx<^!GEȏV_+Ja?ƭ4;*.ٝ/'qZ.sILqQ{I3w, =->4? Yݩ2;{J~2u.Ϊ7PUkD A1 ^Y2+^'hIgWԠnQB9acrp甇WU).sq`ٱoHtuG gr]̹RM\OV-ȅ+|pC0Gc("WUb?5eˍNv 5-1hJݝ1l6^ (OsRrԠ'YpvKPF%Iy? ۽V$wS"99ho#4"Pk~ #mJ\O҃]L6cDSEsS>F{^=',S,$K)1N!@eL !PDLZ:s+*N/-O-Fz;E*~ W:% 010>/ bM @yI'C޾yU_$|G̉ztߏWgJl F.O~ˇQy,èl1MPL\T#e.j-g"xEt?7&Z{/ZGWkBQEL/&êQ=AW3)k*a#!ʔmsJE )Z ]ǚ]v z?Yśd1 6wyO<\ R{\ 횈Dud_DZIΐ MemOPӥMB))c"y5*hlK.Nx[ޙdaijʕCg 6_֢ kmy&w˾;+fuIem$ $&Rr})Uuʬd]НqY&@ZrZx.ép5µwJSrR6,z&HE uH58,|" ~:ԿPԅH)Wc& 6 Ht 9]bryuE]ĩt^kb򾟑xb.l0 1E2(l7aњ?39n^/J{n;Fc+Uȏ|}Y?79\vK ':{k>%p€he*4b;%J!*FZ :Hz Ղ0sFxyy:(㬱JȐP 0):hC %w2RECPޭ2^nj݁#_fVqExiJLXÈϭwcԎFoK]@Ri%@)! ?Abu͆; GFݺЏ?~e}?s'z`'E-9i{,a:{|;"~znj5lGwO8Z M>Z \D1?|}}-9mf .^ -sl,6#z'1]HL_ 2[""W3Y9J*fwjd:_~JtKiʿ%c&:z4)6=r0k,%ynul(gBy5cʋa%^}oy1G&af.nhfҖ7"_ [ys=ƓǙ6Xo>2I:^lh|p MxgW̋CհKd}ஶhݰĽy]w͒ цZnto<ܾ\x,"$"\KII(5U 4q)j•:ON+j"(cƕyȢ'L^+ aRJE&:*ĵ1 Ԙ`3FsV%r@r)+xjF]k "9T[zKp]X#xA+Zor23U#^.g<}K]#h2K-Wzp$钾bAN=y뒾w&RBB֘ʔ)QkAA d=`I$5*[i˵܆0AIC&$T Vs;[#v9+ǣA?\gȝ%mrG%5 D$##jI<)GIㄢd(|042+Ti5AxgX\FfDAA[f-rXSIH E#|dFc.4RVwpzSR6K_s 9HrjD5JkmȲ/cr 0Lt!]toboI ?|fUml|Q;qQnyyM~he4V z`r46WO?l%-%l o'Z8-ǍdpHVR?j5qZ/L$Gϧ)JWO!~~Lo>S}o*y糝vuѸ@it]'/ۭOW~9|7^0꼹\x8i ^PzS!Vf0I⸂5V_jEm`- SGkvZ,c<$t(0t"XU3cx-w֑Wȉ=*sPvzDFls]g_:Vd`$"YER1[+&>R1Sd5qq.]֥HMj*@=Ƙ34!zh& :UgltWTA0 @Prc{>q8iNS(V~}wGWSd-x̵A"du:;@k'iu =\i,!pHD8)kp%Q|P|Fu`@ t} n1YX&um|<~aۻgAaW[| !wy{iEAGFZ c ƛ6Br[[%h/|\pO~n$߁~~]I 2'Ld:Xhq54ÃwrT7B>IwVlEm-[k; v,/kUfaQ79Ppa:GJFg_ldes2H)guOcw~Tfo*YJ.qL28$l3,ՌɠS@\;g/Y߀Gܵu~;SGNdwa2K)05kh^NXtj,p9!6jIufՂX[-EaN"< 'wVάyVPOWKBQ_O2u,ZF5k|C򼣾 ?7r}sg" #Ry*)lХdt$ԋKpra2:Mu6 ~?w ]^J]UgW_eozZWx}^ m39`u1ץ\10!t !ߥs,,:/-߸Wn7WQ;Ɣ#z.O?.E-Ϋ%mJ:08FEnmA@RYcBX85]vl߽ҳ;뼤ٹiDӌ8˝D&y(t$߰I<󈞂j, Ga*4 ^7lkݟɷ4[TAJI~4E %hC}/HiE`Qܑq&uQZKs12Y,lm/ )G2oϗ.7綅rvDc8E}Fôx00ۆz+cF!)RhU<rR@<&m}*S|FAG#,B~rmz<}9ٵ7nFw*ݲX=Xn-sAxZ*cJA!`ѤE\hR"OOAꐂ>*xDp62_we+Z"fJhdW{ս}0vu/)%5Wp%piSb $ mr}<}LʄGBGDsIggߔ{NJ J}eaMQ`%57jq@]/9-ll~ ܣ}j&.|GeQǏ(犒I5-V^>+4{ۏmn5 YVZ8jL4 Hx&qu^9!G)!ט[+RR- !%RBZjGJH%?D')ԚGWLb裁+K~,piC \=AU*3,bວj/k)1-("tUܬNOG5Ƕ]5"_ο !g=6vb4^H}$>kZvUUPJ%6:F!s_ܬb#cڧ."xoa,\ZAuK/~on׊f9lWmk*8YI[.Ԯ -i{#QW6s)VؽZ:p? p+nی+:s&x3>413y-{ocd m9QextFV(NJ5{hP2+ qKie/S S!mjx{Wjph3/ ^1}+b@YE+^NrJx%uDrAeD(mo]D- j'zMB8+$5}d!rVNHIT.l5]ʦq4RR\ٛ"sUZ)(KYj#RR@J:2T{*Emv a k(aF>)` GBRD/q:R!ehF:S#uqJ{:j r}9겥 3tREdY&L%`R2iae%BIְU"D%̘fP, =b$csV-DXa,,R{k]bGi0hn=#H5^ƚe\m aM,:szU@_+D:^%9ڕ $YgĤ\B V [B0lC8?u!]o21ڸ@Y-AQ V;8ՆOB]7;Eū$,.l⑄Pm5$+˖z F-DXudԚ>CjIkYJOaV:!$ȅs.I)S>jhI{RaOTQk\ <nCH&*,F2ц2h@LV"\@jBCt$vA1!DO3$Aq*Vw kPV)VhB@XM`t4[V\ ~j9()LXuf`D+'RJ<تT^ŢJd{NC [pP^Ѡ0.VڬQ #F$<hGHԯ*D쐁Eh`#üѥ%NB KX<hS = ! O0WK咤ViBBh>-~;: Jhf4~ ,Sy9,r&L:܀@/F0!.%(O@lU{CeRC ZpEB;4@$Oa8T|Y&16\"aHB`KC&;ˆCHx7eD&n3U/ Ad%E[DB^F#(#u FŲjV Rl i-? T3.Jyb.9u"z+ۂGϴ)XqA;3kNL ĜV f^U0ʐ>xmBV5 tnƣvӐ&>y4_V 5ȤuՅo138hG}`*YȩՏY_"bΊf lZA^֊ANE VE덕E]dN;=ZMW:CX6\'*Uk*E!` l) ʽKC$`Pr옅jIz7VEH28D )%afip/Hm,?!t+<!Gzm~>[q|\miJHԂhRthArz[QDgt19EK߿}vGvfD;Ql%)V&R./Aa:=ܿb>9YUeD虜dh*z7sUt|DI/X*$GN#D1{V8\~-łJZ.{.N7ۯN^[ [7DGMRor*C]gɰ}[ 7G ^)BOT pv<> :Ai4Ej rΆaУUb> }@b> }@b> }@b> }@b> }@b> HΩ1_%-\"!HF.> }@b> }@b> }@b> }@b> }@b> }@b> P}@S"nD> 1;J(a!Q}@b> }@b> }@b> }@b> }@b> }@bt> L^䤶j<> z%F~}@@< }@b> }@b> }@b> }@b> }@b> }@bnoǔ^um5Zv_ޥXlCtc-|m p?|ؖms{P*ͶC-&(o#+CW7Ʊji%,l0ѕ Vf,tEhw"95]] g.F@ɀ X=,o_r_;?b~vuvyP>602l Bjj~/Oߗrw.N,'*{l=!Vhsdŵ5WHnNh:!kGC(TC> W}^:q>,œ4.ND%[KNW-N 8?fr81fr'k6z?kkuc9e=prśY!^5_M/O_WA~Rq)DAǥJQGGB4B'!r4i{RBi- 1zDW,G*?"vWeIrH xrz/'ޞ0_z򺋮rT1]}jSc 9""& ]ZeLWHWZ {1"6JtE(|2]]z1]6q4tEp ]ZoepLWHWV v4tEp ]Z%AҕZ;""ASAhNWr=6cZj'~6f7CuM3E.^[^ʾa.k~'|<)N򢕗ZZgd%:ctˋe  )FU>k(:Cq'C1xm ʒ3^SѢv[ӕT-ҩȟ(/7g {isgRt]L$עE0-Q/IaFCpKFChPF^'d4#d?'~04B+t]/U6p+RpVҾܷ)0]= ])-$T ]\#BWֆ}+B%ҕ*9""#J :> tE(%!ҕqO\`mhjBW֩}+Bo/bzFl"֌3c+Bvut化rL?ApR;UnPnc: a|ގq u㔤}=KAUYYza*BIͳ̑St{Yz}x}8Ǵ+BUj64nj"|tbˮgfg٩8QjjT. )4;|#pAAapjTHb[tDP[]xjXgx;?}.:}~SyZW_?4dZ H''u+*9tL 9 .{*$i;, fY]U4lnݼq{oBtj ф<7ˆg_]}us֌g_w[m{p-n?%&V|?նm?lJ*'|z|kVQ;RTkYVQDhdN6|Ӥ.0܋>X8jS?.tw̑,^$LPV*m*:VE81F[(@r W9{OyM0[}b8/׊ gi+< ⪣KjbVWQDk׷~z9aڿ`\~0R߮`V;KrY|v:ӷ{y.{8~wʳTUkܾK'^lo4fڸ\^,N^?oM!MyGyՑtH*w$V߬>@\};pC;GQ۷?=ۨn  @3tR}yA_:Z*$/-_N] PH;H団MTRNC'?#}OJw~Z<]l osEsl4A5[7m9y 5j~Kf7G 9@|S#93*3):X,M\5M9{4E(?E<ĥ`:xWY*/ݪkfV効6bkg]tFvVNVaf˟͖Ŷժ*lؚcHM {)U)aI˒ZYfʨsFb772R̮\/Ut*RM*\!^g;lyH.r*o~%2Wk=Te'WSZFygvrbuz{d/>8m_{B}1%\ -ܻTjQimr)}SZ}.Q>kQJ425mRhіR*PT[+e /"x ·G{u֯`_F;zkw,vݑS `D a.*[!xzZRw^~6A -eg}g|R],vT3j)J!zRHG)ދҫtkuF[֥6BtZ$Cy\qDcB[nh쓊UPPwUC5\M2fj.UU Н]>INisyv~% s7?qڔ}T~r2V,cw!c2ٻFrW1n|?`1w Y%$,ba٣d e3m`w$5EUsȪ)|1=^^uRlW[E+ l2Yuxy}/d5(#6JrB)4Za9A8W뒸0opS@K -] ]X|5-HvIoO oM,כg׫ '-Ը1c[:zޅYͭ޳IiyPVJ" ( )Bc8^שGx߃8))wA3f{5N|9YZVveu` Ey«#8\W sy_fc{>oěO3Ĭ1j<# IloQ@ k$<+長ۏWwovg(v·=тTI تwSI יRIF+Je~J81iƴ" S)xSCCEGFb)%4*xrS0a) lmiwrx_!Vj)H3%U^E>LQL\BAww VGgŶ'"N9UhNf  R p,1 "C4é&S_Jk2"Z͑kNAD#-xr}#&%8,s@1R +\d+f.s!`{0Նs雥m1ɰ *m"DZ[dGGtqt:_CS{0Nge@׫U'E'lr{HW׎ޡK&6O3 oh_u n￴nwOs, 28xzuy_6hh42}I7؎Yx\H(V@hd .#ilLNlX^HI!E1hfvL 1∙Ejg5&E1jWSleo{X!. ht FH 6^ "h0- 4)Y ( ˒3 q^P * ^x:os-{r 4V۬j=I#ص%((s%KGJQwFC]\CwRmƀL12e/@IXdȀe4 | .Րݖ䝋K7SRJdڞ0|F Nar_t2gCL6B4MA5`)Pus{TI;+VGRc暵̞&SCU6FVII .ؓ\:>vLbl<&Zq(R{ YN \KX.yABt@h)ߒu`S-ŲMfa1mUY kPXe:f *;t4|nI#0H]4*sU  qx];u+3e=6p̦bhN! 6bJ`44If,Rv8VKQsI_4#Rk9$RA:"癧,dÕĪ.3gs'zEH,Nr-TxF%sz9dh,7C`SZѻ#6m˳ !\z!w㏃΋g8Oe0QnûqqmJ1E A/Yw?^m0$S_ IK;{g\2q3e7wHe 9>|A |AȑLI662ppf30ߊ n>n&'!VBUE]*<ݶ~Wy#Z|CuoMXf~i}\~k_Snqq_xD*$% $ܗC$QR)s R*u-AQeK;+偽fCHh/{BshNhJ .3xZY3gD8 (j]L>3Tͽ5-Z՞?]y1zaD̒R!ܧ( V<@@ C3Us\ |<;/ubaBa^m[,ZUZ7M]S^(\9kit΂5{yWo;( sp_i! qw?,?}?[7}9\63Xl[uD6`- NEyךKV9i[(5b,V*B@|s|C&: *4@3,kVh,'hB(<mcBe;DnELjXQ0 ]4'S ~WsSs))3zE:C"3/7&@O>ngkفG_s{"*i;.jq>VY<մK+%㥥u菰EFLФ$kq.VɊ%F~3?; z6繑'Hˤm|`x.c B4 jKиc iíтǐΎ yz!>G([xgrd5o(MiqḟiKhKGNAJ;?K2`:}Y?u-6m>r){f7 'S^tFc 4OJ2e|P!% Z2[yjǙ* `1Ȗڭ;RMb?{cώ_eGnlܼ~ۿڭ՗8G?X׆`-c`Gt:;Z`[MO'\Ph c5jv]=?;!t>({6Qlѯh~zڞdy^}z6? }|y8ŗhx>fN}8=D(gcNNh;ۿގ=m^ш:k//iI c2̦'?PymQhO1:W_ҦhTLy%OO&o.Έ5lx_ oy7z{kcA6MsFCoݜe'tUM;'R]]bŢ٢qˢE64oݸ|' 5h/'?:C4-Cڜ!vP | 73JR͛fb6+*r6vѦ/ rcԙ_|~5*iĂsUOJܫG<e-wil)1 t=v=E)-9<ޕ^|dI7ܠQ=ȮN:48,fbU>i~.V_؞\`@"X!)+@]Zݹˑw<˃c{WocTeYHR7!E#eR)ic v2^A@03F *J  g "s{ܷS.쥎jWfF NxϽi xyr$>_z NynA{ƨ75=Lh1B8`*/#WZZ̓KB,DB}bdt[!<\bj* C. ^1G{3)PX_NP 0D IRiXmjXD{ Z"jfK JDDկ+m/l`Is4YD5-jV;Qe)k0tǕ3VCYC{wC\X-־\>R\E?o#Aa' %dO&x u[ՅJ>*Iah*筭p4UAzHQRmm:'GO#rdCMA1jgSӀK&Z!zݠH "VMݣ\۳-(ҀV*cҁOd۝ߑ 24O | |oɶ+[׶qp,~#ep\6. W2 . H.d]ZÇ+Ri!XlNB\]\ HUn# /UU>dV>n"͕hu]5" rFA^%'YjVF~@je$*qhkVFW j[L O}eP|i39R p&V6z(|pErE6"jH.z>r[63n3}ҐછZn*Q+WM:'\` :&s} T Ypum|p-lpEr5WWҭ=+:\Ig43 [H:H+!J1TNAW$L." W0Qpu" xW(X6AR+]J(:D\i2ǻ"2\qX{64|$Ą 1Bڸf)!24 ւei *LZ3Ti*>@L#8pq&pY2kyYg+l-i6 t{q(\eEM}4 p9.6ٮ' VZd3~\4jC?I6e|㧝`ƌ+~qY~ ;Ռ3k\gbyM:TKR-.-zHsp~uՉj5 W؂ĕ{7\G|pErWv"V\ ЌpEm>B|: JUpuɘ siU:\ZÆ+RiyJI2 V*\\#rKj*+!JKeMNkW(pg+RkqE*U pe"@lm6b}傫'FM0W+4ϵhfD1N^n/t-p }1o] w`@LXsw]ȤڐSL6]֪ rTW5kIfR}' ,8,)R+?#*RfrO5s`)h [c= \Z+Ri:߲)θ; V&WyZn*6Wmz,#\pEr>HeAR(H }S視ۡT [puqJ0H0lpErWVTmMՓ 4W& MI +R)UJiFe+,! .\Zg+T \ 6ՓAx6"&\^WR\ :Cuv>gd[49.gx3?۸K Or}f6ߋ=tK"c/+_8m//\~ntҝ+ounGm_-a7-)n]ȋ/>p<] RX8Kx"rWrF^/VϥғP-ڣz1j&ciP)d㨑\CFRΝTS,9j 萶:g8Ur"ѤGڼynKSQi*nvᖦfk;DY͍W(X|\i?Ճ&CXWpĖM/ZMWe{FMn*]WmzP6"\\uZi+T,:@\ 'xB7xWv_ϼtq%8w%b{/QWqE*6+\V]`ʕ+R _" WG_\W(`7rՃ4qetg+lt6"+À1:H%/OWήgXi*Qq>RaߧGW⵷n˴ѐ-3DӨV 5tLUӇi*Q&+-w׳4x%Nf k9^u|*vq+ScS#5-Q帐d4A,<\hP_''`ˌhM/njTU7z5{^)ƳJ4%;J\=9Gg+c2\\r0t\vWJe+W$l."ZWrhASWO+) K'N]\%s5|"V\ @Xx i=B76 UlARmJI.W(X|+kt.B0|\J#WEXqfE( H>\!yȾ$+K$]'Ñ,knKDӞZ/19$Nyteod nk+&J1Ut̚jfZ ]Mެ&ڗ'21UJͪGC )yg[,T%Zqo[>вC@v+ h= '\E6C "1xL롫/Oz;+~+ŗWW;fz /l/ Kr7vށXCO,wf5t5"nhDI sIfC+ɬG]-xt5Qc+ܚjQWصD¡P>؇DW,.&ܗ66<]MOAִvѯg})V^^JkB7)^]MVCW &ZJNWӥc5,Yp[ ]m st5Qs=O*Sk+yHҢIVA?yh6\R ͊S})4;E<7i1<;R+8ZM%7j*Oj%w\F;N^VCWn\ͦ6ѡUdtwDWC_^޻旦4Bj7]+]=w)sO< 0Yp&ZχNWeJWGHWsWCWn\ ]mCC2y#+'lX]M1&dBW@D < y_t5] ]M>:]MlJWGDWϟϸ_J埅\Z]=|JkWGAW{z/~݇ c~?&7D1a{xm+Cǿm[veb1\n>~ ,&'"g#ڰ|1?+S㾜O{t?.fOb\mkNO}\]|%޿A/["ӂD0+@#Q~c~- 0t7_g3O>2>cޡ9>yiחg~e6=_pZRϖeHY/Sײf 88S:?d cI$Ow}>A;ćߡyvꪷf@]^m.Z\L՘D#]Fє- \BwLJ$>ϗVb%&δT" )WLnc1lm񈦅\;&o|yP[؇i0h|yB*ڜbr2V8m=bFɕV1;Y48ȍFy\NXlXD]^{j)b]"; wJw܍)aN"=Xz B 8 hq1ZZ\M4jلbD4!"~kcm"ڥb(GtqNcM!:͖J*0JHY]ncH5g4V cS@HЈO&3I]kykmDɕl-5PY(/`TR (|c{鼹欪M/m>3fjm3\3BɍD%[CUu}o)>&S)wYz5KGZGH}m-fQM^BPYz'XJ!DQ[GKF8{hɆ"D@K"X[5/VW6-f>dk Ĝ@#[d0ssv`K "r[k ]{Y24,REv4I$c.@N`dMȗ*%9$S,ǂb ' h**u(:<h^ve!r2.5 J3TT~>wHsSrAe!:SB0Wsr}/ĥi& A͊1*MӉ8!a%Ӏw&^7j357ټ~T ޫB}ō# A I`-$ >:g@uPe0}Jo:@G*%.ІdBmV 2:䑀d2XQ|^.Ao%CJ$JEV2ӗV,.-1/E2}fHV@ݺ:;Bdm;^XTuqJrC#[QW=!V1l#xaQMZuY 1H_kc6~uy~żMQJ% VlKе2B,XFX;K$rbuԆHT&Pw}Td`8#(na6(X, mM1gT ]\L:(:Azxkljw ,O&U31be@R=j )(&EVc@\cwq,E$ fZf2̅jʐ 6"(o*"6pYXTV*D ǂpNfcrzXF^4ݿ ^{٧&jdg Ԁʬ)ix6Rny魙+ UDtcm{-w;$a3.0|wGç9q473oPs|__t>_|L{}vy97$y_`>޺lѳqmj0f1h\֚%58τ5rFCo bghʌ4MdväDy SηJC6G=T3ʍ[V{uR"=TP=`( e ۚ Y e( =7zT}zC`v+E ĭSN bBX1x.Ja5èsW(eQEqQGb$U?r֋\O` tA౮Mic1tq54[hcs3^Vj G׬U:6 |}^32[T;*g|AꂔP?]ɑ{68B6$;P nF]z!@6(= @PĴ ֚ VP5>o5(Bf{@S k|TD{=-z5NQ!,6:a$قn-m;Ea2E|kKu2x\*iBC,K-ٱuPn") %wR@ǵ1[ Y\+ygAyMB#$1 ՠl5Ab\~ފ[./onn9JNqH] (ѡ]Z#w?=6T7lUAMs)#`'o|veANOh|~>s=mҫOlZ9~ 0 䟷y|:ɿ^^Loon_~M62G q&Jt> S> H}@R> H}@R> H}@R> H}@R> H}@R> {ٹ&{Є+m<|P}@D}@R> H}@R> H}@R> H}@R> H}@R> H}@R>c1x0bV]z> >c^}@R> H}@R> H}@R> H}@R> H}@R> H}@R>EҪMOa=>HfEBf}@SAGH<=nTS> H}@R> H}@R> H}@R> H}@R> H}@R> {}э䇟.Vpkv/vzX@$zd[ҊlKadWc[BۖD-mwC]oGW~e~? 6$X k#3E2ʶ|~U3|ʤDR#cLOU_WUWຒn.K,i ]\BW,s+@eHWCZDWRdk pD)Mᯑ !F *9}o #4 Q.0dv{W~zBDȁ{NO|ߛMu9Y<# ϐ%qKl& ]|8I*insz M 8A>E=7~B[oF 58Pc׳Ӄa-WoY-儮〴e)6,ejNYh7& >ۿ4;g]#XHH9l4l@ʶl@5f~Ng6BZchZKl{ R֚ Q 7CWW} =DWXx: &e8bǡgFWR]:=tpl ]!Zϝ5'F1"BƴNWRutJPJ [CWWJu(uGW$FbMU-t%!;]! CW rۦP;~Jtp%m+D{vD m+,ik *B;]!Jk+0FkӂcA\58s?_v3<yq.Ǔr> p‡q x '*l:=]\p|#ׯ R~ գ4x_dX!8Rd1GI,7$Ҳ@Fz #9iA]>JzwiV"iI/.lr h=jdro74r52'_(Ko N1Ĥ-01xrRҏC)LɽBOrʕm]`ۓ4pUkSM^>8/4u\E^8N~ZF^Cyѕ>tGW=3"Z5R}SAIR]BbY=X/y\-BWSOW8CIGW[E!5tp_zqh9?wBHWBX 8MbBWhع<#y;zk& !ZCW=+DŹ4]5ҕJ6+kx[ ў PJb:zt%MvLm ]!\ٚDDkDut A7Q:d@e_ ʄwD$8wT-Oth9B X-a@3/.V^ղwUϭ?M}3J5wSW5kp:J0}uM_JᴲzᏓXϓpl,ŵoP HUᾛ7[ pR-S{gprq6 f4~yk{[직O]{ݳW5n݁O| -HiUO> Wqv u[*/fޅh jJ!HK^^L´( *j<*% i!HS,Y9g6^-棛]%tQJym W6pH.ew7P T2JfP_nB5s9D)c3(#nOQ T68Lrq>s YjR*C-Se>Wեtz38hw߭KK+ps*-4}{ ZXt2OeT#\)׍0KG`ע=ց1:*+ʪ)i' SN!SԞ`=?u')Ȣ> ~_1ӧvaaٰ;ڤW$@TH{Qp1*Ʊ:h/0>Ӝhmh^.&]PLqNfq&h{n{mh=^,$jJL/5u=릱>HNC6H6Tq*E@*>MVؗŪqF),_oXuf{7c 'b2{OU@i{6,TN]H6V$J~:u^*/ƇA 8t*Q(»߶4QDޥ/Be /}Y봻_a7Tv`CGEɵȂ0~fFqJs TeʋFeC>RVB-!i kmkQ}Zkrk &8B3p 53%QosE,}UfxJ;kݢST]gXPDbsb9fO)XXRbOD*g2DԱhv`(B Zs#2o߽zDĖGD6WҾJAҢ%N΋NT85]kʕKkkNÔPz;TD`2 R٦Cb[pY{ET*Dɺro5crjqw@~Tǐ' C)2de6x;%QTgώztNKֿς}8&rm|BJ'>\{ KW`ǓS*F:E(t$҈YjMh!9L<՞Jb,;W,zVKD_(?v|FmՖ|WĊjw_~ %E6C>%z D wz:"NkUĖf]P6 g[]0~ l7ӊþ1/1Dd͡|ssA;1΍y-'RSNnQ2 n]muwCi GcM:[OZ0)xk"&DFE+@9oѻU+m-Pw-ܟƁ+0)b>0^tƽ9 7Ә,]w720|Y`BܳUwe[vѬ-մ|m1nZ_OBv)A>oR.1P UЄQ /t" J949'{'!z}'ai>cA{ƈ2^c09PiZ+;h4.Y'M)I08FtJ.%%33xC= } {K1ye3-.{g3OATÑ8|qOJݓ:¦Ƿ0aM퉜m?d>[Չb},|xՓ鼪zmo. ^h$|HUW1uh_5}6C (]KuT*롿7 d%^}UWH{9捒l2m>D'|X6:w~݉?awZ}OJ4_=t'Uy(<]}9STW3ç`mzno2ľFQg Rw* . "@c!kG:\vsF.9 =:CL3.Rx>dL>;rlUvXSpěeoeC_eYJ=c62at*KYI3 02꣥x0X*a"NjU$  4 I<̅eI 4$AI4>AkMLhXQ󰢦K,S,Mqs،ؘQf7#OذVP(+p+e:" &jS`%Tҙh260JdhքQFHSԓ F꘭gѥhrN'<”;JpHifl ؜6%r!k YDžOʅ[9$7 @e_(h44o fff#<̻i -vYtY ht4QΞpgH\頄W"jfc3PIGR&A183ㆋy.ݘtڼakwv;v1da,$(5I0$M]Qq$BcKMnK` n1MHɉ_ԛEJ-O;LzzdiXIg %p:;-)c-DLT Ͱl]97aԯcU`?a@nȏ>(A^AAmOѐY30X@JdJ:-"K!q4uRU*$J!l#r&$JغZ3sn(Nu>u6ify%]RHz 8(Q+@ HQ .* srѸwͼxh㡽wT"7'*ՀFdmLEIH2]}--}A*C2؄Pֵ7ZAIbJvTLz8ka%; 3Ƃ-uMR)r*C% uZ2z19(tb '}Dlqa7 qOF/tr+:)~Pg'%#@-8]@)ҚGbE"w-o"x{t8{!J(9eb(^b,X(Ul[>E)l9˺A@Y'TQね9;omZīrf2OHGOc}BW& Ѫp\eZP`$( }a! oӏuNٓBgRK*aFm}ӈ$yxқIV3'!O+JV,+ x 0QLDJE,DZ&+e e p~]3x"u)E|BAwJQ`(֎o!,L&mK'2 * vÊH*֖fR3c)6΅\o! -hb6ڬr΢JuP$BT.FPR.<^fn!wƻ/.R~s1!g&lO$jʺ a(6yb&&Lpd-y}RZ0f҅} LdzArkPT|\EEIhǓ6&S R*may@Ǔ"'Rjmgjl* =u:RmMxOA_[[d1|7߭Z1 S@RY:T+z3]ySgsO Ʈw _>[nhC^>{Gߟc>M~; ~5x]z:3W}5b֮T Eo-S2tqO9c/Ƴ%z1=%u[Y$1WR箋lgyn_38;yyavu0z<~-?`u~ j}veʿsr:.{_0W)|E°7W<73ȟzA[C# rE#?:?\O8M&~ʽߌhѺQ7I$[JJ}i|Z‚3hV*iGm 䢥OcB[ VEP9z'aQߌ43疊Ow^d6ZZ X?m/ 9~3~ ڋͯ6OT:9,<\ST E{2!Z7DR&Җa<7Dh՗o 0|`0`h2z,T=mF,Da3j H.AjG*@9ʚ8tIȄZ2j ̹EEOw^lV݂<bu?qfr \݆뙾͡OxHNE:#" x-’)kQj6NdWd$g`L-6mG PMF})jٍz?F|^ kboq2GNsUE4:iX R&LA "7]vU7{N+hoڨo'cZ߳oxx|6xJ/atW҃'4y짟ՀYͿqtQg}%Ee (႟@\vfTȆ34z<=Mkm/O-ތY~.}9Ie'֟sǗ ⹒9<@ϣ~]uF?~szaYY̿N8O?? Oir2iS٣;@QAtt(: ?NOƓї5Z׌NW_Z,y|$hRIcxR@՛ժ5y?=>XU@[,_$N5~[=x=>hp+[Zj:k{N~Vjx\^\zZGJuG~L sƵ~k"eY2b1Z}Suz~}-+6VjG_fe-eX8wXLo{un^(״hCoSt:f7[n̂юMXK}XJi{:.ncvwk8߭}5p&Mx6ܠը1t]}bK7K1ɚO-_^X}a;mk.,T#Z +@w]Zݹ,g9hx wsXA4a,y#dUQtxYۂYic7:m & =<2O%wolB;R^5&V:h(I %@ U۫|ҺޡVC^X ^Ijψד ;,"v|x#\N]֘*EKKE!#7qDr/v7w\e?"w- 2 1r:. W&Bh5"C ѐPQRUk+b'P0#Rbd څAVB9{/j`"LWkvMUs3U]Eg19p[e}OUP'>E8ɢDZK*2E#X) P1gTN> OZl)+v Po2ʜ[n ?¸!e9Syԙ({ L!ŬF] Ip M 5nbnkFغðE(}(I%\]c *S,R07~Yy[ICOhji&otڰ,6ʒ9wBКhTX"{硒blbl.)SI_&|CV]ܲ4,\\υPb] O+dj,30Ny;"u)&y<{\p=aG wJt陿PgaT)HZK%8g|ɊA&9g"5mZhc̚cRUC -P,A%P&XTJ(gХfL>Qq%mUW@/鑄HԇQG=MŋԤd7VZh Řb;@SC(F;Puwt|ӏ>W4z{?)1E\Fƴnxߌ.By(S*%W)^f]*3$uR89:xNHx+֏|4@.ψӋIL\ͤ[ֆE&}3)W Lo)\$TvVYGTAo|WmF7袭fby^&/ow8J\Xaâz[էD'G~҅/XLh1&gft= _dzy?eco-{9,PRT~>#.Oڙ?מz*WY`E*q;U]+R+++* zo*+ɾUV],\BKxn;?`UV|0RjB:zpJ)NJ \eqJ uRR++WY`{WY\7`]+R.|p%5=+XxWY\IzAR++{v}cysסa4J߿tzPz{i]>ytju0 zuǻw(qӣ_ה"s~KoB;3ٗBڠʊJc.(] $..ZD)6?l…ּa70}$MY2z͝s)\gr4%@pO\'l))!%X'Kt,|7#!0qB`=ep9\OfzX饊)E$hሓ)9gC$Tked9/롳2jQ%6d^qa|`޾{1nQT^3 <%3K)XAy2=uP))g KeQDX蠬a+uV:`x`h#A NIF?ABtVo TRrdGY e%l3 ,GyfS:5pfw`j#NcTF8HM4'C$ ŕQ 9 Jxdv:iy^xsQ'-,!xƒd"$7BBQ#CXQYOnYdYOxqMV^U%ŪxvyQYqjqx<Ź8{TC&PE4(rFc\Bǣa@aG[,Y\d~^z<F~4K~1=EYԏ}5A)9dT¡=⟣).uXbf|W=!XiޯWSPEѕ5:6LCJ9URE>gAxw1XMoGD >~]>OX_ë'Ezh(p^mŕb捎o}y+~qPIȆفzs2d4-k~@`ѵripݍQ+|*&X-z 1ًc/T喹d6M:ϥ e t5 hQ1L79QJ%Ӣ #R}U-MՕyH!_k&vz7A %)J`Pپck8y cʋaFz2-7ܘ)[aWK}i">;N91ۻ҅'5,ПqQ^➓0Ϙ\M:E_H _XѺnf%ah#CeKO>#烷))-.8ń&>r| ȸ\ cLN+ UWyJ^l[؟]sgP`mAeV'8ьO|PIM6deV@/kO'4._cؔykq-eN2_4>qQ1 |kٴuU>q\;r`Q`}"J/dYF) _#&ԯ`*KSq}Uizxy~ cPGxo?wףXF'Y<ԒJBOop݆pvE\rn`tH4NhOe*fmRDPl("ZY}.YJ)N5У+YZ&l W"Db-U4yVY碴QBLG唗Z)}6*%B#ѸV Øb1RN+Ro[&; e 7J3 %)~wgJWμQ5 W×,oMh8}>tEqy< g[@kTA( *F,GuCpV΃>zYf;HВwvFN*Q.*e-*AD'!% **zʬi,I+F9R 1KL^!\zϒЀR1ޕ.rb oZvV/PԼ lAOщVy4LjSYB0R)R @!D%WЙ@& g-p'MOJ8GM.הDU@qIW *|jj٠8ii]3 4j ŐX͘ȑ8 .:j-"{jIw pTFk:fj& j6wH3=*NJ DpVS".S"N _$Q߼ˈA.&%8]1TdMc-?ECY¹ {"1faʜ* 8⌵F Ḷ̔=X:cu!Zm:kL^LMSᐲLu:ʑՁ&yj5.fAUUCڪbj,anPa>ʶ0^a\$/0(ƞbzmr]zUv hT霿UΫJ-@ Ly#.^Jo8g]ᗣ*j~1:qw˘g]յr/nC$(H+YRIl MfVP`nwf.&!M>7d6&y ]βXX KH"1 w*MV啋0(hNBUU~>I׃b%h9:ؔk [ػ?<ዓCqF}<2Ӛ#>W:+L3W:0] vj[x-6ixewIHK eHF)t8 \hUcuRυsa=76pyI@ekxZĜS)Om_mMU7Qh6i;/οEoDߓkm#꽽v `Ź}r:ϔdÏѩ <':D#r ;D4Av^`+^ "ty w ܀#{ewP1N˜yj]:?$E;ZW^-Gb>) 1uS\bjī$ۍa pmdrmi %nJ#\&s?_|'EVHa4Lm#%tM9y88aUdݡk;Ӻ%Cn;'lƗBJ81R)VEM\:2 )iNEdnܙkՂ"9P*Qj#bRQg:EKbLx-UzĹ߂o^OFYj][z} 1EQWt3E [[e'lur;B(ܕN U4BL λrK!$\w(;B*%1 3.:CdQɉ>Lzi5B)FGۮje&P+|m4̇n Nb,y,9Yp[KҲDQSbwUWzFq讖&򱱖2CBkZj6hЪ1팜tO r f W(yvk+_jѶ>I/ ceq_=ƪ0v<,@LW_ z\qƏ:36 7|QM_B]Aߐhc0oSW9Aa <}w[ ډ3)ufNH"Dr)i#T˺XE 5,r5ɀW]P.$*H0+qkk$Z е+rvykf<=x˛Ȑ߫'oܾxͼzOhDΑ)*ֲ:GmbE1lx& D"RҎߚVɇSl,)eȲi.fPt 1 ((FWy5%t$׉$ /E"DƣIs6g CR<2,ذ"M BnB c6)KL^J$B^w%$a,yE)tZT,W-c45S1m-n2$%mBl;e)5`s5>#Z\/c'$ckE~R|ײCvUJڄp3w;; ob% XZVz,Vj]8J?`UI97bS )oB~a\rrp7jetldw.\pWV.isXHM'MbjgLF¢BcR)FrF^֨S7Mv*rO]4=~ey|7vSI4[Fϗ4Ly\Y#er_Cnu_GGϣ uw+T&{~_ /v93_X^% ( Rly>i;O5SWmok}z7o `JrvCl9h1+鵖̨eżY3I)2kYb`3NTuhkGE `-e )Y{?*i "T5 f%Ksׅ:#g/QBr*+\}8(+vX_OOOR[`s[.&oZt3U#,ۮd*6Nю)^6nH(F3ZK7j O v뷜N*~Peo:V/V4*q'*S٠J nb zc-@:eNpiT}vR9?aK#żQ&@˅wR2i:gTL9lvv1^x;9zdwi"L!/g`w5zTWoX(s4fրlS7nЫuFZ_uU;4D=@:"E%rZB@+<*:o1HH2& n6w-ʿ 'ѮA7iҲ}セ!_>ՌmEUH6@ 4XۤdSH`6N 'ژ A%)Ɂ5Ax jJF@) #hB " K ө'E"k'|LTS 5,W+ B,-_dgy cO\g4CSf| *UV7[d6x֮12mCɔJ$vΊ쎩Soi%#_ۛȭMj96mOΘ /6 n_G7ۧ9t9ftx?vBfέO~5ݓϳsyev~6Ϸ=MaxT+b:鎉֏-$Xb+6BLs>%F^5L; V,6%32G#L8l7Yڤ 78V,Gd7Y7$PȀ(z5OF?!ݟ^dsp1`*:/*:O[f,%`S>'%MM *,yS94};jt5@Acf+ F0I3oM @ ̰`x#&\p *id2 S'(3d҉Iel2>w,[PC2,~oύ'OLΥV7wv6{M%evwJ .Ybh3cּASfԢr Z[IZrE8) %0ֻXR&ɪr'֝HKZ* +5cgS3ҝ.3Յc]({]xT]v3xC$f?g8?^ klLCHYE)0v31 cq 7!_fTc׎ښC5"JRJVV'ɀm'|(HE|3rvk4኉y*Z38Tkv`Mq)KH4׌ &YO"JVB Jbfy2#ClEgEu Ȍϖ#("&,k2,Sc3rv֨_cUx28+8T#55b&G!L"ڣ^꒕)>yѠUX1Xu͗JE e?!ULZNRL”&K"3[R;\Kpwݟ+r$~ՉdtJθPh:֋׋^kGk}$BgE, Z Qe+v`IzzTagq>C{>< 0MWaۂW~ zCØ`t*x+*ڝzJq}sX[ɢ~Zqv>q:~n>}ގmtq {0L{fUF8Xg33¿Հ-D66E8뢍Bـ9&pIwTtYעe;st@'“9JJN*Ub<Ō Cp%jy'(*-Fw!lxic-@ @?8&ƇS' Y$i-1A`ͰԦ'*Ww+6O~2͛,.W[l)R/wt78Z֛Y񦳜r};0YbT*ҍ`/JU%I0*#j>HfRpV7_ubBѶ̓nt|<[/]:K9Jγ]qtmr`DPd*S%iI\ylh^4º&{5WLuq9ԷC=~B6%U 1Y#&~N8XUr?=ִW:ˋX%f«>[U8_wl z@#zAV J$քxF2ְ+D} %@~i,/ FA7Kcriߖ8c}`uR8WItfraT<n񴞟ô7 GO}pWOxR+t<)ϣVIxJVJGxFJGS:}`LXI=u浬1Ps<0/T. iiJv|*f\cM0#ieE^6~8M<3O߰`x,킵?x|3?b.zBr?ac̮ ЪP_3_8|E)M^8H@5!hڱQ^1ƈoݧ|KY現GP^"QZN4lB  DT̈O {-Oע"Ѱ+F"D bl4#R,H)%%:u-"[d8Ɔէ9_/]l3iCZO؞?{Wȑ oZ]mȇ 3FCKCuWJZJ&^A{IJdQDigݕf͚Wl]{Κ>38\*b GE`J*XbURW Q*X"ŌXGXτ**f_t DjV>v#q g9nW+_㤿*)V=qeeT-=t]_LEǧi6 |\ 3xH9'akohBȄGnB]S(: i„ R= G#+d3S0U`2DZ># @c%YcK#(N csZ kX;rj@ lL9;ӯ1A=sssٸ;)6MaoؐsXSog#ci(p l^I&5|x q(6 Ĥ*{dx<#yf-p?3 (Q98yT$V1ٳ*̅㿲iU$h+P6άE ΍8aQ Eyg8+M|3 7Vtȹׂ {UU(FU[ZX'hMc&t> ^D) d(\#LMJ mm*:{ݚ rQ䅢Rn'$4!FXJ0TMaR fʽH; |2"VGhn$5W׺RBdM\Is F;"܆^=k^@־jNU>rA;܁_QO#غWS}Q9Oipu?J+1 ݣW{uŢE>հ;hG* p1T1枘ή{W'+|gKBT,ʛZ7b]k c*^(ʖRW lZOFzBK: UpgN:>yդ 9[w1Ls̋>g sy'n,VMjhftpt.n>օ~P(QL6g)0I4a:_gcїZN&ZHQ̂sVbpV("g9e@*aD]Phᖆ~# ͷs_ !P`Áa뛬c/12<_#m':J%8B.T;C8eycarޑ=i*r[Gc>цZ.GkH \dTv t^[QusxMtIr&KP1*aa 6.6Bzj$U93U@|B DXSͤ.\9`$XV8Ȍ,ρSE, 2N@:NaJCj3Q|IF A'nMߍŢ-'MRQ QX!,cN8|Jli s).c]v$(NQNPF} &v^[7{c34C1B9,Mw~~Vޯ_ [VɁ\:|+]\VqH>š~3)OȄVwGJ^+dyq俷2OՒWy6g(odE\*`bsE]dKgǥNSbUbhc`+oEVt-m\R<)|Z>I)(_4rR\6/'Ɇrпw7m~n:8=2݂9q;gwz R2Ŷ֟Օ{ Ej=g׌|ׯi[l[o5Ko`<ӷ~l`;^yTX{s hZ"Zw.g'# ho P.% %<ě;x7ǖ;Pdw/ZE DP9$2D/ mBW"k-?Y3^p㬊7lmYR֑|q1Xobi0Z¨:Yd+Y!QY̵WbƩ7XІTRk6u0Q x$k*鐚flM%)FB| GվܤcSEX___DN -9bE^㿭menm]U,La*Xv΍?kbBlipauh;h BS>>^CE:+Eiզ)Fl9\8zAu\ %LcOb ۽HE_0>;]`c݌َܳH념LN_7Lsyc#b. # *JTpjW=68FI!S 5k7)DhF*|/ *}JRW jNNęRڹMCն<{蝒ߍc+ZLˋmP)OO$zYspqxeD|~ֶoaֆ[f]Ј}`fÏm׻K_,kt}ng"$mަ-  j@۪q!smo{.z߮VW?t]}oN۸}Y.ݵhKd!df\dJMGdQ-$mܪZX{ե\)c<kf.]HZLJ8M-*,c ^v|JaFꛯ,z)r2Dʲ98)"'eab*ũA$@Au sqI$-Ӑ<΂ڍ|-m ,-~c('{oNQD h@}G Ni}) BL V4JXqZU` 86Sm;7Fe'BdXٖ ʂ f(:JPYe1K >\WbM1 jDY#Q#!OűCF@jbmۂ+\vdZHU羃%#j &qYrrp6=8et ` ,qR.u5b7rvzRU'ы̥cuv}Gۚ*R \FWͦr,͠\Sns}!nܱ>ӇPaӻ_EM9>7f?TcC:qc0ʇj)&7P~iĮ렗o}ޭõK{ sGq*|:]>`ù[Yy??ZѸ ^5*xfzY\,)e-T꥜--rAef߻oWNˆ rQR%ʹ$t}׮6f9|Xt_@j澾:wK3Y-̏KE{W/d3ϳ||/'ҋ3jw4i#^'r1)7eL'gפ^:diǨ+QVW4MKeGAcC97}d_[-[VwҚHn*Ra[v`y{N7m}-4)-O6p@ӢR!JAi)1H4m.=Wt؈G7͚?s^R*"k 㸧V8o\Zc,hͬ-Hvx$J/jp/J5_4>{9?^*?ZGIgX"cGIpRc`:mqd8 o#[ko]~Zsrj;[-4}g%KZ־y"s,n_){:~n RDDo5um~c"+lvxtJa[29Æ``Bp\BW[`hI`E*(% ) p 8 Pbzj`Hґ6=Bv3fgͽewSPxlA A5dGI7Qb4ƔK*-xA BZisCGZarLwD,,jϝ!h{XG-4$GRr3N`Bw f"z}^ĎGN#dƣtRlbv:N~t<0=06wzl q˗u ]մu:,~AeZQGsS$zJ!Fs(j% (88(>y-U8dR3J(0P$~|f͓^RΚ?i:qWcFi>A`yc0.g3:?O?MuԛLuW? M:0&jTl-Gk[rt <\ \ucq/M\AU9dN0Ec|}LВQ>/@/ iʏ.xdrT'Rm|y>0IuuE[wT[vmf[wV&wh>򠅹]?0wqoib'">ޕm+-UFK"fh= `U[睉Q*p* W* ^R*NDTXM}%i=۳E4<}ZjŊ2!3IeVHHr^r+RJĩDZp38r?j7-.1 {UϸC)5_nm(YQ]h63O2*,uԖےaK+ҟZˠOږ~jK?MMcJDLX'4[k#qt$x4aT 6"iJJpp^PzD&ٯmP)U"IHArfbЍ3M9w+ CX9̰ "PƗr/?RR5lZuT42,G i^3G* 1C2o zxJnc9 Ip+g=rYAHHr˜f)FF\˹ dabtrfP4TJF9`Ť%B"iUQ\1TB@Zqg9k lX'| =J' q2*X39%BKIs@,ȈAWi Њ@"s7't''[6mVǒ3HǰKXqXqC=US|ItnBtdy9Ea,Yf: d1EDGG#W[{gRd ox6s ;1ƽ5  Cm)C0 o=o9~q<0A[jck=1GR#?0#b`;-kO O8Ƙc|:6$i/9-# dRKJIfx$$ۢ&tX/zlS9 KP嫁y_{Qy{qu6ߜΆ%x3P$+jRWJ[J!8XNv@fR/WlVoFb+sG훗/: 0`^b\Y`gDWB<-1j *W|60ȕtV}i!Ә}WSGn|v6%& XT;@w>OQ^rc{Uwhy,N8M=5L>. 3Iyk} `<0?ǧz:<"uzms|ܻ1ö́rEp:J"'C?yqaWMYk/unS Ս ~W₹F@D%;e W:*> mrkƷJw$A<1}ካ0؅bQMdBȴ8_ cnm+r5N{+vY֮ZSV ڭr :|L}&1"K^13#o,!1(<2f. }R/)^%sQ1yk4M:d{I2LP`3ޚ!5N'<#!S{>h\l\Vي0ETe+ZϬ'-!vZ ʭ&2:̕R3e֜X{e68"hf x7Jm"H [$Xr8 YcAH[gyF;# H֌F#5ۀt7ݬՃP|PYd "u!J ^o%g7n=}LTmS,ټw~ZU@~o\ MZoazͽ,֏YbRx'W ߙ֩&]kulJnRNxJM\Kv]vYgwϯs<{`9}6?]vl䛓|a|0ڙa:loyt{Y›gޅ'[^+޴P?`;O4K{.Сf[nC"D\)ۼvL6ꄭ~1su(_\Ǿ1Wgհ1 ṉQJ 4'p3! a$|>I"J<ߨHaF \ i.S K`Mx[ȚÓn00ncr<ImrLݧS:db F)*}H,~K/wbq=- bX20dSFOJ2e$4k"puV١˛i_[<&?/dϦ/иr?~ 4>2T~Z'-l"Lf:AX1*UdIIaR[:,D%FL`!hzrr⣢ɕ3Yiu@( RtFvJ;,lEbjjƒbo[;?~,- 'tb|v#c dQ>{Db< 6*T1")YаfBc4!16ElP@JV6d!Z Tj֡2Ermc#vGwz1r:aX(ſk@EMւv @.ZH9VeK-ґIn)hU?8ޖU\V_Y-蘼Q KY( T$v>ބt(1R)8jA# '#`%`< ?isfIG8g7OA|㐬 9t` g С`X CwV^;vd55Ͷf5hjiW%2 ۔K.:>ᜥD{ #ZIb5Vh 62 d /'|\pQk-eeMQtP0 X5B:e}I1bD!r..,87-;PM;2F-HBR]v%)N+#[U֠+¡I2b`OlDT&2z4 Ogc _(^JxPZWDN!VstRw,T k3O{1ף dzߍft7qTyaa.&ZCB]/Mcf[MY"S dFaLwS_Q颏ҴciT~>I>珦NFbazI/WX7ϫ~Ŀ1z?b>[Gz=Wwb֔}a\43z,mz=2gF?g.Stuz?;[N Ft.FWrggu?~ bCg=2wzfuD8ͽvJ :'oj-u: Y~/>̮X̽?[[vV%>S r7|tv5]/< _*N˨W@F󮆕O+E_^Tgu=;JWJ; h .4Nz=_*[o<Vo5^z.d~f5{[*.jA.2Ξ4S iV&Q$ "1 `s,_]_Kx33Z|}'Eb"A}!wcK `PU.vdִtsºX9V}.){CC=! dR 袋1b!& eXk xȦjZ*"HރS^zBd`4 B-i SҪo&5jctr\aCnYkb!ˮ+pMvNg'!3wrc$| VY؂$LqVQْMle ۄyWt4jC)So `(z{=s{+voɾ1^;}C͘_yI37E45Nk;v>P*ȆBI`<Ÿq;!'ݯ3{Ƙe A;'ey᳐fşRp)A*Β7 QV!M[>T>/*|Ӧ?]_P1,hm*0?D;|d4}u4(udnZen<8౭sܷq9pρK^\rOk=_5A{K\3iZly\b Du)I)ۮii["S!-q^C2ؘPrq%L 1fH2Ңm`FL!ȜcZVcv BK@`QKԺpV1ZT1/g!rk- a/& ,P]vdeCJғ2 8*M|%iUu x´SC18_a" d0-EDDֿQLoj<" ="s"KdYhhtDֻ\ X r[Xya;R$dEmH,{#% }%9P7zDUI+imoBKy JבtMxj CӐ`Õ 픦/,i(TRZTf_eUBWWN)dHCU M_1|]U:'OOfw?1͸M/~_6[Rx,7f@b͋kY Y/.nMgU~lLnJ_TR$2WǡSo+`ϝ>z--%8,8zlXs~_}/FG~꛶/o\Ѽ;_su={7ek;{0i|^-:=15>1ȯZ R1.OG YOKક80gY)Xz VD}>A NIosWK0) ;.%4z=_;h3"M0gZai|ɀ$k1kyt^(ǽ䏋ڌjywz8ٻ/W&2BarN& ȦFHkVȿmJ2(餌-8F a@ BKL* \)N'Zw'X1 6YZujۦ~89G|wm#I_!v0vm2{;MfI>dsF?mMdI#q2A&)YvLIX -bwUѐ/~u]|Rwշ6PWQS*]nL:#36 SS cϙPL95 $O;g4N4N<6L3'mI-a!xJ< 0 JR|`u͓':qtϚK$}94XhOFZ ɚm 9 HHgl'Ica -dvSp`hQEm[v}@DPYA7W 8{j1F`'% $FEr$a8+%^PX|D6YՄRα$тiX(녏QcX;.%'yg0)8usV58Zq?L{_{Ҏ{gvx&'G{] B|atokj"z" qou?b4.#)+C'x kq4T3"iz=Q cԍ&ǰGdG[{nŲ~˛TǓlGq{7cx W6?N8歼 >WiDCzG.0xwa^~ySi9ϱ8w—(xqLGp4'dHyF'gAxq>ɺIGg9D?@6\zWyٿ`/ɡ!2EjYc~8͓Sj嘼_WWG@)ȆځzՊ5>UGwc71_ptr3C/Yܘ_PK`9Fb烛Ǟ 9?M:?WhW2+w}~[N/Y]X2zR'?Vq|}`bwpE7 FDZč@!;xu]XᎬkF䞸 h1 dj:DE}=l`%reTͦҞ wGu͊G&({˦Nt>_˺l+&f:P{63@VPT{:Æ5:fwo:gqYqwL\Im^aQ{f0&l&Vl-v`?xS\?ŭy]/9ц;``yhwzlTEgnO.(smJJ L@u`j#Ǡ Xt9tR Hm)*Ȳ'xePƀh#q)sk:tĜaM%XkCq.zlr;lGeUxcԼo_E\jR/5C)0CW71,iGArmEcٰ2 sIW"y@ΛA3Q 58`~+ZگI,$׉u1"H-O"9%pùNI#h]$KM Z{-v$[Ot42Gָmn㱖[i1HZIWAl$\XҤ<#xA+[ Z2=U.h^ B >GU1ĻW)h3br̋E.X 6fKܤmA^i@#G[yfV3Ʉ%wKGF4h3tcld>%bLċܴ;=N&XNNJ)d^U)!v 1j84Zzqgkǝu08}y?Syĝ69{q9/v c8a"((N,I  m$Y,gBS0P *X5Μ6;XbLD fA2$v̶ٰJطЋH#֔(Kg1J;*)q62-*Gk$&<|C4WMuhαZ`+V'DEG*aVdh 6oX+@ErU2~ݝ폌a+sp6^1L|\Z"NR/H,pdldl}$SG񣖵'[A me gP'w Dr<ˮ)YP JXDTk9``&8˻;ɾC ;mUk|k5X)S#4%d Oh'r&H4 01򐈓LB+X@)IĉLJD KmAK C\㊆'xYFe+(wxNtw-xޘYy >:yOS+jY [Ldyl4D|x6`dd$'(rm)ӄ[ ?1^Dm&ip4K"1<m$'E$P9@SmtαN> Tha4CWwӺ&aG}u#~K0)oDʨUrQcFCCJZr!{2wIݬ/oHN;kd4HȸoKSrTx,Uض}9%Fx2RSڃ@MQZ+β˒cM״6n&GMZJFZE 41d(8ς'}=O3w[[$ҷ&A%%a*0`G6Lk&ėVԥЦ]Dsj`1,2s$Fdb&`- ATUqb|PUNk>e{阷RPXXt:!u`Ȗ5rn+ {RE'r.Xe"Kr_MA|Lre2!Ӓs . yԹ!w)EI [v0gD)( #,88cp2vwa3&8*hǣl_ [VG]kۨ|$U6CYt{sݓcLMVwNQ8'ԓW%U_8(}vPygSч^Ƴ'E7PfIR|wVRj4+=VwKgr\ÌE{X)g&u/ٗuЫ|{?ծ.oyE[DvCךV\jӼs"\[}W +^U~Vwnޫ68?/j1./wfR=x@/{Ӗ֊Ysѽn,>]:>^_q?mA YBZ8 Ny±ȥv%(^v}5 gP:ϩSQwjwH eRȷ?t!?;-·+52f9\,]}͗gx6 A%WV()5έ!,C1c(D9AFY:&8-tk' `igrV0'?bS}8{ ]ЋZa.t& F??Wx]߀z1Ρn̵טv<[F,8gPo_V,ffK0נyaHTT"ֽtk=d՗O}TTMcFuM[+7IdaL<L-2KQ{:XLRX{Qob/!.) u ^G&(As(Hg8 G^kjE4[1G[?:E%s4i nmn5=ӣ~pSKfikX'}fG_[#KQ 㪥".kZY~v[󤧵B|' &+N$W&a np',6O9W7+N 7 K+ɃԖO޶U2KauHJd/XMiu(<͐~7j17Ac^^OlnMwhũʼn\]ZyI9f:ݍPbqBV+޵(?L>!Keʍdo޲VlGtEWߣז%^J[UシZDc+u%o/<ܫ |Lxȕ(cx8{;Nr7smk&Om7$9ïB,|9fFD,`ؘ+_E> [[2nSEGΐLx𱔏=xj->yz٠qhr3QS8U?QSU&^'hLJ[Fl6׭)׿^<|wHΌt|iKKj ;5g.3$&<|z0VF4ԆGk Ŵ4G3ll6~.nGk~[l&"7k7:2qu~O,eNF~`LozNjX 5?۷xu)ȝ.?5O<hsL~ijbw^Ci5ļ63)y޵^#YOFT"(vF:3JFj[zF*ׁ({60I٘jܭ1yj?f‹•+MopֆaprGp}>TW'?Wo~lo||NnI ~񗛍==v_f?xo$|1璋?9Y]؛z6]vOsbOC0lVߵ~fU bd3s +N rJ薎+f+f• >yr+U?P*:A\9 ;.Wl#+']ZJUSĕwZ?T%w \ZJU_qu qWmW*(RaAUeZquF7]݈C襑 ,Ս3Mo\Շ+TC[_W3:_:gYsiKrN5r2I:Ty>5=#` "CX?J? g^7$w\\ApJ9Cv\ZY#zUi+ڳKf D%Wȃȳzsf0\ \ъ6pхapape}0+=pI Y#y~K3q%)R +(Z&t\J+NW.DF` 2 TmZ|2de ʇ0_{` ~\+UqGEndY0Rލ+U{Qfr+NWHzж%sg/0S%3M-4y!hsͦ\ak꺂f=f5ġaN'K?$kƭɴt E`f fg W؎ɩZoAu)fr)&xpD~\\汞QV\0=/L?N>Oph1Է˅9l}_߽eI=)L.N-;ۨX_x6V&f|c;{V3a؇^__kE~o~qb\|J4e[A[[));"{)8c.\Vs="6v6-`Q6Wfaoȑ F1M )s}/v6lQHgÕ4ӰA&5 }2 aڜM-KԷwd :魢YP!Sh ETzϽB)XM7VɊtK(FKa<5N)5իQirB[-&!:.\\1SJ0-$4Йьf>nTV!ZRMi%gC ^{G,˜% ͯvٶT[2""'!T {N V{U2kSp΂sÕlZLdK&ck u}G6C&&nN'ka;ΒcO N?9Aj.!_*o4R`!JX]K/- h4)xAkk(mE nBN)OZ֠ f %xXkZ؉BsN2ҊŒSD4"JKltⓀ+x1tS#DQ#x7O#bޚy#HpP3D%K!ʃ:_|Ӫ" 0 pl5N]b7!AiVih,(wp{ 0 Bn}SqN6 [J1PCMPc. ޻AIT!B]bw!N'p Zkl\ jY1RL(E +҄v+ŦC-2()-wc $V Lob㆐oȱLv@F>lD=5DB;x7]?1$muDbh)HA (ڕId"X'*[K!)}`&SG`CL.֤ љ(͇,-OyD`<z/d7X:3%\pn SN Zb`pHOp_H:;4Mv \q&D >hԅh}kYu <K2z@]B~H-\;CDׄb3xY&`AbۃD+V&N/0;A}p3:~Qw 4tlĈ4c@Tԃڡ;m_r0Gû\5ۗ:/r`b*z=z~pf戇ty@1BK6s ^%fp 3TA<^*4G''Oi@zvR^ ucc4X!C|o^yE`1\ƒr8Hbr iL.:3A #v*ĶC!D b RŢ45g RO` T.Cc#dn҆EB;zV19E@KkLJ@jXzPpkOLA I "tDWMIA1X~f]r5WhKm5Be_j%j5"mP j0XAa},gfC8 ZV2#a0_Ԁ@\DbM t`(^cOZt(D- %*hI%@b~]"`0XӦ>rIbURD 0ebJv KMH"8DԅK=)r -Y\cy]Zuj+=X0B R GMWKnZgT8-ש0}ۜa]*vZP]R@.-ӛ~>t0֧i:%0Έb?󗯶n_Zf>qt-) vg g[>%^fE:7Ӵ?t\ 4g\}?[,vqyq=g_zhi7-ړ׋ǽ.I}od<{h2K(,^ ʸ*tyaP dNUNRpV1LVyAkFΆֹՌ|@̉D> |@"D> |@"D> |@"D> |@"D> |@"D> |@"NN^ptO>v{Ж}@?s|@"D> |@"D> |@"D> |@"D> |@"D> |@"Tjkca,W֚yzAZ|@"D> |@"D> |@"D> |@"D> |@"D> |@"N1!?d|@rZA> |@"D> |@"D> |@"D> |@"D> |@"D> |@"ЩKL}můg_4mԫZIʝJɋuYMϰ bL%v#:%=ؖ\m (#c[2lzs\xq:/LWu/|Qh%W/CWǡrXte+CtԦqFDW  NW PDW'HW c -gWUC ~>]5] ]I#؛r6jh:]5] ] VXUkFCW |0PDWCWڷKGCW .WcV^]P]"]YtLK, ]5!v퉮 ;tuNeHͫv,tJ1tj(':IJbTnHm~=kۛ5s;ٷ{Y u 'mPgL_jyٿ3uӏ~}6-׷f5gqqs4ޫo?; n9^"LsYYKbw|Ni^Y@~6boeg\Im:[,'iqqY*U@-gK/:ǛEz^aj$ }n=;R5)+{sv777%۷/NSpgW[iC~O, }}jO黿}VcWE,wj%DŷR>5UjO0Ͷe:)eT!׮D7ԜRY-e@VֵpU;ùYՔ> ƴx:yC+PjMG6=g^,}a:|ᅧ*2tuJ=m={]Y6=0?"jxf,tКUC5 ҕ (BUGZNW DW'HWOA{+Ϻ6,jh:]5S+%C#xGCW˱UC骡JCC͈ UX誡5ltPZNtute1Zr?jp ]5J/Nڋ*]5z<`kX誡uft1+۹TKs)׌#-ތkmMY'iƞv~_̏z~NchgCНwNܥbX*5N3%F$`F#\7E^6`d)^BѺ=࿶lٜW#^:wݧ<;6,$0{ݡHw=Ѩdlͮ+2NI`rbfͣӫo>iݓ}xsDbioKv-tQ{)ϻvl,:؏ZC{w2Cx|Mp=)D]s}"TOj3UwכF[6V*C`{m]^\TSQ1'ce!"c,jkŘFZE3g[/w:^i_N]5{vͱ9{vv=IovM\ol>Zi3 JmRyW_N!Tvly}ꁯgw7oay}_gƆdPҌFX[RS)<~.mÄv)}SF[qd*ƽU*1FXy e8fb m-p!XʢՔ'kJ]L.B3MS=|| 4R3RN" s B%EJ֔c(2;>ƧM^=R}=ifRaAQ⮆0D+*>bdBv:J'U!1Bg aS INM NIdt›Tzír4{]U~KkjX~972rOnٷWO w뫳^tQAPO=_?Iq=t#x(y1]]@7US}`:M' ,v1DTf3DއmQ.7s_|pMEG=uG.|xkaV>凞}Ɵl0%+LZVha1(yMLRT<-{] ZsQP)Uh+KUTUR: x9_aezq9Wftƣ5"=Xny|X+5ǵ+}N cz^%bf[L^ٿɢ!Cx͋EI\\ɲ`iLo :N:UnPe_'$H煤ޥO..V ҂kTRgxSkR'̡"N\hTrȍLΧUϕ~pl@@X<)CHd$+< .SL +ِ EwzH2KgM- Eג>zGݛ:7^yC^/$12{"%1oBӭS wl| F҆[A>wJKyxWT,ʐz['#y'S,6W+ZdxaZe](j9g!4)dODـ]R*]UK2[/0 oDP' w\gop;!uٛ_}q;ހxثvqwwrW{6sV5ZԾs8t-vh!pc5b5P ޶CxI'fzkt1xS<hx(RWNA.PX]$+NF X;/8Y_8fz(b{m뤣m8=)VqE/Yz/uy+*6Y,o)to=蔲]4vrS"7KI!' !CH9R(!pFRq XL⼑, f0xrj*yWB |%8]:!E!u>f *+ڳ 膗$9$^¾s1ҫ6/T|x+sXQ$ca܋ҫ/W=m>k%58t U:v%()Qe3\$j} p? ,=១'b9 0dzd/mGι1;%YEJS+YPoٔfqu*{Ÿ FʷJ`Rs%GD1E4S>ez ::gCpE) b/f|e̘*U"HeK:3=;@j>jFƵ% f&,2ԝH1 =3J,b2O}mNQN cJ^FϽNX)+JJȞʰ=+9B{,;u;?{Fܿ `GC-rY,.9[ꮖѤݯzH-ITӢ1`[4kU]IBg8p= q_A¥ WĹ{UYi0>q$}iWRf(]1L9sE!ᝎL骦6NR7W\URkM WQ4& ~`O ;ڢurn8svu+c]nW̋)}X}AH/a5U@ 51d"s-=XNt297Zzi+چqA5݄ԇ?q~ Sq{ f4lhOHB`nzɭ{nۖ[?nFO-M8nλnzrW ) _lܲ o6IkA1F]vغ-d*'q+XcG_[w̺CzRbQ^RYt_'n1_oYm:_͚q(R{ I O,jυ1U<䆾>Jjhu\W%֖ƹ1eh.`y`$Ry'x#사ºI.pbh N! ҫ6bJ`4If,Rvgga*©UE(Cb,;idR\'ryIx0\yKVe,cY/HuFJ 4>%rB)7:#-z%u+>KHcD!H8LAryd<@PJOO$ST1;&PpխSWw ۽uQTe8\ەv j~.厇1F_3Y8e  p,5@c!#kKéwvA5 v-`Ɓ-xȵsN<>hQȒLWYĀht9p'R h Ixy `gE9{FsO&&h-q`Vmn?oֿ7.Vml.?o~l õs8fz ?v +],Ito hnCB.޸Ac.om5{zz#.D~} R%`kڈQzmUAEŬ%RKu5/?% " HHZ:ZbeAFりW^3!X~S'f(^7t܆f_|Q%j=(nM <%VAc$"c JY!2g:օ0ٜR`ͺq.e.:mS RrVb (ZLJfF㚱V kqơPTօׅGՅk9wt@O?Ol8Mǟ+؎Yx\H(V@hd I$ƖN,^HI!EQhf0*#fƆ(qǡZ[Vڲ=݉Sps4#$si4 `Kì + rEVthI$dВppR#HY KԅPYV#q}tOC5U{3!?8/4*UR.9Q9^u2%) a Y2cݪqƌUEⴒrHB[.F2XXnQ5r}@zq\:%_g5.9T/zQz׋8$!Ttɣ*+EFϢ6L Hd4)rlж׋ЋqǡTև0}xV1E0]Ap}sE?2qcw](0H0*mcgE}v_vV7̵s] ~>j~ۂA7O]|bW{fwѴSa^um(-& O87>&KRK19˓jӛZ)kLRdLYACx +1%k=w{=,ڋ3YPV'Ӟ& h:/m%F˒UșV)12 HЁg a*Ϲ@c%'Cg7#.t)Umr-#+2c vS6/‚*~eP$6Y:d$RAj+Uɸa#"^rI^ΊWEA'*Y-K+6lV11LImEP7+R812$e"+ˮl=y1"uJ~U#gO9;F:i;v0*]>SpVhI8Ͻ\KI. I'O $J$ Lu/"PS>+=3.zrl5O!1RB )k< NT<2\2P:Q#IzQEqTjɨ., kz'K0vRqȡQzo z7t=Ґ, z¶@-ظg }H?hW\G)CT*7>kmhiAnދmS4]`@m2+JN c˶l"--73|sO`603́IB`v `aFJjV!˘ 9$VW8oD I܉fw>Zr=ùݴI2`u^[$ek^O'<Ɔx=]ȵm&3:9ў#Cȥk£ks,M'&$%IFIE>% &Hz$ _oح9"TFܐW2l2&~L*qדRrl30rn9kiYyVn~:9͇` W+ _OʡNrʪӧ;f7LytI1w7)9lGz*vrf 1BKpA`e@.hP((TH X̧ <`[Lz&b%8 -LOq%zg%a8014o%mb" gCz^%No<]_[P̹Ш|K{du8t̉ +!HCU\{YQME,@Q9)TZb/퉧)SV1 QE03`#$121`%Q<0O4׍G?^\!យZ9HpcQAcIx+˴0a237\h %57 oQ˦ф@d5vP0  BoíJwfϜ)eZ2 21*pYBkFA*K0sK~u 4~mƐkb&Aĥq܅4sR,al E0<9x<}~ggfTUGGxx>}P#4"D Ӈ0:Fc#7W1ꦜԑAeO'y1`,F$Bp:M-#B'evQ係j9NI]ɢ^=ҕa \@\40^[}K@1:"X#MSh ^?6E#780QW0ߘ#}W'h>_Ė` 0D3(1P|~O#ly,l+p[oSZ=,𾸈~YG3J&z0q b#s'QǠW=j.#Ԩ-hƦp %Ы d%G\hZVnhL凍E07VwA/6 `3qQƇחp2Bu~lL6_k1Lx{ͻ*7D^|Q M,uw a_ar6`>:٠s 0-Hb' 3>#,vɹn)|SDs(-a&,c<(N w>it1uF#AlͤМ MTmZyp%=ߖx8ZH痿oJyyL7tL56#CXlnr3WڡXQR2ҕL]Y]|dH˔Îtв}KuePMy5Fb9|A*vfW7֑#&}1칕Û'E>:s-s,xrw&ˤ* ^R圂e%jFV- H/ "x}9ZhŊ2!3 BEP[Rx4u[fH/JˎVEޘl7z!sk㔳sUӣ(ų9ı8KY#`!;=•+Y#Z۞P._6Y,+:a!5 ]\qW*tQ(DU+YW bDNW%=] ]1%+ ]E3tRv(9st+ ]Eu"ZNWJpYy(]xAw*մ+t% m@Bwi] 8 ]\,;CW_oc8 i"98;?_i~=6e.hM=&˘N>qA 2LFJc*PXw -ipv֌.@NW.]WDk?"nCOBnHW;Lђ숲etEjQ =Q:DW0ǝWȮUDT P*Dz:@p;DWX)> NW޺:Db+% }nw!OjGXtutőzڥ;`xW *%NWRl7\OWOCWc\;ݡWUDy Pj$z:@#`ޝ7m%G+о-l@HJ8|P̐[lp7"oD3Kv Q AtLt:dvQ4Y75+!he7e^<nd1+vYR'sDL0֡bY[gyxVeAI%D@Z3}1 n"θf$3EsM {M_cn5QPٴ[T'GN_eY -|L,?">|upZtV}RȿF7 O`7[B= WzL^fxwG<@bԍHڕ\>P0c=H󉏫>f'fNVվU \|Q8j 1\;XBRZ?,)WKqwۊny0ExumP6DW^W.]}!)۳Ǜ<wp5] ZΑkJ6/w Ul1_#3)y[Gԫķv}pkqu]Yԯi<) Jsȃ2kS6hǙNi8Z6{s9^4CBt4.^ĀQ0ZQd4akڢ[_+=*?nގ?~+cǗxf:.VQ /n`[Y0KjzoL7ޮJ^6-]& h j'i.wِG^ZHҝ3F4;o,,/̊yn§ŋחȫ\71Fտ~Oᄐ87Q_KG?)\,?gf.|f<}~gOUuj}TNy l(?B#ja23>|DZ}u? | Q )'u&zWA;1t pOR, AG|yӔr:ywR0ec nVrPUWOt8>ufn+V!<V߆{gz_yh~1И v'X`7"ږeǒ OQ[v[,Q i[֡(|UGR#}$xMBO~[}dzg+zŹJh:4_R l]] q_1L)d ,Js$cPi dRjϚIzͥ~dJ _4? H $C"Y1Ġ "(0cv ƾ:J0GA3/`tĝGcv5WI޾?n_'c$ثq݁:B]=3 ]l -UcA8шVƊ|ό3"^8{ѳrݰH{xɔqjM'|/i1]ɁToL*x u5wÊga#ss;]Hf_ ?{;N%˘0A).NSVN,1 0u\^&IǢۋ:jt |;<󬉪fHX$=Q56Yƅ@6`^ik+k\:VKOL{$7*%uFU L#ŞoTwy3ڸ.[&L}tvAo(7=Y7ƅU0YɆU z=­r~]dGJͼ ճ$|xeW_Ѩzxz~i0e2l.Lpz}t ܷ:r/$ibB%;8 Ncӹ:t4Y\V:;ahcVÛW1ѝ-\3t;O;B3e=HQ3Kȶv<ۚ+"W7IE8 DmS$>y"t;Xv.Xci-[єۍB1HHb0k3U[p$Qpc}P>/r]9ENhk*o2!f_oNQuJ6 fyV![ڊ[˰BNY7+3|fVk͗,/Uǯ^z|Or/*śqd8ѡ @:) $r() *'$%A9v0g|vlQR4H\%KH1; ]/ZˍU^]v9 )/T.EiEP \ٔd1^p6f!|l9[ٌ|ln  NW(}A+a,@Ԣn= doz 9/ (=yUq&FPb=IWg48 Ri$¤j5Zi:Nane)C^)"Č( uB<֚FBJ"4^3hoՓho1^qe@vJ`/-Lt*Yʖ£h"/X[Zef5+׌Dvd>\].e=@6fe[]xMo(wU֊ٞ=ڱ c+ьBvј;TŚ`tv"fCn_Atm H԰e@BlB&A8k, s o0N1AD"a.,vie E&dN_%4cgGx'Fk( >:/Lx!;'G]1.{!}d PӒ^ZQa"-Jݍ]lQ<@ꐧRPhC\ŭ2Ϥd;@fk7dWAУʱGG/Hp+vh7Hڦ{bIvh$.=Μ/·=via=4C4{ju{+tF}i4t$<™@yPE/y<Ҹ?Ͽx*sΞVFQfPk,Ftj$M$~5΢qSԊSjR+jڨJ{HBJrn+/+etUNYSt~ ʎ-Έ~w:靿修 mŐG>;kR/sbOTq-׆g_~qv Z+Oth`kL**A펼|g97f`L3HBoTLZ$쯠=ġwsCc7ko=Զ}3H͋3n{#3=3"eZY(6*``YZjo.x3\f/ZVdcM(F"!xȂ\\'EcT}s؃R=jwN+^>_v讷?!颡6ڛ%~4oʤ`ʕѥٻX% U,2mM98aJAKI!ЪD,>I_24* *j9A|Z<\{u W\gM1sm+zG 9Tu~4,gdzy}<{6 EqfȢ5ij4iR y Ru܄T3!ULH IoUX_ 9Yf a0gM6蠵/ň`) mh(  7 GZ jj4`$hUi @-g޽3N.p~\l8bmFzo40V稇-` c* Um+z [LxF+3: zUL+j0`4"x╲;B<,ȓy/^0,=#&k)+9ԚV$ɖHC"ED-[-gco!H)^,3R8TFxt>zl!F+!B U"I_B[uux_R7ܦЊ.GtQ/+%ȌP `enrg\=u8ZW&&9eHJ*Z[%픤LրD5>#Z\c˜j"?4}~JC[Ի_h٧/2V=*lhlN[C;:ˤiEy OH♪hBYZOV R̍,$KZ,9㣍Gʈ79k,"Z1BeMVVB)ŘX-XM<(>YUixrSPmw,6k@H* Al<\%#$f!(&$]rL'I2 S>yFuw (5HACt% bK,`@ۣ&($4#HWEeMXM~n_#xn DַavwG1_bi:Ld.]hd dMdt`Z@{0&Iw@P]q/Š\)Nde6%VADZ YPG+,H|; Zz<U1k_,J*A@ШVx12ZGJ b|gyQ%\d,|XέQɂ1a,6ffy*t4g_\mrՊ3*dED>D٢$1'zf9 gX*?Z:bOW=Kc4y%R ImLADZPdZ8mr(P. 4HJZC2(\:rgԵ^i]r ,Ïkxe#ÓVO#/ɺj#o{fNleKٮRY]` wY|r~.O[|.O[(eQ]w~d9SՕw?)l.ΔRi*BDԻtW^Ùj͠>QT̎s){Tt*b0db}Tۜ!7pMF;\Q(0|!H/6Nifc%e"ha-F2Qls~}ݒ-޵q$O #wn ik-ZJCmI$51D=_OWSU]Mi h)V%iCqNrC>;9+TSv}Zn >)H!E)3LQBe2cac]f@vJ4_^i=ܽ9ur8}svhYnm~ \;ԋl47EMwR' Z3H*ƕ "b1 &ki{d^+p9/n(- h#1dBVVh X޵~eɒ7ݗcoWO)EVW!;Ӑ8Φ8 ,(|g3n,qi!dUV@?yeMbPZ :a6vQ;FON}љXcYJ5?6mNY .-lK|AZf<3)p^ ȁح ))֦"#hU&Q;sHorҠP$38LW ]w̅|½rӜ5izz3~A/.&]Og_9c[$"ׁ|,(ݔLz|1";>hm .>2|L'D-pLAێYW|D8¢] &fbj]Wұ+kY[ =h)6Ի ^e[,ge `lez+coخcWF3q`ć$^yRe6Ym,Qb72EH ZF9&T6vo d$/3!heYƈ4Fx2RE͎386~qt'_ggR+/yQ 8C(hm}6h"E%;f3^Dͬ4"Vf}b_3ؕu|w=PXUUa я~L0,k-?6xAأsPwI2t:hsI a{(*UbU]UtP V}Rs"Ur{+iiF -s!K&]Vg(Zn .89ڃ)gej{qx*gNXIK ˨SpԅI*K}7)6dRh@&3MAͼS k}#uэY7xmPoS_ǬU?YM;ajn2uiS͒lc.!EIR')&T3^ dFw{yZIy:)?\b=rgD/:'yB8(43%\bQٔuPvLzdo&'m&Wê"ps37FhFGSeNRu`pGdȲv^ ~l ]y7FGXYLtkX] #S Z 3.gzLfX;Az3?׫i\/Qk $(T @* fkΠUYØ6Ri؈U@&t P\F P8/9@E3޳=1\#8+%NQÕ-`?&r'~v8)y4_fznLN[ kI{us9;.ѿ-^886zϭb^7S_6ts%w}14oھseHfU/+ F6_{|>5~6VSHr9i^4IP&J:1\_[vt=&w"y_h[!|Gߔ-."^==[9Ϩ^9ϴ%շ/!M$Zlh/GsvoϨ8 2|a&~*$ٛ҈,{vcO/;Ot? _sOk[jD(*6ϔLd/t<_'x 跫'F×$ї}UWjۺȞ=(ų)W-|qfw 9Ԁ:ۅ|k"j{_ޓ7zٚ_-B>KY*~jF*P.-M}CgGnt*ru;˚EDa ;z`qXׄ9د6'\ T="F #\d(*$ٺ\aYLa,Zj;M{iAqc*ekdW1ϓƣp+ϝxyv}ٗhdbh.`1dj[b08HH0{wgQ^~׫?:8Ȩi LFr}q?e1VyklSlijx5Jl1bVbAAK6b Ey;ټ kػr1uO)Ko dQ)hw"́4mTg?࣠'W7O,;&(~JWip:Ve2i( t~.erD.lYge֞99G'/~s7\Ⱦ߾C5vnÿ!ˋed74 ZnXF??juM惟-clrs;tNgx_I^#'^1mkJfr}Jh'J芓5+++d-tEhZ+]!] z+Ѯc+B ]#]IѕDa"BWV40ҕJp^]`•BW|#+- 50@=vߺPy:2im++lZ+ t-b=Z[(vt7f!"!lz-p ژ| rԙgLj2_i^,cgʜRbo>o.zgVGS?Na*P'1nzNQ@Q,_$NyޛOgm/tA< ^k^//EhR5-jw~O6*;[ɥ+.ZB+uBFpҕPR+ld5tEpZpͧ=ҕT3S]` U}+Bi@WGHWJ9]EtEdVȾb#+0"BD5tEpmO eKz%S"BW֨r#+ kz .:LL;soz抯o'dd:; .^8;|'S&#5E/.wJl1>ڱʐAC͖2JAK kTK'\~eeېS(!4gHR㌒M{^=6DKȧ|m G|%ؼK'ICf_'һúW,d&PQ`}j󼎑ZrXNpŒ#tñ1ZrN;YEtE]=ASWjvhtE(4Еp)3V9tHVp90]m3{2L[u<[Еjץ#.HV ]!ZU Q|tua4KY.´ )O=5#6WU8"wq$ wzdl{  '(GuK;g8$6ɡZeAD {"#+듞^Q=vHHFww\*%L4﮷}\; EQrV!g9*(#QEx_0&pnsh t}^1Ozs({tnK8v+~Y z!n(ѕفLv=T vFt75 .͆{bI_V֩_1i7/nh;]1}2/BWZ礮XH]1ܗ~vZc^!]v4\)\ʽ+Fz1]a6tp ]1Z%b5ҕ%$͈0ͧdv6ϮgW2tuʑ6nN ycBW65 Ϯ^#]yImLa!H#HXM;ܷ2Tý{75! (af f2 Q7ή[A䘓z?=sXnBzv jZ]?NЌ(C~XY\=X?]F ˚boON͐<S.'euz~U7 <ϮN_YߚWn8 ?_j[5i @0miMn>/8eP6f H3WL%'[^ڴqC X$J}87eC8Bo߭H;f;`jmJk=О&Y) ^HrBو5;WhԾ5FiTkfPPPXxmܗtd7/V垭ow+]˶9հμ ]1\BWև}+^_NWhFtŀ ]r.th_n(mNW_6Vƛ.||r2<Ӽl6.۷bAn6pG_k{7ŸG%c0;d-ONr\tRoo"_mzvO qVkÝmQ$ K?^.c>^]q {ޯzu- S3T|p?{=xw'[^f? oiz,o{rwl˧`TFnQ>0g4(GJuFMʤ4룴YFMw%p|5^m )4lwdׇ+irby|ZO7p)ZբBl T}KɉѪ8-%WP4A o]Ϻ/b"%$'ф2fE,B[2.TT.Z5hZ$5'?BI?(4R iK!W(iqmMT$]jɸHh5BySuZ N%3$R6JpMVdl)I%eGѢiAӮTNb5crIJ[׌xQT@F'()SS%*D c'c5ƦfHf hj$(:LYTZN1 늳@x$%[}=K[!uHJtw4{,4)hR!TT2#>劎ƨ2:ּk94bH6Yqxx/dp" A <4bL~/XW'1׏oYxc*;mJ Q!Ec}J:'erZΛ$,U%X3FDE*Qh:xѬɵ STBf@x1Oɇ9Hк1DDk091į#Z)|!-8643%XaS!8Q+Q49KBL }T[2IeSq%B`1zҼuړ(.zцD v%d)0rcݠK"etALڲRȦB@*=+٥$ِV!_&eHzC6rJrF[QPQB m>p h-S͡mG :8FjAuPqLJ6! ?^J.G[ W20s )(yE nq(hR(ZByPK$Qd"BU 䓢Z08XrOxL\th"$Qku37<om U|XTunR% 9UϳW5 V1xa[QM uŸ́ Hcvrcply\E4Χ,}XR BF"G!A]JS0 AJ,.^+@J'%LC]I7SPA>ᡔ4Kq-Ah' y.!z b@ l+ jw ,hLMD0j!hk j|ӣ`2Б%YgڍEQIIbD" ~PlCV@;Q5\-J>0l" D%M ʉl\j%@Nڲhʚ& Tf9͠R)ԤWoET^ "n`!-!MZVH"t0U>oW^KTL!Sj1D:O@e ,|OkӺ\˕I4iѼpơiFh%LCGs%,)(CwsGAEME;T-GQ"e%6f(G'- @rRZTHpPaPD6HmJ F>EӓJdp * RB!!KHJYHX7mb0l V$b%*cIB\$\a|{?C^2[0jaO1 ,ʨ1@jJ (.FB,z΂ U @  KT.~ zPc@sjg6j1rqo jITAR䳗;d$1rEI՜ZOk-kDo)?>@c@TBmEwhUYTڠ@8XA¤aG9 Z&]PZ8hQ+Bbx)AH Gk|T '=ZMڃPCX4T*'In57E!8Xg7vҘ%UX9ed$hVCM"Iǩ %:)J-, q弩\G%~!(x)Xo 0H5φT/i;V멀0-\[B$%-yJthn]8}wݣ yKY/*~P*xZPޝj\]];/jlz>9rv™ޭVGoHȣ o9ήVϴ:{]]hG a?| Y1qb*zLSNlhZ~/tGJ/gef`|VlV@ۭC]VnkønvnvnvnvnvnvnvnvnvnvnvnvnvnvnvnvnvnvnvnvnvnvnvnvhB~Nv+G^n4ݭkt8 (]Uڭvd:V,KZI[bֲղ,nA2QSTY,~EWUVnUVnUVnUVnUVnUVnUVGיn[eqߔ"UȮӭ@Ju[]"݊/a ݪЭ ݪЭ ݪЭ ݪЭ ݪЭ ݪЭ ݪЭ ݪЭ ݪЭ ݪЭ ݪЭ ݪЭ ݪЭ ݪЭ ݪЭ ݪЭ ݪЭ ݪЭ ݪЭ ݪЭ ݪЭ ݪЭ ݪЭ ݪЭZ[=/Z& ~}#z+C3V`>0_>+ 62;,aQpoqYaUv؃D鑹KsU/*K{U¸ 4W|>'=HU,-S]7WYJQ%+$}/5W q5MΣ+f\]R(ݧB fl/*K+yURb.\i"b]gcՀ `+^f\2qؠj7!-r榃1go> fŲi9Y}G Ba?bLc6XkRlS/jR܎|tưp |Lư@_~^o~O拡_?7ߢ9̫HdB&pMG : -OpKAX&0ۚl:͇v8¿VUu[wj ?_~2A CF8%8ZS+)^4B2ɯ;; r39tTc!nDƆrZ""x( -i8O >[%*+ 3EePm"E( ';8u*/TyrY`zɁ )Kǐ/䲔O=YM}U N2W n?'R.껹RGN>ԳVop\=S\s^s+gqꋹҊ2d)kb.\i"d?90JaPS©h$=cTx2TCQ`nEG#:9urmyo$kE!n >Sn5~ WYJ& **JQy5-"'#;gf S7ӿ~nS&kPro$Z3-vYf:NzfܭUz/t6w zԳe*mkOkײ|1thG145ٖp \GY<;ܦۜ)_7"N_'qm&0$ #9A>^eϓ+m'؋$-bIİh )V$P9@Smt1*lb5qB9so z:~Y x`4;@7mNw> {YYyon'WH!ŚNyc%RF$u6$8RI4Kʵ!"9YCi[\=)ZKfaQ%m+5q^{q;Xdzd<ޮ7:=6Gy( 8WnԏQa{lo gr5Xj5FqV^7|>Kal 4Gk#+E.X 6f vZ΁%bLċ ʴ{O b*s{V@ Ĩ5gRf!JھiKqst25!$Jw,HCk5]0w'*ĺh1w6n4ZZNu 'TEʼn%)pX-V O$6FB)+5Μ6;YbLD n@A2$O-92*1)+;+.҈5%RY !jҎ%c0F漅V%h%i+&xMC6Mu}OX-[yTõNv1J&CcI3-)O^Wb̐/kKk [iPDx(IsiI;yJVP#}5TYcZc'_⬣ȏw.o65|x08ИzöR}ŧǔ-./HtATr*dky7_VqG}?/XuIW/C:~zNnhFɕ0:z&|X%25<>x1a L`hRhD}M@{ v !eX"LUMp~<ۇMeQ$a& h7ͯ'g`mxɕ>~97oǓZn4n¹JTn?@_-:l'xnl<65I;y/᷾-Qm bhy+[Qf\)`Vmy{_y!MQͣa>(G `9ht#wm䌡DШ-J\Hĕ rDc/(un]>鍽=IBcUFrZR[fN}Bo |8-d;& v1vXwha`޿v7K@xH#!lUQ9%U3r@7TaCwc[]Dw(s$P) RIe d9O"ޠUFO]yu0e~n\ǵFOx_/;(_z^% 5:>KL+#׉ZuӃqLCr8M Y% γirƆD+h߮s'rX0" q$hôx)p'`VT"0ۋXҝ0N L%,4s$Fdb&`- ATUqb|Poi OY9^:-$6V,R:ILXZ@-/yh/l/.XL,2/ MS.&(eL abyRB!u(p<\|:64)aK 1“]2:eesa`gp NZ9 :ɃS/ư}A=o ;Cx_Г~isN∷>CuDwx̝`sūLOїُ?],4hrZ$ %%a-T2E-)B s8WV֋fzrILd"µ WI S-^o6y{|cm|WszC]V9q+`YI%hlR\{*Eʀڒĕ,V >j4Qovm .b1moM|5& 1⡟o[0\c!nDƆ)x ~\qn'rW6=1qKZ5!s&bbIb.*No$ց}2)Djv;AE]f*6x-HLHKjUdMݗ,}Uu_/X(k_/sF)&)s!q!pe,_)p\ w#z(R;E0T%OZ]2FPgm4Vc2,Z8P1HTc4 悷y$4]1*Y,xvakzx1> <;Km[9~8>vUdGt|/>X6D׫]'EէNzJWώh\g DS]^o Ā]<Z#h]?Fև9?۝~UAK=zmt+g刞k-CQX혏y@DžoOt**lF} yᯛ>pih[g^oQbzi5ϭ%(Q 꾹 Kю byv1Gf*BRb1J sdc(OB>)~.{w.{NR#I+H~0n'`AĘ&yR>ogHzz5%fOtU51GtlB'u.YyT6{9cv 2Ba2+3Hgc4qexC FG ,GnQE5Q!,8*Q HgA-Gp$q*N)}aE-65W 8wڇ͂䁛$.Yp†f@EQ rRj*V@6=Z-} Y$RFb,Qg3GlL4xdc9 ^rn:hNF Qc"6 waӻ_k*q[\LُF)vI2YGch:S15s$$>6>qG|W%lv?u\V !nb?3&ڢ[jܐ}q̥,mؚFڼ8GTK:%8iJF &%0DŽ)&:^gաߌ$/\\v{P~gOfAۻ0%]}GFJn>QspB$ZIm5$kQVYTVZzϗ@YGE|9⑧ܡ= >qlʁRhJA3/M{ `^brraZ.Y,pѹb˹1RHꁠϨdngIWU].8~6Bѕ5ݥfAJ7 h{կ*;v;0"|yTV6\18?sw[c#bw~k׉tƽMRa6>o-ͻj~ޣ[ G 0Ŏ8s$ΫǶpLq?eY>7LWv?L/0ʁƁSg[]sChuAvTz7УAS<~~!W+g҂PF08B#{MFw/rDI)i7tqpG^[֒D ɣ= Egc/'JǛ+ AR3N *'㻉 F*KԶKEۢ׫Ev0*Qe|/SDvTЀ62;\ŚerNb~k<^`(bwfF}7UDr:+e#mSϯ&GHaOBVKF˚n͹\G9- 1yZ^%QgRB\.@b2D#\o=[vV{ўW iEIǥh+@ELECB13jEM/-^L&-&D85/K'KS=MT|>`K]ڰzëj]_ q2O%=B @պ vQIET)ӏ\%GVF~%Br Hp3p~+/ҼaQoXJ{L*apT!n;G: N $g!ZRBjN֖X3 x4ӎ3}&ҟE=Fb*ē#N\.c8hshZ-}Ph$DgH%^(ƋHM*@D jB|IaꝣQ1xΒj=+ ņs.}BG9.AKޟm?0vwso-v`dm<̠mY" h>saM1*u=B $&|!rPʅʉY%yz)<\~ό/^`&1MaLGC{5>iSɚm FgmŴ~/ҷzٴؠw?JmP2J{vCwa~d!{g^69scRK$1Ai\KEsG&a[KӎQ@s/(" B`!;唴; pNoջy&?g92 #$U||w@:f_'i{nP&q)P~10nN\菶*[2ٿuvp&6b}l,]&Zujp|*(AJA@T8<9u18Ǔqf\ rռNN~qd>QOq>x7S<ܠ0֞sg饁Rr:J J菣< ~:˝s~ɜJw-R.bׯ]4{kOHV[X 6D 3pohP}r󕾞|ȉ>~۷~3hEo^h.#Esׂއ"3 uZp+Q'ģ+bC4'IT4qj1\e-:p{q5ȫeFeugBXb,׳--|u!IJ#MR[jƺZ__tGy=[xSfyR3ۜ~7L#!= ߐ4-箶u%6FP/Ýu`4OwQj|6%IW1\&<2.G*¥3s}F`ExFp}&Wk!y1TC.\(ދ X)nQeH6U(Y8Rg[Ls@! PN*Q.*e-*AD'!%.$TEPQEO5E̒dbIC( rcR*+C%਴.rb2?Zϊ zBq%[58-8ŘWI8% ,E "P(DD jcPz(g/8O I h܊gj{Ir_1SZ{X$@p^ƾe%LVKe%k- xԥ6d=EdRVK dL \V7擑'"5hRS  ;䚔-PiVDt1RH?GMAw}zkf/&\Olw\׺C?u ?lÏ8Gs(c(dQvD?Zڶ=[[6\?aX&jf*PYML h61P$?v<{gSyJV}g/ʻ,;LeH92TIRZ _)$J Rr}i9yd",+Ӛ P[뒖%dj4sijﳦy8jЇtG4V喂93@s^Qe=jh-Z/^{!^bmkw>{*4ZL5fT(I}ϽOier1ɮNgH6cgsL5(-CS>I7>[7ȭ5yn{Э,j~lڹKҧY}v)5fH!ň/*T=%jHy(|VZpUKTʳ4 %LrY*EZPK UMw6F7q{qyݔp}׻f)FWe+3x:_~Yc$gR6$Ȼ l@-bYPȚΘB>4N!QHs@ ):B*dR.@F>hPSх`S}hcV 8I1cFڸŖTͣE%{S;nxoAb מAۉcnx]hٰ۪?75|05ams1#^}) WGz{Քl>u!ǧ6)r9M_^qӝJ1π ur;L,(*C(ZTղ7"t+}O{JƁv!(,Ei-ߚ51YB@ c\hƇLb)jry/qvJr8~,]$hPc2%r-- #a1Gbt埪Ucɿ|MCrjpOEcIPHzA$"֨K2?E{ՠb mplLoaZץ`+IMZNI:}ZeF; ,;P囶wMPّMhWl /{};^Ӌy ߜLIM `EIxxj\2o r [BWr_ࣟv 0KVk(Z_zk yJ Ln(KSJ*Όc)dli/[SoJ|3MČO3&\%lAwܿeLwsOu%lgKbrg=jht:/l 'KJ4uw{>޽fU j@Q$Ce^&k΁P#欬tE!ȭ}JO  X!Vg]|:.[ ѮS[n/V~= Ow5F_RskʬbɰnVȤԜ$V j*j|^JYԂG0u1;1U *瀆"0XX7uj .YS.T/0hcuȰrthex:LŐFqq~'@{)ȡ5}#y7am ό ixgr7GbQ@#k+h'Cε(VaEQ5d5Ygjho\Vq׃>OÊT~ڨX?ov̫[c]}_~ "=oPФ.xgsEUxI6J1u&C۰ԼYm8-L\R׿|j/O_tL<-yMe6OwM~ngizys4"v4wݾzfw -*/Fnֶx7'l^!q)zo]wZA*@rv^GBPzcÓA[/Oů_u(N Qx%uJtyDxAxQ-Q,>!*4hrl`!jH$q)eC!O;d2-ܺ.u/ٷCǼUԵsIۜw3칳fW|2("e!U^md$ڳ֐k}lJ O!ڪj&?o]`O3 iC/ ];OEBt>|;7dOo)~5a> \fʐІX0ƐhNkN07Hz1٭k<!XBU `5(%dM|R. h d doP7/E[PvDZ/|% _>^%)2fBDD(HTy@#$5<-zPoBi-=PP;I9pXVDt1RH#]Xc(Kj@dPNHEʞ3L +a4@3NL&Ύ }P 3\mJcv ɿJG?)Nŗr51)䭭l+t5<]ժ ]I<(J3zO.>^b.0:VK*DVjRJ}=; O|V~>ZNˢэ=Xf3FȹJ]C3Fwdi?3N ||),u'Fwp 8r%JA_NG3KwQJ;.2?;+LL.V5j~,tlݴ$hҝ嗋M䌟_Vpåʋnߺ _Æ_fr*;z͖6NBnm֚g6wzRmrsyJ-xmX~{쁛^پA399ܐG )ںmM7w͗- wpy~~ۓnygہ/7^ww.qBV㯆zmþ)e\@D0`" X`DxǍA#K|qq/xwlckB9St0)!Y TTk"K>A肏oFg=Òޖm mo< #b{fYF8lRk J*EtډR+S ff=մ};;\\l5Z3Om0m߰Q(rm{#O G}P9V5hJ jto4—Ѵ4o39;ް,,pYk~2h*ǹJ*A5QxcQ- @m,CJ9;9܊k_arjЕmrѱ˪QޙVi-y B4NPD2V`|Q"Cʢ2KF*{ACf(dܳ8M~]4{6za$aIWSTp )|`:D$"rOi#dLq{碼HFɩO瓦=)n88~*^Wh1eUsߨFf Z_eP:×3qᚉTFdRب*()\QT}\b/ĖoҺvd>XC t5 'lmpIj"r>RӠ% ޗEV NZ2,N_FO+66F};4W >MWIBnGHW*6qzs>@ubBqJE )tĒ. HUڧ{^k]}k)??_vPϠĩtM@HOJ\ 0xJl"Jhΰ=IX{fr&WoJM`ˢK 2 q:_0=^ϵ8{8PLT|O 0Z P(gH5V!P(BXZm:BWSO{` +B;k2$:hC JH.#U4eS~#%\> X+ ԂF|()ʝS)"Β10tF=2=}BGonc|8_ Nwl5v{no(4n/pM?nrYbԎF}x$&B9IRʥh(NA+O73ϡ#ϗ< 7^Y(%)<"YzKAF+F!DQ< OV]e?~3#K&Ѿ1h}SUNEr64aj[sK}gR{#^V-6}. Eԃ-@H_ijOj 4M21i%4FnDȢ9f I;G7Qh";~61吴{Ƹ p w0q&?g>Dӿ˕M> >黜'NP ~{7o#p *qu%OSnxO3T6=iLu87hŹΦi x:GL"",qxv'Sw1 8e$ͼHt6OdS[l4Շ1%g:8$Dg<~1w]㇟ة.a<-/k]<b1իw-`crN+d6$x> oSnɇw}77(5~sG EsqsYZeNQ'gē+f@쑥T:OqVŭ>YZD]cyZ!/ڈ Xxlƴ,v؋LecGxfc蠇E GD _rrRHUWK,au8ђ"+t ?odwȽvC>i79e:b7/I$C'P.Ýףη˃PU<1 { E96qظccvwKΆ!1ImeԎ泚 1 U&i|˅n]5k΂LDޑ}r˝3 7<ww@xa>@p)i# ͽ4scuМGƥXMC_Yԑ:zAW<] FjkvB,BsWD"81@Qc[YW%a:m>rmzcwb(>%<BGp@J/Yz?Ҁ{RwyX=B\F% P>h guH4ME`i}T.5$L_[%h2Z AcrMDhreL 4ຮG9w+C&4-/Äo𥯕}DEڃ3VA⃪͛' F FhU-$5AȰh50v&{ۑrvRŒ'Z2!>ju:xJ].7TG2g;3Gq,ȑ=S$ͻXFpQGNuȹ_iR0jݓBpVp*9K(V;,*ND9"} t)o8OdIK)DvF;C.]\F:T^L Qo'vҕ/2f0 $Ka -3 PM4FF\ pիx YX׀Y(d䝔L# EHQ\D =_ q䷠="zD'(ǒBW"|Ȗ:~~5Ƶ Ih{Xfaɩ*@rzsVK5\1iՋGwR0~CڱtЯ fM6D8J"N }3()Ye;_ g.$q(D*3wWz%@џ&i*xQ}gxHU PeMLI(fDQS ڑ$hh}TYg&y*B%y.O\`y-?&3!u |gۭ7?ڸ93iG_ҎT\F/eI3*ҿb @]#hk.'}C^xU?->+a2(9=o&D\̞Ǒ>qr[jhG`jQ.jHYJc@0._^+C/neuO"oj$汍GOW岐h9[A B#o_5ukGxyB-03_U@ͷE45dWثUQX|,f+RGY*}K/ F\!\er= pNWH"`*KɡU}L%W=\}pR\e9Br2L;\e*L2Tjsk{\eukOZ(eV /BxlTW-=GET` 2Bo&|8o^Tn ~9?.p ^?{XO 2"vA<ݗ U,mYH"=#ɖ},Y>en.V*lZz7ؿ۫{K'_O %ﰘXQWVS[k)yU]דin+|Zof2^K U`o?Iʈꢅ5D ΊW<IWM tLjDthm%>ޤG}Xi.Y֠!U)1(,$/P TpN;; jсE*lIziVc()&+lݳ?KQԱ[r8_%*>#u٨gsרGAŨ֧2Ye>hnֻQ?)<ۮPW*ЭW)*U#8gm:ruը}FQWi2{߶]wIzݓ /[Î7-Wu.fJ984ߗy(@BU D"Y$K~'o~Vf]|ֆ_i-t}k״a ӛ}Vȼוtli{O K'4ڔHhY ?W NƇILΡQHf-3OXe> v*+Ra 6A6 䴬TYW[.kBV.1WA:t#ٝ8_ cZz- c~3˂\f6d2[bc m))wΙ 9S>bm7|9A9MҹgMVH]DV`l~D?~J++wFW%9M T,RJT MUV!V-ըT Xk@4EE>Pe~F~}!>侽u{I6r'|\Ղzxt[b'+[J 46a)(K KX},JƳ!!h'TsRbDH$6PG,lyӲ7u`:9{\c0ȹ<í|d#v.o*adCXs;`C6 fīOWMr`Nځ5M&e.\ c *EQL.xS@?}HN <={=7p7'R QQM1jkbtVZUJŪH2B֦tYD`27CH&$IJQ:q2Z4Z(Ўϡٍ;q1[ J/^}5ZHWH>BNVBRT@NE^)l\l*(Ei-f Ɣ-BdcW+9b\ Mԝ ؜1aA@@~y|z5\=dߣ]U.q=Jޕ ]JVXӵؒj E%5NظAb*Ό g$}ѽKubyusAq/!1Do>GD'9ܛ}- {E(xlYLN|Z9EW뼰gC+9P2=m}^ Z(A,*xe:yY RGpEF%RAT%Ya/<+̙jκ̰Mo~XNX|xİ|6IiB4:Z13ΆyP&6jL,z' EYBƷtP$E:)FhǂdwcDiFW І`KԵvߝܤc Sֳzxk͓.l],jZmVnmק]΄gTܳ)Sq^Z/{mT*={}^l]&`=(@V(PQIX] ̆&@#O68;Ӂd<%j JjBE1#"ڤRFJ]"'U/gv<1V'6VϬAڦP@,%W1ZF*e zbVf&P2R8Dblq5d0r* G%Vhr2{)j Y(,!,c$d>cC\5Ggv9L*Jj}A=HɁ0Bl)]K hIBe~~6͸cq)- 9EWX\g^BlFxD*C<~}p0~qRN>-OO>Ng?X+͔/i6mp)4dm dVŽMj(LJ1JȪAk(@p~=-'W>yqBr2SN>ɒǵ3 T:%w߷tyڃ9wS>]a5Hy[֟f䝍&=B[3ަٞkB,W#aV>GDPz6ݍOoeߪ(|NDa FJ?hDxAxr>x/h-9vZjKY@-Cvd3i8T&&1xP`։q4֢0Lwyc?v0B/_Q 6ʧL^{Hַ,B:6@"kdz5=b_2؝} n;7Ѿy"Zξ3x?W| CAx_ %=D)CBRB"Uє,T07/}}.}98[x01*EI"j) ek$*0Ƒv/=,ʿO&*oנ֏툴i9B ?~KLs}9D;( UX@(BE e4©Y%t:l{0ryR-}ĠK:Ԭ\ёXSLiEJ nT{c2bߤIʓ_ Ϧyvד|QeiyvqƇN˝]tɳwyP-yv~Z&];.}ftFD. b$Jm[ ZtBgy/*n^Gm'0٢z~-'Wy`Z,Q>4C'tWJJK]^o8\fd]43 3}g ҭ{I26TtDu4F(WiJ~W: xȆQya,˸gpCkv2Vgp>6]‰wX-NuK>^l9z1;3|x1tY&_wW=F=cKiu,MNeɛRY{N%tlo"o%mv)Ysmmzrm[Z-xt6ZF>=bǤcnoG )ܺf{u!=fy;?g[<=pj<+;&^tߔ=\֪oId7CM$S\fq}c^M7;k=mgm3zڞ$[viZOt*lw!҅!3Je+RE#Jʊr G+u_xD]Q8ZuiJU} 弩s dlsJ Fbpwwncf~,S/A3#G591jUd>~5TR"!c A\L ) *ʐ4T%Aq5٩Yb$hLVyt!Ș-k4ð̿;Cل }(`屸 swGkz'P3m.Q-=YuޗPeк*LY+P*lU |TҬzJ9@5AXP+o@)gZ3Fn8VNB=.ԣ.|P]x6g!T=Ћ,_jdyrXpYUu*IiY#+e~Twn7 6#I.!)Q9H{g=5UGF OQhn%_LDRImi%ze2|L ؔD AC2%Sжc h\X4놙]s7b8;S+x,jˎQ[k)TyԻ ^e[`exˈ*h ɑ)h\v\ (m|Dy2jЊ2ĚYDmIبQqD>#uws7F*qgخ㱈#"#5uBe*dH`V@RKՌ%JZf!TS"Tۖrƴ8/ ɀ(3!h!묹ƈ#(aLj9?͸qYE.풯3.y,.qQm,hm}6AeedǢ 2jfJT.YqcPwqx0{/^QaяhmgZUwe<{(︀ԜCK_DG%Ue>S:E.rYZgz:v˾l)s LZ?Y mHT DVfnӑY s.)hQ? ۵0e.yiCjҒ.'틡|j ]qr6 sSL@rBkAsJ|0 /!'4.NEoS||39;8m?dnSh{Lg$i||O2;&iA]CκCmkB۝C?-_Es :1lc.!eI *Ιh8Kݾ۝iߌ4pò\p΃B'pER:k!)cOdtEeSb dB1wUí/xRl8hFe_V.%KcьCmܠDde-I}w >'yL %8ׅ>PLtP5d[] Q S Z3ba&TYؿ4=>EyQ ZCbRKe Z9/<i%kΠUYØ<) R BDBJ@Dr\VZ!^7NK<9~{rLbhRbۆ_= 8@ ?P?>4to&J Op4mNkI_L;4OhfQ妋ӏEeREY.y?-5SosUq!~xpIJC?qZ Q;RsO5o~8sÐ 2 yEsnLp%Svpb\4}}:hͽ?#W/ϭ~[ͲĎQbC9LDfl\>'7}9KV:&y\/P|m'ovD/WYv0磆x |.ߚ0L3uP;Z~+_n8Yȴ?7z`Gxբ{?E3T3SJ@ u5 #yE/\ 5L^tuM/Ka{_6n 6z b|5 ghJABQ!Abres`B}nhԭ{MJ`ګ"[)^,)$H." Wy{;>7y~6ݕhxoҺe`inn1rjp-yxI?GՋ8%Z]zE/hb?@C^xG曟 > 'OO"7zyUbCi|w>ZeK*iT굸XMFY]`P"W3?Uz~xShuQBn\4OY:M> .V=]_z34]ڗJi@y|6.=Q=D.xOް+ m量3B;#^V-6}.R"AJQז0M ";җqSwyøF-6ĊP,VYb$({f {0񫭯:]IgJ[ShϏzP`g G T|ӷKB ~?mdvrn8NTh+!Å+l4?^}x_^_lSdچ_ֆ/"ѪS?Mx8pB:PUI i̎qߝN iOJH&Ywi;b^$}:'q2+烷8 rk|\yxhiBQL%ITxPJzmʆvFe!$u* /Q$7>}NTnPfsLSMF߅l) aD$H]sJKE'q|/}  hM3ah ,m*TLy+ e 7TUYh#wIb{8v^g<4٬k]#2N)T ra$! *'Ka_,"ߢtAڨŒB7"|#Ot^K=jEkW>kpL"*LQz(%Ah*ţ;8W8~C+c_-.COCvyU$(d $Y"H\${6~GfH`{HmϴG%Tũ5f=x"Zu@]E2^4`ѕZߋ֫jGɫ+Oa07{8淳w=yon7Ow[³8m76lQן\]稜/U* h(*rk]A.7]OΖ ߾WO &T6G7߾?G0o]nn^h}T->S5DQvn}W+x׻}t/ݭAu7zv _+fRⴭU;ټQgYUiCiOw0W;8neKZR~yP[/%oQUoԉ^z  HW#iy]j^׮zJՓ&CGѕЋ@]WB9[^GWZ[;#]ظ~+ dWB럞+PZV]-PWFO+ԍ7^t%+Pe 殬Q/刺\{ѕкgWBܪb`FWK] Ҭu即 T7.nڅv#)ê%*^w+ѕR7sWByʽ%H\iے—9!7&Сcʭ߻2?F__wh{+貵C?֟#w,p\q?g3~u.Ѷ(fy;hs:zNW z~|wq}sV7?#Ke]W5d05c&[NuniE|V"nM'[Zk\]ωX*CM Jΐj[kTrpv[b5H%g̨TZ*?Q~qqojsKWlGx_ J t{{ivmvS\]J֑h* 3GGl~յMG#ѕ?#&btEz:Q6Z~m8`;uv58Z?#)̖]ѪV=9Iw+܍כ^t@4w] ѫ+ǎt`CJpu7ٕZ?w] +\P=eWu7Ѝ@R%ʺHܕhӱ] ^t%S=8rZWq{p_îGhMs(ܶXu*D;ҕO'8h{ѕuJVuv+] n FΠPZ"qP({J;YtAց\2T>*23t+~t\O] T4nuw'] uJp+Q~n`z]t`ol7\{ѕF7w]25ZlTt|7\mzѕоJ(=Z\4|G`ܕn殄g?wJª 7=࠻ѕNx(0zq{[QZBd{j؏YZ"Ҭu)?Y4shiSf9ǡt/Qn4Y$40 ~׸9:e>4Fլ I=4ͷ˳Gt>LV4$ nfєomŹ3c6yr͢)T{򎣤UW_#ޜ(L8u5'^45ŗx8J3zf̪VV5;FWѕ8w] e+Yu.+Z}v%zͮ+ؑxͪEW%}v%sNaի!u2X4+#anVo)<ʽ> %qa=yƉ5EDƍmGʮzjՓaOÍ8\ =^2rnV]h ܋g])Zqz {FW|/Zvs( uewOsW̮]7(׋Vu%{W]-GWj{ʮLw+] -uU+كUqGTsgNJ([u@]/Qu`ѕ] - (:wH]E${gxޒPG݌PGr+Ց8s[OpY+v^M UW_#ޝ(N(`|\ 7mQL (Jgɥ+UOޚ:^t%/y=_WBV]-QWGTGӾ] Jhժw+Ov%+qY諮^EW)9HWJp}7sW!UW ԕ uO 3\Ӎ6J(YZ|ՊJ&? }n ]iAj LJtgP'>ynf 4F(i3D]ETΓ2JcTYn"cɮ'kOƞ9vw3ک6 E'7J̊{8'ohIQz[(%~V=)k#MQnIqSgWhB|'j*aOQp7dW+Pj2D{ʮ̮Jp>OjmsוPrXu@]Ye,]~殀UEWBkf+\eU&2ͮ\t{_V#J5|&;$mw#2ݔ7w/|}wA!_w-?P?{D[{"?j_:^*zEI7)BoQ vη7Wy[߿d/і,Fnv.S9h61ӟ }O|Ǎqkn/{monKI?#?mwDzh?v"rF"h_O,Z<:xZO6ė'?S@Pޢ|{ާ_ej+붭?]Dm-?7I;%mK2F84L>FFx* >&򅒱UA>ۿzs;k.~G{~lfw> ؐxoD iUjy^TRsϩ$m5jrҊJ hNc0W7oLI)6v%ń6nFpF7 22FNvDŽU>X ]J$DU56 .68fc\iǁ5 =+[]x2G)DjH jPXiH )ܔ CQ|aR\^YR4ht!BFd K4fY*EF R+6`e001am[<=T,2:j-Bvs?Gr;oC`eQ "V<ĪWɠ,<*EXBmkP}ɇ`w)Qm .Ekhkzm ^6-cT 9 mvpܘ krVr:PE "]zZEeSh2 6 Ir.1vu]\54FRsd- #̆ÄDŽ #(kDW%LFU2T|C g AX2 d4MLiGh;|j7#UCqWl%WB 1)փP tLXDBCYbD&hDʃ%?A2-Zc, ,,5fBl5P9":PF6!-g@FۃB7Yx_0jX!A'2gF& ʕ_hbF\Xf r>1&(XnAIМ0]}iyMʴ6]CX>+-?83ݸ &B$6CdX - :da 9;2+aR|b2Z2dEbX!O@' ӪEϫ Z[o[G aXut&3 KA!) }VI٢GcDpZN}UQR`H,_h=ɋ PhEE`YR:[ZcM#q-ضW' ;U1 V''G[EӬ(ahdNI!/+I0~8Xaww|6 Ys'SaA[w5}m6F8a,wp܆{^ @n ./룫FYuE^Kn&#)۽%X4!x0*(S`ynjfq!2j +N6GCfڜ׽zF0hX%YhQi<NHѠ6.`14^Ef&Ybm4,CۍG؃a2!wGzcb;Xdo p0",0́v CU:|*i5kN_t%XfH{-Ygi&0QI! RhƛWRrV-݋Y`^a5-ID@ f|tQBeڃ!KmhC:x8Xz{y:!wTx&1b<暖 -º骅 C$G'+bŸ"0r\B\,\194^'yMh` ;b{B!DKr EiQ-IUk1QhjB-P ͏!!s6,ѳdD).XzndSkO 5r^G=_85'EiLA I 2tDSMIA1X~dr5v@qG|5o:k Pٗ9@52mH Jp ^;XtdFj3`2\FxJ7Es|dh8{ H=F!lPb.(⌚Tr+`/p\dj͆bЪXj":ڲK1atBf%d;;MH"4D҅K=(r - 4&,?՞Hw*;cgt0X5a7K2u[gz)jm=a]*v0[P][%ݒHNy=X:L8;#T j@[Y_a/A%,/eobkN> ^a9, ^.Z\jZUm#w5}a{K/}~7[(cԢHx}tG;mV§0ex6[۫8mSGo߭ob+-ӓ.[?#79ifo9[+-'˽/q5ftWm!f;e9z&c^jM+MgBE 7c'nO?U'+Bznp n%>z QZFkĭ03IV[nEV[nEV[nEV[nEV[nEV[nEV[nEV[nEV[nEV[nEV[nEV[nV!Vn EVV[nEV[nEV[nEV[nEV[nEV[nEV[nEV[nEV[nEV[nEV[nEV[nE3VA_[!`;ӭnh;z1!Vm݉p+­"܊p+­"܊p+­"܊p+­"܊p+­"܊p+­"܊p+­"܊p+­"܊p+­"܊p+­"܊p+­"܊p+­" 3; +[!\cPp+qcǭ4\ ­^#n%` "܊p+­"܊p+­"܊p+­"܊p+­"܊p+­"܊p+­"܊p+­"܊p+­"܊p+­"܊p+­"܊p+­"܊p+­ĭ> w V:zmwb>>Jɋ50J=06$: +1: a֪FtWL}RFՀ=|فX9Bȗ]ʕ^!#WI;Bǂ˕U9Ir - Qj/f1BkR_GDئ)wm軽\}9U 5bTS&G矾f4-o( *gq1?b攱SՒn'ܶfbwr5r7ԜH--PtVnU_LZbmjb)rV]rN]|sxrH\ 'CfީE+>kQjJ^c"}8ʕW9BroJ jwOT+{ϪoY[F_\0WarhzXܕ}\YV=o[ذ֜ FZ Eq QZH^\ =W+aÑ+ \!8vjQzKr Jŭ\ɶ;\!\E Z]$WQ̿\!``䪅kZ\!J++mf@rT{ ڡȕmCc$z=rv@rTC>'-JH^\Ykg*WXU W FZݕ]>Fr Y>&3S[_o-?r~9Ab e9E${ڽ}y>@gw!)}*hx^CYNU;'bR1,ޞo \$DHnWc}V*hEmIU\^v Bs^tF{eZ2ey^ältQpd̖i͞5|[}p۽6Ymt>'G_w ۍr[luEҵC`rbfͣaݧDž<,S~9tOC4mY|roqa2#RN}XvLQ=Ц~Lg%_N67׎fquk )&nvzrRMeFŜQYPc>jœL-SWgo/rٻ5V/?q1=?zw[T"[ vՔX{o~_4Wj?\3m⇰<^S"'iF%.oy^{W4;}\]Nx\ɚUfmW^cdǠ]Pk_JoWX)SN" s B%EI֔c(2wNj| >vj0R;>ƂHᰩvh$5ULൄv4HF-I5ոNR;>6/ȉm߯3bp/r¦;+~}/ngO$wWWg/^߻{%\?B2bʖlKʲ^V{ gZ9YIJQ^eD%]R*]UK2[/0 oDP' w\¹y PW+#8oW7m[OrS ]y]\߳t2^E+{əIMއ)Xո"BB ޶"eͤ~au.o TSp#*Kն & | |6[C+¶ ɮm80u XMݒWDLyf:]I1b9~Bwu>!4EeѶS6'fRHy)<RQ )"YJe/`b2"F*<5$Wժ֪] P_;pwC9UP:Gj&`d#kp6]>&)g9Wp~쵗oO,kB#On1!ۇfĐ_}!ʽQyvIYl =j(Vؕ|D]Pts.evGwG9t=2W; s BQd#mi1F/4I9: `TQ@A3L`0դMx2{!6n8t(,ڽoJ 8N/i1 `] q&_?7g^g(]0F9iWB;q_VoU:rNR\t`%5Fr"5huD%동_Dn7X5q~٨.Qi+aő.1gBY}/~] FY)hkc&D:H[zR &er3nYVd lk(K ,mR4 op\vT]⨚QYў'U~7Mr#އoh|7ڂ9t6{ ГOzRb^[hl%mx*}OZ0X6gR| \׮wr|$[Ry$|c'={#ss70WjZB"0*H+ᥙ]wEp-Ѕ:+mjeAA+[| 1:i1BQaΛ:f-͙p;^c9@ܚ07,d!Rʸ9 .C8Ot8SpUx8Mm7W&&mQj; ,f@L^/${Ccl 1CSutT k WX*W>Q; PT Ȭ ]< 8jA$nV1{KH1xrA3yImN<8x Ƃ/o`FH-CAsZRn%oKg4t(wA;L6c@&T ty,S2dDe4 | .uAǛ%;'#;>%K>Yh{TϭXYi'_4`(}; WVÎvYET4K(E32$<T_T_R0Y Ͽz\e/$9YH6q:n޶ݰ0gm siw^gW4;fv~6Cc+Z<`PUkJ{fѴg→ڨI(oTe{}xiY<=rl?\TO_WJԫml89]*5&Zljh`SԼ*5L k,_DցI̡*13(sRHqΤHF +lbwWYJӾi/N,Z6gH8u)HцGŘ!K@ER⹐ %/=sWhX&ym/jejNXmT[ 424R<(\~lLGآff}R- $j*FP֐1 j#F3PH˙dv2,ezkB[M/Y%-XJiy)K'p-iUF[9Ǔf=#uHr@k'*pozjzYѾ{ v"uxEs GsPҌYp10ICԐN:sF3!YCRgJУVhncTtJ2҇笙s8רvV%A'mQB8 W>|3{Tr=߱q߫[PwtR.ޫ{㢜n0W@Br$9v̸Za5HjÙ N}TrwW~u_1Gkɗr:VAb҄I_R[`<ڒ+ W:$|7}[w%t_;,|*~ׯ?Y%fv!abMLgKm2@VX'YMϼ0};iN Xmze\hapUوbע7^y#F24#AO:C̸Y[#%j)%*xSr]l) lܼDq wj5BZ޵OJ%Aʫ=B!H8LAryd<;JЧ'\)I^@J(8xvm5+*vuypqz}JڔUqUw8fԮlw#w\pRUwO"އvmK߮͏nf[ǎHyy'):v vi^׊m]GF>l{FHdG?Y b Ss,X15"CF*,N}q 3q49OKft&nA+KՐH#&`Z I5DU&ѭRx~@TDN3l T6Gs:/JzO.ӀQLoO rݡl(/1V&/NΫ6(<ЦS¦Mtr:UqM.qKӲ9}RNb"EMѱVe렘PY 5] zam/26'30y5yv{EPMMR>(mYP%Z۾0k<4e4D oP ]H2Y cE3?ϧ@9BOY=0Wx1TMt5%3OVpAcgFWDTg =% AY#z [IrKFzbDΡMĺMJ}Z#v 'ӵjj]SKm{Nm9M 9>]I_軺_Nz;[\J{O'^cyt/Nj3Y'앁huBY|& h:/m\kmH@\FGpwkeMc_>x%){#U(ҢDJ#4ӜwWJG֤QNyirqΣqQ;CaLR8Rq 6r.O:tvZ pоtR\ybTF^(=B.PUP&d 9ˑ] Ȱxh#InEHp#gű;PXrD ` ԪI\SAE=e}$LB#Ր_a VY¥, HQi)].?Xj#gC9?x+0Ho0*@88ŘWI8% ,RBTHKj_m0d) Oz?oᎢ I hi|XŽ NB+ZXU&Nj%|}85bHCfhHAQk>C~1t8jQ{qRO<T.x Ք)|>p)˔. !q7/j2"5fDGDEjc-oỆW+[Qʈ!c̩#Xk L.$0%O<ؼl6V:hA_b4W&֗ᐲ2{b# Nf M(zgpJ]mRqc=*Jc.(I9jjž! -r \NpgBFǔ@2 DEu V&t^0$^qgARv?19wv 陠 U[ /ﰀy&xQؚ*|Fˍezw t:;yii<3A(%njc8-xX O<31o]Ur5?~z!:"cQU'''WMSiK0VV 1g7ta) }֜5S]&sßAWၬkFEMQKH JC( CT |JJy%gruPd,C>3 'c-7PXS/|Dmn`ܰlȜ~J#5[uѪ3ڴ~fOn]Qޮ<[WY&E߯/H)QzeFM#vקh,L?,W֚iʏ3BWb$L2;7]s?ΧEv+f1Ivj35/hk4_ DŽ̈́wzlF6y~ٰzS5 ɂLDّ&mgf3u[6<%oc;2IoSRZz)&4cPGƅXEt&SK2~gUGUմVC P<-1a evPI#N4pR #J|Ge+|xcKRf!o:fKRƋOj!"< pL0Շt!24/IEZ]a,USXȖx"(#9\\GGulZ&KK'-v{+L#y$>ӯGDY 6h'O3ympBt%<ЗQ<QIpJE CE!l+nAF>O* f4:n?9nOkGA9l$erX⥡ZcWe~䘘E]a'BY EX}(^Wtgcܯf8V"s(/u:8iJF &%0t`,oʁMIƨ^P""B3jHȭ _gՎseL޸  Pw01צ{TK;[`[3-7hܫ}~BRgT2GYdTp:GL Z`4x^^\q373%#ˠ!ɃLU 4 I TDDEE" X12yʼnO/qTCR!!T3i 1'quK*" IG_64m("0㊺@H\ShH\s¨J ъi{ 6eЎ ͩ`RܗUdSt4)*)~ZKiu4bBh/>,,1z_I:-WJ('t y/礯joO˗tQ\ve0ORUqjtxUbV qcR%dƗK\1\ xۏƎ1FBc7٠_+aNw9{4V]+ 0|~EwIZ)vpEwQ cÈڻ.M><(+}_g]y]-x eZ>[UiDбbŇvVoTZ/#\~ 5$?r\z2jN$/Dj *aPN+9؎Fuh.cהsO-I#yg,/HtByt\MF _EvٔβRy:bd+|=FTx2keor;ʼ)O<1GcH"Utӕ,ה/<ȴ_(_Ջ]|inesCa9_z[fْ';F 47} 3ޑuZe Kzo֛/ܛs*%4wm~#LJ蹀KRL&ctkfNYIѐVDod>q4D[)+j9> eDeQDPL+s-ZOl1tq%mUXmzZNVD-=tZK4POS,mK2s;Zn!_SA5P%KJOM +F#So-/@JDPe^K3:D\뉋ZЊxbmDL'ӴeI[ƬSzH<  NX[0ҔV 'nI%0I2ZpBxC GYk!n($'68+DM*4:ǭJ1%2ȶ4i^N cߚ>Msqhep6Ye¿m'o , {Wנab(q\;VǒL"%ܕ`, J֦j7zW`[CC`/P5@^\ukM'I+}l`۽X`tgL`Ud"y$9L)pCG:@ڶEޫXEVDBZ}( Sd'MXʘ(|A۶jojw>ELƿ]h)j",E<.yE|d0ZQ{{_sۣ_>Ww=*o;#ݯv_'I>Ihڸ9n~mM/gkcj]l{Oi<-V*UMx6팪=vFUZ/BʾΈH/tct7l̊Wq\FKkl\YA6L8)LM/<)kDiNb0SFe)K%HMr9ǜ) ?*g$Ɣc0WdKAGJK6>8h&NyJX\U4{CGU$CN(I>M`] sʒ<0eu]WLk^5pOo?]-rȋN]mgj6=Mq@vzGZQ5?b)^]Fv۴;puE w:X>D7*)[d+Mf`u[V?{5ǞTC5>"@o^8Io8_8_z?źo,߰FˍKr1\KW<ώ:Nv2X4x>s~ʙ#GW;_,E7 \K ?>Rzwi?p-v70e |oVzҽd뷄)l¦(.c݅OiγF]JqdLz=0w{9{gσ2̛i;a:o}uw|3 P3/ߺgU[C q[hrp믿zh3%|성'<gL{ <#EZ:qX ;.\j~a͘,ߒ`:0](Iw9$",.~;݀sV xj$K 8^$`p1HX] aN%֮muSJf%JEܗoٷ@%d\_wBbqSwT6JJ_ů& ,b⬃ f.~QZlj?\UQ %fHl$NiiPPۚM-NP\LOX\ mq1c6lmϊ)*ЗKzi)hc^3ע6"REJ m]]!*) o"/'GoL.W4 Tkdl&~dl竴VP,TPXXx4g~U=-gˋn@>p>m]&G#g dQ!2ES)bV3컙{F8MɳElM[{]KSs ")SRZU֡d v‡>T17FfG4Fy,L;ں1j{v>E}fgE׌Ȍ&YSJ&)c]D' !g(h1Pޒhq̕@DPA㡈zDqg72IQ#m7b[jԈIB2IK)ˡ,xD:O=eɖĠ"P 63umM`oʤLt1lcA;*Ryvz8kKRCFcϦu)J9!H钁:-ٳz38$:1EjW˰[b7FPV-'QwqQ"98 !z$VXHr;j'؋ja腔*I8䔝QFWxዱkc..TmD%jAHV,kN*UvU1X*֦E 5@;;>.bu;/ {7?85)d-V"(ksB% 5Fa}} 4/EW@}N?W${O^]IIR,ɫr DI*@7Q4ػZ\-8nkYV*@2`2)蕊52 Y$gMV6bgV8QQ6ӉGyJtuR-v| 4``2i[H>I Yj4aٮ|?F<6[5EMf=m*$IcpF$ .y3kg쨪iCq5jl6Wl(8<[ DCY'$&Olbcs^3y~3>vƼg t!zjd{o9?-3$E)s)` I_ȃBq@Nj"'Rj[hTcO,~6 j}MRmMqM"Xqcvy6߭Z|L)}|_mRm]{艼#O ML mO/G-@0:Ђ:COAA42haǴ=__'~_ҳ4Sq̘^,KD.Cs&)3uR`"BtS ]Gj.tm6bkm?ut2\J9݃["Z.e/gMɖRE<_&Āè Zi[0@%MւoS\r,\Рl*XAuEc,"4jl&΁-)bw9bQ:L\h4/Y >^v^s ܾWXW~%MauIIPO9D };J2*Ɂh"m3:ia#B#D||ui=0|?c MϽ("9/< *1F,DwfǑ\.$w*@qheMTlwI$ dBJQ5fp*(#B~vH?ؓNj&ot϶+uDr;Ö3Э^8($oGD$oP KEɜ8^Ւ=t~A=ly^B{lH& JMJ)d( Q.mwVI,8:ꑧ Tw=[{}s8N9BQa (tsd{CHt쭾X~t-G ǡ:C"ud7L ^ x52 qI%$Tq(*5,)aIIp\F, c{6>R5p^, ~ְLhe1t¥rrlALI;Lџ'5_|y3&#tdzu2_|H_{xxYLoO tҀ&}~4?~p%^o~TZ| K-&L.y<_f .pz9/Ϥ 'ʰA!`o/dzdL,x2΃i%|a˃ٻb,t2dT?NF/Z[`,>ݛE5Wn?p`3q9fb;ȵy[4Ë+`^w~gE1m]O^4EiG%8ot_{YWnlrZg^X'MASCd{vd=e4:jp4q7q5.^#%5X[\Y^URh}Paq$=5XvG̵n<q?ki=V=c[R-n[W2?M%m^|kW[MxFl+们1w#~.QLħasl{b|_u׍n_<{}r9X3QsY ,J]]h|+eۤMaj9^r2,B>ɖ2,+( 6[s`]wmI_!igQ-``&l| >%eRKJv[x谨|g@WuXYee%)uN)[lVLN.C*[1W 6˯p(_h*tB]\rEaj 2EQ⣔ VQH—Dȉ`x_L<^UK d?2&}fAտ42X%=*gb5 gh4)SEG:Uw8ws܁py8P}ҙm h޲A}ܷWy_T'/l;'ɷ'?b]ƶ"W nE-sòVTrY.ZʌLYQȑUx+gxZoܓxfX m;P7d֋!V7$tugggi|擼7ˌwYRV7,W:0Ծ~>ŝO/N/X.謽_Ldz=/}yŴV{#~6 {v8»Գht͗^'hsrv^f}bC8q'CtQMvnƺE!%MwgکԢ_;Kpb\='Y*,ɘT`Fy*! sTd0mFJkQ;FUA ˈ+ Mb0*eZ G!K)swS )"oƇktyOHA{ MpԞRSV99ɐT)ORc~tѳNc:&P  PgG{~Q'+<'PNA9mz'(&9AMl$Z?dcz`<P`'(_}[x+*#rToSs:Al={5 3AL(dǼ, /l7'?N G[X6j8j)%|s#SGX,"J,Xsg)RȤ2BCtx!6帳Ƽɽ-a&15x;!G%|Kǒj\ҘllE(fSڰ!e ؞ nld'ajg#بl1oF= Uw,9K8t29/"VV^x_(+,KT:mU1ۣ#,d1yո 5+0[փ #R܄Խ3!uLH 5:&`tLYe c02DRP:$lPWrIZ-2dZPw!LM5JdtvFC4zi:xWstwּ}fUa*cn{\,ÀW D'vS%e@RwntTE 04كk@xaeST%0}(:JSV5{6UTGV&*dT ɬdN[kC0 C* Td*>tWaM$r9]x2I %$-` xb"T@0XG a !s*,Blh6R%Bc錜 HO+6za'IAHA :hTgK*IY[[y&ұYaم ^tm8[̪F cN'Fm(`r OJC {Br)tXKCMP5kb' p̼EMaVx9JFOΏGvYky ݗ:A;/f] + :䑌$r2 \Ot+:YpBY%EVma|pʗzs]:Σ2qrqZ-tRw4|ǣs~nyOHoGo$߽[<0|ۺXmYzR^3 s䣍z[5|I]1oغ5d͵4XK_sVbǬLP -ta荎~X؜M4 Pm@50G)f~7[0F3ݍhA9%/DT3D+S`,6d.Iɿi,&o܄^p>}gNڜ4C'q֬uYξ A=eY [olɄ V@ČjkM神7! )\8=~WkQz=96WS}a IGŻv}H$QYJ)^%ܗZk^VH)ʪ²$5X#uQWLUr_SD@!^QW\7芩uR]]U* 5+Uխ{4Set)`uΌG\1~[w>B6b ,&!o!*oތgx)F9-Kh|{q/ONہG3?Na8e{y^ϣ8K'h<ٻyqE|:i9j\h\D$>|`}50K )~BOI-$֋TgތfNS -1 j;~߾?pPjXS*q/gJ#Vm%VcՊMu0@Fk4ΧS#_RTʒV$} N\+ʨ\%xڏ2F}*JݒTj5Xr/ƒs\zV(EϭE{nujAg?J _=B]A]mR4=RWLroԕTVI/ꊩ$JpzJRoU%ި+&W uUUR9שQ]v$.rGg~:l7[jy46_w}]A響*0Z#hxes$ ڈx3֚os"z_/Qf?) ?ǫe}?Ӈ`t%cA!f 9U( D%863ڠ$((G,޲[x/Ѯ9`Ul2LV0׼̺\^o7?*7sϗo3F^,TS^Q}4dlБ(LY[΁ZɂBX D +jCSWT;G[]fz"+㫤PԼ*7S?;>~R6RF fEReܢ "/&r< zHnzZ1W\gp*BHTOh\ЀLS+cL()erܵOFqOk &2 S5UӣI'B6>[)3qXk)K sL Zg N ẁ'vD3rح:OunU'uny9)LWjnQ޶Gt>= V&an_-_(gY a2 1G{< xH>J^׮ jN=iYr,$bH>TP''UTe!B(]RX8:Q|~q֗.%DߺlG] ;V[0& d)E_"ߢ4xD:0m zг޴E6E΄oՐ"F1m' ,xS]) %tŦH ݉ 5qr/l*c鐯 z1k•>iG#} X $je@P9lYu+S>yA R!ģADfeJ㼵 &6#sj+M+F3OȺo;rxӎkdͥQ$mITL/lGHD")2x2 :0&P[ %#l)2&9 RVawЮ3rv7Rub0>\:bO~4*@&~([1̤b4?=Kicӟ{_& ;M8i~N[ _yef8ǧLp:Btij x6ós;º;=qQzOa>8wS\DiQ5hmi8E(FG#vрA~fogߥuy\~q17[??WSizY Dx8Z~Y|J2Wv.C}f;,Lδ "v H0GŬOnr}4tí~~P~bdC3"X6( Z{xQ?9#&ߗ_̇Ƙ24FP"V1ɚ"`]P7K@ ݸ1=65ml(}-lhr) uC^QV#oma=mʵdRUխRlbϮ*H)t$4+ O|(ܳ0?{M\ɾwm'B8!aɺ$sJ—at!fvʐ™L$$_U=@PJ,X?.;2׎ S-Ut/ſ6 hcw]OǏa+: xӛ)#0eI2&ĜKGMyT>bb/q'7M6B,r=~IU7j0R]D$XvR-fuջaz­n쯵.q2%5IXu l6b<^>M LIw#3PT@Snl{podli~Z7<0IIx3aV*ΠMx**6QEEQ-Z,x0ﴻhcC߰wþ%ó Oըh= " z+ W* ^R*NDTXM}i1GQE[h(n2RkD,V ɔs b)BCKnBJiR2BiuWz(ā.*io7U% ;:,z5 3J)g ϔ C#9V*{Pgb>>.ƙ|~C6˜ICrϭI3ύq1hrcgߋsEH !rRYtll@҂""V "L3bXs)**RNq,ئRaʽ38aFhNPƖ!nkU5.)Mv{MS+ >ȶ-$/o2hAmS%z[1=^}uA 06xd*ԅQɭ7(z-,0 oxZA^]K t䑖@Y""wB0qb6jc3cn9sʂfĂ:C9u<.z\ =-THe4Ρj<#KGQ*(*y;SgP;ڈj/z L)ni.Gl%T%S[yeyqwހonb>LVNnRܫ֡2_a3oTzT\T/)}SyΥ&o߼+1OC1K?qEwbE.# ?Z~`j-G7fQ\|Jf_Yf"gV"7 ׽c BYX./t poMoUd%jWe;wSR7,k.~[%:wޗXNnVeqQ)< Yb1+QfiyNRRed1[:M>^e5ӛzFhq0uҎ3;׽5_6` s:ׂ<0][˯ZދZ~Zk1xP yN<68 cμXc=!% ZڗC>Hb+ڤ[U^n$FVf1?K}c<ڋ">aSFi+`!l*4r#A8S( Ruh5Zj 4wF Ta#*&7Iwi;X-r8l2&սoV^wx=9Qlrl]o=Fg`&괨qӍMQ4̓a:g$0%roDhS;mڱU[{*M^F)2(`hT)e`cnbؕ<ık/nM@0a443'L'0 1#%R qNT1UV)>ia 9a3ࣗJX*bS+T D`舖5r* {¢XZ9$Ӧ" lSpE@| TaXrP!S1 !e{ѡlx QrKvtNSe#J# eqz ["weLtOU̹53]F2U ʤ[ggژbuOb2 ˺A0{KuI\qXg-jbҹV瑀2!yênѕ/<[fAN˥VJ>+>=9pLLr B($<`si`İrWL#,8ZI}#j8MPv]*7^y:6a $(GEΤL)*Gy9o͜tͿw;^ V,A"SA^ZC]jx@ZS21Ol8ŽQ2z{Ӌa kl芼b;J~kS֗Zp.%1=0J9ތBr4A5LBЯ uթ,x_a~E[W=\}**n[zlWrO>v6+̉6 "iy Oh -[˴Nb D99~SR$ -$`u@G$ n,{TQ >(oj.6PNv|'>jp¹Wy9@ kQi| &HW|d$c$HEPnR`tGGGlu^^?LN3p@cYc/t迶vyYhD{L`%\O'&&,66_DդX:\MCEb.^ݴ%W*kǮׄqH{;"ը-aIfS m<֜Zк|:Zos\nͲChC˼ݻ6_|X;r;?Pw߼.< y7^}H;(k._tWoyʽ19䯴C=NG }6+iFQ9 rF8 9:ZEiPJ7WZu(\r VE$YSh@!-aiS F,EVp߀n7.eI]o;>&B%0U͑[6_WXcb qZj٪",  g' qfm#IeOwGnF^"g>d _%E2%){Tϐ(D5-Rer^j~D 1ѫf&T6j|PɤmDni;FjzF͛jlܘzzڰ^ Ƚ*_/G4CthNJm@bҙ.ApB]1fY, It,ԣK N\R8_GzJ=,bb­bjxkj /~9/?uOg GlSz,l\@x - ǀ2hErge-7diںݹ ŻJ`{!EM: x:Slnjͅ (")M\:b(1wjvlڲ2j^?p:O~1ީd<kGbAy4Ism?yc2fV] {@2y(2išX oI8FǬ" T'jWx= wƏ""TFDq@{Y8cJ ODF@ReS"($`@D>K&UUj(@Qs5\pG $LȵL&h4Ph*#b5qGwh=tQKjZ).ʸ\p>(ψ%mR2 r,(d x'r>F\. VӎM+!n[{ +_ⶪc s}T({gkc,3_#,-1bmQn_ĒO'IeJ-22nk1nū*=l(T9ץ5[neM.P>5|@Gk\b2f+P fV9H}\gb =I\w\_$H>va)sVF(8XoFG.0c )a봃Vv`! V.3Lx$(E!eL2G c0Hα:hQ#7w`uCo~2њaYIQm(#\֢B/8 Yv$6L<uTMtN,yaoMtФ-${Z@ pp(Ss"#JU,@~@w'{=\'$DgЫnc}c&0rh F >URh8i٘AEvƻθ؋bj#cek^c,d;`5EUhƌQF ByBBJ@ qlROڊ]NG4͑9B\SC+zpuW,!y`^Û!t%s6LWzLEng'NL;o_Лl\km>GS査 EǻئQ˦&5ixwW<}H*g. _!UMr9 ǜ@!H4-m{|ߑ5NCre5]~g,ú+X:}ZGhtbɾY3TMRND!mޅrP g/2yV]D1#hyhm~nzGhq&f.9>VQ /jK;_V2$wdp {%g\2qxO4Je8ti/>Ng4}8L K;r  %Rt`ڤEeq)nL/,i4딂`e"QKM8^߫ųi5W/]}42uȧ{7O-rMEJÓަ$xCWP z즧.IeKOI܌l 8b8pߌ+:Zoz~(rCW}2W#]~j;J soMpgx89jeϜQ`Ľ͙<Y]fdJA7âv.F;gT VYTʌC rP!&Q$iD!\%F wXK n=t)5# ~ sv<tFl\ 'Ǵ\B J,C~;] Bb ,E.9)ql5S\FY`%*>xKM*ӫ(*>'&0ڭ vB}5IOWo~Ikv]l|]KIsŴ_ٗ$YKS 쬛WYد K<At}r0wwYըiD{բ\[ή 6,?.yԝo~7X ۦ'sfWtq4n޼qPz՗'euÆX#dHSơz9OۏDBOr|kµș[*Hm;c|7ߒ'1G)aPrƧ d{#肊c.hwm҅҂j\u01NEFP[I <מ+H1H+uz\G8i7o&2.HؗqqKprw틃?|mSqviLsai!lìw(46T6o3/wZ A[R[AV1ylܾtEJk<+5PT6svHiKltќ>(էi:~땄Q{,OG @)b Z>M ֈ"Gn=rNy {i߇Ǵ,G ^v֛z֒n7ݍdOS.Mc,,N8#4N%f79&)3*:nh?U_}ǿ ,du9qsy/ߎZ1PJ) $T (\ eT8`+BW.9sMx~vkoCa^k՘_xP2ĮVѶ"p6mqܗmE$-l׷)}ۊFYTۜB" ]2`S!3䉠oub!DNm$V7n;76~"+R deۑK-Y"eRLeX<|P9j0:}uJ.%TYf6ѽK%.($BSH" .HnbnI1EjI>mp=3$E$EQM29lLSUO{!QtvB޺NDA3.XNV Ө\J@ B)%-1JWH\T%|_;clyp5㛻oP5P41"SuteSiqN%Ҥz{s1mR@㌎6.;߀bBm>-|,džɿ ~+|0,> N)>h1N{4źԂs>v NY~MkJmԷdb$~?%2G(@Gv}BȆW-Ip tvkg]iDۛT v?R=29γhY qlrw3ڔkH-ÿ۞jW@ta7ÀKY?n:KŮ<]Q$}ovzE7>{;OrԦcRݶv>t#u]O%vnMwqO f{"ux?v@+>?=<+WnxeŻ+y'[ɽY>sGGzj宷t,EsqKwٰs0kH\ [KHa`nʖjx/tδv2h+b ȘGutTA6G1[`}7VɦT)w R&kȬB]!i听ʂ[/RGZ4 Op"nZ\!݆.3 %-qBWvsAƒN ||>˴qFӒJ iŨ$y&Ӗm:zNnf՛ctiQ<ÿwI#l˦WM~-IB[]KR92|C(tB-_9}Ǘ&N(F"gqDCK8(7ɵ??..kGr9bbVnG%X/i6n)SND{Oyo ?@L|Fz6Ls6%xfSSӍ!2v5I Ytv +(=N/wzaH IMˈ; Jڍ θd䆝\X#%tw'4ſΰlGg P JYOho;zW\>;FV(9_.ƟƀF"7кv/*gg[x6Y᪹~[t"Tįz4.K7lEtsk!-GN/3TA5,G48{pwvz[*ui;F ~8 s5UhC8VGs"0Ƈ 93!,ɔSz4âiZC#gF,ɕ2pB:b@>Y H a`s4noZLY-VZ|>^ 栏X%n%)0 t~4K sv륟?}TOG+5To006K(e㥊&r 5}ۨ7x o> 0OK!1KSC,$ \,2:( u9v#Wm7X Q; [1h %s9esG] Ō6(!ˆ+c= XE+%o?z(k֠>zկ^(k_m>M~%Ta9q̸F%MÙ[65q>{_wV B.V ȤQ)^Rd WPBUmʦ̐`+'fB*7lzVGKPO:ZTQ)U[0F)Ѹ$EjNGM'Y9ǀf-3pBR%2%#s: M ]o8&5fwq|$ $;9 1#(UÎ /%[4[!u*ucQj"rP{Ѣ4$cYDSRY`xYY2*A CaMc ƚc:svbv::7L4%oZnQƽRm7ָu!,9hZJ/fF'äTn<8(tC`Vm֯fMc2xD]YTr)L&Bq$@BUZ^y:6Z6"*yf+^ʶh]r_Esk|@=s#͙z4硉.R*f6?{qe jf*U*U׮WDQS!i#YNmIp D o>qo&D<@rFCdS|҄ދ瑮Aw3hH'dnM^-30gv,sRKI*)BKQ^٣]RFasPΌ!*340*qrnwNk]NC{\uNA['dVJ{?-"Fk1F;Δt$Nix c˓Gnv#30r9?dn{Ax \WVby_yh>>16Foҁ0 )ak*O6#j=@ L;HczF8 #$[n_^ZO C h. 8bT((THN,XA Xҟ#[k-b 7iH?@ɻ5V߭bmch oOs"ng +!HCU=,(iE!.()rIA/(RОx!tԳzzynBTQ+LQi ;B"gINrl{xG=4kYD~A kORkD +ʄHtZR(Rj c\֡!d.[ulc{(ua빪 GRJz?τ>Q&4u B?ͧ_q&_nF $AVp3iHu:1#?}i˓i뒠i{ [`I!D.T*( :HPZPdXJAiTk.%sZc:=V)nTE;[B7zf)aa:/?d)*yTԡfާ%XʛZX(v nhEC8ѧ*bzz"*qGF26xd*ԅQs#ɭ7(z-,^^-;FW%<"Gug fId +V0-gi#kv#P>il6CQ{0rHRIHޫ0RS-8>.C NK;s0~Ƶ@j1[EO(2rkCFriۄ;I9It1ap ;9B-\4`5 . gd?Z4Ǧ@*?\;}m5֩)__kU&f~pQ^/fmG?a%;:7?Oٿ?~Ah1>r.V> %[a@&Ro^̳bA|nտݲNrc >(x"h)[5CKn歿vϢ2 /p %ݷ:{"N.["3q@-t4[5< "(@{)2zlru5f_`5:?/OR9? =`2Oofu*iC6 ȉR4g b;"txs@GnWCj/mY%* !* NbG-%WVXa㠥zDCGlo\9&r6TEmRבaYqGkZ #RHDcԍ=f6Y{m66'i} Zߣ=QAՋ`艹SS*ݸ"8 roX79&[dΜXP'zs㖇U -<"ƘzK4Rs(IQ ]JN(;ڈ@j/z؈ 0td;Im8 d݆[>@To~Qf}P|4S XoKLyh&ۣϽ 7QF5ls:ׂ<0N0))?ϧ!ǃ2\ps4kL aOb M3ZKGrLg*BEI?^<ɪzpss;P sU6ڏUOQtZJFQ@LЄ p2H'SE(MRuh5Z4ʌ ;N.6*GTL1o50Cdl+u3xE}u[DYS,OYJ1PY̻f ޕ0V궇%3;03)%HƂv\%BW`vJ(س+En+y m)M9/|ۊ,䢅 FX5׈f@_߿}Mťzs(s AH __n}oSǎv?83 K0.?3G66^n4 x6eC`ulr-c1xPð'2Csc̱|> zU8_лt1EXR6lZ`2︖Pzk%}=S;prw.LrJ͙"<flw1y I?MLs7< f1grH8+:hr`(oݻAF5Qs-9˝R 3GjŻ%tƓS8uO.Ŵ\BIpɝ'K'5t֧c'Btl z|'TvL81]nhwC6;jߡ u`E;CWW Jhn;]%LtutE4RHw`:CW 쌺Jhl;]%ZtutE5eK ?l\BW -k=]%tutŴ$Du0?JBWV vbZaޫs+0~7W&;*坡+@K1i;]%tut%w0#ٻJp ]%&*3xtOPRH!g0QW hሩV&opy ׌_tN*ѺM+Jvs;ӿKc ͖Lj|<0i_?nO;(ݧh&EsEl*& 3Et&;nN+:Q@J^7<@j=D ONW;Lv+N^Z}lPفHOW=ƚj!U+YW rNW %V=]!]4+B89y6npUg\R1=]!]Q"3)CӮUB+ZOW ewOWgCWhՂkg0Vd!Dz:CT1!wBը[Yҕ&U295}77Q5~[pQQ\MK0.j9 :\.Dy qάQ6;R=0i?/o~ ~ oȮE%7i0M%w.SݣYTU40W( (& WKkQp8a-}u_d;׏JQAN(326/ g鏂%sQ6ތ, p҄V(za>,y%'B Nξs ^/;pwϣeֶGf{L-.M/+ww=y40Ͼ mVORl[ -?l8ͳqx U⦳MPdGiw1@+jK(Y&՘LT^uTcjmZl[Qq*zLğWoTV:O+ʛ6ViXm@6~¦5o˓k@+㊱pmZ~ uUIFpE}ʈyۯ\N$+[GNI|3VҚ*=f-]5n)64hoǷC?y\,L=x:4*zRZOjY:#5UJ -m0Svhruy̥7;h_6J%چzmY{C lH?h8J,&01Ox37TKZI wYm LMG!C:®پ:n!KF`ɯwHUV˟ɪͱvnڷ[]+7jK^Yt, Mrgunѐk18U+q <۔lf{nSbiUת{刞w[,_w={4 t䈜thspt,|cMQ%U}H0EN(^J) -in[2^N8DzAjIygunє+m99ïPD&Nr8b|RS~wPJdMuaB H|'|-ܣ lͳY%?j>Qn5M'o|ȌM.^00 }bW\j3>w*צY*9Uܻr=SaαKzLG`guӬVa[ʸo??9V &jWpJ':nJN,W]Oΰp4Rg[c a:¹C+qW6 *Wg+vL+4\k,R+U{1xop}W*7,Z6R$;Wγ13]` Jզ *3ĕ.W*8S Bn2]$[ǕT9*Ha&\AyprWV6TWwϳͧK27~hz{||s"x \/:yĢt%F)_f݅ IіvQ7&yؿWrk; lphld *Iv\#8:ܕ 4 TnYpjqbdJ#W*XAnWoWrk.rDr ipr4ŠKdN:wqu>Ilp,T~;6 * W烫|t3]Apf\\?ܕMf )v\!ϓ49\LLd. 4' Ygm!;yRkݽVEMyR3'&x݅owy;d8Q%LL<\Tro~\Unm^=szRGnNo\Nuj㉶aZ2ml qWnc71M+&NJ:; TmJU&qu&m)UwY%7\6:tWg+& ]|ZVu&͂+Uq*SquDR1L+ ^N-JU2:C\9J&4wWJipj:Tet;WRd3Tpg rM3ծj}:TeW:v3]AurOs:nS:G\E+Tfhq|]ASܛ>eM@&dKkw„j+c^ k^uQfQ *7,j}9ZL+O+Kq\Z$t\viر?9-urge֩môNl"W~cX7WM+k,RrWT:qub[%4 7ς+:`J{F׎JyܕMq\Axy:T;WM4wP5]A=UjT@RhwW+`g*! O+,RnW2S+DLW*Xd\\'JՆ];W8 ؊W*M3wjO5*L;W?]wݮA;T_5?Kbzȸc/~O7A:|//\zP/ë4_ (W݅k0{ɡ?C|W\.狨?BƼO/Ƶ]ˋ]!V>s%VHﯯ*_o&1^կ71}-_ +Cv⛘}3ɇ[lot|7JN;Ѫ@vÇ jgnG4(f#9q9=Gt7л| tߟ& +(E]}4?gzG.xtܲf ij) { 3J?1;_Fτ&W?n>\2Q.^mɕTI4|K(%cBd/G8r;dK.BiBR&7Xg]ȥ4[xa#4cRctd؏ 2$a$MZ NWWd) 9⚳AZI⏩6{Ghbzo%.3t (9qNP_T@޼j)pbه pqr=Hr$wcFJ-FKIha0ͰU6\8:*F-9Z'H' ,IV^%S hNCQ{`ÍA2,M!:͖J*`?|vt4Fƈa4B֊ &U&t8 F\2I 7\/s%[JGTL%;KT}g| т'!T {E R0 z2H%5SCo)= 7l $Nw:&k"7к9;rgɱ'WsHHI?% !c5TP )%@$@mMc IU oݺL@KɢX'G͋k( Śa|7!'ȃCK Fn bPGۢX8j˒aB,ANI2 t!kr@,)MVF'zdcP[ qEGG<x-~[s_vYIQ ‰JқC:_|Ӧԓ_\ S2BckN%zTB FC׽Ď,(w=Ӥ YO&F@HDٰQM1\kt d{zUq#Z)ch'p Z56JbHT("ȆeBGk4>Bb+(FB%En(iY-p oDXNuvLG]uSJc~j +Z'.|")q ".ڕ! /5 |u؂ ּ_XZ bcǣ.`H6#&P # f{.\Z \qdX}13VV 6j i(yG aY (bQBą򠷒a$RQd"5A C`JtiYq9=K A 9Ͱ2s [aq2q XTuNv#[Qh{Bb9paQMZuY 1H_k}?zwyo6bEMv!Jq؂ѧS6"ZwprInt$:DY$*j:s ā FQxA/(Xt, mM!gT]fc;6@D!`zb jhnwIi3cey h)"59@ l#čαsXlc:GLB(- ˀajQkƣ"lC כoEaQmXcA8P'Y3 RC9QScN+UK:CzԖEsv4T–Tf5PҚ rz뭙*"X kn%lA :? ƑI43zcȵޜ6vٗepn;^_o6[zLc  h憳=f מ5`![ӿ+ogSCE]Gj^ky&<Zef0c eF\ o:H#(j7 JȀ=T9 _Qn$ܡ|A,Wsw\z@RAA*@5diY8a=7:T}zZ ;Th0'"X$O: A-,\`|GI*[ 0j( k,1D.u#-g:Xc-иgrSkcMЏ5<'-1\1rew jf=ɴQk$ILQqZ+XG\sɁ[3!w*6lLVRxs9BZtFuڡbM=:Z3FpCi3 _BV O p#k|TDG=-ZJO`*qK} &n}n-u30iYlp1:Ԙ-`Ur.I%b!3m#I_!4e}h,f@3F/3myZSE-E,ID$Ҋt{*&Ȋ/`wD`MJC#C۬]o q18X- Ҡktw^;J+fw"A#x"oY``tU p, Ԗ3 DZ .:./Cξsz+)mCe!WJr#X?7 Z[̇&z?3b`0 rkC J<䯆O#/1ՓƗaޥywZLnƗG/-(a{Jƥ`m/Տ&u<| 5MznЇglu!;.C"b/G0=kզ9Ü@v'!>9r@!r@!r@!r@!r@!r@!r@!r@!r@!r@!r@!r@ZtQ!q@ `g_HB+!z8 C%9 䀐B9 䀐B9 䀐B9 䀐B9 䀐B9 䀐B9 䀐B9 䀐B9 䀐B9 䀐B9 䀐B9 ^)e8"S*ڳ&O]xHɰЫq؍-"r@!r@!r@!r@!r@!r@!r@!r@!r@!r@!r@!зtX?%]K3\72ucZǷY'4=׫Xoʭ߿YžXCbYJS |LKUzÔ/0tR%wy "fA?HJ \ PN{CY%pw ekcQ9X7ŹU}%UcJࢃ>jda}ڑݴ l4x_e]$VVKQclŶ8gϤȂGl`Bvb1~R`^b0n+38*Q+0q N<'TAA\%iɗ ,Rk@ UjVܰf}o)^۩\V<{=!lScG䗄?_h87iMwdEYEUo*-%8Z.}f,eiDFDiDFDiDFDiDFDiDFDiDFDiDFDiDFDiDFDiDFDiDFDiD_ " R.婛Z߆?RY}/ E0#Ϫ`QeGʳ@ZeO*h9ZײB lG*⪳QWEu>_Z]Y)Au+Ѓ+0#%.g/%eգ`xuQ]:T2 ;uE%|G]&O]])W z=II<#u F]sQW !4W,D"07gB**rHQWB1 |+4?uʧΒy)Eg5+R礮@`M٨".="-?yH)WJmH]u FE]SWWEJkTWZSm+Hv6ꪈF]i8uuRZJQ]Buek5tAi2qyr=z'PTZ,sL}hYuULV" [c"!XuD )ЊR_rVK3\l,3њlt0_ ]Zԥ]]ڣHLR%5Oe ^f2whŏ54S]b/l :;N^*b0?IG/mv?}OzӺ}ts6Roz2~)vKU{/akeH`~:sݟu [Ρͯz^ˮLhw9γߚMƒh?V* %CnϮ?ie.lϮο;7Z4rqF`!]S3f|qwkn\'xn\Y<+~ Ѹ~ZܖS)hm53ww'TxeZ>L߶oԿk+= 10O0֏aR.sjx[W+7~Uu`:9,FW)JH%Wc&U 4 6NkLwO `c>۴p%[?\]ID0ghimukŵ~5ͅ8Nv7h'^콋?0pJl%gc(ePJ;IړG]Ykܾ8L?6XQ#BɎk)1-W,{%y`Y.gXF={/zq~d~ܙ}`9>Կb~U϶e~|,vjstDJS*4vKB|%P!RR zhj'e.*P YS.222'd:BԆHҒD A)oKu ءbL݊+IN@Ւ%v/cin}k;7SVfjo7k/aQqxQ<*٨zP|k'pCJ\zenFVX~ wڻvO1'xf0M3n(}CsWGϝnQV"D'x2sLs3?^lxf7뱝>(2PuHjLW99OKɍp&\y#trP>> EuZW@̈_-U-e-]Xu~ݟp 1!#xjPvN+mm !j!XF$VA׉+30Usz?(q(y}fz]4(; O^W\ vC_[ Ѽkp/m#!}Q*p$$_T%?rU \ҁF|r%QXU&RT*qG NjQ$:I8LWyT&D)V `\g*rt2ªb-V$v:E6d\*nNB'evɎkƙDN҄elٕ%\42)iO8;^g:k2r2Q kihƂΆGf,rօՎe2ZP*Rt& X s@K[COOz24KlE ;s)l0v5He@S3S+L'XY >:bJ1soAq'"BV@Oeu6E\|0 6!'}h0A;R |K,,;U c xՒxfTI>̴XcU|= 9tuЎW}qX\ÆeACD06eb,%J'+?AD eȕuTBYn*`gX6䘊k+A3ru_+vHJ0 ĬǨI9 Ō6ijr4TH4l,j}Դ{Bh/,yX): Sh [Ёr3jokn(?O$s :w3,vu6nhuĨZ: *܅ O02'+cY֕^U<(4r9ヌgaҽuȣg+m%:D)=F:pR,$-\Iq$Q!2Υ ܺIqʮd*r1!35˴҄箹mhS pNZ̿-*s^_烽s`}´41_RXӹyl1m9ܽϣKúu]0a(f+!J—54@uɇD qRI$J2"Zk\DdM)%42F4XIV~#g*5Fÿ9{b妓>麷n4`q+9X `sTdtI%˸GJ9׉?YJ4e;]RΪd)U u g RQM`&'GfL,Hב#{u<'ySDٮW=JrhU)?~C°R梒5e}!C8$TC5MTQt aQ\1?IITV.&.{Ijt-b;ec2L\pMz?ج?@Q_}~jlB[M|w FAYLprq\!7e gG0w r`HJ$'W3XI#D@?q9o6P3ʿ%5z!7ʖR&tvJ8pvW ҹ펕6_ݎVeA cg@p&8Ag>>/al}1 \4,lԢ=􆅍Soaa ʄ=u/4CdpzzP1ڙV7]kXҚqN faЎެhTJTB=Tz/UsHӂ>55DTES, tʽRyXQ$BTv&0֠~!B; 螿h~Ƹn-\m8S|i<^o)@TDPLCSMy tW@4PLw-6&v9#}i}dl:1!$,)d^k.˖Ž9GKQGu T 䍅 . !.*  u &0F}]?g^ rh4atmqLH\N)[mET8εLkP49C5bd0Sdɮ 7;ȱE=ٸ,72/ꑫ=95AJ_Dz TɜG ٲ\QּVNLj9kt6mj5P>pK}\nb)ZQ+,?t]ʸY=;:ZLw7MKX9{{vlhssO9w>z@A5SlkInF~֎x7lޓ pog[X];@f?r|YV\ i\;A?'pVCo(:>(Vآi}" PȚ4V˧OojA-3DZyTpjٝcK(c*)KlaܹtǸUԵr w3SGۏcD+>WmTcgZ&YM7¶O'"K_h`l=`iep]%gofIzs&S|֍@v\hz $4@8uhJ6UOW^goxvz 5> CT(tT<)FFr\rtD&GvG }QTtvp^[t3!-0{h~u*_/βGTF_TQbRYMpjjyI&N^Nw3'Qn)@ *{S˅I9SRkq]eᮞ8_J*]5A :mЖ8'6UJU:UJ&yF#udyQkt !8@5FYX(99BpqrcGc n1Xi/J}NŇ_z5*"ʼnl:|,t5jjf4nX0QVr, ծ^]%.Z)ULC@8$PZ :lBߝFQq{lg3t>/-eѓ8?Ŧ(ssܰֈ{¸G?JVvk1t% YBIt >Bي#Vwr@|ܳ,+p@˵\L=x|M&ҿ|N8zOM|o]ѷ~-;]>\l2|da4f\|x?t?Gn&ory>{9K%[fLX]F"%,ۯQGb`RFV;و>Z@hD-FQ`dHl*΂ڍf]!M';MwGP66|*sU+?o@{B 9鄙٠Kbs hhB,*oL̕H*WCslͩp/mAzkn<a~g Mg]h&]W]vsyՐTG:tytzzd~k t5DmT}7r՜fD q2Gc߆\ҌFY"cMS}F &b۩k bRUw^U[d'x(w>TXq U9n )CqqmV킗o_u{Qהb[q8C9\*Cf {S{ /'^ENk}S 拟+ceD!COV(@\R-D4 LIXaV!g4y3՟WFZ 2!5ih B/P>訯܃9" %eA\!6*`KTbN{䅳s} >ڒh'Lmkf&k_{xcc(7bpc\W?Zt3oC~x\sTȥ5}"۟Zլ|! ۗQ~]A/oKTb گQɢG975rmo_WXW0BV?у$;IӅr> ]].6,I?`ܻ)T(ШVBo0G+ZjVw$l*rv~rF'˝ޞ7 7.jHmE/V%Wꟻ/} 2=.mmezÂ:vNa$a^Uߍ6xvJ7BV@ ȥy[le$OșŖgO[0]+ZPb vkөw^; ˲ޚJ6aπk 'kcD:R`Fc[)F;Ye[0Um|CX~˅!^[&m[ZjEލbwfxȾz jr(I(+h]p ;޶.b]w $L.`\meE3 =UEےM:]:ؖtH~ͮէ/q( ܷE啃nDŽ!7k(Q*]lɫڶvVզlu*|>ES>ZN@*oUW!A=eUЫ ZJCz܆t҆hk2aAEN쬳$|Į*̓zT)hѓ9.k°g<}9oy8of?ybٺ[ ]˧ N>v sKO2_AE;L]o__StFëU}iٜ_moHNnl:P+hf,~l5..(G}.mH<-{ 6 ntw:0ඟӂ4ڦ[k TWQQʕG}^wW,ny8qq᝴X]O~moψ<3\̇e>(wкjޑv?_}1j]|6Mzi\Um]o c|2ΙNY V8)s&T=mըkkmQUU6awKﶸad%SݭoupNKiu ު  ޴H QOgc[4RNJ )WzCeMuk+h笠E(|HY1PE(LMPزeP08' 'FWJ̚9!w]Vj!D4tCTq+/S:,LJAb`+'d\iC+Uv$lth$%ldXRtEmbʃ(\I/qALi]WLiKj+O>I+ ُ]1e EWSZovPuu֗4}]1pV^^t_%kҞUo/.כHzfkI*?pQ̯eм}sgpSk^Ϯy7d4tXWxkg|\k]Mn~۝6X~89ܾT7}:VviR?+?]_nT+ c`8\ڸ@L';*ker #\ <1>cJSɟO&P*/FU5cKu8hǚD q򘠫XtujV j/EWLkSz[t5A]!Ї+hq+2@uQ)銀btEފ]i1]MPWrzW `qQ{/Pxҕue1+i슀+1bZ)+&2$btE!()"וZ)k/k슀1b\/FWDJ+D,ZgXQgwf=o+R/ă e ӸYռmMo<`9t}kOwQgN1@, : X {k;>Ex F L)A%銀3qi}2+zXorlKpw׸4ZoDN :AdڢSL> ;+5btŴ+ jB ]1btŸhҹ)3Ut4ҔY+HWǞ;&F+j⡕銀+uA֪ǮH+6O^tŸ#KͿwŔ]MQWyP 2 btŸ^L2ȴ(u]tO^t5F)bژ}2RjZT|0?~UpJhI Ӄ˼1IXKĊ9$1xZ3?](-z%zВZG]0Ǐ*7-ۣ ^P&Gɛ$ڠlSB>L:,$ <1+J濂)]|tGV=,:? ! ؓpȷhGړ7e6tEWV=+ފhuEQAu-J]1蓦p+p}[t5A] ]1pD1"\ bA5aueBƮء+ƕ3vŴ]1elD 4`#'d\RtŴ>+EWSԕ+rAƍbzWD}2ȔX)g<8ʙ@FY)bZuŔ]MPWMyR*_ꃭ˪ ϓJ Hmy!.M\Zx\D%Ǘ;*x[7C/V̨ {)/go#JkƿENspiNS# "QBf]aթUʹ ΓJO:%EWL;i*z] ]1pbtEZi)bZȾwŔ&+ \tE0wi:HZ2DWJ1b(&$ZOK28E]Y:zA"`]13δ6+<8j:rZi+vr{)b(*)BAbHERtŴrSꢫ)*x@ 9u cC-倖0Vy)$E3%b%В`oDSSK :%%|ᙨ?Z/?v?T֋z>MX] QQΆ~_{M}ׄG6Ay8 {jlU_/}I_Z/ks}Z lnu~˫%]]?}?3jMq7~!tn~̝?]6m3v C7G0h߿~UL~4~;%zމFdqL3VQ27dZ[2Qf,-,Yv ot6 iƕ3_i]hҗ;GȪ )3V}i|WFGUeE+]tujՃ& ]5Nw6h#+ Tu_$]1d7JFvYՓrAQfIOcqdic]WD6,(>Ի"ࠢ]1.*)bZrS_t5A]Y!I]1A)bZksS:St5A]QbAR2H"\bzWD\]WDC38E]yk(D銁+uZg 2e()* ;Hd(=ؖ,xlS;HٙRKKa<}mQPAd bnmPߢgJ()I|XЌ"ƍ EWDk 4)uE|tezPN_WIz[i0r*v[i&3]]SU$]17btŸKX`% ]a* k1b`h7[TWL Xt5A]i]wSrAh^WDi*F{pIvpIїdpԭ(HW Ӏdp3δ}2Ȕj̴(3sJ.UJ̏2K Sk6o4Y%匶m"hK`bdG%(Jx$ZAX.板[PJ"}+g'%cZtdrb4:]㒀cJd+EWGWȪ `O$`PcoÔ #hH+h(Mf)]٢S KqHӆPRjJc銁F+EWD+D[t5A]唗 ޯFZ%FWD]WDmI+luE>]wOOW߻bXzWSԕt!{q+3e()ʃS銀a@pQL2ȴ&vte} +r&20n {WLjBlΓG:Pw'yR ^ 9-/l-"޻rGDQPGB< 7INeo`J(2ϧzPhU* أ2iz]hHJ t劮Nz+A׃+5҆ 5tE]1.2ŻEWO+-:I+`qIӚAj2ڃd"]1^WLiueu.btŸiuŔ&+g0>NCՕ3zГpA))bZ̾wŔ7+oP]1pTbtE(EWLkbbJo&fx@KRabȴ`>9~(̵rNRZWyvGR-K-9'GG *Z-Z*XxTʆH]V$ZP&^O b-({&ǔPxN2%i~4].񜆮L4]`ć)J_V+.1hLm6ͻ݅/׻-bUwNAk*ګ}@m/pG-ލP's5뒡_yU%zJ?˻wTj5_o.)Vns::g,|(4@]4gt^ͧ_<Ac!NU-_5ۧ_}vm.fg精ڞmK3>R m[?|;>S8VceFVAg o=0!5ޛ3[Y0`m(- O9dRm>+h-i0~mMVi AٖTb%,dp.C^ a![Fh#22+. E6tg 1-X=ƳRMWLU.(հQ ȡ+"]hջޠlLGֵNqA)9x X5LGjD*Y0VȬN̆mBGkl};-9G1:([ZD2 S45W TW7"z,8(&`̫MO J+G_YfH57]P?w(c)hduƐk $4( "*ڕ"MÒvxk]ժRA]зt,/Ϛ`&䎊CC\L݆Z8tR 3gm>dt já<l24C@WJomMp%Ji7 ] B!YG֪NRl mZA!NmACqT.g =( E>|eHd^$t_[lJI4/ F 6CZ(rʰN:P}E׊X{B}f; ^-=Q,KT+f&'cEUzP҉8!!`ۀ9;z0¢_7r5¯zkU0oYΌ#D1I| x݃KDC\%l07\qih}0Ͼ|>_t]-k2/ige\ϝIvic &{hLk0˿cgUCŬ]ǰ^ky&4Fefcfrz1p Mʌ4N"_ADy ؓƵJC5G?T3ڍ Mt0&%\!ME;h~` 4huMUځ4PCz[:Rp֪Ͱc6Ձ'݅+EēmrJmE5$\E~G?E*aZ0zhs;Zs[Tc$4tGq3#1t \:'XW!hXڨkǤz4=4 khc1[Vj|ҡǚAN%_#Lzl5dQ Tk5>?MɁGBTm4_>@|5@>w1Uvy(S*fM֔0hٚf@2k/O 4%尀hGI4k1j4N’1-SAb~7Pilp1:?kK52riL  XTTǢh@MMH 8iBe. V3P5wD;ۢ`|J[ R GR/^b9V7Kqj-ܮ 3ZNaΠԂ]*.,9{⧟>ݣ}Hu]T kYkzo͵[g^CX$nɪ֫Khӓy=シ\~=ٝҳwGeqrsyz0!|e|w8=~H/7WIg 4A4?c/+\]m%#mׇ_76ӓN×K7F|w雓q{v 㧖@g파*.©SKOE7Hb/F}@@Ķ}@AI+H|@$> H|@$> H|@$> H|@$> H|@$> H|@(xD> y@ 79Ey@RIGy$> H|@$> H|@$> H|@$> H|@$> H|@$> PH<Z{4> u'=E/> H|@$> H|@$> H|@$> H|@$> H|@$> H|@_|@:dĺh~> Z|@O%> H|@$> H|@$> H|@$> H|@$> H|@$> H|@S}^[w9%/5.K?λֻU'uu|%c-GG"5hlK@̣-q(m)|ԣڽR_ 7}]ZM_ҘEWt:]#+hx,thc|t(Szt1+Ít,tVNW"]ׂ #GCW ܱ5Z ҕ ;" ]1\w4 Q$tWwpGCW 7bW)]S^ ҕk:"bx[\b7~kzt^P?"bGtgz},thãWW2J3*j6ڜu^zW+^ZD ׫e{-8j@[fu9B}syV׋A]ůo..}߲`2w^YV/o'_^owx xyg~z~S^QݫGcù(om{/7h u^+/ޯgKb7מ..K{fQ/V޽9)S]/=PZ]$2vaS2dl]55j7q_Jxyu~~ύ)LmԶƫ W?p}.3lrX߽͏H Cc?eoWIomV@|CCbuC˦^%eۨ5 {mC!bPUܳ+v8r῞~e8$#/vӇ›xMo'zRg3uxVgn̻=0[qDR4S1ZH|4.H̪GLwPÓ!5IY ի/̸Tʡ58[wb<뗷1߭.^!z;\wx˼>7l7ݗYڣw+nħ+V|jn5z xsae2cv[aRNOkm$GJm[|? " ᲋vXj %YL60[Mŏ_U߯kg܋0Qy)%՟ T?wt2nE8 8q)>C[][3>2',nlnTOߠ!":^gV0E+vkr]؃h;-Mm_YG2Wٚ٦7: MJf%ZIH)\y8(vR& @vT (R g?ADLNp:#I~)6s~{ eL-3Ys]6O}K4ΚJڮ>:/|an5n/rm wbZ%(>nnl~O[|Yw"n8[=^XyO0Fb,ꉗ~_+>nvj=_l{FLK TRWmVN)+zUӘ#R5ZG Uib9ث;a앹d\b$j8BK9:$} @$THQ)hat`e,J.+&΁r8Q<6ʈ2Ń0őd40'rs=0u|?rIH0ZbpķC\trrӆ!5A3*!OoA^9 Gpֆ(On._8rQB]$pDA?M|E6S LxU_KG]=f@j$s!smw6v|WHc<ͺpbU+-wo)][J2]1\H {P)'zg2d>E96 r< lkgԭmgHm\KwZsS΋|>b_l6|;n3|pjTc7͎ϻhgoY`?\W&#dJrh,KWߺln:tO+Qkgjƻ7{]޳?ul\͵fwg-?ʘݵWvt2]iGz0s(Y;3ev0IC#V%'SFi zL{nɾ(JNeF:}?5'R2:dZIFkZB4=5CȭDh*Q&wIo!:橏ph%7FP28'jΉ)8}w:SSHR\܆f/w@p樼? !mVQ|u[=s@BH:X`=DFBHH S}.XC$AQR UqrXXJ3BV Y'µݜ]f[rtt@7?5_~wCFh̾q6MFKGN O4xŽݤ"4`pFJqElnx>#qa+%f6脶16!-bFbGlÊy.]L;Em^y=Sy1d40L, JC&IDDLO)(&8@ !hE{5bM#[b+$QdGQ: a1qak/Wf`<[J?EDQE="Le2 .! L00<ɀ14o2<#0(c$8h %Dig\iʃ,eވG=ZLh48п8Ԁ,"YLKEYe=.L xqQ" 6J-hF8n] |qx.xXL;CUqxx+.0]Ap}E?ZA{59BY\Ky,KyPR公div׫yu!n(DiK~T´W%::;T$j?mvٻNVBj2/XE-cU^纴$\3ލQJq NEDL)eXn4%A U %x3su>s :ل/xEh{x\,xw!fyy埅*PuB%UU5)2@䶯@y+Fm@bg|LG||@rlIhJ 4lQ4Lj()c$pš0z8˨Bdj\Y=MPz8 K4jO$C4&|3mIJK+vVɗۺM[e4.:&BMgqnzA3YdTr&{XJa}q)/_u&d|l/-!2%O^hfI*H -O5ΈP3@ HY**uQ&NF~MC5"ioh TǐjJ3d&:MԿ<8_$AVS<7xJ N u ѪRRD&$'RىtT]&Jéd£86J%R.8ՖRp4UΑÖ&\mmҳpiwVSWՏMv Mt&R 6iYo٤mW]l~z(Bqpئ_Ff ~wc)>u_;Hᚉ];>?;w]ݲkT6:wJ{U6l_ֶ-JaEBa]UwNޗ8&~LqIs81ߔu>%Ystb!hTBΩ=,j[Cv' .0tjom2G-1(~.AJmz$ٳ5heaPN-֊sk F/.s.k" Ďh\̩IDY\%Z>_KƉ8JAr0gN#[v.kkvmת^^٨ɺ~]s!<1MO9;\R3ݜvw-\Qv{ݢ:4Ҍ&C吤ڻ:O#u.sfҾRftw>#&m44cw&xZ^%hQ)!s$Ѻl"D6lngϗs4ylnl~O[;] [*;o|LèpҒjP+npetN\J{JFJDqJ^wێїubBq%T혉5?+{ti ѳ+F^A14]W8啎I0rTZgF{QBKv{J8$tNᎬP!lI:A}-sqC-o}/9 T4^.RH)8ahez*4r; P(gH5Q4D!p óD="BOsW,㬱JH|.-fHAb'(IH^?+ʙHu.C:&h`Za6h'3%yT"Yr: C1qؕR+g!ocs[kL%« vtj|?zQӤƁȇBcԎF>B1b$a)Rs 2A+OLg7#Ϸ< pdPV8C%)Flނ% PGQgӥj$,ZaXx>w>B)feۙ;{s)ՙq`A8Tb:;_q~Uw$(~2M>x|9wh0}.P;[•Yy\2Vޙ4Jt̄`=Є)k$jBaOF{ytύ7;nnܳH UfĊj_M]<o2Lfj3n:ok^4՛֏x!XWf]izn>s~{g!i>rpBmB9hLB*%QI0_aie|ұ 2Zkak8nSZ& )Z6HBB4kdVRŸ^+Nh`NkM7,f|bvCqa}UK-p R9}{=? T'+fz( m%fl}8Tw:xU&;gJ`fŃLGލLovN3Y6b~ǛPW7}s#T޻9=t+G[#aQ|oJM*kw_hv|VCQðKNq&٫c#'S$!ʦB ƿ??3=%U)Z?o; 5jv<(WD*S(=tlKpvxH?03xۛWha |/gKن?`^#Mߝ]В]7֖5斶SCo?c  D-јFz^VY|<+lOcɠ1iLVH uVb%De9D^}.&f.0_oFcx QRF(StL',GFr>x\ٚАv!Ĺ?/>K4N_$_Т}=Uˤ aFM*h"B IRWe$DT2j 6 "-5`9,c%dU"{f|29L ҚTZ"7 L2)plvXv)e9qEX"SUfec!H*Lɠ>Inlgu/HQFQ邎0l!(@-Hd8N?U*R/uLf1E%]MM\q' xN (ytٗ\9#C&tC@tr2rz >N]]pHB )f%ep%21% JBBeG+S 3[ eB !Y!0D !'P)J3XY-M"H*'RURh']1TdK-]nP4`rIDeb.K10ϼużhl.Ōѽyg 0VXGku𖭃EIX;=nN ~Rշ,)=HyBBkDv)nOs;<IN_E#4+(+pWyL?KBs@۶LWnۭBva^[ٲ '0k@LL7yZlQ;'k>2RתݪO}fl}2Kg?uY4L/_8>֖[^O&CnKwߎI?J, &.^Z~\ 5|WSh2s' f.?T7WEA_ HƗ Q/Ɵ/_̏"џ77z|>?dMO_sygetNU c 2>2+$@l%KA^X :LI" G$+F% >htdgFᫍZ[ ZCWTysF ({l+m@I[Y JKi.WeϭO,Nie3̢;h-do(<佗=I^(y ,krCk; rӏ &6GG1ce7Z M p )& ' \6U ;\4% UQKRq1qcPAKGFk;5q;q3xWew-}n+b:\WYoZ,kW\j.]|cWZ!uR!D+8QbB+)rO!x);B!eJYi,œHhd2a0ڂk섍1" ݕ6p|y!UtGp jځV eښ8p~Op=70.]^8BHIB9EUaƶcz{\ALWJu"`kƦhU!U ѱ4HB@RB(Шhx,I0R\Xe6%Ά|!Ho)ݱ>Pp 2ŘPa4.IL^dWf6KJTs 鐘S$Q]$/< SH0ĢHm@V貊"oս-)Z1ɷQmK6Ɯ.$aʲ F*R2v<'A+!; eVXl0}NE~=k[{~?YŠ~9s_4IhU*w9ՇIhaUPPK9^%l啷UJ07w("qBT}v]42~\٭Q?W*kx9?<~z&m>v}U3d-OxՅYv\[Ȳr>DŧV+_W-ey y4h$XzJD+>"H.ryO9s^›*I ai(,0kR(p18DjL)ᵡ Ѵ)HgE,h/cAbc`Qp$rr^H%#C !\o,0>YBB|\8kQ4MrO33BBh͢,ESZ%٣u7V՟- 3#%53-Yˬ|r %<1TCVP =LJE5žjhۡ3%]R*r{(H+eUT^ i+ﮊ]q8wErywU\qW A**~(H`HiM^RG++>wUEy(ULu]) tW ]i.qW$.PUvuwwUԽzp@]" i WkW/]!7(i92y8誈ա+Cڋ\z9Je9qč]d7*r_QG͒j92*q>Ɠ0L>jrҠ @58?GK0Hڢ: vU$D/l"eN-N\,'?;{~L> \FyI~F}EH+kTQ8vdi6yұQޤN>X侗!Mъjث^ҵZͬR Օu1WN1S-/NJ>^AU]ddf)NC& 0u ?03xۛH#wJ|GUp 8?Q޹ƕ#F`a]",b`-g&A{Z3v!v֖n:C" Gݷ.wAWϢ)ō' QUPOY2 n$X^Py[]іtUE%BWUHW\Ѝ7+=ZԨhO~t#G2Jq֋h+j"ɑ pxѕJ2]W;IW pzѕujgw@ 6Kʚ- 8'7=9ōMOh'L}Zo=ES ̎v.Yni=LYpr8ypWd[]і(z&QU n[ګGRLɍ 7]4+,j‚#]p,nteFWJ +$ڡdHW ,nte)zѕfi]WJIj!xʮphy]%=*t5vFWKً6uu,]W;ԕJ ].]Y([{KՃ*RJb}pBg2= 9GjGR`?ٕf7RZ غzgpJDauTJi e=J(ŔJ<)te&2Zj>2J=ʞ3菪r君>0ܜg.тϾ15+1UPz-{9ka5Ƌ3CHGcl:7JKw٣a6{Γ tŢ,1F 䆿]l?'M+0'piō^4xRJ!f4MgV=]vO!l ;hh6֫6-]QUb&GR:\B/2Zue=Ate'gGt+!^ntegWFɱj(Q#])pFWk]WFډ]W+NG2'R\ ŋu]%*J3hя 7]V,}jJBّ X؍ 72[וRg+HW 7A+ZוRcW{U-tN+A9&MRēR*~^VƗip^1*SȔxbXRBZbc £2e11%z߬nJ|/ y{Ï~y飿>W^y~^Gպ>V?_{6_K~GD8E.xŀy@`šO'_1;G@p*+KMJA|r@EH4ݪ,f3f@hA}b)kH 3G]eDHW ph ] )L<>ä#̍3~{sjK?9_ubȪ^xaEr?Uȗz0xr.)zZ4eϢ)-nV ޓ3JL'U/D*o*򶺪quGs;M*V*v]ݶr u9ѕr@/2Zuez]GRۨ*گJ)tH~tER#Π~=)+S¤u])e.uŘs*t WtPbuC]d4Lnte%yѕ4?ve3K]%= k7JqJiSRJ>3G] CוPntghzh%]*#B \KnldPApF+ͧFY{l?YΓBy e֞ I_$qigp͜Rh]5h:Y2Di-Sjh#|N骊9Tuu۪XB)tFWxѕҊu) L:`nteJiyg QRuEIʀ ѕѕR2֞5u ┳xRFW +El~([;OAt5br+.FWvǮjJ+N~ 7g/R u]ek DW"ESgЀ h)+=ڣ耣PU.8z^g3w\ʾϙUыG{T[<6*bA=bASU˜ܜ!U!#44F+W1\2J:A7ߵ'gR7'g2ES{iSt+g2/2 (K'vt%gV\,"gWU9l]Ƌhyi:6IUYHW ntu-^teD(~["͠Oge(<~w~yXMt=>Z =yO?_Mûa||,7h ^/j /޾{Fc8n^Php5ᣆݯ|gB9M!}̭zqv1'oo45?3m;8eCHE4]Hm-.ˉQ.j􍷸6Ɨl'P@oQoЂ;NmOï6qJ oyW닟2\(}>0,r^4a)~Gfgoi|Ow~.ZQ\<{9/S4^:r pL4FBK VQ Ȑ s0aF2Lp^%inWZwO]lhǶDط ^)M̲aրE YDEXwkAF̓y` xK}ʲ ˢ91CYYky(j1"+:ðC1Ɉ$^Lŋ^jFu*5i1׸(2D5p 𰄰irF,} aNQof1FkRkC:dcx K.4CH2K  \<2(2̢Me`AIW!SI g8ZUjw9Eiъֻ*0ku* c$9NAC^V )IPE,w@U{R/7?.^?Ea"iI)˨Q>%8ĔǑ˔ҿ%Ddb<vWM aWN8Ӱ#¢M4V*9),_zGӇSJZ Bl8h<ۜ,`tYNPxf1r:L%b@0uq;V$L'Knc+q0`]ӿh ΝDi 5EZSRP'ՓGݮ Ș#r:Kãa#[J3bO*Vh<%2`rh[SP/O(7""UwӚW^r%5p hJQ Q%fdi SSCaE7n֣bb + IISL1. @;XwHN_*@Q 2=Pʢ#Hmr3u#z@U4h+~(]ڈJ5(Z58)pc9 E5j+M3| R5zfҼLFLPha3\BvZ;Dw*UaLMVk֫Y88ڢ6VP8j* (FW69EO(Is!~~)H 8QZ{*5]QzB%bIm(6⊑4 p@77^p\T"6gSMh`Uf=]"/ c0&iPd`N %;)Ւ-0 YX yۨOp]g銆358D'#@߶zQOn͋j 21&(3PJtp[ޡ8}{ 1]<&h*#߿Z`;uJi,J9l[Kͨmgū^\'A/imtn 0 t6o:9N'?/+l:_H_#O#H ߩpqA%i5[MwHN"n\//ڶkɛ4:_M=@wRl* ̯ҟ*yjoT()G"C8ݱ耔B"+vb3(g:u@bX: u@bX: u@bX: u@bX: u@bX: u@b뀾Uci:?`5뀴pV9ꀌ: u@bX: u@bX: u@bX: u@bX: u@bX: u@bз!DH6Px?Bykj76{仗sj5^7׷m;Kku麟|8W<*Xu#[ =FD%y%BȲ [1)XGW0""M8"sm8"kzpEV^9•DpEYr:Z&+&0\!;v9"Zl\O~ԁ9•J_]Oi^⯵/i7j4wm$C\붺Pk>zz}G'_:B7[/djz~WGW7wy ߨzo&/b9bfw ՠDS)Nyatˆ`Z w<_m]`,a6_ptr9v\ʺ!fZgȎ^Rjſ~ym \`En旳'.xOeN^\ZkQ1}ycS4ǙlVǷH?]m/d_Y>>o ӓU*&nk͸\M Pe/t\j!j_$fZ\Z;ө^w_r zFŝ+mt7ؾ xq5r?qv);&c5V5JPe-j:r=0p˶k+=gQCtH8|wEKK f67c[ծwAQ %jV<#2lhߛ|*7<27{ᑋ,P{DPMW`1ȴfETx#cCY{*1<6:!$kI<~"2YZ=6.4{}Чh>qn F3.a^W6X&Wr7zMXk\:RU_Lb a0F ]m{T;R5CzFqgWwp{cZ3<|v? 푴K:Z] 7ؼ l*5:R䮪9Q{HF3~jwLcPͫQ]/5|e#--oV2O\Ӧs͚nINaJK\?u r9\Է87 ݛ}%vSČ͎49-vy5lKK-mbl @͜"B{Sn>MK긻B_MOP\=8h2:z'YkmpdBvƣ Ԋ6D+E&%jX0SbD"@d B!:cKzL δ,նI'Kqf?Vz;ߺҊ&(c'Ҝi,m)YZjbLE &@9 7K0rO:g[c4-§3oimox^VݚO::8^7 Z(UzMNzmAUeN30ײm^Mߏm~yXg$>Oa^5T~1 y9ZQH"Mx1m{ q؇ꏥDi}NZ-3D ^҅W+Fa5 tbfJ/Z[otE16ق/tz6n}!㶁f=~RΘϏ]LC]4 b'|`NN|͹6;q\0{n8_3V{As7;tz$CyM3󢷸+-gla{N}1J{_kfF0d"tu :W[tdv5o'](cOw'nYdQgugݯK*=sÌb fହDqr͍zw֝~("D!2idΙG7u g!R93K YHR琢2 D%\äV!K|c A=Xxvf22IK^)ť6HBB4kdVRuʸ^!;U7ذ@(\>Uzą;VpU(pQ*' ^pTa{ȍ&7LX'R jQqʽLUw1sxFӼv#YDڬ3cЄBssd, Ƭ2̭ML8K~[m$#D$E팕zt1%Dx(1De!\'CĹŌl:qK?%mܘH1FM1E23& ƭY,Zfw h%}2TH.%.D74k =ޡC/J:&H.Rh,ђ/q5,ew,ED"_rL1&T z%U.IL^d-k{_?tkJԋ%tB]L钂H1I^xD$0ĢHπ&;He鉼UnI]5Q.'a]1tf%S*HIrw BVHkNg`c;dc)"?^me7h*Qu!2@%7K{V;xNzFk(kpy<ڽ #>[Q1çԲ_dCR)3Edkw O֐s#)%{MJqA5yYJJBΐ5 ii@ Yxo00~ۛ8é۫.Q\6l>xGxy/ٿP7ڪkBl┌ijssY]ͅS5Șko!kuT znj*A\ܧ@Rt H5M42*ёeA#>3ƒ6ioл:+lMҬ.mo}Ww#o~+ 'u7k1@XV! 'tYEWA[GN&fZlkmAI[[ NK:+~M||^ɀ)lFYT"r#3VE8l8G!4%sl +J-/hMCkf mfcU~˙M9uB}B$\Bjr0l6_*{d9$ wIdUpTg s!!TґQ\۾Ĺ߂Yϧy^EXfM{%m)r_1x/YZ%z,l~~پ> ?BB^+SqbK'VR.\H.9R )w…4L)+ X>yb A& fA[VbrP#T8U#g >=jځ+/}YZhoHy|O=r;URfw@OQXvɀur`kYPׄ=e "{r_: ;^etXyG| eyQhJI* *[#Zk QH 9<4.3<\MIBai24fMآ.9&(9C)%6U!@G$3a`A{c< ZG˘3Y*1bD,g=Ho<@vbÛ7.rWΎL2hKXd>$nPXa,CW]z)uQ% iR[xBA'o(GGhe)z +ݑjr&Mڕbν)ȆŜ}s^Uz%uK|ha{{a[WB;Ӣo⣮_^d2P[8*!uk^y[gAOM&ruoowoϪOIu1ivN_W*yqEdJ(>#=|z6#\`D&Dt|.mbJB[2q%zl ݉ ,cOUHbU޷DFէ4t^,?ўSi{{_5ua24Y9wu\qpscu, +ē hrOuн>u+,iDK[UK7l-+;^n5 MRmFZB=>>DBWgPUBNIm}~߬Įq7-H-Točvyfw¹9ht*1c@f#$Ut[nH}Jn{?0jRhi9Yfuwz|ދ/k&1Pʠ x$ER;Hͬ ] eq,Z\ԘPk^MMg5tow}!UIz;abF5wՠ3)iMNް*薼dya z& !eL=Az4!d a2yZ#JXtt=k-^OWLˍ+ tŭpDtRS4Hdz\$1\MRe˝Oumn/e A1J`A2Z(OSN%IT ؒq6Kl `m xMe6ԣ9d6 h]5!bMga&8/9XFYTy1K !J+a GCא6azP>_\oh۰0ⰺfy5էwy>4;s!}@=~k,g5.! ɦ?f+C-npPőWpVzpZ$򗒐Oë5w dpx[IJdoV{+a7&8㒉3wl\=7S?ܛi;Kc[w?Q֎8/ɻĬf0-VV/ō4pԩQY'd(ׇE2#g*3y|u=8X 6כVL^v"Tԯp[.etK74鱻 T3EƥHS gCnu/7Oi6\>QWX1FHE< 9ќHx Lk2gqos&>+ 1̒L3~@s-NXϗ]tv [t҉oz^ֺKPtӅ`Z/DXVN_#XCrRZs& C6Sj\[O7r'?V)BhKz1SyFW.6#Ewķ8`; o6 >,)9Uad[LfΨguYUY(fM"*1D 9,yJhzd/4/̤I]*~wx|>_2DVF  j*`-Xa+;mC2m2y0w,wH]j' ,tϊ]/57owNͿ'=⓴nVPČjG0Ėubs9C6t9ֶ97/tjخJkCP!yዐ@Kl*':gG2r]l5){5үsgj}mց,-ǩC%l5ǡ\j9?PY8 *Ah6xJ?~c~Wx&BҘy&mEc1y b![s+QڍQ'm4 U1DP:L%GPʥ3+vjg3 eK 碓RŜB^rto;ucBWRc# }`rțqr^-.ItaUSO.Г+_Y,p}d}˙Xُ֌'gϳ>.f y'b|L7/(~^+5}5^s U̟~x{lSbL;is9?Zc@m~8ΝT|hSHW<_g׏%-fw~_dPИZ,쪑}qُK#g~+l쬮g;]ZY?ă@n ߖڿ@W KFɍ~RpmZR|VsylYLÇB7=y~kV;wIZsz~z<[:.Ok&cYA;čW/7wZ)rp;4osOѸ9ݿt+->@IPM61n+m!p|AV"sy cc2H<5&=]T9&gǖxr2WQNކBhTcBr]mtA,贍2v>E%k pw@zXb2^c;H LQsY|H4f/fQ X".[oE /JC3`PBK$Sё&9-M. sNJu蠅='վ R_m]T}V|UYs9\%**VUb+z,=t='(s㰎8{&HەP;0I^91D+ ],dr)<8:I*#AA'A\ø0 n$QWSٍ[Ke7V+ʵ1*M%T+,W,|-bzb&+ H_XjQuj;X6 Ws&{W3%b~y+VVj:2 W$gjpj JlqevNT+ T+qB!pb-OWwxurIJgp{)odz|Dj=1 $Yx[%R IZ/->%jt7 x˵2k]E㿾vM_0LPLÔ4zt=wF?ߛ5L߂I*sD:iTDM!vZbW4TN#F?9-n^F ܇&m'lQW ̵y-yr:RȣaUѰ\Wͱ{u{wUJ<)z4ٵ-H<_F=}۹\R7z&ձPgjuiyJsKK!4 ݷnTG6`b5&N~[UB[xz9v?ZŃ㪗`x`\~j#;loڶ6j F-5j^Wҭ5\MWQ"\oGIY H-H1v\J WĕB5cW\`dkwbYWSĕF#* \\8t_?0+Vqd]`\\jǎ+V9g![}EbgАq KmJ7v\ʱ24\= Ck \\W3HJ1ʆ)q*焣M6ؼoh-hM/ Y~rNfpku&֢z0^GA#+XpE$׈j Y-~ϋUE?EHL=2ܫp6r\Z9"V _hmR +'{^ק^;++G Gɑ\xZ-FD*MMғZSӶ \\'Y'gIg`^ ]/8u~0{r =a\TF'{J6\mlpŽEQ X1cjR@E"Fԃ+{"ZT: WĕւW,jpEr#$u \Z;zU NW("\`#˵\Z)q*j2Rm} pj;H]+{p5\YIT++s.>H$hqE*W@;]H0z֮Xd^W NWF9+c9OjmL'g@rhϓ"hN/w,Q_=^y=459$XzVeH.uW-vv2ҷ-d7]G8w:t~ŕ{U}WTEWWpmK"\`5 k=Td?V7\MWW$Y XZaǎ+V jRʃɺ"J@5bP XscR4\MWZUHz+kqYuE*Q6gpBFd]`_Hպ\Zcj2iig:S Hd6\MWW,Ճ++5Ԃ+Vkq*][j"bԇkZe׮wlkljbk9աWϚֳ#ԫחeZ/sPzBh8xձ&KDx0O"Okxr,YxbZݓcضyꃊpE%jprWvQd['9Rv:F*e^5z\seRjd Tն]/W+iPx?j:ʵ䆫O5YW`@U HW XtjRd]`\\}Å;膫 >Ei* >x^rƺbzDp5A\!Y(+ vRT+{@~jo]J&+izvI\Z%ǎ+Vm} ɺb+qI8v\JhkWS\I5ZԆI' Zh1 Q7ZϤ=)rZtRۚIzո9ر ҶU) +g]E"l5bkzNbZ4\\ ^ oXy WqE~rDWpO;zbd[tն]BpłWK\}o(+V9x톫gཬWJпXW,e-bƍWrl[/Fwzz?ONJ]KR+ݷ^._Ӈ/;(%/&(!,Q"hQD:SfEѝwҼ(G.sx-4OrH1D`ۛy}Ӯnr6Yڨ TVdtJHKgrW0&%s(w #ejd\,!xzoІz^7{R߼ |nG2_os FPGȏmIguu'Y In'(W$mЇ߯ǻ‚\[_cZfa48]Ɗ/Tm0xՅOq`t?\Ge3]/ \oaUaB-ZH"p߿/=f&ǤH t`i|G Q3p7 ZM%ƭD Dž65> gCyZA 6+\t]#°: suyۙ+%}0"5~ {k p29,'|kNˡ RMxEYRmP3yH`g4\T0(xx tM& My;4ܔ}rd9f& 8'ۋ@Zy5b.&-aM JژcnZxT;6IƓAK(Aꉍ5of $skHrA1)\UJUU%EOe:27/Y(X^>~+\v,b-QOL4|\r!Zp \ fs7\,_s#`OTd|׼aXJ1UKдZ:f& +EߜB2a|:rGbMG,Zn܎ZUVH30r@R W1LWx"MIPeў(HI2ښ,Fb9>W`2.)b`s͛+(ul@`ʝ63yhL""tphla>8RBXLS,YSO iu pDrN!q ûtۊt0] $JfʊӀ ,Xg\+b:J2Z bup)NO/1AOxsmSh$ (QB-&Y@@#aĄ1YIVjl+z&8$`%R$&,䕔\*Oݍy2Ce"iP:'QH$SUpV)ոHv`1`VVrk,( `Jf%hBM_YY7ԧ8~(7{}Nh~|`7jx+@WSU FyDWdЕ+@_+rC[D"J]` \|+ӕAٴ^-]]Q"0LAЕ˨/teJttutՅu["ۗG4 gk2oFV6~OPR[@0<+XKo r ]QAIYKWHWrGte +W!_ 2=9.$Un?{r7ѕAZn(U]]"]);fͶ>=NaXrVܟp h5jfAId?/rN !"K;{,mܚQJG4mK MʛMi@oiд(zCCwZ.k h9t@WSUR{DWЕ奜R.i מF2p1 Zʠd +s12p7ѕAxʠl+S_]`7tp?{W-M+iI-]]q8ѕٻ2p7AV7~1(nK+!hy3'weR ][%W-]] ]I &' .fʠUtePꖮ.LzI!og܄)<2jx7?gD aۓjA om<`7+9x3h.5Z}%gPrޮ.p%%{DWZbDyp-?(}[dI5b+ס+7Вj<㆒4]ɖNU=JI]`-7tevŸ Zm:]tutEA 8S ]BWbM+@IV+#a. \|+ӕA)DKWHW DGt'#_ *1%ҕx?ֶcĜu8:߽{Mlh,p\%=,rO`= ]uWyc-xebFT@.X*('pB> =B+}½i qg}oEIeL?yw_1]#N9~ go?G4^<_Vy]V/|r>.+QY sXAex:O'99ƣ/i`dW:f_At_n4R'Wi%u-vdCOpv񱵅> VLQ$ tH5qs-M 2?g~ ͺڵ3X,Fvϩ~X'm9 }I*T/4x%Kh&uЏ\Re ˰)2́VW_ U.jWJ 8sҤF5-2:Ɵ`Ųlu:S֤6<B&Uqf"VWT֯AZ(Y94UHڨ.?f]4.uӤ^Gs֛֡YOssj}QCm.)Xo3]GW׏u%5sL =ql8gѷ} f0DxT, sl[Ƈ4Gyy;WDжQ{?w/x/)X?:a ՁX;e|:kׯ_gud#6/aңro ƿ-2jx>( 91P2C(T5p_:a8&!CT*(Qw#{gZn6أg[{C: Z3w7K}Z}gT Vo^AAx@45hhA'>ܿXƔc,N%8'yM/)|S˦9Mt{?(tn_ȁ=Ƙa~0[vOcXjQxFo?x?V\)ûTy؍? !|#Tb&T+ ԗQJ`DF2V1ϕIf1ITKC!x*LmU?RajM;!)$Gea݄)U%]KҶƦ56],=dQd_ !8!e!M4F`tt\J!)*4qiDUd4HfBH4Vi3f>3ZD{P%8xќ>p2\܃S!‡V0fl~[mKnO[%I#0$* fnʨ@qxsڱO{sUzIcAe.SM5D:, bOy?! [֡8ɲDfsF9ի{Jߢ-FM!Qmc2)} HXt; z:p]xwl5L⯷yhMLl;՝dnX|*Ï\a-8.\qM߭1+\|ΣhU!Ҋ-Tm7p˕-0-}ՅOq`t?5t5kƬ啬1MM-鰩](8ޘ4jsn ]Đ4HfWpZ6:t#żY7ߵ)nKhDLc_,Z?!2̅<A;K&MyɛJHq7;.{jor#̛Mv7 DrHNiZLXJR5KVX!80NMxHDLQE ⬣>| MMCiZ+TEvHX&1+p˦4Mx4ȸl-4d.FH UJIΚE4廉Yfbi]% ˮԘby<}%Ӊ fM_ &YWw%sa0myNr3YN3:Y/z4p,Ǖ}\y!\ўmjvOK&<W70̟|bZ2m%n}/^d9"ĩn+P*QmҸ1+-KV&Ғ)VQ)옵dF'5IT7S10/Vz%KdS2+Kȶ7f9f}fLx{L˅IRj30i^E J/ww1Hz%Ƚd Ϸ݄QxP&PȥeaڥŖ%ٱ4p'm?v!em""أtyTвtөpYbɣT0јJlE{Os5mw6G p謿b0C0Iɠ..x` ;V D+Yvh紝M ZԢ 6{8_TAOaW82Nά ^5X_V+q [&jXqݝ?fY%VqĐR1Η,d%+q$0'H94AJKVf7CJ%$hF5|ȷ=p~Mj .bj2ԗm<N{K<5}( }wPH"msЅ@ʟ/tu,8-g>fNPqƂ /| |[[qHIWI<3 3m .*J)}ON`a-y7U @<<<204mr8eM9hOR%e*UWWEQ>wZoJ_95g& WD%\ ƞ .J)v Xk<A|_iJ!-̲Ar/q +VG&r"hf~HPV^匈 y2պn@ˌWXΈKPJȉI 08/Ѣ! ;*7%LѢ/336ox4)W =O nrDžku2 r|͹,2!5,ޝλˁ<& `r;e Ɣ˔)4O/N`PYRKssϹd z]dTcQa-s[+\?]H^9s 44\X[P8QQ.qBhNDAt H?gx ț]VN42k.2I~}}tu ΋*u>黁y!/i8lX.8\y8[,N_jBKŰ7"z$ɕk#6َq%="12Ƹo@HMܥr='Gs%۴ vW@LYuvԀEbsct6_Q3/IR.xFWq^3zm8Kn^%UR<;O!3JI[&IF8 ڳ|BF0~CۢU0*G^4z.s=j~gm?wu(4ЗQT#(oZm_FTf!fwN'IJiYZԍ!# ^܉E@m*] ?w拍r40s5xl8DXzQV*6 *},g)@ 4OvFJA$T0B}-:t24u"tPgS 0p&}/BeD`;Y#xP >srJ7Q0 G~X!N/mJ5=Nd$'s\pȧNaRǪm L3XZ_Bm](n4uwrGC '徱_',b V\(ЛGyfBܗ3i[pT~3/ N22M MJd."1aP]F-϶$Ii2/ɋ(!!Fdž!=rss.*ȚճdEBU!NhאMۈJhr!0:'`9t06`LTq=NJ&)b j0䎄0q.U΄7TIk>/ZRUKUp4pc|AЖBH.hdN3\ />Lg#l=8)Rz$Įىۙ=w-{X˅Y'S }-xB#r^ 1¾>"d3Mak}Չ$p=G棞,[cNdmۏ\OrKJP=zhMNQB=ĕd,W<TUU\XІ<1]P& JՉHa:έ9M2l\bロ'r$Hʁv`ۆ'Y0- &qϖA!.mhԖt9h\;@drRn&sѵ^ PF@t}֢ 1ݹٔnn5u<]V)4ֈJOlNADΌ:ao(fs{*a.#neJ-#.EFJ),d\,fbRx^lnLu<4LYEofIEwx0E|簾&+CwќI%rC4~<UZYqÕ&}*}%0|l`II"euƸC$(e: 5H|"&J0C ۵wa@ \:cU;zaJV)$ X]I$G\f_{g/ {U,g$ UeX'eF 0jc e1~ i1kY+ϧ@GH(n i;rv!gm9[+ٝj9X;\sΈ+[@Fsa/C$)7 z.G+ݭHPɏ^*-{ Qw]Hƭֶ~xmI՝h2+YxJa9X|Y:*&$1{?[w>Ԉ!\~] ӓΤQRөdWB&BW/DeW5f$GW-nH]ı k~5٫@iC/&-T;MUPAmx?Bi1_qX!(PUOWg¹0im ޝT+ݺϼښUzv]I_fd+ݕ0(\w(nq>j&,e]-@k!4JANyCoba;k^FxcfF蚖wa52?_&ҕ8qJѭh~",W^<JQ]c1^UHQEX;bQ>sY]:luy6<4Dj艂Ո7<7ɮvzxM0yw+gLCݯ^}0;nDlq&uchNV'W5.mМ+~Zs b'W|fX= qk俧HJ붕>y^H& 5 'D\^OyЂx"$9U΁s% ϻdi[[GE?~ZNa:mFp ꝃUI1ȁK+I,S5B^<8f_lHpt;*V%A,܌[Zkkv!(*񸮏[>sKΉ_l,7R/Q1.Hԭ,UYܛ/f4bY^=( G/uW\ -rԧ-ᛆ\ScMC:"(H1o2<[Ɵ@"b Hm3y,W\oMuiի,xP >8;YGy-]&ۏ79W2$A.ZHE 4?{Oȱ_!;ܫF% ])gMS^B EmV̴R2GREEJCxZmPƩl0@iZ~gcq]˴ݶFsn eYLwb 劦q,A%'5rwE(TejdFOOQ-85J cCqg4ɈuCSnLʝ,;+w1aiCQg 8H= vi#$jf]MEbt҂>v c1Br Ҳ {GkJ:fɆ۾$Hp(-"kLR![*<9H\soܒsRઇvΘ"*:fIoӈ։R Նyzr/t}b +trK\uxszHIe V&Cϧ};33N݉&ח&~c]11nxk.]$UA[o25-K#`:+W5Xh6/X;.uvS̹]ZUsZF`,Rm `f1l51^K[*EJy>CC zTX=k~j8 S=TwjSD3ָ=7*rmx᧝DD!m c}nw;_x 9Zb[һ{x˚I.DpǷyDԌt#u_ru w)-Px,~O|RBπn0ӗx.bu5B0,iKډҨ9\1oJa D' {N}X=aV b-o _z?'%3&0iԬ}K~>TwCO~KFtwK׽}6dz뫢$<X`?),r-9gWOj lX?:'8GƇ>O\'P%t=/0̄|#\;=,aF#IZ ݬnۆF\jU N nǼwXڕz96v]"ud׋l׽?3䷾Yt16sA`i`Vfsp.5Gӯo]d yz3_+􊑳oi{*3 zAȖ@V_{B8)`hhWD_{ ٟ?LnSK"/-">5oC~} F-u]\m\Xo ,g9Lmf@-:5 #1C q `Γµps??G) zT]9[ UٷK*aӎ2*X[yrԍFA'7*5&-h+"*lE^Onhh)VXfi\ Q 3lUҒZlj%|>>W0>U(\hxXPYCpNĩ]*qK0O־nW2f|s7 E FRmVS$wr|8 q?)C˔Ldzjb [G J93LT8g_ CJe6 8P7Y=qybC=~?$1F(qrktW]uem뻻d,-J.s#6HL[xoD0v}|Pfoxl: =aܝ' `ZWMOb `Ep7p q)}감 Biۦ5 >0vOHJ]XDrYˇ8+h^WZKFkE˖O~h5QdJE(VxD')qE*`W6?Y5OQ6áGpv i#@; (PІK`|?P ^fcvxm SEM3 DRDh>e6U%/G ; N7ɴ}`. 3T]nE^h3maP (i` jhI0~wcJЙHB Z/,l52bLDeS(7PpG$0е+xRsTҼ/b1ɹVaE Q5"Nw绔$;d/uVQwF8 "N҆&upg1e="z4%k gZ,;ŀi:5Wv[oX&'U]NlM:q95bz\ inT4OX]~w0߄Z 0qٍ=xFuO:jD;zc)0x. nAV jk8i:͎d XY_o'pڲLtO?=|rjKOMS3I2)ÉHpj2ItJ'%^ip3l/TOE: OJ/'Z5l 02/.E$İ`06onyHZȞg |JqF%ί# Tjv4".=a ?obFub杰4m1a>=']_]\AgwlՏ(}|~\}-Rƾ)c_唱v@ThRSg0q.#S8")O8v@ǰq2e炚5T/{f]X֋t珳0jW94Na.[-->^EDjFKe9Tv48z<<)HKTGWIEZ p(G[?g]a/y^XP!l4WZΊ@Bi[p&&zfs\e ˔s8R$h~[N Y~Q6-3*!80;^:nޟ0odFA<-bDR*%EA49]AN92C\  WG Ruhn!^J1T%1)4ULK!phw%JHZ!Nf ]կqicM۶rztzݏh |BLV:10>$o)w 5h B!~F8;`?LI*Lr`}c Aech撅!˔j!\f_4!k׵^Ž0qij17 f,}lH-{$AQGzñeSk ?LG?˧s OwIQ%)J9Xv!C0VoY\dQu2e52@8u/p=:7G) d-u(=5W˒,}BAN l`2E>:&2pP%|nا\(e̻X{y 8sXZ1ƴL;SQo"1h0/Ơ+o{Ku?~^I1hĀN abWZBe˰Lj(q^9J\@{T\R[fú&VH]!`C DSwX.0~8\5'2 Mew uQ#`0ʫn<$MWsn ȐDI}CWPq% +-J(Z/ѝDJTiP[ V`8(9 `|_E˴ XW4-?@ų 4QP hhY;~~L^kikWZ&^&HRt7 :MJҌTZtN.RWM{'P__?ǽ.RO]=]BCen^NDPYK?3Tltލ'LE鷆*`\B10<^K>M n cg[T[{u+|ҷ2X oEukyN ]7Gs(!_ ׽=9ը q+`\^Ǟ\zgFa]mH =g FAt11vS[,XwL,`Ffs!/<\"^ޕ\7vB+fb5(/Qc8Kbg8QױI1Se%&xaw.Hg8ms/g10 ћ<%Zm^.uV#>ȝX.iU(p ڗ/tv(t>}xoos݋ߣLtCs5~i`i=A9|ȄOʺXYF&JEvMf ]0uh;"u&?el33vVP"MQ9,9Z 2,dF՗[~3:ZJ'rGܕs"BVw%3:fNKѠ缭/!cOO(Uj:穎A9{:^H&5']6Ic>1J|ܼe+E4|EAvsч1_4fbofiX{ʨ6'7Ak@O-ÅURjڼw5~e\X\8;f؟[i2reξ.ꀍgyX9F݂Y폺}1VJ~`5D02Y`_?D MFk* ]YCrޒ5|X^8Ak/ o4^ػqlWi43@zҹ$< 6(,UF{,/)k)Gv^]G? C"$E2lh4I+34*%_Thɩg*w&dYWrViv"\+N=n.3s EIva !q[1K hچ*>/X#ٸ/yP$[sTXᠪ<@ޜqt%%#At>щȹ([W XOcisMBʹUǡ5UXyvpZpFM0#܀#ѡڎ=BE42B:xQyP-ySuSK2TJ1?n%q }mYi^+q-NЌݳ듺Sθ lˈXM奉V1򥸮:8G,DZ#[>[:Pݒ=UF;JL )iDbJN64*CΛV-?Q}-k˺YtҫF$OU\I˔*5B``cb,Q!xWLJ_Y 演l:bݰ:.1Gf-$]j[(aC`Ia:k>-AI9zeŗCtn|hgC<=$`>B a$qbmzx+^ bl }N^/ֽG x[qeP%4Y%_{aR74g*uw0_\ւ:+. e&_rqT ̫ƏՋ*B,"V`=&Xtd|%4z$Cm#[Y7+L2h!tyo3կԊ1ET%"_RDpcBE=iwlX^Т/ckA~; c6B)m|!_M_0f .q}wbeenHJuG~#U^* xDJp,Ŝ&XEǾ)f1/O$}h9wZ=&yҶs镋-qY*˩$sS 4 Q 2QbP, DX]n.i<8+Esjn%J$A:5M)z T](XBm ӖV$溽B,lx+xW^\BQ*`KeQ/W9ca&ȼp>/삵 R_zE nzLfz,o[炿KuiWMYX %4z "8[9[C >)c/{^y &Dr4PDWBo \2%-9{cúESB}]XxYq%4~yBB~2QCE ڊ*k0?ǿt !eQG)څ[Ӵu2Ƌ(KAh_vEƋi_R0^k`=ks0\+Tjpȴ& zNV G[1b$\,Z!V5|v=1D/oEIe{'$WgAjo & % 0m/v+OMvUY!JcsJd ƓH8A2tw[čs}W`!Xߥu㻇ZSu~Ag@gUƆn꧴TʗRS^5/[ EqK QB2SCt''vǹYAe, KhTR*|&_/sroEXmVڧ4ׇU;/"|~w>5xPV[RvVK5g`u?` s|̍z/. (m(+=H:ӻB$y2ut'+^EJh\DtD. w::^;.a/$'191T"~YBG$(vQIhIxl 'C# Fo{i gW  8ƣl2{| 3_'_m+߳غܯZ<}s+?o?=/ꬲIp݇ ,AǾ4 X* `8H``|Ţi ٦V ˶>$)$Eh !)Z4̞딵FlLl C Ym==?m`΁%}؞?aNV4t#VTߝwfN~'$Bjb) `"yM ưq;0 YCu K"M5 A%ɿ͇W: recPЛCf4`8wSqԬO697?D MaE?h*8=HZqNx WjI7rDa]SiEeӦYTS L=67f_>}h¡K`58#tJ /uB4JZBᒒߝ&a꽈)0 v\X,cA& 芇v#GǮ#* Bæ>'Cg _D¼HJ!#pBE#Quf8){􂁛,Aʌ}ч`1u8F1o:PՋG k<п7Tu5+I=i\b+FIk!|vY"8ȴ xAs|G-S2uy.~;F"uqECl4MR\Wgf>21nr&U-IScZJxw^ ZTUotuv.D)u9EN_/zkgQbro87kXV kQĩсE׼r:c=;xJ;eT2raW}hi U|{NnuY#4v=UZZ$sLfjl0ӌ֨{kRs8npTfx/^V|Zk~\Ӕs?lDJrP4_UΑ޾ = :xu<8E Xa/>zɛUr ָGs#0BƽqpqէrDwPGq C=0 n+lJIMH1$oAuBm5ܔWx_afD*xF7IgIm`bA/Y̝+vq=э,wbrQ .2RJՖs`>Ŏkd(y {yR$psZ.PGANZlTN+ġhl|{)!YHԃQYRuH=B/%>(L\9UE !a#0-7Cp xPօ{f.i<+tV Çe^o,{)pmp22S02e0F{*>1)&]nS͝Qc G%-a҉JMgD.WP|Εعg%Ō[gaqn;[ }hȓ|skaّ";0ދ%4 ?p.:b{.SIĀoĠXƉ$ _x5$I<)%][Jk{Qd.w hX%QhDWv(Pab,QF fLRJ2 \V.q5LtX=D2L"$U8 -C a$q @؎YHos5ݐI -ֽGS=Vs(1",HH#SjtBQyhU/S1jYRfgz8cYPp"9ʴ"X$6a%Xv#!A^l(`JGÇv-&-#{Y#l^a.!Hk;5O xH-k>4x:xDw +L%4.//Î,*Ghagkᙥ w2S4XJ ^gTh Ob$˭)&_o}JǡC2',%Hc#~B$G/ -X21 T'T$[M_-"UkZ^q?HB084+}|hTۊ*3_2[]8cN ZVoj5$n&-hpZ=ioǒЗ}{#Y-vl`b>-ںLR })Jf43]z^"㓩&jTݷh,竓T06c;.7}/ 9B/.-ʼ;`LQ05XT,7.PǓe\r;`<0J=ñ>S cjS7F(H7Y7Z0=2 foĵeM[(lW4тiBPTʆkZ "Wo-%Ɨ0[0rIR_{-K]̦i>;.I2=Qb91ߗtb7דWJ C?9+AL4c''ߗ:7.&{z:C6QCFt3NbQmCx5E"z>}3~L叇|59%tL::1ǂ~q'(ML<<]׿y%y=c.};,c} Xeڠ{8P/R~EQ;`Ѧdu S (ttV0;ql!_j5U՚];%7b` 9xr%oxI>'dF8ѳKV0U[Tn0*P1MVwD2.Xn Ln} [~ /=6"v? 1Hڸ&:>n9C NLMV YkV HMYZL%-"G{K`֒q*_A<~WtlOЉhw3SGgg[e3{ Ak+td1yCLN蕰ITk_|!uq1'LKTBk*1̿ԉ(A](dD9sZ !,v8`(eR~G=)rs/ 㧳ܝFyrN'>]VO~JE7+U78C>ܢǨF0^Br L*}%dJ[+%A(U|4ܢ 0vȅkL񳩗4EH|(k[F@Oz0ﲣߘK,mfW0w{um&)sl5}SkׅoR;_r6X%c vko'|ƻT$=\e6pkqحRuMG;irY.K6_1&պ_9tJݛ`޽чwԻfטEmZ,S.rvM[zm`l,Q]VšGA+:%7-D rTR4+5jR%J9dV:˨ _XBuQ% }Phdf?" Vaj*`W^ȕdΏi<-i=x,3j+H=]_ox<lho.޽=wx1$e# km#6z. U ȇ.z>s,bghC"F@_'Q!R@KVq!n)Ȥ[򿽖9vC h/dU 4x*X-pye 1*sc+儨 *Qq/@h!Y+שe 6$~{%M5HL ՟dAGTF _'[4=>ћWw ;Fr}p%bzLH% 1WE pH?XD;"K}uz+v7a4}_.Oky"dJRIT oQդ'x܆:k0+ZbzES֥0q S T<9sZ` "{yL&Zy@6b"h`| R8;b:L߭ 9 3M@szA?3Bgmj|Q &dTNK\n(W0֦;un~^'^Xv'R0aɃ3E*nyH%b9ojR K&cfW4 }FJsmNu53l<ς "5[*@:1j3N !.X1l}3! ۥZzF>kAhCcG//;0G}E19Rz8Y)GB쪈NM^^.sp3Nd5WCՁ31t2O[,n4ik2u9%#{x;ldu${ 񲹵Ӄ_'I# EmkโKcJW p^dWOtmv lx;`<-h$ыw|fTZ2{qQƿ}Id'H+l&c rZd⤯> ?uZ"j;v8f{gH-osA Қ# EBBADzh;\ =kd0K{gIQnZHQ A&h*g% |e"ƺO˔`<'I\+ƅcfY$l$s"$}.^xźl}KkYWWHHc$~U}pS="٢Gh;^㛽fޯ'+lz32šwP˥qKF-˄Qx (=$KB3gL )Ox2MAu¢UF?DH0rN\|[y#>r5!qT"Я*2YfiFޣ&Y GEó^ m!kl$~@\CȀ<XX G zI҆bt1f9<E"2RCtV&E$߿0Lw.@~_;dH[Qw&Kpy??t38AE/yr07weAf3RA\}AM*UH$UB;ltѸ wYɿi~YKKY ]NꋋDf..،F?KKSh2iFD-`37b㧛АKYNwkG甮d6>yv!y7  Ką7VɡR%?0NS2ٔR)y0VF' ]D5g 6dh g>Ќ 1A5VJ0Kk9megRBJXEivB'+[&Eyy>-n:ڡKmRfQJB*ɤd ȹc&Dr$ɲ,E$!M\NhEvJ+q t|no]KR-M쫽0jz "@Sww^ߗ_za%6-AhR#Gkbˇs~^p/EP_bq~xCZesX8پ>$]%w8{$x瓵,),R Y]:˕?ڶsV<Zq|ů~~ }o%xHP648&jP?=,S_y4QwS/*|̢פ"FN.7Y%b%PXMp4&aʠ^ᙤZQΨ* BQlSr nި1TXې!ISeX&+%ARe# &]/TƧZ"ȃ&U {f%Y pjԋipE2S[ O%fDp20ɐD)Úſ")cHj48m4P ʂ42KP4XǬ HG,qq(%f d\8&}O@j#s-Wm;2ba!և(KC#r".+eeLt5ɐ&M.r6 &KA֥3K%l1w-0dI*IIMsSpRIìUʁKIjoYԅ/ zA$7yVgښ8v_q9-u<ʩ$Uy:,ԍ_4EJr,KHpjnkz_hX X49*^^ďhڡCU;>@d4*KG#P3|RH.d ZIC? uS%i0M[L!ӢŎGՋ00lI{) c 1n4zm]o^Ɔ򦋠m~{7]xw?_|/$xo=J^Eh+K˝{݀S<*3vAՌ-Cm*֨^J5C)uŌLZJlHL2cXd}I#QMu'hQLQ)[opgkR $iU*!+V bXFb`vPF, j7 _E qN8'2 k+^"V2#o?E[%5'ƛU]" u:D$ ,=/0\?qv !t$ܤBEƈ?I r(:;*G>vš8&fDh4e$)Fw;3gh yM*>hUX7pyoѹMf]\zlfrnh#yqO"h^`ˏ[h FU"_;.N6n9y`<1}O]W%-I&_^@x|ljp~?sO"@w^>yc^[ۅ2Z)&>Y~|n]AC ڜ;U AjEV޺a-(T8K wK*(d c)̈i|#~L$X:99 8$^;J 7ȧ0#9wqwAIR o?E skSZ]4FBXA?r ؚmwyzRH?Y'$Џrqzi:Erg"4syS#PxZ;ִ"R׎߽~p}u;Nb) wk^9=O>z~qWlF\Y}*Qӟ2ghʳzO;bpp`{c$D#$p 7K> ,75Z]:Mh/VZ) >/VXh<ھ+Op"U3otfk>'/_\t‘JUc4]z@ @x-m(F:>< XwߟXb8(fG|8G^)^~hZ6xP 񄪽j/@mЌQzяISZXZUm#{~JP >4:[D=1~n?~1=L|1(鶡&іwEjS/[ !JK{[4C~JMɹv>/0D$M;$;4BџWKG[oJBczU_\;w8kO&JaN< 1!a03oT8< G@< .@=z~mp3{W9:mcxBR|TPU\8o (ߍgXs*QtQ6s4}xa9AlB*֛aQZH1Rg9U-C;. }*ڇ[1|@- jjѨq` q,l@-O +aLi_6>eA=e.лY"[ѝ;nmh6L̐*"1;u%-7];A7i~':FL^_t[o~=vowli eWYͮCnv\8ר-s Ca ,CC!%!n?HZk*e#z-P|u_x8ؾa|_-= D1ɞOIjsP\0.=eGv mKˈ 1:WUHmp[ cb{T{'c1Jk='6^[78QV4'RڤLY>nrޞGPG<ţ `{'lHH rZ'3::j) TʨՏ9;LgÑHS4|űZ+X~B~#0=ˈ6r+=$m3<z, u3MŽ"˒KhPckd䇔XYOr7sQ}ɳxjzfqD)8$7lNSi#쏐"學H&p3d7 f2 _#?.$?>>D}0ּL̦ < tD[&d;$42 3Zz+1@u6oFV/4ԉ+B yeMCWcn܍m-/oC=~s x;ynAJ ?ߟ6]Ϸ2WU91&0.e ,Q&%Q fΩ XB΍cHfӀǞ期nLS!ĺ tۛ¿K#>Ξ~x7,2-LJrzal]niKtg/{_np/a•",aTTژUd}MG8j6wjfw;F5x%m `ڊ;i!u(FPqF.I2f%[+$@6z`٫q~j提u5TشWl3 Y-u÷B97yK31E0 Qf1ePuXbДS\ hXG["&F`Չ,M5;O=TP.>aJaDjh9fqHJQoL"&l| eof V9o(x(8qdZNE!B3(f&@|aD{ 12"F{cܥ*Ms͆tg:zW Q!$ NCG;G0IͭΐMYoEImbQ g|{lsj/WNfFƌ{ޘi{r7a؝ fGQg+هae4dj*A+Q;ͣ{0]3Uz>%=A\%P]19<.ֽw){6rhbj )SjSr%#SA$=*&ǐ=~OfǛf;feevɎdYq9 oTmLX66#o|ؽ5VL(]>: qH0C%7;M++CS7S 0R0n'ͫ H01ՓY B2[. FʷiXp-fJcQMO:b[L b "J=y,oPO(0dSRVHJwnTl=~W܉޺bҍ/原'waCOΎ#(-ۗEJ$v߳Pivm L3_=lބ'\Rs;~:uK4 TL:|$3!ojU @vF$Ig R S9GߤgĆ7Pj"fC]zOov xA3kz_nӣNtc[~sKdWBgS]rgbkPBdgPjKa\Uhǯ j֓'=b' kȺ( hY:@3E"͔qCMޙi ipL2ٻnWRyC@{y)ęZZNnKrZMs@ƓնC]9}aۅy ?44DqoLG-IBs@TCI;C7);Bv'2䇽c͆OP~YMT3eItpM:˻7 [tD(V yY1wC3yzsQ#qIn@?<|>$ -`I!I B哙h8 ƍ_@nAYgGfB5 qɈ/޷ìv j8jo|7g*g Tdt\t?{ Nm.bJhB vl;}hȓ~ONӞ#~i^~fsty#L hi&3:qr y=7%LBy͐dbH ,'׍ In7X;J?N}7^؀SQm~۩?l#S-S }ظWel:;~=>ŷ,l n\f2u2:!L>ͣٓa(*7.q/>Eū1xi{OW>NxOq?޽Z[>N͕͍w!_|zn 0_ ;l9vYTnz{cxʯ.i'?S6"uu{},J8"& CL~j":iC_0oBsC&s?G{c<oh|CM_"4 D;{4sF`f*avj-MhgwO.$fhD]펷oa&hw]~$h1}G*9ݠ>d;C7hB;d kĤ)Ņ Y)lnvKM[=^7^78*=}qZku2jSucjM D7[ I)Ia.*R$ Rc^%p֯rLs_{NjWoOݦ{tE7!iP4p\ΤA0IX}udVH6&)!2O޻s_lHt^ߴeo<{PMdQcG #0zTZgñ'ʚ m _r@]W)DM߃$1Ȕ9DE$krQWL?9gg^^[nͰe^#D[-߄:b.aKC5 wӇwͧNpM@FVRe!`>Ӫ[1߀NL~I5k+J"X6(_jR_:+W"X[kɾ%}eg6Cwf@"6E'ny5ExgcUO-t0%m17#f(HNeaOt(RQܘj @r>#{{FA9&+͆g@sӡՙ<ӽ(PU#\rZ/M(*O4p~ƗuSL~ASO!c,UV8BaxGƗҿ~pہϏ!o&IĎ~bfء{9}Ӯ3vYڝc^;j;t6y<ۜ=|RW`̛T̘oy->'9d6q'#B$F=t373J@7{eLL]Z[qT>H1H$9P\CkHCfD6逸n}v$ kbRG_y~_>\DCv_wclFC .,r+Ijp ϷG[|OCa޼zqWtIxR7hFd5"Λ9t1r5ssvT[-̑ͣ7|햝k>D2YX4v;^ɼ|@ [7C˕!JC''N˼5ɫpXz o?꫗S("7SO;k$E Aل7x̢1TQ6 -`l(]}>(%g<ԇ h[hw19rtW9tRoe'7r1rEWRs{RyRLr@ʩ :hrԚ]M˃CܒㆹRH5~YKfϪ>EXK8Tfcb4wJy}䂾ZHjئzE>̪֒O~{hˆ;9g`L; ,1/͜pN@2ZjЉѤjolll[Q4{gGMh\t7bR-Mz `a6*7S،BwnQ/8Wj/j-l7ŬH'>$Sي>P`L9N3hŻZT>/Itj~ .VfBr1UFrjNj8޴sk`۰1;$6oxp^'ѨݳRZIj%}x?s^l9]G..::S$Y(A!0Z5 zN0ifcC옪^C8톫hd^}ޤ''y~B;f8Yej*dYf ُ]eqę'go B{kdkV>VmFa' \#son^\[4Sa& Q8{*@/}mjVPM5N a&$rAM}4z,]yWy9- O3n14{uWi9}< Ӛaڇ/Qq?`{ מmz>ь9w-"b%9j ˚1\֢qT~xdG1i꫌a.g M–758*6kqdQ+~" flMttG`0Bh-@F%`g]I9آa )) H"Z’gIho?bn'K)k`su).b !@?3r5C%,gj󡁦&r )t:r+%\%6!dկ$YUR[HTjZ BG%nɪ -y&<[u0cuwGwDC&H3$/"cS0qnJCsq'ι5DCp)0:qc7G3!T7|L 3vu筲ŰƞuvS3ny i'+2с>jz#'ʼ/TMe)|ӵJ T+WԐI"h/Wi|$Ju~ͦnktQCwƩĥ`n0{D %C Y̦7QO[*ڣ`kk|EPs'v-$D]_`2/|Wq ߞ8" \GXE oo2ُ 3:y{Ty3e{+)WR֤& ^θ۱Y_1qޯo*i5mO- RtsMJ|4>[yr3 l )6#PRXz1lRj#J'W5*_h&(QC_8f0L;F Mju])ݬ_Q;()Hu5,{kX0a!a` vŁrYݬ[h8~܋wV}F#f]dU(G\s'(#hH$)&lw]qU|9㯮h(tM:,̅%1B.Ds"b,*#Q5L:yۨ5l\TM&Rx$+:rDNSa 1y$^y\@׸O9tDw5i"Y}u~ uswM&tv %wr>! 0뭠2L6Wr>K\)WsKcoWonʎ] [CS =!{H „d;h L#/R&Nf-y%r$\r JgX L@}̭Hk3?e8p<6z$1 Ԏ"~s8M7|8U3[rvsy)lFƽu&;NݤOEŐkWxdG`o`d)y d3^aK9ϙ{.qKaﴊ 8yZ+sܲOuT׌m.v@kuC] ޭx%؅9{j3tu ܓ[PX1l>hXҗW?Jڽ嗟nW?>^߮bdWm\PYjj5^٬8'i'wRsw1XTa9>5LzU[kz|PgڧZEɡ}-`p/y5:דhB<)%G7,N`^V0yKn 3jX1jݠ Z9٧}jVAu%EcƔhDԘ>c?|8[C9IJK}] 24@b+$j-)([7w@UPGQ+)^`lV*lk*\BC*WnFc7@:RrnIZr2%@TXJmfj1Zgr P'#a6r^B ~Q5-Z"庴✌+$f$+9bm6ZPsMΦ8nU&1B{M g"Ib=ߗlL nyþ; 'p~ESf lDfV~36?wF}}_ݘ1g7]&yY=Ɖ6HBseHhTP\<<6bŏ2Nu()ނKVYsd;@z&wbQ(VJI~5l- mkfo ;CQX3:)r(kc -g`o{ޝѣX.ԘP8S+ud:&;6cA-Jy @4V6mPmO*X-<*z2EwS(lA^)gP | !(~f@@,|T6>ф):5H)|u ÝdYqwz}B˔58+@Bc+1ymFM8[STL PBY\92N*ȁmv eE<\pN<^"E䴪M @!Zf=u[}H՚Z).~Fxŗ7&q.PlqEfs6F4 i$rfrAڪ*dz??E9@m &hZRj15Fܹ p xuOl[lkz@y wx84"$\[N(&[+њBEch&,LQ doZ.*9jXI"J2r%:=. bQak4x@ɉ,'9oZMn]l@D "tM6؊E)';ն A!'QZ) BrA`ϋX>9cbNNSHn]f\zwWӦӏWX} ]D.1O؊fA\+:U0șM/pSS6<5I0$Imxhѧv`FBpiJ!1xh}L"T'r"vJ2_^IR/O?E5=iESrԛxs&P9"8@5Q136f>>9'LvfA7'Y*_w` g/&mG6ރ^7fūdRϘ> {z '[%~TK $oڢ'YJU9pQ| 5~cٺ'd3IRnpv0wopv Y߹[_ݧ ̉'*N枠?(cDJtnX+N{b+>ܦl4q0qewפۺLT0zS>7kzy T4h؃w8 \V z]oG>^}wY/y5>V읲?N¡!WhO٫JQ'F3,FK^9,{ 6B>j Fw1? heX U1ޗˋ?ՃJSlU{w ؞`yU-YA[ɁW7#0-'9V[7BQ / ]s3(T$7x: &L0?_~|rNK(gh<<c QvU͉e3]~s_CΗ_O:WÒE_MxՆѠ}SCN Q᠛s`<9*lFΖ^'7MAEXw̜Y/)Ze*wx5m\~C56 jJ#qm;^[!z>,ˎM[>Zj=b3v͘%hڒӣtoY<r! zy)z)8i&NN ſZNUpyG;ZpJ/\-:WcJL\iڃëD띾FR*W]!e7_tX?(;5z_np0jy'7`ŭVo;3eoȷ5nnj=ϳcJ܅><ȗ"?7\п>!w#[GՑˌ0wz;H+zH qÍu x亙Yޓ>Fs$k~W$kwLwa G']?~1#G3WCću(ݜz֢n|ϋ/ar wdvLo1):{v`C!?W[n8wՄ^u9еSnӐ!W`@ Uu͗=D]kNLDJTIFOTt??uJA3,)xzjՄBm.N< P4q4TocnUv=N?"Q,/-^@( #ĢWX N9NY1ڪ]k#(s1ʰR#M{9}^#ZhԞ/y5(i$=-6ux#@Ⱦg Cl7jqzx*Tw/OӶfes䒔isF&Qˮ@6sFsΨWqgVU~?G _zP=Zc<@jƼ1\oܐ/yiŗa%W?N>?ŀqΫ9:X+FkA:5a~,P(_jBN#k!*@q#B>-t;ĸnWgIgr#JVƟQ(TٻFn#rAZU}+v.9|Jj%Zݤ_!9Q"El%,9 ܪJ]$JՊ1 ỌQ\ϻۃN@^?8-TW_w2=lxJu2*UN9!UP? ﺿ$<~q{}Y<urhLG{轰2^t[1SdжRG+3}GI}Y"33/QH$d̤nE9:_\'Gƽq5"6_[1|O_,Xl龗 8BhᵸV y ohZk*p,RQO_o7Yb,2y#y4FY73:6'ou$~j$} jsv^*_t #d=8zqc5~6 \詔Ew4]睾tp:ё]e274!v>:cR۾'E%bDILosF7Nd; "  =h On'w:RoS9ERetfW?sv*n*xnQU~: F,_F۳SOO;hO/rtqCjt;-Dx]U@fGg1Y&x/_ՇG-zqJ6cSeUǾGb+UYrS9E}<>1R] Stw>xF3.O'4pUu!QW?5sA]_J̕Q8wp| ~A+r#JTZDG%$z*vͽ{n Ś[NHa4" 0ƌˆsfOL|{mƃxtQˣQ15;[ [[OM"`*)q e+Sy'ic zLB B.6&Zu?w]ZX-UUǔjY-Tґޣk,W8rԬ"@ǶV5! sʡQ!m;eʪ.GA_. O#H}uneMVt `Lضa&a "V;}<sWӏШvbs%J #jȻJtaf![gɩX/Dno\=;Ol䝷gG?O7嗟OLbt!*ѸYf U2DuTוsZN gsɯ^:kx4̵/bzap/.J9%<(5jiByEXQ+>Kϰ Yu&Z[tjY+W0J *|T\KC`R`YXcr﹣B^V,ki <ŬO)g r\RSq!_@Qr%+r dzؕύGI1Dr낚@gZd#* mx}߈m^ < toy ^Hjև`{e,8&N]oҏw*CRx6LCq^Εpj=h|בwZ,Ѯ6#rc}+F;+# p[@'x8hoonH׭q{"wjIVxl DGH.XZWd β#pcKŮp~ bѪp@1;<œh!&@9fҜ2,9#TZ挵b3#d-o>y9Ibec;*;T'X_КBHݎv;Ty6g*}ΛБ%T;֎BwՎujbgegŒIKAMrY zxql77z`G__v¢BEm0nU_ fޤBB [UؤDܮzvx:r Q(9*\6dRX(c1e{9gNJbU *P#(,'%f oߴ9rFdL~ߜU #MqkQҫ/w+5x<ώmHN" UTXT&gk #O4DN>eM 0+ [V,Xf֙W 5cQeU؆輬Ճ!4|):My"sLjʳ<0hkv'mt#8#lRtZِܵp &Kgsy Ϧm㜣LVbT6)U*Jl*Ua2)Py߭# Z{<$h\Bz'%(GYj+\`.o^M!|;WxdGSo1W^ 5l4&Y;b ާ R*$/\*(%c4m✳m+rZ](Fys904__~xcHRY.00gn{~;QP˧`$9WUnNtjkڄ((v.+jUOW³]]-JI[a/ @b@eUV}~WIx\fU c!X 6p3|Q[3ш93O08Y6y ^[ 7aKvR:"2wa[-imS#:nzU?xHƶu> /j!KS9w1{&l܅nXv6[0J⨉A9tX /녪\z2E*Phu"ȨbT!K+$JE(uyk[СPҮEڠ%1&W\5ֹхqd!XEȫuVڐk,B.8>q;\1Ca&[//0Fu骜נ3/wF8m\/"v篹?eb/%CHܡZ+[0c'h. @]2ΜP?hj\|)LPW!X/v)K w̗ ?'8_)$9UDv(LeD TszI#Nkc-8`H]龫v%pm'nFzPTno@Q -T6)x ^soRh}aFˊW=48JpOq>vuRZQH^y*떠ޠZ#FTrr$'G;{Jg컪mjY۶<86Mvҵ`}YĖ<< Kbg3-FD^3Ec W4+5z&>GC[¡no2SK=?Nt7˟: ?Dz2=GaMvt%cC+&k)Hˁ*u/Ӓ.G?"g1TXS鵋jN/(n9VUt AƺKSCk4]F-::(G`2Ž4ENTyTG]oci68lP|5jHl1] +[ z,@>Oy~|ioVdb~zӭ:Hݞ P2P}4lGCGfJ VMԞj]Pi. _ej[V!e^8v`d(Jl 2ʉ]oGW~bGww vrd&!O'ZTH*w}Ci4!e6yt]E ]_ISx 뷫bI>@`B\%Sy-N 㡈Q# »p 0Kso/nyǭ~7xI{?EGJi4Yǧ8]׿ɲ4q?y DY\7D8):ۆ**IY#;8Md3<J~Ri:%OvPceYIW8cAĮb)!,zKZ }6hYem9kYE.YM|vslF&Gփ0ϊxVij"'ijF07 VPnA-uBq9JŒ7Vr3 FFc׌.0ro6ui})@[~p M^_։dd.PpRMJ\[F֨k5 d8Aq=/^t|A@P[)v۔j\'hv"#jȓpk7 C% 8Ge98*uDI > ũ2:-8Q#]bN$ƌo2☵TMܡTPF,NZSJ3ک1""Sdzįl7Tp"2[jʒ/˒?V?L g}FגvsYefSX}q90(7cӲCKH ў(% H(,5(y~eDS$z}JIW>TDLH:ᚢ5kBQڍ_Q/X7 ( o5tr5>xzj5"tv^pR#( "2şߔN!&R\Ű(T0gB&BIe$8S!R 2hiy@μ!me1hi ttEwY }f1_)^9\],'r<3Z%{6(jj ZRv8M<{T:=r.bMJN7F(y`u`rkA/(ˣc5{ju~ż|~._ XlW JCπ~V˦L+.<c[REKvHXpyMgpL@CRᶱo'c촤=-j[L[>U9ޏr7 i<ϛqXvzE꯽dO't{''>H񇏋"'U/mEˈ5QU/;zA{gmEn: ~? l:;tObL1 2:_T}k3=Q~|}a?_]yv5]ux*+oj'hB_N-_HϺv)O|`hiJ~@pB95)ܖn˾?D+LmH2i }.s5ٽCnsTg*|_ُ}vzJ3C 9uV/ _6,46|$edcz\M~W xߐN3pІ%$GEǽt7>R<HVːەiL{}{cqO=S"2N+08(D>V'r¸>w%EUѫ9{8lNn,2T]q%w#uLə.WyyE)hht޷:R@e abop^B$d9R=ntGeT}F#Ɉ (3"6"@TtBLBgD<$ZlJ@~ȢtXgwq^qW;,oy1"L2@NUՇR%-r5?E+.<'ELÍe-2c5fˬw;cdU< \A(3FsR: ?LNH w̗]='$r+>zwr+RʹGiXF֪6NO|z~?zuqΑçONNLBroic#\YW4yvpsVi9$VGazS'<a Ԡ$D,Np ?0:;)F@ G;\r/it.n҄\S);L% ;ڙ>> Dw4FJaORڭ\)T@-3/W;N!Zߨz\?,J} ̂,@H`V&9 $jL In䎪 PIvXZͿ:`(5QVdԹya(m4HZ`\\-FH^H(>sA(T  ' =s:&52qEx4s^_ҨooY}߯<32(ajv2՟M_'!L|ĝ#94$FN X?O B MΣVґEvBc*Zר'sk)<¤H7%]WΖ"єenXO|ZoNB|(Y&DܼC$# O{$P`G չ`dI @Ky(V}Z!oک5ko9wW_ dCk12vyÿn~+fyVI nJRg8ᐜ5v2!.QQ,\/ۙG%9ΗYy96tvw]u gS# &O8޽9yp,ԇ\\B%Pr# %nd)܎Cɍe;־K+(;>F1ʎQf;8%!γ`bJT&8Sr hS&&RJG8826:{,\yn^8+m␒ 9Ej HVFYזD'i!)(ֽ?a]A}"?^Jl:]ޞ٢gj+hFwPȝѕQڦ/HY+R \X*U;NjrQ"T]"%*^[O_QvV4y΄ZB.oƳ(G3m%~ײQ3šSn^S*p͋Ĝ#űo"ƌQ䅈i` (I?eW_'w66r+=ciw挓k[Sc>bfBw/rE69`-a21ՌR\8ba_kE$C_c{J=/8)`5+Ѥף $jH7:lz'ᢐFÔt{J)LUX v9Xw"'ZT. =oXRKWX)/IH; ֩:K >CMTC,MB! f74ޞw,3#ɩv"sĚ;,/drПA죆X7W;I=gx flBdEc@cFQa0 6ISJ@'P(IsTqk>c',WFkyL˂1<+mU:Tܚ9dPoʘvX_1+{o-ǟZ(Ma 6 p}$CFu#_ RHAoPȷ8XEQ\_*Wun.lu aHy2Qx Q@HېpΏQrI@':7V[Pq`W佩eg DW ֝ifdyvw$-I7!ӲIɤ=2² j,"K䲘bjXdUVL6S0"u2`FFUz N<)4&H9OHV'(QgI"XEwjXM'P2$D}l+ b0Q녋ECi.?m5_?TPT2?OtܧtuGnA^-3Bvpu'!WB|޸ $|yZIVRv;]#\޻3 +L4!=m9~OܣO{%u'ֿ'62D[np{]v̐'*-5cN/WKmL蘴5l>>((R%#sZg+a,0i{[[e֭sqw78]!N:rQn)##GUt۲s\~[W5)juL^E'\wLj.5vw8!G n}~ŕf}cZ4Qu1B+s=x5MF?pK3|d42ma!XH9I~r2[XB;9PNzY8ɫH`A֡ tXհy3k: wwO(׺»Tc!:WnfNc?] N^H)Lw{ FI4P3zRxFqvl#Hep#($Y|79 M%p]6iиO%.wMV} ;ZS s@R CD(FG* #:0|8IL|%"2&$f+d1L%[y’eU2zVu,3 2ӂ:+gM;r9`l'I""ba0DQ :.B0kIVwsA;|7ޗ Ơcx#x(hĦQ!(r0n8%F@? 1C֑V>f($S@K+*@bICL YH#| VM)NNxqbϗNĸ1ϗiN Ĵ-0BGIٍ])^9I99$'CD>xbC(!-9d @Y IƋqI],cH7qb&[iISJl4lsj>$}Zr.©%SH !2%p'ɜ ?_y/gn%#6a߻/UqG ]m_77ۍ8Mr$,*<0$pVZEnKlNX2U KNXr’`IBr’d+,yqKJ@eB2Sn]3A Đ2MXa,+= ї0''%jEcI’R:a KK~\_',9a K6cN~AKhLIA2fڌS(Ȳ'%F,a&ǒ!RM`Bs x}/hAܡS{9+)%n'\)e?̾2Z?,.ÇGwOo% Ń.G,Rduc!0˂a@qRZ1K3h%ĬRb B 2Q @HJҜA 哊LIJL sLђEK-%S1ZRDo(i puځJM  I9JZl0PI ל,1T}|Oq~lc'8 ?٥ Q`C&L(C'1ps͹Ǐ!r|T0Z.c9s6H"!g{|p߰d4JH$Ai"%M InCL[V #XԹNB !:@XA2D!XLK,t 6R(K-f0 Cxmǹ^a ` C@d` f,ոH4M% TK樲[a27H2Hsq%b#|9c,@&5QQ}"6̌"@‡nY}`M*3OΐTR)F 2#$VKj~MKl3#b4?Q"E$ f80j0\($!$`*8NS{ (JQ>y. 3cQV 8̽f! 6tڷzH_xrڿ F tUWAVES*~u8ᰗZjk+9x/!x[]JAZ]k_}7c]ůzPHX"d 09DKblsHr`Ya+"r@Ágv@d)AH0Rj$B#gՠM2F9 M^j :uhfFΜȟA,(9VT @!J r"JkHt0F1_Lk˫= (Za"P_SJ'=O,qc>NzvR*'uqZq{J.y@PoU^]=3Eu.l.|{'(b7 znYyD_9sVd?ݸAdg[xfZoBL?3s7lcw|vtk `WBrSYQm#xL5;^{xflK~hwjѦ>)9xm ׊hޤtƣUĤ(Q1:uWGK*fY!߉Zj1_s_>mƟ=90!b8P]iz"y1oAVvvLb7y5ŻOf/vCzqBOm7{><` D")`ّn9pN~56m̾|Pܖs& UNK},>@#Z=d?Ưn Iҧ) Vxn׫?˲^49ݣ"F`OvAMp^UMu~v-<6x04NW%2Q_۵RY۵Ϡ{kXsX.Ho,?Y12Zl8JJk 8Br@p;D1262'Zd5b#zWL\@™6 J6 s2RQ KLyEDFɪ !.ckSDg IN\u7尅ԑH 8ĎǫAkլ jA $ƀ`>g{oMW_r9nU7VdC`1}\V_{|b,>l Y/Oխ#$H؄To:8fpJ1uv-nmSHjԷ76J~RCqcR>wlHQ{YAwT F 9:|ÆDŽS.N.£A:vBݯ)#^Qԍ@&k!apǟN7|ts_e71Xkh}2*o`euҼD@d7Ki:k} o<{قκ8_= i/Ez^oz3}in?D:L?H- 9G!=hvṔ@"hyTOK>@>9J9SI5n)fwŪ1LX90?~?sBPR)Osc$ jP[B~5Dx-^ys8 ЮYm3%@ BTÙtP,H"S??pu5F3n3iJ 48RjMzj9P fqd5s"t86G3̜U&˻M֪wۺg  W=!dA\c$nZd8<aG_^]?^Dm:⎇2L; [1Rٕ۰* Go6\y\g :8yBFT)V26SDh}Z=CNT %.jJdC{iH~z- d @Bly&5G;^tq3HΞo} MX;}MnV &2qbVsef"A4OuXqo3 >sg/K'\8͊-cM㽤PHNmYK&|աAE&fz,&vy$ eVf1Q*QM}Ny]$NGKQ'F]N*%h*R:$P:d!xuyZ PFw(*xsA=A@32B7!$n "Jǝҝ{;Jwt9 4if ISqA8QdіX17*T4uR3))4By+J}Q:lSf($>Ãl`AkQB_ZSpə&k-P>NiE NaC9d4Cf<ZIJQ;H>GEo ̪V$|q\#bA0~4` @61*x, >*4 q1lvW6sGI6]}YQv:Nj*B8emL ކ b>v}ٻۋ75t}9+ٻ8r#~KZ.~pw \ J`+4 {3֖fXd^?))<(/;(N%=a0*{}z 1@@n2.D=nzt&kG2͖k,gdĨ]2M3o/t9, &܇PIċ"z/dH)y&1s' 1ou&*]j -Ri[ՙ&mC/Zծ-4%6N @\C \gO8Pѹ"Er|I07aͣAr(OۅfS׻B;DNb|~uPH].Kg(!Q6'@%P pcyc `M!z{5*XUY: "J"g s(UE"/lN@R3x!frvT5U`ȴPQ\]qĈN^\FޔE +X#Cy4\@(j¤t2^=ڊxc4<ARt(F;N,0DIHP35@a]1:"QrTB!r)EؐS-"wYʈD+~-{$YTjN4ʤ|Znlr:@+M %Z%j zmɀnfmCDOi[ %"#}ϔj671.=r\Nsq}7Y9Ka1rmI;5`t"ji梓WWHD7+~s TLOuZPM_jyPZ6^s+K6YmY-)]q3ޞ]5Ww|e\$sS_}iV<} ﮚwoV\p&&RaF7&'~:ɐpQՑ7Rxe[۟T'M%״F9\.Fogz>jk#/cO TW*r]{WTRɆWl|n)F.]lM(hcN7c6t J哧'h`?RJ'A7Ha/*0$41NoO#AW*ŗ-qes=u€[A~Fʒ`񎓆\LWģqhˉY~0 U ;+3Gj2E EwdXj O0[r5_55KiMٹ?k (7fqڙ߆qsqsHmr 8WJ MatgѓrhY!)dA)\uq!U}2^ϟknla΁n#׋K*UvɮL!U3:!UjZӋ\ Imwܳu* JW)P DϷ)0[wR'8\io ,Gi)0L?W)0(!}PK?0%n'AwIfu'jb+ȼ Di㽊Vb0uK#I b9:PcpLav2=ܭPOrX7fֺ}rI(1tK.B*=$o4gރ'HZ1, E42Л*Ӟ`$&$\FArʂ  6!\K&J2ꤪ`ԸJ)/9,s&>#fg ƞ|JgP"&P Mho eBzsό5B2Gɚ`%1i%l!gl$3BE9 S15P)T! 9Z}̈)^3k RLVP}\O) )t,G g(!~ t4:[  Rֈ/DW_)R ~p G+#raqW<3\ȁV( SE#+澊eBxw2:Y\ /Sť"0AmU/T0 W5}DekU=Ή1㥋zq)o~ [G\L@~ QO)Q9`h#W¨Yd+| ;}R0"P&у`Q mFi*OA{Tt>*g~4E T>Чh;/PEB/-iگ3wW3:jg7Gih$d"zQD)E[~hp{#UJI79RYa 5F(F8Jx5OD\:nmK᳑z_seG.4^]?3tenG\r U;Աe׋˟޼; WloϞhG>Exj\>I _׀:e.A%/ ;yGz4|/d/f9Tt _cV0\ zo1(f}|s}jȉg*^X{"]?qç%_FaPIZ~oҽր`.Mư-"(TzܨmC7_H: "e*f$Ihv%VZShavI7"O]_M-3{Mֆm!.SQ8-\V dzRѥ`~a Y$c1SHĊMD)qc,a` `:-O5qsDtb뛋6oveW,;W~m%7w~y(ޟm=~oòo^??UcMz_z y?}ǺspWh]3e狵~~k9MV>gb)=)kV.HhLӏU[n~ ڭѩ}vӎ$Nݪ -ncH7.U2չ-pv'GnN;hcF[5m E`|+A}dp{vsw{;p{K ߼޽roW~wՋj]xֽ>ei/_]B9#J2>RyA XRuVΛ P%ȶqw}V 5r5p_=w*G+#τ}BUDp8UGY KϭC(a 1ܞ+`X5ԇ˘`v}!4pBT5v(/qP<@=n,B:a=EKIUsޢS;;AN-%bj"۸LBK9:K6{-TcpzX"Ę,vK:@c$SX,Uq)} 9cd8D(XKNK3{8fLK YIt\}xd۾,nsNJ?o#Z]YZey}xcAߒgt .;O<6&$ۇ}wqQ4fVUWhoA܏ho<7:-bcz)ؾ\xwd mQ:&JP"17F8&iI)ŖuɋzlgD܋دܿ%6jJc-*bw|׻rT;C(ӱ7C.5uvW_ᰪ_4xgWTq₊C%qU_ϧ{$ŌJgga4jT2\IeμgogKeъ̓lt`VǾfEt vAE<Oijϼt"F~oaS1߸ 껉q0?08a I0~9F'Gփ3)8HqwDcľu9sf$}&`̨Q4ȼ=@PrRIw'|Q4&#lv{9,Az;NrBr>;L ۏ?i9ԻRNZN'ԻU369 iUPTzd)fdnQyb;5pvm(,4񔫩ĐڱC4'`茉!ʼn!ѳ'hN' ֆ%Ƽc&h31+tib`andg&g2)~6Ǥz!9ĴH(AR1@素H具lз*k_;8kYQ$_#OߢxޡAtq/qCް7?  MCHٔē!q/>&jY` ]/nrP@WS25i~{wc._'~>EV92}ݪ 2n1 /?XBW ;4ɾƲD0K֖!^'%SȬa< %_FY" (2pUqWW=| *ŮZv*SyIJ: Pxp KZZ D U$۟ӛj>CrD0%%k9zy>rqté̻/ ~,N YϞlMɄ!1kDuu`68(_(lO/FTO>ԯtyt WXR Om;O=ZoO2AT!~=$x@IdG0AWĤ\(!RZ7seYfQNTbWNj)&!oٙ1.F>+J@X?׋GKz;!֧<yrÏ_%"Iq8H-?50I,=XNoejŨj)%l@7Pn/@ᄂ XљI]%I 0|ã󶅌ĢvH"G-\#_ES0ǥn|ၐZvMbBrlM_]ω}jOz(cv*D P =bHG;ƀQʥG69pؗ1ȀSЄ)7aӞ^*_}'o |=/ Ic Hje.y^R4*,a lhI!Yb&9l+/\jU JzA*Xa)g I*n XmuQniFkre^.y{qY$]#\:L ^2#e4+ĕ4G['gz|?}5(flz~vVX}⸓\i,ReHkcJ[XnzHLJB ֢d Z :Mk߻xEB $CtrR|;fw&g2 ȭh{)8AlY)FEQg7k HDĽ-7F>Sd2G9GCjN,\5pI>U'Wc]R?im){rʥm1OrjG4e oi K5g %@n s9tu;K(z[@H+qF`mrOvnKf<!7w-3άLj^!۷4Duqdz]Y7$Ydo}nonXXuE1[I[ aBz,< +_-@BڜW6'{Gq7ڜ4AT 66fm6M ynNMͪ7nx } !F5lPi"C8'%ӈ( e5 o+Y#M,:gdom4!#]轨װfZ97ql&7{F '?1S$,tdqU~ QI/uX:ShFk_Ss#R"DrFFtH Rk%7r+eB2;j䁮K5dbhPa(wjJТ1vZcʲ(Wa5+8ƊH%b@٠O/b>D\Xa3̙1flhPt",qL!lV.}1%Pt&M^ƺΦ [QLݘs'N4갋9MPw \DB펛D;M(j #WVo1ܱ^%jYXPܞѴ.4Mr҄M;DXTpX_-H ,(5cS钵=e,Mvj> fؼa%"|j)h]T!NǞ!_I7Xr]tRhwhQkC UQqd+nR=|2A~wт,Þw`ZVCqC*pu=ҝFߺ13h?13}q- R ) l?,./ [tcҪ^VwC@YFC? 8ա|HՠCJ!Թvzla;5`mߤ5(`q =i>a$"5x;GA0{QB@0D& G3: I_QN+2T BHSBGqD u* qK'^, uaD3Z lx %|ރD1~%LKqy,it4Z0c]XBS7ͱƧxi^g[ɵh,F#SQ`)ZQT^DP9vچMmX o7VP̝L+m?܏fLUk@ɝf';1yy=9^abK4ny%O7zBK|OA;{W9o~~0aʉZ벸x+Tb!8T uhX<#gP*?u/gJ~h_y EDhN6uY1 Z}I&v")a6_3 C``I1a X9Ѥ#%m(lX ꘦7D#p靋(2#CL)YvOg_꨸UT@>Ds&J!)z)#DBnCޤ1D"khsG#)&7ʘߎiWNKWNq73bk52ʉˈ 3w=LO?TgqiWe*ٳ<3StytW IpuYv1k2.^T7Gfө?zg(ag~&sR}VM|,'A2 |! om1Muva;abNH_;Ut H"Pңϓ@֜Jp'" SD~ōR׿|V ),["2)-dr 羡 ;k!RH!oJ!uXm$?$[d2qQ0-MiJDƌYAgXFB"؞v.D D.BJb!vA vW s5h c5^1V$k;L3Qw1ھq/d3֓Q2X 1ACq;}?O,w"cw}#pX:ǜxq6Lqk,Xvq&W^,:pm#Ћk踝 fp,q%Hgq*!P b^׀9`?W|(1@Yj\|Ĝ37GAT ?NK01EtvM=84 2!s܊{).B:,zO`DHe 猉"SBHɿp"CN EL.Y;OyO sqê3H޲~ӲR|Is4ؙg~2vsgN'{!s>Z+oynyг4UN,k&ɉ)hiT!Ii4CJ)S5 Sѿ<7*Lџ磷jU7u~(Ee bFZ?kYunW藷VoI[yWK^NlyMgw{۴qdƍKO'g峳ɸ8[̫LP{O" Tbê<|1秀[-6zrGB@J,]D% ]%ņ %D nI0 Q*71BZs Ny@3AT9*)E),%qYTyys5魰9UR/|w}qҽՄ1 Nk}v;&d0H Jya RJJ:%J1K. CA580*Ey_\c݈8rƻiVhMqRÉ: z>l, i.o OZ ?][ʗ\ _J]QUZr˻GalT?=>zz1_]K޾y~Rzĥ9}ϝ=C.OjídV;'*ΪBO;~òZwK) ER0O\|G:eJ . rq|G1n=rV"T^J/=]Fף_yۄIء_ hsz"N$EP~t%(+cgqxQllv<2GR#uDoYRvksa]m-z(-ɿq锏#e_ܿ硿w~<~qM/3PPTАQ8` w* 8}MFXi =e֫t*S`|5GW4C9&~f%|=Kw9BQ $9IAޘ$RR҈(e[XҼS'}h +eou&옵:G# xG1'ƚҍR/T.`BMi77<[Rzb8ٟU-W_yg(ɐeb7+(Ο(νhEEmNл/"?G'/W*KH bיSnj):ɛ>=ޮ~;6(EY[ D #$ps5uiruP-u>osd_ LqhC$ r'TwHUY$MO_x鬯Hѹ/NZjq/FNP%UOIhm0DXhM3-|֜(~؁AQٲW:/ohҳ t{ࣟ҆ϱL~o E2-9M eb5(gM "F^' XN"$:iViYU)D097 <#kA":611̥`F,xh&SZ#*Lh02ު(j1spuGw|>hZo[޿'yt)X>H&i^6{3.]s._T_^/ͱr!CD(`a0p2Æ8Q8*IZkt:zZJ 6WZX0sMQ[?Qc8?x5l5p~iv0ukan=1l]qg% 3/w3?,WqBqCdd-aq_N>>no@?',[*7o=E4I8;p;n6h(T̛n-ꐐ. 2Q$ ]jT ʈNjU[w~&pZi<вڭ y""S@T(3ԡvKŠ褶QEu%nQ%vCB^T񖅾qj7{A)IlvnZp{-u[ E4E? 9n 4Rfb<η%Í{- !/\D/)E)|?\:{9雩7Qzb>9m83*b'uٸM%Gź;NueMd:41l4%439jV=dל+IXņbqLu^Vz !C% ǐQA4V>Ƿ X2BZa"c3Yp҄kdYX}ABY#~4┴ Z!_Ú kDA}%RM9_%{3U>x7 AԒRQOGܔjnIIΏ,s|R ˗HSI`J$-S,1a058 b@0FDz‰`*EqHZ_'fHD D^m^L ̚+ZɆh wCbwY|10q"Xlˉ^zﶌ X̫ }B[5v8퀝ݤ%B6MIH{L>CL$k`&ۡ̄MS@>4-}V݊x/] 7GTC{Vì3}^SZ`1%˦4ѦźNAs^#[wP= p>jxŝϦ;˚ׁd!Z_{A:n\5u+]?F!׈6ނ*?h,4k -Z<ap料y"4gB` 1nu<)w\n}SԗO/ӫ@~D JVxozUx/< iګ'_m|M|8b'ʴ>ԇYM#qhub̧u4=< pGVͷ[x7 oଔ=\X>GN?` ^?#ji&t_>cYy|7.cfZ`EC6_Bq閩eKmJea^ez}t:$䅋hLQUױbEul>c;j}aF_`yI#@ xMأa&UtAYBP;|Usa֏{Y}j1V{A+5M'THK<Śmf4lr &^΁ #ȍ7eItIY.nv??vV9$RNAm #Byp> j7.&7_̯}\:0asDYB/ŝ?<T$ĤByҩF0!q>Dfoog|~GpYofϏO'o {3lvdf {Cw߽o~|e~.^:^jm/⧛OYCony-̯" $Ncxbo6[!\u%)HiL*"A* J3nK3 'Rf0/2;cCWoQ璵79Lx{o 8 `Z,W?LaJ!"ݙܽ-aj$\#nۤ㛸̏My3ӛ| ~116F|#ń'(~]̥^㒠9*ʼDVVkTJ7U_b N ^+t}nGıFѢ~:Άn/,"oQLF FA&6"M]t`I*F1 z (~Ԥ[YmOd6@y':t/^ttV;Q"1LUN:=}sZ N\08S!^h̴+Ckı[YYBcFB24>X$}8BzZ谊*T((t]л+t][_eI#GRz~=`]\j$h(U(wګI&Osmz` J?.eZl&9$U `.q_lqܾFˮ¢Pg"uEsHQHBJ"|EܢF8""5xXƽht3`Dd M?,vaݼ*'0N݈h坋ӉOf f6'1JePe᠛oIzL7d6$h9t&J+=ʨr+`T ,ҩ,.\ H6M̛\c85R6b{@fS T~!Kn|O2hIɦ#A۩ZnG?^5)-+t7L 2a\##%7W~{qYqXg1;www2YawÖ:ܨ6IQUePk@\!-iNծ=oVonf%)b#C{, SbhКjH,U%>dM7Vc.Ď)UT f+mKLaQ*U2S, dSˌV:99be{*Jy&p=iC"v$/4mJta> f>Vj Y7ipߞ<62D@1Y"ʼn炮#ߢ EybHPwyn%*(fSyʤyQ@Rz8(,`1e@)x8in}Pssf2)IjUS뒹4/~~/DkG՜~JM,PI)eA2n2h/  P%x!y.@\ *jXk͈m#+>m|l!xİ7 ^G"eݤDIrf,6%Sꪮ # L4Sm0>Tn1#"Ѐ(%6# @7?%6 +FBNt8.ɕWB(P.P;G<_>BPgּUD(zA)A]&0;֗u5i:^{F9moJ`Нg/g5]TwFO@=L0舫ܗ4ܕII(=HsV* Gl4KG*52S|_b6.me4#[Mbā= W~sP S63L?}k) FnpaؼD8JfK}z.0.ּ;f$y̧: ;sԷ NPFP b',&"hD!9-kƚ({lv3QD1YPWE}1"±hv;SzՊE,;-\ei?(21MPI!)[)UW3؜Na{ؔ`g+wsjۃWN544 nB>V UcHczrI2 ´=d \!]J8Ǣ1+n.}!i HRjsjm_>ݬuԯa=:L M_~2۱Bk?틛~ltbՉL!D%2zVbM3!(^;hG,qI㒿y/yDrQ*:trkӿ~ۼؐP[oFmI3ͳYt5MuuZD+SVdwh&U%"} [,(~ͽ8`ɕZXVŗsbh{ D-2kg3 j Mk֕R<2x.ƉdN6O%cA{}RPc;{nɘ^Dsfo.p3u_r@k-n$,oj\9]shAlsnaQIͿ:AO zO=x96}.[u>_T|Ҹܨ]y[k.aow Qyo[7;[v 줨PvXȡ} G\3FH.(wwN<3|9Ąsyv6Zs 5b:HĴ [Ђ=]ScW'w˼=) .-hmDscnPF;";MKTv0=0U8Ύfg#h)|-´=z ]ς_1D=@aTYE @#ܷ)_$HħXq</+x͔KtQcJ'K(Q=CmHp;~rѷu0a'tup'4.eԞkN5!'kg Rg6>gf4=u̢c3пʋ+SŔ;ƨOD04as쀜PpmdV/3C}b$ٜAr" YC" ؎Y(~YE2y>(vŪgrf7KW |I,`FQW/TNjED p\=xj2!/"ցC:ӥqcu{S`evb<йTWP!~\E۹Ұ 2jΝGNe吶SgJ#$u&j.W" >v_u.Հ%fAudL(=Wф a+^y(B1pYk7QQ~)@AkvO{;#Q$LߕVQ0;@Ki5?st,= Sn% tz*d:ȪQ,5NQ>ŅU>nlD+UIvLQkBn=%l hNV(G`}4G_Hz7{HOgNGeJ<AJNUz6@v㊈ ڰ!̖ss*L L aIFmy= +ӟLA C=ςSU yog&x`gI^:;+ku+nX%S8a!MbC Sх}_pͭNj*҅w"`IPJ. y'S"e Кީm ٪ 7~Ѽ݋tg? V'' @džKX0e谄d,Mlqef!8&JFW4,p.Q$$ qeک;"܁߀r1 Ǿ}K;hZqbk)BZwk?!!QW$U69m&apy( \0f+{bnN_ҟS.59'$r5|ѷɤ ӟp]B:8HBu؄:(ܵw)h@P=X 1fQUK VNP,~,G %Vї)"uk[JjRI{Pw|ʞXSe}2 'c#A;_?!rE)УWqG io[u3{Mz_: fCjc]ln]lf`6p^䷖l(Ɇ@7k 0pc??,V3LB2GjgWV}o#1fof3 %0xY,4ǜ3P//s׬(i=4鮂Y7UN$jtkR}y.:TJJ**yߵP@,|ݲ 2 lzv R '"}ы#'?'$Nnݚhcȣ!s9Tڤ._hVCi&wR;~!iq4+9YFIWvסwJ3H W1X`[&a$6Ll=2b*3 KbvUKws(n-qM^L FL0s8Ukz qFU[PrD=@2вmIgB`f֎)&PWYҚGT >P#a! 2vDΚ!tGE nΞ.T6BYcGR#*Br_.EưAr`Ȫ7z%Mi<N @ c)s?YF1H=)iCO`ʨ:nC8Sးz)m. ?H?/E^戒~vBvl~Jg t#8ACTS%{։ذ1c`YBU;Ӆ{M xTpcnjgoH`^:Ss9+/t%=Vl[xQϹ͡|7yz&_CmxQO>(FU0Ȇ Ot0P->~gW۝ 3@ a3~`D9Fp 1:lhX@|ݬ@2 "l\CF//Q 6@ ҈31=]{MZu6խ6( jq|e'8*>UZX.%CI#Lfӗ۲LW>-G=}]dDk+z> ]gv)`=r'3*G; QM,ZZ|k.[sMR X&O︾d2o| , oV-<+ʹYrgzi'^j8֬:+2Tn7  AN^)Cqy;t/dmuke),ݛp\fi_+p%jP׿ZHe}ă(}_'_+w˗yS?Z/u+28'z `a&qV+e5)'ퟱoJ~N9:g:AQ_U]OrOsش=Sb̓ Ǽk}~J>?=l6V,fk]Z3"2y=# `{ͷ?"h}7^vejYhXͳU>dtE_#f@(GC*Y7BƚK3dR+SJ*N6U+;q{k^9j M>|o嬚&4 $)Bd :<X XlR9(!n6Mp[J,76mcc/Fr 0[o xK7?++&R%EJ44(;/OnN(RrTbss~3T v$-\eQ}-777P db@oCwsf(󗣪/DjPBY`\XCN6 ˝Cl> 6ۙqU6};x7NT(rH}EO/3vv-捵-]`}:i\ qu5U'هk_(R ^_7U.k}^EMz <||Rp< H ;V[Ni;K9ƦDrNJFW(iބeW0^i186ڊqpLšv 0kLW2Em^FjZo] [?x7p;Y\_^&˹%{mnXvS۳W3 aG-C2j btPÍ{ V cN"Ҟ},A(sC9a[E 3*^82=Hj 4:RVPP1N`<@3 M_IHC`>˹2I[r7'Nř^.<qQiK_J)J-\Wg2d<}KmMܿ89mG߿<x21??N]/[< |̵OݎƯ󟹛c[9y8B#zzpDs&l.9[Ԉ1HV$d(~6"(2]ju`QXzi`Ѭ|\H17>Y+N[uy$󗍹tp5.¹p>$IUǒJqm[ƺ;Ӥ] YQ Hu롹{w=\),e+O+vw`N \1xK쑿hAa-QBXf͏v pIHk.JRR ѣFy!,(X]lb.q+[o#FۊYƂU3W"P\ HZl/v\o$f-ݓvsʌ烟1x`Gws#ȳawYwONӫa/ f= Å:;JWdU"D}FK}p/uV#:jSPt@O":͏fD;5n)^pBr [>nRV@uC۠%=yjM#)i_[;[3w/`?<۱5.R(BܶZ¾\m;vwb}_9RiMWd$P!RFV;H;W!o ̧(hx7v@S/ ;oUOۮ3Y?hĪ/6W-4̿\O]W+eي #V1bwb ~~wwd$@_ِ-A*#ë$7t@]O?<[(HwJ XҧĈR3My~_1&#w~Fuvv({u`W_I8 gѣy*Nl/qw^WJw+*aM3-v<=dD/I[~ppzs:;SkbYSUG|ڙxP՛Ϲ4 \0,+m/&S,y:,K뷃tgdQJ\4ǩf`)if2"[Xo\>QPݖ}_zy-g~-KKLKoCfiGU/ZK=4Wjny"lJcs+ZƹZlG †1'^QEg,ZI8QP 4"\E jw)+QDR5Bh#+ƛzhf(4jTa(~)JU//2S٬ݧ҃ R7ѢiDrjARxh-;Rcpƴ3W?^[lW? Z[o (n&b K eCa52FԄ(9V ŸU q3O]+5808JP!SxT"j$XCm.X|bM6+_]=ܿ/b]kM°dÐA!fXY~-()87G X^z[EDJi14!sP4JU֘D EhM*?<(R[ |)Xz5ug1N q44!SSAl1Ϯ)pƀ( Pe=4@TyGy51xŔQ)qPJ4pCG7p \f/KNH1J4^!P 0)f Sr^PVtKD".hUTk&$RsSnOGWBZb&v^=sfܽsFa{X!W2$+8x4bXLo|{7!ygQg+F9}=s,wT%q'5? ;^VPB^ChB&S6BzN,vpR[E`twzR=o Eu})RHl3r6s*^Ff$fʕ̑M.KP"+ZVT# wʔcc9 pLXZ$ѤblVkPm1.*D8DH%4BVZII_jTKLEٚySY݈(ՐO)pi$$XHY\pH "*FbY$D C ,ĄPL6!,-cFPWCFqTGۑ0Y3bHyqP X;a_F%0(?u" ց#O2H1 FkSHOG8s12JzM9%fX3qMNjW[ Y;XKP((V(kdlB@lpKf^9h<5^P1AW!F& " (m IJZSR,@BtJEP!e%즖IjP(׎aƒjAփyr)-%jU:K'{7ewkZe9B`rKwA7G^b_΢ͻ+,(WIIO&e<] M^fyr+ 79h3GQNI!U:H`5Wi(rRZҬF\YVP"")gpԈt{/3g9͙suH$_s4"x5(C&{H.S}T2=LKlcRp+X5akCmb$`* 5Oz>>\&ӎUK'z>wN7뽍wo.aO~~'Di~?ޫg;?+w/{F"5bso7r=EwR.Nwk'V&mw/OH|,ZS3{-&}n8a9HPQԶt+4W5!_8ӽqlI7+ qu0QKw;ߊJQi]w+4[wkCp-)Žnʼniǯs8zvWX[;/kk1=QJ*G~pyG#!AC(2D.VF%J Fu,XN}\Fp k/b)g`qh6{v ٩Q:"M W#Ws҆vmc$!%QI$kMI$䙋2%%<S V*ε}nIXmꇸ[~(5ulwqkLOB;BUNlV*]ÁGVE297`/\L/?WO75K酏/1?j&cjWY5u@I]#S^3Vyt|޽Y U]=FW ʩ)x0}dЎޏ| _A y C>g S<%'Rfb!6Ej_6Ej_7 ihq3U -h.8+dB}zͲh w&BoGz=_-^N/\pP`=b3 [a{H$mK6hwC;ɨROme9S\:Zjq?*~jSH2nZWᅤ?>0xMo6F:g|v6{3t6_Jd|)E}՟!?Llrrzw.?l-1GYͷg?0NJp;׼&cÚR]8rӱ'O-7 _/;pXMyyy5OA>F`^5[dStH 6JhC}SQlLL?x0l~0$RQ[5x6~/*cJƔ=edIަ_MgwbޑIj{Ci UI >; ,͹§dhQBK? K89%\K:tW~/fzJv %Z<RHEl\m*J(d#b Z/⟿I~ X6W$v:rNhUz3RZ@?'3/ Ml.Oc}\2BY}!k kLJQN,j6:)&J$ d~,&aRQRQRQRQ] E2qPk QX\ וQEL)\?6*&M >`“1-[?c;2C^qY#KWxắG^ˆ6$Bd6ngEL~{&埭#~0m S?Ӊ椎g}%> 2R4.*ġ7,3u$.1NݭDsDsDsDsTzN09i"/c$v< i4 .WW 5ʘ~9s,v2۟r)\Õt}5Kpi+" ܐwѝ}jXhn HXsƣF'E TSi ^1[z:C503*3}fg^(m6}$p ~u>h/߷ $ց5 j5o{6"΁$j&I?MvGA#MO-Rۗ0B@-s.ї&rkp3~x+ֽ& [4㏡GSs6=jIJҫ_fEQ{6<)HťӁ ovDmhI$pA%Go; Nʣ [ݑ6clM2d>jϕ3P|_{'!'UHQ5K+RM(%dJj82.>Z4JEo%'*B)gTsډ! fRs]D& PQ8Xx)0FBE<"(*M_ɉ-JgLV$@?55[i tUQQ+C^b*FŸ5/̗Ȥ6451_w5ԩ%niAKM"[h8ns(}1b@ øf؎J>_A!1wcRU;T谕`q]L=Jce*RkX7 IPz~55[7[ EFձ(M://{I޹jFũN&Rև_ޭ9JHi oZ?D0Kg:<"@OS_*LgVL矚:(vL LS&v6y],O ׁY,N4` f8X׫s\*մ\D:4bA'/w[mӚLB w;vI,v ^I f[z-&xKR4 G6\3Zo#vJ3<-GeDjr]d'yp` )DSbqхF( 6У`t"Xt%"O"+Gk2+ w)08%tȎ(Mb̦B`Fhm$"a={CPp6_VOV`YLk83A8AN@1su%2"&AwNAyCۻk 7SdFdm^WߒY/CD]?_c[H$ykQ[*B *qz\E#e-e@*QNDH yQJN4-u q XR`7$h@KDH`g' ӛˠL-tq0e:6LRU Ϩ<7qMT.?v(>Mі1ՑX[s* A{Z8ȿp? ms_iKUhMJ[@uZJfQK P rBE++5- yǣ2#+(p0VrʪDmJ$âJ3_#BN Ʃk 82@[$VZtuy 4ݲ5lϝPjΊdCbq?4bޟ]0jU#-Q-WM&O?z_Nˡzv c(ĆHКHDRvGUpRѕ ?5{2/܋$~?#b)$ꝅoi<"i\Lg˼nϱWdxMϸ(˄G/$L>}ak7CbsmsU6Wq\UbxGoa{0$yRpw?u!b~+#jf5T;k5Y*yѝ5Q¬X^jlg+@f7go[9 ߜ611_&q1~c& Ҫ E<[}U?Vqj/F/wTL!߬\{oLRJ븷]] '(ד{pD Z UPζ-%C"GVHu,pN ,v¥_h5褉!*9 Q7}iOqVw7lm@;u/v  jGx.IL62%cja=C!ʼn':٠^vwSi'ӓۧUheZKhMAf| KhMyz$1|{JNMyџS`p)0f5yFbM!\v5EKz5ؑ!\x>x,4ǖXs{[&䉠VFFzju /j]nT sd8 65<Na;IM Bjw 0YJ o ';} SX):xFT9oSJ: {3ƽmي>  6A/|0I`vBPiy3 WM_+TlM;;x-k"W[boRkcACdg'&MՂ|O !Iޕ(0D#mu_;#iAܡ8$2 Y'Ȳ8~0xYûWn/ES)-ưNmw6n_8N#iNFlݽn$MʻWIN~=;oHCiWukj(lD m[HjwZ<;#/JR;m2Ix *#]8~wG{g50e"L=5iK(JoUYM0ޱ8.;r"$T}{})8Jg!8&8z(XR 2%N0AG120yJ{cl(od4 %%'zoA HR\jjpM VhO,\Q%唫,M :x tpS;s4L.\.XҺ˹č$EK**6O?VZ3/zkşW,BmJKM^jhkW xdp;#qЮ8ӤZ+8ܕ30[Xϟ3G?>EM/&+[YV LoSV\`c=xW- Tlb݊Pc6[XiL L.ҵ__4%Ҏ`Ydi֩+OX`5^UpZw(- #ґf:̈́K'OiD,;H !N}؇ϋݏ.\%Ȍ8^Y9>s,&cӔ̓F#g;xt|S_TSkUևnqǭoozC ɩ6kc|Tv9VnvV*|3k-uso{g4k߄Zn<"cܓLc~^cܥ}+3g-o UcSPkz0Yj8rM-tVtMX"R{ fjܓ>=)~PWvN&1Jw"aJ&?+AƊVU?/~a招k 3/v{b2w/_u>wéŋL |3L-Ӝq@rj8Zu1k1`Z(Hx2@(Hd ֠ݙBFJ$Pw!X։(`iD@W̉6A\!LPLDK&atX@H1u9;s1tFؗ{BrB`pIG!R0)5qK MvF!j'NaLnf]w. X j鸷Ɨ$!оvh+*iYQ# s0WJZZPJR(NNf_{be,Z u"u@iI$Fbm)7Yjfk$hm#0e;pFQo!K עn ҅BbwvRچAP2R+Lq:B*ÉD`t^gN(`և#PӒD[ިR:Jl  .dO@zi JʀlFNBv+|!$@P$6Oxj@ K3 q>Wi{M16glɔRჸ>-cI@hpݽL"-lDl@6Շ0|HhS̛1\F?=cmLer]jLp="Q)ۻ2}dGcyfoeGM@ Pr#(ෞQF8c˸[JX'*rϼ!,rk08w3%8s{Mau; PE`0lf988/!6"W;Qrk!_丛R$^wQ^))n!@ίޏ^/EW3=_ͯD +2 HWf0njg%҅g55R)wHR/ *YdpB2Ō{^C -^2 uWD$?RÕGN (faB\a Jg N<\><7$,jUQ<tvgf3tGvPF~6[]$?E~d$ ?9~ɾ@XAbF$H2Hx 9("@$`wyz~!g8aJ9& T~vScgRzvտ$B~X(᭖L>8i +! vR{X^ ŽJccKAO>O=}Ě2Ǡc9rf%Q8/#mIzRY&#{ 0-[ XZis*c(~t,6^&Odg0N`gJTC*H"Ր36ҟ&|FVA\)/!DTfLCulbG&&Y9<?ڴbO<A]MYYhꬭ.!D-0UH~b@&@ZDe_I:BTeY7ɻ/)fڶ`6yr qsf,b\aj2sA:߬cWi*%f @qRC@jK;a]cAdgx9@2ۛzcAìiNƙO/C`;)ewcSp+Ś=Z1Tu˘ea =h+m,}i3 n;uzc!,^}oeаWt"O.iމ3X[+nO9 XwyqӅX!ݘ``k:{aG@?>huwn㵱 1Ԃ0_l~?u!!03][s7+S|lH~qr,'lqJL$RdSPҐe%Dh455k<]!2iY)ORKWyr-,wJR,5kcyAtwl^+GGig _q>`YhN¨hһ f+-xOZ`i5Z^~uVT]lnd(;լ?OtǨͽu&[dxjϥXfucM' 1cd:D\+,zlPcHY(M&yr);+vhjUJܬ9"onXnjmX HF Y l[ ֨:e%;-V0/+e˺q>B0XQ~<ֿJh٢z 퇸BFh[:dn[l~#n6YvvxTh9+|dW:H*rmե uI9Ŕ֫TE^Q~;~J 윬CFFd윮R{!w< 6pli~?qtdmZji{aָZb.J9Kڰ0&RHr`D3Z+\{aNy*X|m͊Mf:] ?DŽ6lNPFP{t>.B5 8١հYQUՠpJu_- &C ܰ?ՒUtc:M}N> ()1:EӲ{ 1z9n3vibR9=]+QZsCb BˆMkfuI-RNWAT#@l[u&`Ŋi/\Eڿ rʶ֑Q MaͩX;[P 瑵OP$N͋:0!c=t ͰpJj7XD \!AKvҿS8f)α 2l2,d&/09 l?80QEjNhAgj7f4>v>y睳;3:7ϠЪ^^MnǗŏ\ H0Ntq ˰]kʀ ٴWH/^Y/)K< s4@zYe>鿊' TΆC{޷rѩ뻩Qx#{25>].p*s><`d%4i-ȣVQDR31J˃"^PA;Y?ytstg 0ܛZesx/:/u7J w.ć"W ~;~\ p 8ǹK7s0 6pgS,?=SޅV/(Ts9#l@ &&,XHt{cp5) J&-Lc4FB F6UqC$sŽi@ Thd$XƵ,0sPpN6cR|U-y&MF$J ~mj8 :8 ,\a0Z;dHK+>;z -K jY@f˭gA(bw&%L0Jr/) UBYm`VJ ]w^B9F B*`.X !9 PBψHht~N\3qx;xx;r̩q%R ֟^mvҐmjI`߲ 8ñnP=N(1_-aCz*O&$aNȠ"ϭ5f͔)HBK )@ G 挔vj Q~Ab+4W{GHΙҁ|Tm5B;Aw ,v蜀( 6 z`,c+g0) ^gQ"0wF7.2E_ۇup/GX_9ߧ?Q7w}N}UHO!JqN)LJTW5$!\DdJуX7M ?떊AIu-rZJѶuKhUքx03lR|JC1 x)ͩ,M/s& (2Z2.S{LT+˵ AUjA&STX!sd," *ʷ ,r3|uLnox|}V0^S٬"EHV'(+'V~=a g@ d"hc%2 À2'#*B(zTo0P'^u !ĘTho5g 8pSP0I2,9oFC^h'̔ތǝ漓өVADc5|d\m=S6ιoo.$ qiTf+С76FfCtoz>tcih;#^x}>'&^%H$`QNGh>wSܝz9WΕO3UpO |ҶMZ]S0춦Zv/f>{vbӏ/0 Y$QW#3 ^¯agQx{jxͥ \L"U/>>5,?U@ ! Њ=a&$d~L4/9׷tUp}t4KHYox& \[ں ' c^-M(+¹Z1!sxؔڀU\8 yPcG`;L>! 20<{삓qKbʞZ]R0ҵfɮnS Up%-*/٫x}H:nnW0;{7],dmLJ]޹YLJ܀w}L=r. sVȜ*܇lUub'@d@/0zHͤe*lJ9o&t}tTFX1]sUAB:r-1?>t"U_ 4 hx'Ē]݊_?׸*+8Q?O’Mԗ 8YC`q'SC[Ǔ=$UßNT6}=F8n>}Z Ym_b蕡aD;u`uB8=f`>dпd>CDE(g?_1`eSľW}^6UXA^+eT˨ Q.0{TF['A[(>"0+LLt{nJKˬX[zԊ^f Q)1>}r[˼f6brPAYLNЛ0nW$8Y%&ڱx-Ӗ %;1,_}I=8g .S5[xQ* -9#d!1U]q-f /JSi5BZ(E%(@Jt1Yd9FΘkLQ -Xٵ hv2VqўߗcHc]`+sx(I1&^*kPn Rsf)k #lɘ/ 4bwr\aK#륑#eV!ןP@J|'C3GP^)~._oM!vFEn+ ӊ}h/zbW|_6WP2$ʻ [" {wNfcfB.n? j%w״+C c7#&>m-759"Ĥ0o31 VDu[pZv`Z>AohZ*ɳ6y4.P 1L/c/VI]J>Vz K x2#~v a8 &$WN)ս>YSՎM <nn/Kk~ P&*75-=٪P|Llu.ΑʇZ>{ûR4? O{<"^p=ypy6߽1 # *[- ko?0x7<=b'*٣O]+l ǥ_uJ@V*o "J??ŁeR7x3zk{>8J2guױf7{x{S!s1vdz7Ҩwn`/>^1F!$[Cmm6wع[gw}LoYT:yاK_r[h!*e\n1 ٞwFYIy_y>Lu*'SdnF)yϚxNj9rw`?/t4?QgpƸ9k)O 2"a>o.;gR%`1/\hs|yn-!jre|sbrS@W$xtD@_$"Ȝy/&gZ"#^y9a {g ESG$lDRCW=Զkn];6@wxjU:TnwP0[u i-`=v0JDv@V+RQ"vsq;0#AFQb[&GEb^GDh ]0,\ ywÛʺiq? IR8~}lxE_1䷓Lssw6n#1ӊ:Ga.{V8h=J>漅ě}ũ]wtt9ϟf,v2gv "\A<;Wd:/@X"{6Zގ6ЪDz)-'EzaͲ3<cV+JTF[Ų )sflSnbRޒDL9w7ӌ6'|2f&td%M=fp:V#Cxm6g/7fOga4sҌ{Agy\#b *l.CHzA(ԣ6%M=xsP0I(.CNc>>юϒ2K+}޽~usзJ3]YD,]$>T_y{'/қ ,1~yZISe e%PEY`:_{Ó;+ w.^Ym`*,a_> Qy!Y_}T Q)E+K2=|2f4(ύaz?GՍ?#AQ5JHJY?a7E`wV-sU@CݚQU%fuU$ h_PP=^@?JeђL) % go&@Sg841@^].se.윿o]dOUG5UkO;բR}EXqJJ͘9+㲤},;jzf粬w,;Z2Nzf粜w>:Qõ;sT(P`@%CTf#9o6L@W%ӛ焄Q^`of>s*=idLw8q7j4ʵq5c]_}]N^h-Q|T "LdQ}]yqZ!dة N\lWqfzb[YrQ[05*;vzXay}Ii8lhU\"҆tRKf&̱pn-mlt%;Η.NE[%.uċӉX-SfM'A,8ABF73^fu/XO+HϧO1FΜ'jhjq8!Az3SmuS"<\y iΆ+_2׼KtyU J6%nlFHoy 8B̩j\"dVt_Q<Ȯ\w]σKtטC똅KTww > ioR!ʆP,Je.:ܩ\ ͯKB?.'sM\\ŗƁfwhtD?Awa[|Ƽŷ1om5oqـ'慥>B!LYG( j퍍* $Gg98¥aK?Jɟ׋#b\\^݊os6#,5]x~;/?Z̥٤6#RW :o"G͉@KY)rID8ӓF(Q)ͬ1I"yrƴVȿȿ:UAL9<Zy 6*RZN TNJ)D>_ھ|K:D Fs_0A/M͋LSӔLۊE Z;aFHIAm Iu77K);_AsgN ϥKHrGP 眃 ~pvXjY y P)]hnH2|I Qkx Ȭ\i ՠ a_ x1mEԣ9SDr&{t(E jxeBԅ7U{A ŹFכ`^2QB4(4ӻuhX'cw[S# :)Bеd *n#}.z9kJ9 ^C[צ$ "8QcR+zd/7!z=1J8‡8Bhx3)+|Q eY`c 4glOg˛ |9.LxjKv T p;`%m[ @2 I=Fv!BS-,cF:1 @WFVBcTE&2РI@D:^/uzčGz(!pB V5$Ʊ$zьbcZ))0T%% ͱa- xֵS q^8:~B E5?D1-Dc+Y;&Ij#uhw(pJ]h35mk($c;&'b, ҐAu+%}|9{~z n~龦ݏ?oy֝*fWiԼSnF}4hqt0S4޺[wy'|}ާE3h]ǭ'c}ڿѓǴȌ5ja0y?>o߇v>BK0kZBW^ quCpxEi/>Y>oZYorxڻK)pKԔm÷7Cd(ØdFrσW:g&QamfOwĒ)"qjBV%q뱓 2FXunONe}73*]3ZMtѶˍ^|9dk~1"mdDjI _r Zz@ԇ3Z [_.?L?zӊr9 ΉV{0rRb:D뵽UgpzeW\hˋ6AJ4ؔ>ؕCXIC b9ғ:xYfx m;7+W[hqRkl޾W^N:` 0Q"5kOrMQf$(Ha9k3@,.~ ]OV(3:` KFNy* *\w!e;wG.(\V,A,-JfYG`֎SU6`KN@NnEodHx}VYaYm`\rxN pjEn'ßΣ][s[7+,N͒и;jkg&df2[*\m%ۓ!)Rs)T*E @Uϭ{{[ۥ^rs,W zjśIn[i*`lT m*׀nYt]\X|Oۑ|"d⸪}R(djl_3T$AxLiTC-fmޑWgvI8n)8s߬^f  Hɞ^ۨ"uNF ZY90VP.5c @2F8­1 KKG:8@P]ӓP{i}_l*M7-T$M PWFՕRƶ&w/W53l8cztjhD:x_2ԁ[ӳѐ|YH-냪fH׺==Ф~tW]_CPvƣ-u;y!J6rl[C ^ΨK}J/s4|Yo]s!c򌌺O9'}չC/nOWTz?hO6JwT5jh!~[,Oܜ8GC:wv4<Xqu;_@KXk^ *k~M ~u_g?~M(W9bVv}y|P{ڟ+ U}vYd1fc&%'ߜMGF/ !5w5i|ؓ9͞xfxYn7FF^+E3>9&=I돁 MSGl@P^BrT0O߿?[dVȳ1Z$gtbzRMPވ 翞XJm\!3fޝ/$/7 &"|jhS,*$,c|ݚ,Mu!W1.bf2AAxX#<r0BJqF'c6Ǹ@Y酅R{&7#\b\Ds4a(#e)Q:boYI$7/+}2Uj3j[o1[Um# {E&F*ef7sTw%b'Sz?xiޏ_]Ng%hrTkI rpi4NyG&?ez~j6SJ%ndUvU2VvIųz-Ѕgʯm_1%*n[P,7U0*@o"[rW,$ʸCxs9jKg YD:i"q%=x . @ǨŸS`I%1ȹ(!ZMLtC܁Dș:0%`!gVOpz$P?[>^=>h7"Jf^#q2&79,E,:t$ۂ&!: 5PEհb =+z{u/d|~#vYvI]FZ]n=k:7`>~DVh€-AG?;G'̞_>A唰iݙ&Qo+v>Gici$oO_I 2S=bgp຅=0 n/.pL֮1z{89gqɵlZ8la:m.Yc&+j* @a{9vxAJV2`8a1$pTE"%aZ=O5-7-AiF XƲL9 r2ǕJVp(MR*f)t=+Чagebu!QzFWI~7ŕ8l2A[,STX2\.ZΫOmMm֘mRe+]AP!iy~osw( S@>K#Y+勺I4ْ# ߔׄ7XΌF%Brܯ.h6ݨ[e&vedW:24sɇ:sc=ԇ?ɥ/D 6b]CE~أA]aQ_U9w# zH߾fX~ {8~ (,nzPBw#Bih1ԛQ۸B)=HbsO_dDA5!9J|M{r{[n)X)M'? qZ-Vt')U*ÝSj)Ui:ϧ|MʒkW{_ZKaR+9cюS&HcB=/'CrSY S+&Y1e\0Ts#J&y܍-a٣*J;c~j_Qk \;Ӆ{9wM 9՜ک/JQװo5K{GK]UeL!VG==L 9PL,:E?Zة,<|{4Z%;Z@ƴC:SR =xehz3!\k"3Ǒ!v\j t]wo$׫$tY6s0fWJ=xAw^qLD3s1:r ZdX5}(ԀSΣn*H 8hnk~ƷW䝂"l)8j6}@/6݅Hk*&c!uƼE&`:f[@B9Ns#6n| vrwql~z%kR+QzXҿ"9G[ߖ"Q1 MS V{H}XNɪbl\,vg)rad|3\UrЙgMc] ?'?Ζ^MU 9nctbV{dswH:m}DϩF/!8 xϲ@Qs~Asi:aP0ETXL* *!:4b g_9 jlPofuqԛϩU/3c#DRRgD(H-1Spy畎iA%)̡6tZ`EBNzov1#{;"8%pcy-tQvy/ ?P! `$U5,]cLQjY(0r6m"l6ji* ԣY/@:oU% 2b܈(DYJPYڡL ]lBвvڗ)]̌ \;T.Ss)jS =ꡅ3c1Mpf&`"gpXfj;%@dR@*hw !FoK_amP3{C/jnI]urV6%4HB&0 -G2|W5eGcU^L Fa[O5K0_7]-9S2C-93HU>_Q!˹g®!K:g5y87ڸ/-9ڏZ)P0Q5ب ylSOʰD264€ѐ5p|}*=ȅ+@0@r)ijNcMV~!A =~ksхAψ$bߕ;@zyVgYfuvi]@mD$e+|FzBc KI"2_IPA)~vV@Ndi-d%s 9յ LC̢,jBiN"$IBDi@s=0 %`-tӚlKf sQMl#CV[띌LD?{WǑ .Z?ޅ;;x=OE"i:lbdVgM2lXT*""(Zj:D21l NfOQ3jiVǧ)Í!Bj)< je$>n8 jɺ#ks 26EdBJf BfQV )Y6 -/71 Ҋ)a539 e$,t)3e̘Ҵغ֑]tHK?D!̥<-ϋ`fľz2QMk) Ѩ%+*OA ICo 30#3ާV:1&m+팳6A(Jf\ڵdS[mEn$NEh.,"[U+#A%åk#6PCr 1z&hŀ  Y?_AFV- 3 }KOmhR-*9&5Urf=0c-훕KJCIF@T%3b\嶞2\}67Ѓt^b01yM,* ^-/H gh9WbF6j&[MB]-mPS{vԸ%dF6$ 9jma) [yx^\mJ4Hvd0AÀD|/]';A ++;u x{ٍ0Kَj_ r-96t mhmAI4JBM 1%E-kIlf-o%m\ӎ[(<i+ i+ h^{Dr3vW0QU@q>2ug__(;8؟qnvξ|qER8Lc osVE\؍`.nek mIY94H-[`MOV!Ҫ=(g7LλſEF0~vXI)U7SVMz%d y#Z^PVh9862 n2FD,@i=7E*[l ȽAPC`pYҊG9h OD.xH\䦺`E~|oSL/5kml0LVEL`cژ31[=0 } ;ez/z FռͭDpݤWjyzܸh]22֗ 1ZM$[gӈ .7҆!8Ob5&0cEESZw^"h/چ pYh:JJ5'#0zs2+p)ئ(⠈QiB]Z$Btm Nh&(btO%` L!r/Yf^ȗkZGFAr5%-\[ Qc1-˨: N-wN )`KJ͂s 4W_ :\o#_N,c+Fg䡞ֲ+ق穘;)-??Tޏ^C[.Lwm7,ɭG3bu?ėo{䳳㓫?W'w/_V~oe|o+-쫙ǸXI#3qGN iLEב;Te]wc@+D{]wQZD>S32aN?Iu񿱽l^D jcY[ ,z_G?ӱ/O"~t< oOӓ5FN~\sw~H[-=|f! gga{^zЏ e`-Rȫ0kM8lR4}׉O Jr)C _ S +OހLqR j~f[XA)aΞ#?ðij. =2PP0wG|TEz5E4}`rP_G^\y?nFg]Znlo`>[w/j# y\:yS=-eBb7[e10@jyH< R1z-8^wTys>|&ZcSع8HrCn2(1on3f蕱[}3B6|&ZeS|ƹTgʌޭV%S>m]Р&no MV\ZevaǁVb;MDa2jELf2-dø[y| uCFr&.*`JOA/;ک{}_-,2߫jh^XrGpۜg8:x:WxA˗'/UX&$ՂRC4e%N'/7Pn?5__T'ӫYb_թcF`[7\j?Y!:6]|wS̠op/Jiv{uur|nDfbaJή)ׁ\F*h5JgaS%:TY2 ӯ?!qfdp\a\׼_Hٟ .>m4.A(rvyE]g7y}g/I۹f/^vI Yr1:K,{btYI85>7o߼~[\AZ4 zFzG1dRBhH1Z&^f7@e<^Hy&ry:OG=f/L'hp Χ Z kH,i;f[X*2$\Q٘d&FFUmхDEK0L)+9+k0׋X0zm~ۣwOVK`y-^ ;z[D; bBG9:eF& [1~&zrh}+9ʐj$id6)-k(ղ&E5ݑA-kr^C1@|.) kQ,[*i._T%GͮSyn8&gc)mﴚ,pETn|Tn luv:E`^ewnUzb۔EנY)A?쳼,_}(~ZNӓkq;C(רp7'p{E0:.UeNǂeE2MX4|NJLKQv/٧~z6l{6.>۟jf{^lDZd{CuMu5jȓa]na+tR/=W| jSa@;VW _ Ǫa @z]yE-a2Dt[dɋxSԸJ{ŐRndPRU'faS%ͯ*;kBtQsd{@oS)0ѝ]]0= KXsP*Ӟ}z݁俑g!m!9a_ Bχg^,?z7NƇ@SDմFH([Ɨ[Jaw.Ky< $T xG.4iKU{_ tR+jsùN/ϧe!HZ~v“ьUUqҡ1CW= gYl`=it'{9Ns3vE?Rx!O|*#ƨ>d}z ;΂e&ZiCb-60*dMA?D-7$jN"Z-XJ< I,9UxL.sZ;-H(X@TI>vTޅ>oU- ,1IK'65L_ XHQc)$yy#56$B&<`j}DlZ4jZ!jəv@v.SӂV@ǣ%%xR Hb_I&=. fexC^7bj_7Bl F 7/V |xjnF䒏'iQҌZ=a wǂ^ w֝xUnVWfE*7[}w*JR*ŕ7OOd RO-,ʟ^_˯N(/UZ R>ΰ|ûiG'g/HZj6&|o^x%;Ie bA6'g'׍rI)%ShN~y!~ZAKtnFù5в3@k)"}.J`b'} :ӠR24sȿni|X]05Ses6],<ϚzKL2I6UVz56*bJF_|.Ũ^ R7xl\ m,f7|;G%0ڪVY2*ر^_`(1ga1*GݼpIXLmoր)bkqܙ7!(OE>RP.h\cs *'|h ͎o|9+tEKhxXtĖ|h'VOA+Y{rTLEhV ؘqR i\G%<)H\9F5i(ٮ.ulLvA_@,ؠCIY,g3ȝdebֺ?j1 FJLjmNtF G!ΗmqU:ǢnJjJpq7H^G0o|Q~[eX@#*zlDKy064 bƽ ?{HZ1@1`kC b vl7(=|<).BImT&l LaY>]^wIfԴ(wUŐ5fd\sp"8|?&3q3smh `Cwmq~.:ɢ$ H`Wk,i,͌ejYc$Kӗa-`6սXeY93ɰ뫛Q4Xӝ@`FTk?K'>]b^mF_4ř/cQ$5~R!=rpαHXD +R? ^=n-#S0g;B;Н0ހkD:g&B IB T2H`RbF'o}S[i wx]2;iA)%է"IR1N)+[!n8m݁#yz_:i7ʙ6]71]i\CzM* r*!: V9Ϳ7DD@\^+n0SNH'ݐrqPd%篶GJZϞ'*!Bo7g'=NEURj1&Sl'j&FZj|יV=֒ ~jSxl(8IGV'i}Gcu#Ms7j5Blsҳj02v$(7  -Xs[xX]`y*s {rh*5+?FB8W1GMj糴чQNлwi-o]/x<-wm.nY_t[-o7Gœ`5^ssefhR\yu4C8;z4BWo9nGQusb@/'2/r1xpY+a\}yGu |*f1!y1CN5\:Ii)-Vk;'uxJmX3{OjrLfL&"8*M0ptϮ\tN*âcQ̙ }AOݭxoxRozuc`W:YN*0dM $(O%n{ 5uux"@UNc+7_sv2p2rNu_QV9#lk&s'V]w`V|;" P[S[8#[.jI#2":zҷ0O#{ѥFd[@4h $@#ȏաFU&du%8H~XxD5n[u!$d'hm(_E~Ǐn_Eޮ1+3gːc}:qP\1cPq7c#7A Q\ 6M O`A'nmT8GCoyxUZV5{|]7~y}1$R(߮FN)fͣIn-e99ZыqZm^.CԦpryыQtSck?X+GxsB ( {0V:eRd.W(ϫQys_꣛PjTE3f_q.VZ4X-XSr|WyaWJ#aey&X"7Qd8UHyC8l.CaDZ-| /bCq@IZ;Y'UXPl!aԅ}ٜՈTRZT #7T^+2`TEPYDU?F쯲 ;qN`o4GYO8$(fwQ%} HTԑC&qYDf9`à% (e' 8EѤKF@JNfd|QUJ)(&qwp{IeX0p;.Y.Tz_b'L~3P98Fb0a-!IKAgGi8h󇛻-A͚4kJЬ]㊶6Wj]ŋpws&,GR%HYL)b'EΝ Dp\T jM(o I(do֣vWYgk#'YD7̜" 湷'RÂ7H[% 80{zx66O9 c4o2jt&NS+Jْ42@l#+-kB*ڻB Q%Cv4~`;IKQҢ`R)G r>>r"9Ґ,@L+*; &{Uqf,`#ŻSIE]koǒ+>w^b+JJ8za"i[=$%>$Ἲ*x x 52<8VXlJzyR/QFHZ͘ʭK*G?K/`k{6x_5 zoBLcu*cz `kZ'WXs:9 ӍҟGdТZ墣*Nع|v{?M1h-.i/N,o\f^E]!`"ʼnA6{,0 jtn׭\⁛/pto >?=8M)|۾ל=ׇoא9;ѫ?rt`=_<89U_vOBzѧ-lu^7z۹nn2vpRlدqi_v/e{]Ip#Nhu<3<O>1D(!⫏y47F)ğȠ 񄱯7oӽZz{wӏM>Bi@ذn=ʋ7;}k =܁~k|O)5|;fUV)%mk9;;φm|͠S ev_= $+KaQw 9<>w>9~9|V0WiO#!s:gW#Ǿhw0ӫB g.MpUgh I 1y+D96..Q2M_| @ ~?vamhw7E+d;u阼фz6f+o~}{/xn# luuSUkXݷI= Nk`Wl?>gMs`? oqNq:t#[xڟzoCw:]uǾ{eGO0:^ZZ35PFl+Cݹp_u2PcCbЉW+2>dV*_erWfCkǧ5y`] 0+[6 >8쿀<=q%[t ۵Mv JNAn ߋ8-}svg5 RUv^Z9*q`2 xE`k@l V"! h GZ`SIpn-ҋ}B}FbkAwI+3;/8 Dq?ŕ+G;kĻ&UT-J+ĻJkykLU0/F85O{iQf1jٽ~`mssߋĘ8V^8p5 2Mpx2*0JDk$ {E7r`F$Ɠ61nb<$ƓOb<ˌ4̓2giΉ,03 b:`0YIHxls愛:* )VG . J/Fh[iꅠZb!ZU(7v6.R[a19:pc01h .6] QŨv'ΪkHtam704 U}w#<K .N|?;*LON`'~+V5q53}ׂ:0l[F oLW=:6(n=vNta.@:]^5h]i ||IBbchiLZwuUF1aj0 ~(t#sq3^BO$QfGXڛc3ܒQ=-oo&]֑|DYJjڬZy;ׂsf0Ή$ȪX pe`8.Hm :)\)\ce> $1<.&Oimtg 3AH`xDDO_KZ2c/"bA #+d2YJXA00^>y 8 X v< $'JxdcVp/6ð7wcd|h9OB*lfRVHz2ǁ0Ed\Z* 휃fAsBPpENSPkh8VZb.23"4vɢWe,IqJLJ6)*gу~7W0.X C3(fFagh<@e%^R <>_ñn WF^iڣI Hq,5CQ׊9U.j6b19EP *ӘXX w)duH32t"c bvcv"鴑tf\."[mٚ*WbVJ Xchp&8aDKCB0HzYX <=Uo(J_RLP\5qgqgEKBZl=%+0<}ImB { :=V>P5N -s^~XWNy5SY3$_Tja^u,sA * =8BN [Yŝ簀Q̳&B+SP!$Je֬lYy Ae˪Y"l88#p"F8DlY* kXmDZ1Si}sثvl%-ZZ =:v.$_{reffdI+-(4[m3\ݶm311M?r-)[LfpsvnkT{*\ 14gFr%ƴi4Z5pƱDk)Ao`ZM!Aj"o]WņcM9!^Q#GX f{EKSPjfJRw- (K}YPuA_U4KpDSD0qI84I^L4 <@LT|sIw*ES5=ޑ=8.M?siJo.#хӳ $FjI7CӒqBX`Ϝ D]å /h|lnhϜM2rjcEL2uH(vxkWaC־w/v~۹;DzK ka&Ϩ7nyۿMU/-TYv/2Enx135ez38ȾkDe0L 3r$598V|^̼eO|a=s˚MpbT{svдfR4-09.L+mĩA 93"@c<3r+9. Wiee$(L"ccfPc5vr|UI3i|f;.[i<͜4sfΗ7s>OUƙSej3>yx2v|zR8ց@Ra.[s '+Zp GzYI,IXĵfT8`3J Ĉ8B #$;eNy$Y#5FYAV!bbEISib.4u.EHSK'Ppd7̪`3O5sBZe$H،Qg2éʴ`sKp_%ق5>^aeϯPz튐5KJj&mgq5睬{Ѝ ĜBcJ0yR)p97+u˅XDIvEI6mⵟ>}nkM6kk|Ji!{O:$±8cd0 1 m(b;Omϯ]>jiI=hsv~B ;F:] D!$U&5QrX.$B#h2$z6ۋN&zm.C(3<,0w,syM2= 22C# =!-#؆\XYB+Aǒo,]x>z))ͅ=(j?{׶G\Ӯ%"2s=`g?qȩ8#4$;l ^ nJ:%>VE%&[c]UNOoᯎ[9%9Y`^FdrFӥώ/o)ՎOfmkś3,KLnxlVΉtX ʐts.Ƅ؊˥FZUNf.R֢CH1%l.nAZO[=| k zX SuB&!\rqC$):k'J 3ZƜm2D_FQIȵpʔbuw6z5vYCrX\j|~x2STe]Z,RTr^8lJM䍙V8UŪ)Ɂuv  @Zl"~ "S QƇ/yCvJ5rHP5!ЛMu|22*YTD15ecAkrdWbJ8bD^{dj}wur~yc 3pvdZ@H l ede#)H=Wݜ >qMn4 !"ȈV {y{UθUXG$ы-HKM ,k&!ˆ')J ƹu@8h;*C)8V:E)`=7@˾מ|e}23PmD!b30' ^j \G&(!i zG>QO( N'QC왂 ٩iR#NblHR[Lִ.XAw.a%tU[H%  g/lbi?PdR*qP4$WiA|@{T']Lљf GYR)fRZ.`#aY%Osi+L842A!$ B̽L f{QHcEdݵyA!IWDW7+?A<iϗ[Ke鷏n >۠o}3wbW=ߘo~~l4MA'ۋg;E'~p5:KLjU b ;ц Wzzvz?!>]XX5ORjMQvs|gZB/t|!^Mokf30inMU1@ÞAd}@ZN1cֲ"5 {a%4ݖva"Ÿ8A8Fo *0DQ@on{ =`H`cQRMZF@D TNtbS;"7^PS2i4kAWD/X:(R,9EehdQHD3Pi"D!3m hT"{w .}ٷ#mvgQKP)f{ 1H҉ )~1R(:q 49PK{@e_,IQx'dXφoGg@R|\pHwRZ!DqFqJFvSsPi),p,y'D6u?CoàrYq ~D.ViO yۈRu ==uD6vO"PZ7'uH֍UL@޺vMd@A~Kfo@9R..6#~[FP[ $g5m 6fD`v۱6:;_h" )^eo]]]sHā_޾jۧRA\Y|u64yb TLrdBMj޳Jb3#I$,SktV`BJJb ՜C/#'î~Ei[߯weF0!=N^!{wˀD0׎K3IZk9c|>inҁ_N󀉣Gت>Sg.Swrt~8(Qx}ʪxbY=44q"Td?M~@͔W7gՂIr47Ni;'@B!9M>ـX3R &@=xt YPTfe 1:F5l1a,fCztgAJ&*DH -h RgߎqhRc@r/7ɾL@$X}4эlc9R9/nwkATl`N; 8f[Ϯã#H}^)s>l oPN,N'„7n` vG?zy~<;]ݔDO˚bJl/4PN2HY1t5p7ӳvH o?]ĔG=qth̗ :tԝn9^xqtkx{=#VtDD+Żo:ktPM- Ƭ9^2C sػ7KxV9^(UY99[)UxS3GgHO 1VPMaDdmZP_\b/}3<ѵet w]$8q|6nkܗի_z2wuWY^֏lF\R> CېjC?,9.]sT%7JlA6B"bQIQYvnjwӾE9˽p}g/0Whi|ߟ>LrC맇g9HAJoi u4'0$\Z.1+]sHV_mtZL5H6^g/bw4}"@؜B6S}4Xș67li䴗~^t%dd^_<}v]}+Z̜v/kmP ~*@J+{Mi4?@ҽ4A/1_kLڱYv9dx"YjcH|D~r"iC ًܳI|4q8ճ|*_{!ʙ{ /4c)\`eqO+?F]MxIR*@5!膠\Ϸzї>w$ޔOxsxA[ou{jX'S}'~qw8&&.%SC^XL-1h(N2ڽ?alRꪜUmϡ*D-6:eJ^J*fB)ű31û4v[*ڭ S\<'.YYn ~7PObp+KQf4oL |j% וrpM;2Gf)[z')`[m ЯLI%C kJ絋DU6PQ``CYCKN I"S5F{:'đٳRɾG3qFo׶ Fvn\sw8&;IMxֹ#K0Pk-!ZTb&5c0Ik J& @|O dkBQY6!3{Xttqi{"s{;{QxC3*\-lTVĺzKĪc:AcxR6.C[!fҞw~}͔oP JM F\tׇ97_m#Ya^"6<,bo8Yd_$F#%gՔFCݛEplYCV}U]]]]$cF'2VXC3Eph)@Bǒfh\fLl)1z"`]̒GPX L3fe:J1Y>ElG߅y Ƅ;h{ fS/~=}$UH4zqWH!䢶VQHOF۠<{TWW9 p}d{.:6o ;p%*Rqe)}_fCz4멢~ֈh/ M{ɘ7lomVUIQFA @hG<>lI9{9m~lk 1r5ȀƝjd9OVܤJXN&Pj4R5De|v3}XWC?qV傚zK߳x=1cxaDuX#Z0' ]i]f A!uG՝^/ɼ_GW`O舵cJ+pf")AgMI؝>RA`igʽlj$Awe<:䫈t^򠒬\I%Qyiu+֪θ8-/ꆭG%(+>_9Tv\>n? /(/> q0Ms<{H!OV6= #jzMP/g7_ؾ}?EΆ h;a NXiCiQ:Z{ -.ץSS2uR]աIҘfj\CɦZRfD8zprDks$W*4sEꢗ͓N GN>_YTX'>9^ڄ,$44P5.Q18~ߎ'5 ptn ͠\V?@Ȉ>G[dh\ZXRJ(iaՠ}h&U1IyeX屛}d~p!@h>o_R ŏZM~V4d><d$ 0錉 KEͬ%Dmv- ,2d&30 0MFm0JӯdwISnȲ0Df~ZS`. Չb*( ! Z-*E0 Nڼ"=fi_;LUY [vF1:fiz rR'D9-u ($KXWKIQ --I]T5e?l޳ei q,_> }}妽 Ja&sj|^ЭMi!$:*4JYi?t-hcb[C41iK;4.3PvfA< @97Y _ ߠ9駓-&q^LRabJܾٷ[S"ss:(cfgK Uʮ\N y-4 C%DWd1Ἥف䦣jʊxLі.\i4g@P8 2<˾DY$X'ćB;L?':Ts¿/P%JS*_T Х[JTnVjI Er94[{m,?klWO|?G@ *{M1T vNe&|Tc' H&*iD9~5&3>lN-V#TxPk=hWOS<o=Av%g}BtS3睇 m+:a#3ӣSDY0?HጰeIJQ,ãzeVA-et GYq>ģmS'M!sS/iL8z)rTIKN$U e cSvIA;F)cõ:\V&*u 2匓R~FGk]X;if:0 zW<jr%nF沐CJC̥XEcĸ,R:dp,SbJM9 $`RBx&@+Ȳ48WUaNG:aJav|2,uwٔ++Rj|u/۾}ӏ豈 bbYYK(C{kTvu nHx,.`T^R0}BkHa eu9rBA;[*FY gP&GV8BHF:IHp8JНjw XgM*PhJN:r{iqkS.fw yʔKB[v캔˱)i)zWda\1R6ޫ djѶTe2>O#UFXNZtŒHɯ^(Ip2E94#޽ׯbU2x!_טƠ(|<|c ̖w45'A J'F@DbT"4Q'ݛSJ(;rdf#F.ʲp\< y)_~_d.4ȟM4H#:ɃF^167B>r7_Lz>9?z6#e.t?Ow.Gx[T/zѷ. "z."?yꓫ DcԮr_ėWxgI= <_EBOy`CFv5RKhId2c#9HIZJ'ωM(dnMY} #ʇ iL(͸҈jSKWb7@-2t{l&ً@R%},_8TĚ4q\&VVjA,l0Mc 5*sg# wkjrop}<p\Kf=Dqci$pYT'W.Jl+|G0d?~gHFa癭\^&hcHI-B53&x+ǿ \J9ѼUJrn>ƒC2#[eh3- bZM{.v"U)6se*Lb)1Es+3 )e4sU޳Bp+,'$4vm}:8(4Չh/Q5C1="#DiUh"'uxo?0ah8df?2{Ɵ?'Uu K->g ?'i}-[[t `3x*BOŖwHRsF[I]^m#E:86|NL#'`aoǟG=,|Nثc|ɶN+/9j#ڄ?i mՖ_} hv_+ւW/fha6!WeZ3+˔KVc֢}MȪZ\DAZd#JA(!3B4:ʒVٮƔ5B?φlC;#W̾}巯H?}"ꗛotW1TQ^Ohk~ Q-if4ka rPxABzt,! lz>|?DrbtMv2UՔ 46+ǾkvI0b <1F.Cȷy^zI$^}/^'8r& k%S))i؄2U=ՊZɥJd1;]d2.&ߺ/ʳ{@|2Ɉx['/j]~jդ2q5]nѪ@DsT3qXap=]UECoZ(kP PYs+*0~gs5zMŹ;Jt,1P4bC gFtJp9]ANpkbelxHgdfI LX`낫Xo!uʲϟ/ʇn6r5+{_;v9Z;s33'%S ۗoR1RIH@RGk81E+ll3 gD) 򃗴u ZvVD/KhZwaΰYRZųC oAԽvK¹!ڲy%-_Fk/=b pq\pQkPLIw*&! *,N3_Ҿ!K3E3CA@c S&VDz{8PW,>>\dԙz>vUB|ȏ*/]̈́G"|/S+\~ǗW>}J~ z⻹]ɝu=\ň[џ7_|\3t: 8GۧGwW\N bú|ZRn4M]2PR 5ݯq6s#oo4BK#@Qh =cp4WKAj5߸pᵗlB䬝S戢kJJVWFrAd.LyPPݨ&w9|]܂Y>.Fx7l6tdWOs翖g6?&$%UKG?}G?0*ȸ6։ Y-CΨ4#i8hUTi~T}Uue _֙,*_%@ϛ~6RӺtxc2wyO_*ȍH96altKQHRuT\tch-MB܆E}ef.\'P2& lLOe%*3^<}\ՐGuUCԪ$F_DKG; E,HFf;s]1?5.ە-)$ۺ̥=^__O//ݭΝB kz,j=_].X>OWV- vi}6IוU&uȹ-tq|VAz=jZ1tcIfXE6I{{e 2S9I) 6rVVR1M) 9srM ->_ZIbMcLy9 I+I<\ycO?bA杭y(]nbAG}^*^[)ZeKw!22'<4סk;]0:6;y=GWj%T0l}B>/!$ICkf"fiw,cXI`4 ) ۈK`s=}4<8sD&k{?JJN)ĩs>t@q⾢%E8J˹&(jr8 D;!wgTdx .bl4v ;Ds Wɤ9ЭsەsrTSG i5=FUw6`!E֟:xu;̷Q{ ~R &  >mR /PZ}ԔnْH755~["[7婒vZ] UF\(q֢蒰2*ؒ7 F{j éjJ(˴ی &QQG+݄%,M VUk5-FAVwĮ(D0݄_Sgkr&P "zsHЌ/Վ0HᄒVL[̐ N Bڬf)۝&>*N=%֕Ώh{u|pS߁Tk{E;Mʈw"=~-`bt[[|}$Lt+4QHՉ2Ah=$}/i[~W> Y-rmNӫ#*ev\S8Xg|ЈԄ/S~15Mm|VQ)@SvڜBNB-='nX=XF/QsXXzEd\3oNnJ{ )'zœMCIg|I6q[󚼟;EoVrjLLg$JAdHJ(N8XI2xxno&kZq{an?.W rrgݧ/_g')ԑwƗ@v2E}".q!⹲gsl@%y}iXoZIn]PqXs {xRy/ S%<&U2Yf#,VyP,\Ҳ$ALzJՌ[e[N D =p?߹!]+*g(mt6J(by5%çd:[ŋ#}, MO3Y\B!d&ThMVQi>A+i1"@;uξE ֑3r< K+κ~.'C8h/# , [Q%i➀a';'ssLBaJZ(*Ϊ @6@W`MDaHDFH),iyRL( ɕ>MǑYR3%ۃ-6/L͜{29߿aDM3zNHCys+eVA5x갔Y;rvF6Ýqh`A#>pq (OQ='MT2ht0bWN 8CuXOysݦ7+ə/{N.)sկ/%p/Uv>~~92]g.{mNf7E{*__}\=7N Bu8q۝Y%7M Ip9b!?{۶!ЭCdK: `Mdć*~rҥA{Hɶ8֑H<<91"1 0Ơ/1K}}huͶhY~2bk'ܼ^YLs>޸67GIyLI g2eSחm7;́ u]'Yw@i?:ۀɷ).f'50 ~wUVK{tؿ(Khp}\[ܿ)99>>xyg_cui/ON88A鷛Ǖנ f2˓<B7U_L\=qiT׍mdƪջ\A{l< yblAa<|V~9ښh׍?ۓ@)Aг4:W'=KO^wWh+p v;All`!~*qܨ\V0s ڟ/Y1tc=HgG^agF2?")(qքSawZ.[g}Ḿ-dͯJFq9%΢.ҹpZ89&^iF AQ| K}q"E@ NFZV@q ,8+ݼ5L]P`Oܹ2캣,qf>5{=x+͈ܵh FϝL1 _%T, /]Imf  JHFG(T&%ZS$(  `"\j6-_B?[db`O 8K1;t4tZw_(e<[}Zs;<ۉ(srl ᒎm bb|m5 ?Z,W̘ ?γ2}Atyww L=Wm/ cX`6/ T!c 8:!T0>eJ )Hbm@PDm L^weYYA+j!"|JIXG$^m9&yS[Ɏ= L^VyV« G",Y@x?9hJ%Bű i0kC1H D&5ʜaЖb"*9L] (~jna8p_y^[q#4~q1f53Om0)h8AHDG!R" "EX%J P%!BG E{Kp X%>vs1B&#JB#}!t @8H Tᜉfe2N2J+!cA"%CQH,p*BLy8js@Sqê={V;-5P*dʞ8!N}QDnn~^b/*K/^k=7Gj:ؠM׀N貴9ȝr|QEzQ-/<:jzXe4V84u-7>c^ϝ܎GnXwq@4FQiD ^գ&*4v;)GIPa t/h~E4׉c\C;TnyȾ;ivYZe?Y &ڼbylĻaln-is%ߘPw^PodEllA0<>Evl`JFj! Tʄl&Ў02#(`+JlwõW\J\KNLDKul$Hb*c3ִǔ kI<"6XF{KRd=mKoaTd[r|n2'-77ZznCTxTdI>tTmO+[`e[(E&/ar%L6X EI$ɝY/DNudj=1糈Z'FTZS:;.n\,!K)|<ҔB'.>m O]SE'[DFnUb*qijk4caqn೽ſp19"\ʾUZďb}l cC ,IDӭhc v|h.2C*\"dgtEbgLu:I%uZԚ`s0˩ wr`S,SY*}>S&KbWB/SY3$;_dw(T%5M}s;ን D`M#9\ܢ`x$R*7+DbZ${˷pE /pW WP"RtEPm~`B6Ev`@ %IhU* T$ܭ /wb)Rvb)vb% &R_ѫ+A( dcf`>]X% gh/SC*f/^Z  b^F;Hk\:` ! p[v#(!>5A@ b 976?"D6/UP֡tݡء`p(K;<( NHD.5 bkJɾ'JASy,5*$9U8w"|m2c?s:)x4)@F xzzx;g3{ [8n UآQO>qg8ZoAACUq,(XqXlv" ÀZƘH@:i"7ó\, u9ho3 /}?غxo{j[!/^c;x(^ww3GM E@۶0+%!!Ro\LÂBNQPZyN9m΂~J9j /U^xSo\ 6Lve7{&c`Pr+~3k89dZ},gO7πy9JW;lt=smrWy6~2ᩌO9y^L^ hudN K0Rn"ieI% !9ÙJPTuDŚlL$ILHɱ2󃄴Pmch}ݜFQ3LX; ۘ;D@hAhP3'ȁL|S$c-Dhx*[yy36._~9?J=u1Rs7`$g8dtX{W=V&sg~.hxD~~f vZRtQdAQ`yyܳ5't5b͏Iw|`NZR2@6ԸbJCA`>xV=R0*]2}I"d ų+ ;F H6 :zis^GQ!I7b{`PMePMePMePMAAo,hDo>:bvΚ'&4}n1x 9 9*T {dY, 6`C!MԢWi bN=l^,lM< a*d NJUBjIg~Pet(ٕM2K8mۺ_݊@ǎ:eT+Q4vqOdu:rp@*J̍x2*o7{BdwJ003Eg5&ⵢ]aVe}Oo6孎VXxQsu8L u+lP͓SBU6Mi"^s (ZA8 &hSV"vFayO|r^9YIb"={NX1iefЌ&pxł(dɡQVжl/hAx8;Zj#/;\^qrnA qh*z{}|}VoO1x~GXf2;c\և׵^߿)pcNI,!8]w1HpH( l^moOQ76Iby0![XlJPA2A L )g62d/rK -C |qY]b/MV6.ޢg*"Bi)9&r`M.e\mqF&H7J+PF1l[iHl9)!]cĤ9\ 22jG5k`]B8?[X??|/t#DtlOg[̆b ix0"N΁MFQ,ͪDT$)BneMӁB^QʹۉY%(9 uغĬ LsB2eCK$V^ANu661ɠ!|iEVk﯎,c3viܘ:9f\αn$&UFFeO:#.^6 Hh_a8F4ҋ;6YD-,LL < ɆFRXWM_Hb$bvb/%:cA"iG!3L2GD}FV1_tO:ʣO.(l&P7#rb"keM囃`erf"/ΐ3v^w{Ww7 *Y{ߠY,}7 7.>|T HqT Իp `oy|6o~sc4PP%69ޞ~s{R>C e`Y>|vȅtH {M40䠎xdjlʺO+x|uŏxRwH^ O3GFymgx&쒍G01)A,e=q.Ɇ!v  Sx,H2[R0ݖ/p b):f1 3koԵ ̃! 6~h0z [a21gZ6$]Q 5P(O"q93I^o.*IcVF#3-vӕ^X>Ƌ>]J>]֔Vtqb 7J-3v Nļ0tdz&wF; 8]63{2423X Q039̴[{T\Ղd+[XWV& 8ШVaaFHʙh;iyur2UiekVq`þh)Q.(j*mbzˍUnZ98Ƭ rcڑ, fڎgj+ĈB gb`@ eU-:1 gb{jH8^ϔX`bqVA6m0!Q!#tFڜ#\U'4ADo5Irnz)nx⩎05\s}i쇐r7i۶qLeL)$ejyMۮg nLOS3F+ȓJ]t#7DJHSk|!"IőI.6?맫!2۔]V1:ݏZ6ywu w.W#Pڤ_fg2ttq|]>~>ߝۏ?^].s|AtugK.LgRt6 rU%=ЦeJd!@+tK[+C+OJYf⦿!VWu mDoYϕ W<7PS5Ou+wy4t|{"׾F < J*1\M5'k͇?eاv߱lpqt: G%\^*Ig휳0K89j죓N2/VJ3]_T{죚U/ky*5 7T@ܒ:C=^ =$w^=~E}Њ'K!A=IܑZBv4g0S΋ݍ^ދ95ޯgt@_y^m|z]YM+[`DD2JNXsJO8̋մ‚`N l@4.2Be $"l}i8z6LZ)-wf;Qڙf]4k}%V'Xm]%%4MҬ+r]YqDYwV+yu-0,8ϛaY ڎgJZ 1uք,b Bvv!1Yf#X+C?8#Iv&SL0!&:.@>v8mtQrիp󔃙X)Zo_-}>ځ?Ղ ͍fW/w|s$^o =CPIg~nvBzn_X?vxꮇu1@P߾NW> XRٹ4:Sg> اGT&()>ҘIl>AMy6|'k~gJ&ǟ?]6&S;R3u֙P痪[bTϨiC 0) ) 2e˯'J!(smqXv+!R΢& +BT{&"M~Ѩc|>lO3ZqmTY)eaes1.k1 C ed 㞲 EL8١ gth勂nș7=#F=7Uw4ߑRM*@]}zZEXiH_Wn^9YLR/ۓdRBB(©Rf))OZKf^mpC)CdBIGctmcT^aOuʽ7KsTHB^zC3n<|ߙZk gzVWEF%?}Qj ,I-8lnEV; Z ^ k$+_`K&@0P?C 6 Hҁ\my5$+m iI$8I]P~XV) G A5DmUy4bQ{5XC [3\-r܆{BR3b&ZcZ-Sgt>nkeNks:d ~͒>hMܼtsxFd!*JxvNJo]!~Kjˏ_F%rû:?^L`:-@7曻f b#ej'<ܙj“f}{,+7E>t|Wk(64JhqڋlmZYk}yu"/FsvW}/}W ǛCx_~c?~pE޷)ίx8eeϡB=z`ѬJY,^F/W]؟' Îxފ@BIt~?PV/o}ñ`;rpoqLwD< uF7`7G 2W]-[;_&|B%1F~ ,6@v ^&Գi~ek6p3} Zm\:dOQB:蔼Xj׬yE@hvdoeG׳L^d/s=8 B¤-dHEk=Y?@7Wc^8@ۢTy伜 viaH%UbL{l's9~v۩TZjUUW'ru]G(cPQxD2ִQFo&as>İBV+yƬk6Ǎej"kq[`N0,pC` (kql*\Yл{O){jyR^z@Z^<מJnoj~nYGek,, 5YWRNT{;?Vɒ%dvְDw燒 ѫç&j\m,j[B)4+͸ }!2SNy`ڞLP{CY($wNGP^ * ۼL(LԒ2B k]8ł.\d]ːy4R㑔iy_ %HƟu;쒉DT vG1i-u T ;ä06;s0)0i΅2e~MHPotNKyFDwׇ !5Rߴ1ickw77ir&A\| ywת=eDZ앿B|4k&]kX\+b91?V#1h#@Sao! oo psRpA谉1bg0fig2m@jqZCoŲtCa\8>LΦHR|2BFvZtNhM_QvaЇ/3]0ß̐ZH&n$cˉ\2,_E+fx[-֑:np4ޑ]]g>Gj$xugo\5WZ3~~srTw oQ(Xܼ8CH WՒϱ|pU3 ,_sC.!-~7dxew+kfPQv,׏'D)'sx?i~~&OtUa:mʏ gU:0z uP-7+o.B9y#Dk43,NQXܢO$W 8Yoʙ!Ș uFɁYlfԠ6 ݭsNyν0M|섓617 *U1 QQO:ds^3;|2!O!K7'lF[X>>ja^z8,Ө2tc >cِc bT؇z:5 ڇLO~Uckm+O09Vj{sfށTǨ.vWHYBhu\j㾩%3%$e_ >iey18{i%^u?qrL8=cNۊʑb_>-X>C` \NcfS}۾˷+KjT p |46[vo71~ 3^[Z'r,{T ,] V, ÙZO7J,ݴA|4FҤ87#DfM$NQ@L.<0uQn7W87VGlk,ihxzHe'P8Qe3Ƭ;26Q# >>ޞݾ[=az,\ݺww]W_XUzwDF^3LM9Th';#Dpk 2{Cfy3{9oG!x0q~LMI+vTO#9.|-)҈TqJ#.Sie4b o7hh9lk"JqF&) FmѴmiT+׫7ww7FXgF}ѫtgCǻgmd{X]ܤc /?ئ㶷ը2JcY M} M ^ϠĆ^ΡeAz!PH}j6(T68}l;Uk’I NGϼi-ڠlA3I/g&:BXzX[#\n̄#y62T6ht`9bLszƵP3iz2Epmkޗ[7,RXe9Smqx (oXl4Z&P42/$S A[QtkWr5"q]hb:Ӄg=|p{W=ztSO T\q5'{-P~7`4SB^2p>1Rt*Our"뻴A{N\\ M97kԳv9߱ztUcW(-}\jع R̂bnv RsA :\Hu =\]NCRزL9d4B(&MnE{hGIwT9]S/^4J=C+يL5GȭbA~BʇV~|zVΫ`YM+?1PZH=BpIS ޼ em ;<"$GXa@INw?p\QKGW18e>ڡ6Q;D>/0v+Ol{8*5Wjpͻn2GJ(^6 Gf &;'<[XW9OX 䆉Wh7.|~i"Uf2*.|N3?>=2w1T9_<36i`)yCK8b0$ jjf5+{ r3U9?7jā}>4MJr%lH'Y!^sC*=7DO=ӹK_Y̨\8%yvwWFS%'s$u~I\?$e?G2Ed@7 q.0h›KJuu@;ԥ*d[ ywWëSTA:PUd;yTTH:T4@rqsGWKBZ8ΝQ3j^ 6bۼcײ;slZ6˨m/'m#-[DNͫmt/3/ztαEKvzH9StsSR}$懦mkCO׷w׾jzlw0{rLɍN}Or~Lf_H9gRrpO?mX-NՊQZ* AdʊHhf!Q /Z`E[洷wl(gywy_3%mH}), _̯a7ZqiR F{qNkYaN( At5v]4 v2,D4@>q1yT\XCoEE!ݲXzLdGN?ޞݾ[=Az,\Sݺww]W__|݇ !^=m0#gDc'uB 2;F&0'BcWiNn l \D-@2Gq#,\1@ 'gFA&6>BJY* 1O`c%I&6393R34(c7T١CQ(Rtb:Ӏ9 gvzK+9*3_ϫ#./;D7 3 -fۧHD޵m+2KUX}z,`0DR'c''[/jQ-Qm+ɁmTbHErx}]wd%~9kT:d`9->aAo;gF|[o*d'=|ƾ1>ou[=>=9Ř޷[Il+cOMh!$@ȋr=h2xtC٪Ǹ*Q@uH 3@魎b(|Y82C7U+i{aA)xx10e'31XvlyhgAg+?/ kmsYwomwTntn}N5SE߂pu؂>܉M 1EJgha}7oMܰinX7zZZkYJ]z N8ХvZCnȼR6 3[a.Bш169"@̶2Bj #gDM @凫կ?fIVxR?w_n[3hڲNh,%v+-/*LMT:@)PҒ]0!̃c2h5(mYjᕬ$VUvA0t3X"AG@J>|*|usÊU9{hSɓr<qqb<@DXYL(sBvF @$W2Ƣ(<)%=,i88EepL*C.l+w5Jꕿ6q(?IcsMZ}U""hEiX(ig 5;q\Z~y #>ߓx營=^~WwO}֣ٿ^^T#"ǯ Ȣ;58,#ǯqmyBǜiGJ7O;LLX jDc!gfe^4h?s\rǙ H5g56No=ٕU9nȯc#~^}o6Įx/UH#儁O !RѾP8_ˢhrݝz5bdTH˔e%M}Rӑ/?v1nPi =[Eܽ%?75o1toO7ID|L+94c՘+e^tR90 3r00ry99ګyHBA&.M22WzݼL9Mffd#ȌUOH>C޼Nbʣ߷v̉?/R5 .gispurb^d#kkjjopc<<Q<%|)+SK jcw5B l΁btDm-AJS[!9VZ%TNJ-1Bщzp#cXX2NjkZ8@Ī ^+As+|m6$/r4\ 胶v tUa2!0;cpLm:eMWΐ-J <08NƴXK28VKtknQRA\#3! '㩑NXZ2bz@zs .y%;-$nMvW}-Pc[ lW z6=ZCAJҚ lR(ִ'l{hLӶ}}c7ܴ)+Wm{(?B<(\,nPypҒ;#C2d5NrrhI(z}5 }/qR8q1>J\_77 k͛k_?G7?FLXw}^> Z(7ri'`NpvvNDh+5Y\Ϝ%J6SJ#(E6xOt`~Bk3벌iېuYƤ26?~󸻋1|8#r'*~<ftlNytr0R=~D%_̻&O2,Mwiv¹ kE#IcZ=Uʂ^eD~z`a 2\V p `~!l*+8]#)RO D"b/vnP~5UrTd`*_/jMkؿ1#ҧlzX[ZBr7$5=Clw9h=x7CZ&=vOC`ta2#tgǛǧ''s#`TF19$Bpm< X+–)麩xߊtWC@(R"M0capF1Kϊo<,BqUS$鈺 n]`c~Sc8 59&1R_~#sK'CCgd<<2O/l* ET2q?{LBoJ~,+#Ջ'a:yPVE:Cxx~q&n2}3?en@3{l-= jDs T`,VdrVCWAT ȽЮ8tFמBҲ+cC]{&W?|P$ȺB] $^t;&tkj5PU\ɺK@)+x:K DЖy KW]XwY_4ל)cUcAG+(KZ$Yc=N?C& \=x,a2# uΆu`_ 4'`oӀP@ˊW/W5fo[I"bvd{_㾗R#eRNc!\lAW޻kHc:Xdv[>^l,\ȦgszrN[ }!D X]Hkך_|9fm fСX' ?0\ 7LZ pf 3Qkʓ?G= ȀqXMu}Ct+2?G_d?GgQ [2> JU^l(:;x),N L6Ou? ʩ"B+mXF,$rO炅GVyx]2$:GᡮC7v2gM{f3|<fַC0lR{c6,*0B #%,$e<!U׾TP0U=))K>~6OzyLhdsq,Wb!,=0lCf1!rjY|juAS!i+Gؐ8;&Q[3nY8z$jLWs>}‡!j$4+G\3 M!)sޏ,Yē&` ƔM)DBgαN P?& nw19򍋜&,_Vz .ҧl⚱%|h#7֎ߏf|JT$/3s1`qPq=dҎ8z LoS<$W"zC輛yq %fČB9B`4(:^g%ip򁽜~I'G~y0>}$1 ZqB.NNr/{\(]Px_B #,)ʪrw Yc{Fty{"7D|IB_D4^近yc9b aFz6M]#rMy2#op 4;a  'v,Μ3xC"=j>ÀaLH6A.qCkpn*-TRׄkY&|ЙeFcd29UȼF>dy߰/ǝM-l7j9?p >2[; Ψlp#l}G62gamsGEf_D|j`H멊h\I>z%y!ݙj4ͳgGK*F.>_J,|si9NY'tC@(WQ02':A"%Ymv)WUKC>DCؙc矯/WKZ-+/G͵[jP4M?U{+}oo V9:Ykue%4k,0k֕!S.XlǹP5/ᯗWVFzy˪--?|7\U?>yj|wWaǷC%Z Ӧ1ZfI&%JN+prReB(`ReSNq&5㹉[JZN nʚ <-&KK%/ EP 1O:a+$w5'}[Mɪ,yv5,2u2ZĭKtVVJjC/pcq)`>jdjPruN.C3!+Z4BH܅PS)a+TmYBfXRŗ_6Ch%0mdDfa=Ythb;I3]54:/f kf@~0L+|0&#ȜC*r&]֤*.!(Q(EPzy %ZCƊe2/MZB~2zZ疠7~^G"2Ԏ*' !)*)U8R0( J!$HG~œ. 4 5;2HS;f*AFr۟7M@3wķyiX+zzE(@<P.% ArY1z,I |yϛDrADg"aHQ~zנ(Ŕ|6;5 4fŭ~ Uo rϵ&sl^lzL(q~5M2ˊN>?ۢ8]DtLs8 {uFLHĹKT@|imzBTH0 :M$ =Η!HSv@Ey©ϳM #IvltdVѽc@S[nHOiH8U.]r[z,iϷnl6.Gߎ.Z~JzZ)$Mc{=ʺoNw|=)6ke|i>?,LLo.6:{(^nnXsgՒBeŕ̬I?o: i'ZĚ$%1T|~{mb<͛dLRTb@ $*MB Cu@JHeDdHEPp -4s(HUd%"LpY, BDt1FcAOo2mdAOo"< dC)lyS/V7` 7b-7{s;􆛒yB`|;Fp@8G[kGh(!R%`r P#(A\ lުnÆ9;: 91WIu/|CSw9W嗭\ǏsL>tȠs1P2[~{a asC|C`|?nh7u7 Ǻ ~9yzpbmN?6b ~2o²˫S-xcsxP&I0<^K;'kH¤–=e/+ se&o-k9]'?30$l?- 7[**<|v|&PTp:W:Eu=G‡yùɢ(if2I2((- UHP\u'HruܛtVmi6]M:?/N SS1Mv*9\s_h /pI )E)Ag)(TϐgqsiR$mp/pѶ6r 33kBqtomg^Oz}{Zo~cF~#8xz L?46l&P8uNad<ۋ#FRMUd*I@s ݲzջL>~X>~Ǐ}E&^ a!^;`l%Wn9]BW} Oo{nֻsj-FUF0zo?8zs"+d8jaQ;?΄6@;fKHy(z쓿 LL86Km1"&ęd)$ Č L2/ R$9#Tf͔(UBGh[a{gp3ɸ%؛h-w|;r??2*'{(EȞ>xPsu@ ع\Z{M'{G=/法XVCla z?rv;[+]V`j1SzWIcE뀎VtU.h%ŵUnm (~ !*H.6ŇU3cTX*c*܊׬~cjYl-ԗRA U"oTPd%U*IuA[?C9t9LoCVlZրu^qf SR ak;6V؆e@`hXdl0;ռGիN p-o(.A(9-}m-6A;8Y6HgZjA>6ӟ֡rpZ"SH%!R2rET൏јrzɦ5bSVjsp'zO,0wf. 7dA2l5&Ǐ;+:vs5[of`$ Hí,& J)FBkqJNB ;J_ z0ds&k~Sᵛn֣/OKm8(,PˉS8!юqt wJB#,ݾ)=}z%2{us{IRKxZ=䔐ӡҝ Ch_A8׋ڣG5q50{f#1)ttځ?yǙU3NNRkiETK蝺U|Y.Fq=_B{ H9rס/)$Bh)gMQ|6ɰ?P֒cLGE&I&Clh >j~.)J efZ,3M7dCq]H%gޫ2t&ҾG=t lZ]Z4x Brhu䷻GcP&0eB$2/a!rB ƚ./< ]Vo[K`"P"(I5^z (HPʄ]RK(q=C s2&#_noO*y '-HT@!|㙦eܢ+wz菞h;=NI7Fx@z/iF4:m(v ёzKyl۳[S䳅W||[W$FtX.-\\]Vw9uΤO7FNG|x9Usߖ+-n>>+dQiH(=5zb(% =@ jnN!^&^ZewWŗb}m[9=ǒOxniZ{bNhFTQJzr?]Db$e$H,~XvP#QFRGt;E.[6 o~҂9vkwݵ;:]0-D82r%DiQХАا U> @7s`A:<MѨQNr'q"筏'M&A81\_qҕ?^>b`{$cT1*kwu| CVRB(&,AؐLQ!(e|d(S<PRk/,e+*%e+z#X\mC_O;;l i%ϕڂ>pr6<ʼ<}x{N2XfБ̓SPdi/ Wxrrve{QW%҉QuzcYNE(7GMx\iC4mQ~+X,>’(y<20X-4V$#TBHu];rQ A*@Q^zJXJz{VЭ=cx `|9ƾ'  $x2)AE0t.V"58JWۦ}"sEҀ@y:0&C4R'm%%BFdӢ p0d:6ND,"l 5K4 IaRЀ͸f8q0j.*JN|P 4aLHmhDJ"'ijEd[&#Hd*T rPDkLAY<1+(@i0d Ljzy% khV5efNn%L|{>}$QF=?#g\8&[pD!HkPw (gLz'5)n%5$̤* ݤ߇u%L ovqҦTR7S|f8(G6/80ǶX{]"F%Ba-bN=rB#4h=DWz4w'K rK}if*ifh殆S IZݜf\rڠ&Cgbֆ֕qkfsQŊhx-W u5RD::s7 r/A6PQՕ Ĝ?c!؇ (?:|鯏A*/V,Z3 Ρx+C_cEore֠YTL"TukcW*g>TH(Ŕ`a 5Tb!%Dr_X*`K ¦|Qn8BZ3K| e0c! @(&\{PRiR R1B+f[X[DBh)$E\kGxcFF ,ǁG[!ZP-,&odBGsԱn |3-C5[s.oQl2΅BD'}`0lZxNb-&ڗa@ [˸U P>7)y[Mq!{TgKcrBWjc0ʮbr`dȦXY-bƲ@Rr,<vA\bE,۔)Lk?ʠy|&znG&~d<~9Z2qA2q{c#%(PRU"l(la=53R aHC? )"4{XPW6y,7 C}t<&M'Nɂ\S# A0]ysh^'RB[ёkHq%S8G9ȗ1VTu6g wg}d0vNr ./s5Ol/7l%ݵyRpE@B2su#ǦwΩi;bUߚ;jlcKO*5<_ZOôOvBoک/]vxU/کR+Fc7P{VZ^m;Xk@v?UW&BK~)f˖A85\dlEmLVQ߽xڹ\ue|kvRJ&kILE.u#LN\".5ͣ.( !杍'QQ8`x8X9츄ɒ{,K$6.I IژJDu ŎْD.ض 7@#v?} }g+ft=\'v\c Yw-Ҏڲfi.^۹8u<ur҂zW?e+Vraw0f&Eh\sNJ'[Z P<]{'E0nDWۨaO6s9#ZĹ12PgܾLF86ri)mc:[F"q .kbw1u]6{u?.A炥_!x=KP妣Q*:qkiDtG`:sehsQa'j{-b| qa?G7q珖nNbPf` 4u.1 wY3܅?G_ 6'OǗ?$;:}2"2]}e\wP]C*).@~aW. EW@QͺZvp{v/k꯴tQ !uFo_g4^^;:{{Fo?~o*35Foh<6JW:ܸwEbe5{yw& )vwtf#\.'A8+9 I==:z2vVE )9+prמD3jCPqvx0GlMe}s#X?EPC#?"p܋_iÝYsVT؂e3_ Йta6U˷pә yuS2:{7qyt28pҍQ5*ul(t_O3tkR8{'_χe&=A8Lh܍Zq72,t.ɇ^ar DTQ y2P)v0vp96tB9n/+ݯdf Re88}{Wy{.0|EO3a0,rIC2G-5_ m=׳"qL:*laU0aj.ܖGɆZ踧=ژ u;N{-b#{!:`L/RHHhMZ־O⁧f< TVÿDSm(hc\O64x#ЎSǍWS?ϯvy`Wޟo`n}Z}+W'(n0]hyz.W֎{:ݵu?ѐ]VHt+LQ`y6bk_'S6MHγL=ЋĎ2F9 %yHy| E({~:+j/`Q!"pAF)*b3ƃfš(NOS.T/zVTh|hi5L{O!ܓ5%FKR9ὦp3X ,=&'C>3AMs!GܿLy6'B\Ab6:a楽L+)2\q)de!Xr8$Db'1@o/)M$J Fr!g"v=B9{F2-%M07Jp FQ鼉Tz#=!J8-$JhARVpp~ c&K*xJK.謱:b*#Bk.Qܤm %QM8I Jꦒ㣅Gwܟ(sM DmB&y- p4f|&kAZ>~MJh f)ŌHFK]Uڗc.E(E9L5+aY) QVK(.!*QY^ +TQx<ꦔ.b-ك%V zRJ7LW]GF޸h77T&1&WGP>D Sfb\!Lf4GtWݓ&6@b >BYw6ڝLNMRԧ3vvgkzp{c纒ߥ/r._qk=F ]c@ˊM~+?)/{3rf8.sw8gW?&uhxL$)4sFFW 8 o6 Z1 AK%J{ Ztl[ah*B 7iaFF3LM5ѴqID]gWyOpxrAKwշo }hMztNy)O~LS('P8B'_QƭY>]V}k_oJGЌKp iU! DZMJvSJ24Q֡qh]e gly +fXxZ!+Wb2>U,é9&=\Ku0|S p䑄 +nV $ys>:j"Sr"ӽ'0 4dsEPׄCcNQ?~)WXEYeP}fk m]0'g3ؚO~6>H*tAoKdt 3CK;pNk7/ڀܖ,uj5 u1N~[|r"o@lxDw&ڂ8=j-A\O] *T)Ih=ү;z[l*~dmMOc [[NyNb!1=1t x#t.Iln>&% u 9i2E^PǨJ]1 q Ɇ<=ə$bXN1.r}\??[N%r%ѥBqcAI+x AA #A#gmoI rx?=e/"ȶh ☠[jN"cxSZ J.cnK &%d!zoĩ{Œzx\S4a%84eV Et I1lx&-D!4^9OIF&34ֈ֨%2YJ$9uEaDAQsL@Q)A"D>8AVzrhQos||ZU0~{Ԁ^e#}vv^խ3E1mִv_P8mmLZAR im.ai &-e 70K㲈hmk`]Ч|%w'[qxܣ j:. rbȷ$U;YJZחv5LBATNTIqM)5X ,X Kޞ8ڐ4)*3b~㇫L}LXɽd ,(5@KOE,Xq4 qR/K6@%S>7*Oy9 $9@A&tAaoj8)J( rqB &5d&=AL!悈<>R()Sap b|Bkq - ujcTS_S *J㮊?">$+_AGr;>zςØ .NZ5~ ˧5NQߍ kOp H Y即,$E(@#_C&Db'1&qϐƢ] <0^p@K.+w{;}rW}˂MrZPmf󺥸; 3:kIg`ߒF|(v!"N!]bmD5.o)h `/o7}$.gzB tf8>>fњW<ٗ> gé͞vzd-V\:8Vа3$:f&pG ˂,,Y(RhqH<`%2&_!/ҔkqFG jfgU@KkX)G)ƅ{ lQO[8S[2GM=B⬒im }B>~N>G?{8>^1|{D{0Z_M&~$YUUD 1ܺ7(^&cRen-~?8˞<3Dr\) TMC ɸz3ETkp:ApTA^Qؖ~PBᖆo >7ѣ1S:'w!$EN2 DT^DJZ0GH0!F"!je,O!#d9 # 'dMwyx1۵y9] @T %p@iPHcpJQ R "H\L"*Xg a>b83;j9<_jE|ZiYtu/Q(܂w퍽UQ1^~'*1=>2'ym[fJ}{rAn ֹ O.ox܎l3O6DYVN(HfVXb D 9,m]=Jd%$T6J.zJ ܚPafQCPJE/*a8ȘNJIrzyOg,ʆ}ly%^/bȦ:䍒/o0ܴCkqL2|^M WΊ<ϣ䆦;d}3G0I3:'ò&@"">BЇhw ͥ佒j+(sR ۢGݹ"j41$YzbӿS+b9 2,KD6)+䱬^!":ǥs&UE[iRVSNpcSICVe挔5Q9sQ"Ǩr"ezS_'~ތl:݋Tt{2gQE 9~CB>?d<~>)8Hps^"sp _A w^5l1W IBC c`R()i"Nx/vOUk W\p8ªbA:QfxS LבDK[CBL5X{uɛ]>j3Sd!S8+elZ2ϫ!ɤy 2͜sm7uJ%h ="EU֖6dz޵q$BeO;rߪ/d{A8ټ6c!)bV )R̐Ð`eÙ뮮K6X*\ǡ 5ϟ=E £qI/#T#;>?mYQʨNOGeuBWz"Ѭ?)N'Cw'or UlwTvt|rw/N~z䮎9O[Ja݂۟ Cyԓy\GȥA+¿Xv tSIgʣՐPȂly>0Fi)٤-+%|BH8mDF΁>VD ÉB,RQ|bx-#U9Zd_>ۗvd%6t [!#@;zaq"e'`=\mQ J^=<~|z㳍X'Iǻ2^NM4sK">$ltşl6s4Y)r>}~.=(? u)- 駸H@{AJYJtm8{si"*/݁Κ x^o^Wog?~7н";{s^7/ŧHg5%YϿE9\cŹ0?hvq%7og70 (9gT$5p0FԺ9vc|1FfzA -鑇4\w}=P/W9QֺLZUwE+=-]5 "%w\UnPNt9J Ƶ.:DvyjyKlXmmFp}Hm`jL~fŨFY* oxF!'6Q3ZU3ư-جi`.ݶjEj?G;y{u޻m7Y,&-e=.|49WxɺAmO;e8^mlh^sV9ssl7R^|@eF^i;u#xQj[4<qKWz<__>Om6zT-4hn P`K)֡`.dX{%\>Q-l;QJDgDKa} )I "$K:ʥr{M1t.*DlbAx x28B >Nقpjƣk08P[xraxz}2؄afi5[4mnr LwFK[dJ:CAd7E: i:;9Uȩ**'T+*،WU#Ĩk;|X7ƫ1}Uu8_P M Y<qAQ j\wx/AtRsFD_=)VT^^ٱ4?3;7C'/K= y\+A{O}Ux/cL7ѦNWq<(Jx ļ(r8ΚQQ_F.^~/6Tj%~ ~=G2z2 '5B̚ՍKNy'A|̊cvI˗Zt$ͭ }3Y WLL$JAl(pfBxqT4%Ulx$+sVWYٹ3 ԧS()u fiGU{,.*~D>DDM:i 3m\*\ƃ"BHym\*!q=j-|ag.ŞZrebތ|ZiYM!]Ǩҩ $|";jg]9(5gӫ,ktO De.QqKT\wwuM:ץBuUPt\}oyr&@/Z'1N:hO3hR:;×͇$NP±tNЬ1guZ/QY*ж40K{55p:G#=7<^5@(l8%teE>9!@)HRq{_'A\*n;“Ս9ZtJ-LAWO@jAUjr"F(>e{^>>PnL \-UCRPYHPˬEBh8e㣚>Q|TӇx{OYMpꬾ ~28+\BG M[CJaq2X/ch(M8v;fEhсhhcT4PAZQ#OSQFR|Oz ^YaA!BL8 G)CNbCʉQXh)1nRM\VhKI'JR<&E KIWLk"AZmr7FZuJs\:`UQ(Ee)'*8PR sVGPKa 05RCf1FGWL Π-RvK`Ur֡"%0md@]SXrM Ϝ4D ,%hVPc _͠M^m+c:)%J <ɠۇV9N ,uἏ1:1nvP.Z]kɯwp"E61\ c#nS=̲^nUk=\tQaSyDFTuȦr&r :ڣFBK]I7%+\٠4J|P^攡/ ʨVDd:H߼f4sh.⸅+6D5W3BZ޽(c{g?"> ' ŠQנ%90ߜ p21v4?9gH WDs*'NgO7A \K3⬑ 5ܟncWغe )q/ߏrtaF-~;xHQ|)6Eh7_Jnd*#RR[UH4ZSX҂mQmLP% *xX&(4#npj9!\խDE>pgYS5ңyAdhBQTrGef:R Ђ谫JhL}<|PdԐkQ'f̡ki΢#Exw?e6T8Xc>:4;V\KQy )z0yM6naj{_ؓw}Ir!AO :v⑽<$˲-RKq2R7"땬ZwWRFDџZg`Mdm s$ G7s"%85vKn2rp); )Bąt' %0KKcČvC#kSA'G ťGj Zk8s!G,DxYVQY]ނdE`++QWBU#9#Kjд:qrAiߖ7ox3"g+P(_1ak+ErkҳV3Yǚ#-;jρJ6C4',PSq9#(O*֬;HX,?UG 4McXe-(猘ƒb 5m$G)|fdUɍф[δ?Z~Ć-@ cT AŚsPK"VA*a3Fe\@p԰мRgI45:u3rOGwwa0٧GIՒzEa"\ r%5 k,:K +0fR&:dH;D =-bVTtg\4uF-sjf"F`,2'n9%(d:Q\> Q $qD 6l-Df=%q4Zb>5hʠ,%@j`t+057Qrb* hfr \FgJy IY8~6e~ %Z8ժ@_A٩u_a؝2. t؜&3J1MωQ5J5t'(ocDq2ŵrwJHgcѦ}BMGUˢGח/ ~ZTǷ..Zo zhuxgXڽ .֗si-c c.}gUa\5F+lgm޹%3D :%ܯ_nQ¦g@a7u%rSԵcjrM8P6&qv${]8\ߗM>4*fP vs>Uf*Nk21ji5eiu#F{!|EBx%N1*?C2@%a 4 "67LU JIOhІ=yOuyEuc=T*IǵH9՟K~ Cd1&q|tx,v063!&ԧ1Q9\6M\jq 2}u&}O\ 8{0S]L_oCk(z Y||zJ Z) o瞈Ef[Oģ6spn& ĉǬsvNK["7_T;tgS;6a]r`fV.uջmۄ41fJ#JPJ%/J7ދǚrE]g4 d>Au4x 2[G_uS~*`W|C4Vubf0}e3;rZ4cU^ݐ A3Oo|@PoV;S"P·oV PXew1(5+y){4Lw 8׼ {ZpU5"]mYC(Y¢Mi qys҃;#x3dz?["W݇ J pJ/^هQn:JAh\4q JJ0~s_Gw13EC:ؤvE3ϳ{6|;%m i'`Foipzm reKԞ!h*x zT)/.8EIk-cP SmT~NI\9e.;唳UN'ˍ*_ Ri*C|6].E KnY.G &)B\ 9X bHLلB CZ V `&j Aq.9gfjtg6G⿜g2Yn?]Oռ7:<-#5}=֖=ZNU\wUаs$:f! 7 0kjY}yYM6v.Q閣٪aMueRn\zIAÞ'%@k'ÙY!w&af{|˵?~W``˳WL^=uLzSJ Zb^8c:]zt{Yɔbϡ܊s9Gbg?a[x\|FpAy&1"Zc1 sy>^"Wi:^r>mn?~%<"\[f99L|"ss/X&2!'gmZ5O?6VB/pG>.;2%Y%"%0/lZ<LN;(ܳڟ7JBD5Qtr^Gò|K#KA`:mQy *Bb9M"gQ/褣34zNM PS4pEA':"h4o1ܚmzęmrWSH*$Q4 LE1FZ*PlJ봤Hb(y" %9:&8 {i>HnQ9/~˘"y0S KAY%i:*nwY3kC&T6諤ܤĵ)OAuP U^nY.5HAs5vqe>X\׏{iH6o*գ8ۭx Ӏ|F$3߱}lop~wgjLx_ o.. ψ{n|= ~|LL`mYJG/k5; g\X1RRrFu% ywuC=- =Tam ʫ*Q8w@7YKDMx4*>c [ !#Y74;r'7G\8j"T":O&PA\rb* hf17Z 'W^BTJ#6Ve!gu#,CmG@u;6I2ZjR `_x m.1k1^=4>k-mSwʢ;oiv ).š h&hRӽd 7"/gT ͨւjF }:Hu$U!P$Wk31cZ2ӁuڮMLnudhq$ &zF' zO$F4L&|ql!:&= (8q~AԴo]j28'DJ2Tr9=|>۝rk妐rsnHB0B* 81DD1uA;P jc2"9\~> S]uv+v:2]OMjZj?5TpI 7~9%2?yӆ˒1V6\]\_wl-Bx91vǩ_z^a:2t2Ȃ)By[ϘA"`@ !|4Dleړ^ ᇉ$9){jQvbg]_]FWrCp`b쇭FR$wyP#Q"$EGC5U]]U]uU7랽 G߰kW^g.ײYKZzW%V3@Z\DSdoڍ.[R%bPKtH ڟvKhmւ|"zL ԡnH}7qSBƂp*YMDz{!#-5 5nGek="%Y\~»|pv=ޣ b6:j܂] :շ̍$z=e14@3@(J`mjT [XymhZn%>Dy OK֫ qpU\x뾀[I{0g=*ǞZSGqGgqbz,/ԛX N]۶vq1L쎇G{_;h9#0/ws {rpkX?rMFwj=6GO:wq2.{qIT i=;'g;MҐOb-wш['$Z)fxoQcDSbr|#=Leyldo+KcZ(ɨ̟G9t -X7yPjƶp֢]$ ]K iyr&,Q< LS$3Y1-m <kNSk0!,v0ʔ'68֤Q:[XRA6O'[@l~<`Rٚ<(0R`&y&c`rzKEb@K)OT|7OC0i(iqF=,Ё'e6xQ,&L,U o/P1z@nb g|U\t17{#^ r)|1caGD0SH_B/:PѬ"zы_~PZqI\}Z)^X/*FFݨOިfC-YT~1by1W~X3ʛ_'|X!,< Gr0@3np@;OCTtΫS0M"u0si$XbZY47!]ƌķY :CĿMS#NY>/] Ѷޤnͼ޺x>/ޣ/?ۖy83`qZj_Y ɾyxs8fUxVקŧ]ӗ3ETVGoqTp?eRn&}n u:O|OhP+*q>La?xq9ׯi=TF FZ9\0[?M@XR XT SwD]*p#/Q%wjmW\}.{r uWťqOwpEteo^b+փs G#>\{njC5EQuJFr|oEtx_t{)Re/6UXsԥ_k/@1M%R#[*NėvN(o0 PGfbMX/;D%Ž )CƐsY,CI\r_Yp3>^"RA6|S6`JʆҴ\k 0?p0<1>ﵵ8P3 q)5#{2xP[Ƌz8̚ .>|J54O\IH1\Δ2v.z uƇB"n؀)d0Y/'Ю6+x&.䵃wwk={afR/CMLbzRn"mE%[eCo-"6 };Tn>-柊y?8-ih|z,^MWw_S&cmOV1B 粜[=J)$_ƶR౛S0@AANjb]7(-ȩ HoI@:o6\ C[ꌠJPPkDͳ m‚j.rbsJ) JLFZHC3R暿 +& ֽ>E-@)Z )%JbJ+(eZZRU-1wl (;%6ϾL Mn}/gi7@[30c33-h\"sGi=Xo~mz߿]z?f9lz>M??M(R7Qn"u?H,K]K oނ[T&o by4Td'UTU GRJg,:RnL%ُ (:s̃nl+y<`:-0l~aXDq[ Q3 cw"")3Z/Z*"Yi 9"*M)CJRBQXVg>ޒsyM9XR+RÄG@s`X[l&#ÕtO6xT=nID Ap#0P,EvCJ.9m40 !@$QAg(O^K1pRjl^ǯR<& NA`BS#6s-E'qLPTR6+cx 9ގs*[ lF Jm^qǠ ePTE&wdldY}!VDeLf)5G)Uv)vOʇ;=ul }=OWtÔ]/_8](&_Կ=S;8[IĆֳby?'Pë&/_/{yLϼv e]ך+$)Ja=-8mIFd >jEfr,FZKjyd&$2Q!94b ۬k4aZ>d&ruc$<H#G :fAjIB:aR)qK8) $Tʲ$ TIltȹَ'7͠P͠Rt#QOܸ>=E}TAjmUŻY>sm:''b_J6v4([w$7E--d_j3Z *6 ,=o3ETUbj5#~;#UZqul~~k& 砺SL[8qZ;>.OI9pl1H!MPTzAb3qQE:Bv~3QZK'aUaHC2HL  Pғ4$Ȳ;@'AB4!NjH3YRa+of_CWZ@ٻ6r$ەÀX$f6A23edneɧGfWlVV?NI,_b,{݂3_h\?Fn^w=C[oW3C*Sn=$^Fm4keb2O5ͼTĽ,{u^=v|I!8rP^n:wqOQB;O{yiACۋ2:M [4B)yOVa;k[#},Pˠ*Hdǣh击N_߂JV yepqlj K~(ZQh(&GcҬuꂴA;vCו޲Ź18DL(V<`% Jȹ}ݕTּm0Xvqp7f[I ː3a\Dǥ$ 4N u$!ӭ)9{8L`arxƙ6'&Q#NCtl+" p{WvCKZRGZRˮۿJy-Z֓@˭Uֹ  e9WLW!Bd{Wl[gn2ڿ";6UoepW %r__+&l/kkX ǼMOl$8­45 gh$"Ukb(7m9L4RtK8l?JiD)z'oO_R"O'Xc\ONmhliQd󋽍 IsnU,sR|8X;糇9|-ZN0$X^X#Z -bm,JyzRO$̧2jy%RDג#"MVLxk0Ҙ#W? 9Y&QHs"D1tҭE1Q1C)zTdN&]}IN3vg-4(ٗ-OGI L@sLK!RewTISRa(Fr =^x{̚diYN.7EҔV5U4YV YN" Sh~ohrE{7#s 'e$05-OVexɫeC u>n㽿_/4/a6ntGIbF_c`i`EV/JT ڣ~# 93ε& BA\'!Y5B[\%u h οns5ϲv*V?>|)r'+c>u*,)9=j)gnե5<#ij#pz|)ehля *)Oy,3|; /Lc+.>VR`4E**:gte#t2"pm d#oE`= P3+e2/}>*Jkև\UJU1T;23ͯ~2_ed[4L#R}) K36s[`l7E YmR޻{M֊ uCy3hU lʣ>LzBZV8H9;+\q + ݈׳QkU(kNN|GҸ3\Eˁ͵Fsquݧk}F(#NMG$~b#MK%I|3%vwJd1jYąn'l?II}ۻ-#E%2 I+YW$TxGE 466"_}U g\Ն*k+Ԇ2̏X+P|mFq߼aD47oEMzhx1n-TSr>uɛ / cTrCP?Uck3qD b㦔Κ5w(u)M ДN_`e7Gݖ`kԵW `s.[:Y>ΨzSSްo*q2^hZgak&v9Ҋ`3k臻_Đ  `x1 (Er+\'pWwKk%zvz*_zO*Caّgl%̴BVYU\!#B}TxEeEܹF\ȅO$r:8kإfjǨ}jǨ},a7` rL{n gYֈx 9#bExQf32{K̰ s!Yyc}Ƃ#y9伽]yb 'YvdUiA+eSUڪo[-HڰJM4/bDUSO}ņÙÁAb_̵~)T**PXƳ60;Zs*Oxܛd11ǭTm>G 0pu/f醓D#uRs8\WTb'䧀VIloS͜i7svc޴C]> /JR>tQInCDtY))\Vf SVRpAA =c1@ٛݗ$4â-K<s 4whn0Z|g-ajshY=p.D7Ğ:vǽݺu&GP\©j^^h:#wp $e!~p|p[M-f\t:^n5l}[<9]bf/v4m;hh!XGw^"c:^C,O\_$[l͠;F,?AOky[3Vm脐Co( KI(o$m&*crVsfF1NQҚQL$IƼxD<%A"x=&`bL[ ,Y2#b,89%V9u_ ;Z̅zȕA 1T2ժyn*tf*WlLg&cW^+:Ę3MD#a:j:%f,Bg S,u`gV8ʥ7955pKb8𑨂JQRĝ(ƙr9c@a(D0rxb EF`_0 K0h.,qZxTH`iQA~p\<0ǯo=uX88N{\+7NEuNj9LagȄF(e%aH?`iJHfR;2YUai(f1aޝ!e`[Z%²a_D3V<2bi7Ohg6p4RzCpr &?ՙC@zaP4U83g8oA (2BJLd$`%Si' bBYG׽gNUb7\&$HNΰaW81.aPMŠ:ڀNjѰϳ,8aF~>G}Р*ݿg{c$uYJ3:x TO8,S04G8H}CQÊg!2dCUP@$>fǛ̏ܗ t;KqٺWyGT)2 >o/ gF%Eϯ.@pOv>Fw7m?!tO}f|4-ږ)ٻ+Ƅ^b-XI#/ Z>Wԯ׬]Z-)rG̔G .i7\2ˢ%3Ҧ$^aa7]s.NysqQBw/,! Ca ]D:޹$E7uiw4Fn`> ZAYZ7h逗fwY:%5瀷{u @Nq ֭sq7ā(_1@/<HsHKF[!@M5S0NZbݽ~8D96@ЂKVs'L9.d!8ݟ.>NgZRTJcL0WwTrMk_-N!q lzw|p:H*>3*%gX e}d + ]x +,ɨ!qi@;/3.ɄW֯YZ1)CdN\Sް$DsDOZK"^ap>*~Q\t:_0 "[}+8HqꍧN#3ĕL.eFJ=W"Ebaxd+!̲mWDy 6=^j.'nS t{d q a~Q9$'pf8Me(/O^b5B43*M<or nCV\¼B]Evnaܸ_i5]ɖ=y-A گ[oXvQ@vFi|n5F0#9AA6B=7ZM=,%|s,ϠmkƐH0L;-ٸwwN@FLk Hp;c=q"RR$Nj0Zl0뮴[пdmIH^8dmؘ/aކP0C(hWMS#kMϢ@xǐ!_ i5퐑9\l]PVor0nB7zjp WDRǼQút` W-9R&=U{wO/nP L:Es{Wn1:(dHcnۃh!Cσ< rg$`YcLw{x^݁CxXB}zn ƫP%%NzE\UYltՙJxfp;8b󙬱zs2J2ˡoWrE#M Y؇fbn ܅,dBۥ'@TCa3̽wEnztObd$z>l+EVwFz6Q P)}aK)r Ji\]>pqQmU5>L>h Z*pRiVcب+6tx ׼SѦCP = `Tn Q<2v`[dhI$7H<}.?ßtNIRu˾ A吺Ө+DXmܕP!l k5Ʌ"qSTۡoY:^pIhig?}C,m))!ώ0Np(`6|(VA;AKC!FQ$Q,#7SIdK| XE1oYE}P 5(죟n$O7p?O-hޟ ;wdY&ZTII|}=Dg$FxgdA?OmnӢ6+J%:WL,ɅN|W8oEi!;I^Ҏxh%`ܕGL)e]G*2ɮ\we.R"W[;}y BvCM%Kہ6(@k+29G!L")2 Sq  -E|j<\Y(}5VX3վ5Hi^Vc+OB#$@H[:.|6G #LG]D*@Dxx 7?b!lz?'2^b>^^)zMx*D°f*MPj2F>Kq,ɝqU+aXsk¿`9 $OPi|wR$> ^(uwPrJ831gTB [$%sF Vf]rJ̕XuIvk=^0)@hy+xh|/ ΀,8+"ܮZ lBҝrBpŪcje y%qi23@]F;was7;w;5},ÉjQN\#%KEJ`,?G_Hjs{AFY:&/\yF"_֭npLrQI`{0ޗزQӄ4G2Ś[(+ba;iAaRĥ#] ff6wfkFM_t3zߩ8 %!D<΂um0BksJ/4^|/o D7Ⴌ_qb>U K$ΚzZe0vy%R GI=iRlZ ;cGN,RԁWTҔ) r👎A o08!vB1v:`ڨp3B2=2x1X.bW*KdkǤq1ZnꗛhYi-ڍyJ>& XFsj}2Z>=֘Gs0$L(V!o^?}~Uz`7 .|2g "Ibא~|Tǻ;/S-_'LH8A]jBrL=WO*k4qWT(%n2t$4I5vMJ.cXy)ӂj.5eRjn4^iz'l;EBjn%[ífB3,~d-227W%K=!rR#f 8.8K0ΓxngdW5)# S)GJu Bk:RpGA6EG&H%ΏQlYo|1$eRErOOZ Q5kK襉5#(,r)~Xg1'[|\ 1N`zͰPt"a'{7e@,J;,n{!8&*U.w!^&bHJO6MC6AWW^Ǭ84}ܺc*~{31w,qօbrJgJ!b%M*=^io${#y%;WGO#F+ENQTs.4V+CՐD%|pL?wHSqLk{*D俋cىVR v(B6K ĥjЈ/{ sI9{0&0{Ӹx#TR۔lc71k H4q1[ Z1Xezf(Jk278 V6Y[-=_®9 _~Zjd*SujG1)MRO20֠n<&ab.Ҷ M 4f NƹaenBb;1 A^cS vV{2B0JNR1q̈́_HXE4$9՞ÌZ̋%lTpB?25j Xэ 8NQO, R@j/P[;ε7[#nbE]-6b*eR $ MNپ) ifsՃm;'ӛ/(U,.EF1 *SV"k ;A3(ce q$ԠHyUtmޯ5YQv60o!1;Q́&( 'nt -Gcw{L9ˌ#Llh4t1c;2 Ehd51볔)E.UXX䝤i+<>aM1g5$a^^Ym9JbrÝB9eSS"iPy/fi73'Ϗ?]n:/_;`& D$T}N2ɯ!/4G\{aIBfJhkm||[k%ݎJu(DB,QFY$M|aZG1 Y 3F$R]U9vijxZhfv___==V8oĜZkA yZbs%N1liO* tŚ]xAxz`\u>1ʦpY`կ6@vZeݬFSt$J k9(]( LuT٦Ra.$Kui^'^kBwWDZq\etPUkxz[GvŶdoVLj-.PegzޟmO͊n][o{7JR/zRZS[](B{#j-nPe?'Bp3Vɣl Կy0`Mku_0kޣY~+2؃5^v] 0zPA=!)a=zUeKՊ!wyW*{Ґ!v.D@<љM7@5ULuodzOZHEǗ (OY0fe1#d* _\ 97~,QgYβDe:K|7cR0ż0#ՒGɈtpRM O~4=~1ߓނXu"|I̿o^<|˷%.BɗA@ɑɒTJ1NP'sT@$0A$2ev @#,""*A!#xbBZIWJeCZ]ZIlf離Qfо6o߳?8ɕ[ڇN=ٻw|wy݂y7{L1([P*yexFd~%x6^#jSBU.袻k9 +n94{z^rU-zЌZ(˯ Fz+C3oa> a>1j1oj-XHa=4NI LF gc܀y:4a%nM{L@4n$ĞHp-a91-Ҹ1`jTn{K Au-M(dpi5@TDKW<x$>8Nr+ح`S1j*h҄ZN :2xtL*IQ&A؈|TZ= @-v~==Ռjcu$r04XL"+$J3L2pH\eF74D,zc8!Ƀy_;Vhi3/{ZGweuQT%;9>Np>FW\jqOksla?OAkPB)8]J>]d2$\=:N*Qp8Qxί{伬9j~W3U}v]3֝Z̵}=˵®gy14\.?O?= 3S+~zD/}x~Dfjfh#}:4 .NV@zB.p/ߟLL8DjEs:Շ"ȩA$L.ÏܯM]@-F'N! ڿ=^_XyaM!ߨ䩔(B:ewT<3)vH %89L̒ebSEdKnaK-?]-lAt:=IQ-0õJQv8PR*v@P. jMO!sKtNsJyӕr%@FH>eZ+"a#7ubʑT^VVo)Itd5}|M*P#pO^47OýL\T$$1fHoBZkqHOe9e U` q*%Iibi>X7AE? 60E N3u 5yTT>GWZžS$Ԓ{ьytl0" o5+c3/ uwcv{\s3pqzv{9WKD"!kAXw6]UXereoItLW`Sxb@U@UVIVڰu+/9bZC'*SZxfCeZxy 5՗igK5xA7(ߚplDγ!eYτ>ކ\yi :m\ 2O}ͳI;PNۖ#=Rp#?ڢ-XAt HȖ S2zmː5ЏsEy"㐙]hI 7B!=gv.Cy&/0I"a*õ88 j@ug86Oڋx!*KN(+ A}Sl25NLq fh:?߼W^g25ȡC N;hΗD%3>>Á QJIJTbWZco+}djwM㐙D}h;p Qri㥰l\+CL2XN ]BxnM>Vg-$(uMa:;P J)c#yG_#۟o1|fw9}|^)*,GF]Gbu@Is4mR95̼h8A\(`щW.UxyDa%*5hKyQ+ gbgK%l9:U R^kȡU1Uv8 9*CxIM05FCmri)* JU\xQikxX:Ҳ`qטW2&_8S>zwGjO/{k(iN ^t& z]mJ#U^lK d %vϭ Ħf;꺉SFusO<\!D3ZMz`S^WFb•Ss$RN pfH-\ cށzߖ ┯U) ɏ}~BϬ<>%{a2vӾ _\ T|,_9||#ٺ(Q(LP6 J"F '$mA, pȢ8䷼OSß'n_X,uO=}xMMt6`y9yk=_Γo-Hz?˩pԈL$ h'xRJ"`씶<T41.)WG` kCeP]Z@irR.gM iJHk{\DMHUVSduRQMC5%gh6cKgG E2'FPH4Kr f Xm 0 >R8ƎFnP@Dd/bY $uDmaNu-}_:@[gH ܬ@y:YT&?ؐng ep4$7#V˦scàH0M%:.\wae]X""H% BgSOKMÂ`fHBA W]WϹ#vz/t9_DZrhhJ:jrG`?z&/=$MkHvB"GRL$ DГ;O%<$ wx"N ZAfIjF ,r aĠvњ"& P@]QD= `&IUP.4exq0E59 Pp˴8^P\SGqS#7 ftpo_I$8rЭA)hh\qSܗ;SD"rS{C&c׬?&"|';pHQ+z9rѹ:ҹ0 ͭ>\k~zt.6G_P陣14G2yⷘ4/+ r8ʓsm'~e-^뜋z 2e@Pr` ӦFwqpikBH32-D0tOpb%RtN*EjP] (JXsVJ/()j7@PA pSAm@$8c"r-4JPQhAe<1! :J <\O=.)!>]}?.FLfXg?.MgDF96AEgvz=$8h=3QpL3ً+Jt[4u{1uF-|0ͣFgH+7 ,ADgUgQ$76S`mk #ǯ(zb{7x:҄. f}`=`2[v(p(C) B"K}_1m#%ܒx/rɢMHv>$;d$Wo>44qf - L:-%Qp^8t%L~@7[-tn/۵N[8[UnP)s8P12hT&Ã+ hs*R@TDD#9t}lC&ݒ=zx ɚWrY\lo)PtBsN@ר]9K\'5ohBpBGWբ޽]V-6Jw̡U" 1J\1)Qx#ifZ۸җݽ~JRwk뒋+SR. 27#)' )j("f03J* FP XBΡZa@0="(5,q3%8ja; -0neP ^S  U,Pլgr{sںH$Z64W:E`To:_؜̿yy~~mNDHGEY [*bݔivzRyaYO◅IYo33)*j*[87z{0-f>_Rje(v.K!JeH*=⪤9?鳄#N2GmDC!Q mN*[ٗ(g{ %C\_q'geL*) O[[tbNZB+q'"Jk* o;K*ԯs))%(U6'!NlY^vHQ[g鸃,q(`FHDGBTRcGEiwHՃ\g@>+'c[,&YLzA,ŀ*%w 1eyV) UC6IBTtWv|+tch>d +hqo4ɶzj)md'5~Ἥ*C豊as8uOFKU kIrYxS'6֐Lp=$ R?uTaB} Bu% ]g4]v{h[SBEHDRs;JkDe~SU;DMƮSLh\^]?A&'ie%:u{|r_iػCAϾT|sXgʬӺiJ%x`rS!"FX 6ȗxRQD9pZj!Z *Xl 38f~sn ra8GY8c{KRb1vsxp<]x,nN5cg{BܿfǏ9Ac5wV9`-ye;eԠS8H|?Jno !0&NƴR)31HG9FtG1;_SSf!߸fGXjY7[.);F֣Lw5Qbٛuf4պ!߸6)a/%{99xZoc6#z{J}u~Rv80>*/Қᄌ9!3Wۂk3MC{ t}HssRIP9"THI~jݻ0tr 3o^ׅML-pr%;Hſ1 KW=9LK%^[G5-nEN_1y ^`a$1"`UCN!5r},GI"iUJX exb+jUR0p!'R#7(>Uu~7O>_^ 3e0<Azړ?9bRJi"l a(&5)"'FGyE2CZJaPK %yi\^ 3e0.+2N4RLDm~r,Uާ~2N!Q96:mf0'exeݞdbRJym~i,UѿD%Nbjg/EH/z^/%;NׇBrF}Sp84yYh>Gqc`tbg(Iq|y," '`ջv6>n G1OyŇ00K]YhmD}岄@g3ˋ1)EJsLOKY)=3#^'*J9;x aQ 3Ý@^Y#QSY9 *KV.<!Dy(=,*IX[xƊ= 3:4nɤPs2h\;i' K:zLz$#蘟 8Qy3ɜ9s0,[֊}]:tB{L'p`lJ;hiq0輎Qbc G slO.z1ܩNdsŜ4PiYҦm+qJ=J~0AQA 'ba\R57 `σFL`$Ba4K9Yv(tYj{ oIie38(f: -)+\B*n6I%$X*A`X~&_`zg#8D3ݮ0z]5E|Jt xLawXK,?JI(Fhf#Ōa܃U҃["`XrpB鳹bN%^Sh'=L9qO\%EF5b%i0]WDh 6K%85 tv[)|G6P$_] \X8,G-/C^~>-Y',f~t{-._>-WsWJjuɛS f2Zh2i䍷7cr<^]N6r2(/xHݧ{]5=U| DSEuÅgwxuy5?d2 9C49 hn3eed#&ݬҊrXRܢu.\pZ6l_vk˲VX7xgCTj*%BQ n8(1*`Zzٟ(5fZ˝l>LʄgLنܕ}֌f`KG0x{  [9V̿[~)j, {xBVzŏpzQizw)Ha,D VG[5nYYEDDbJ,1ZP(KèeL^f*sê)CĻ@K (ltH$xMaTsp2 6DLyDx΅!߂s*g`pĝߑq7mV`HUpE#~~{+"?ѺݭG%>?Bߒo~$Xcgx%O?\O/π+Q RT5 ??>m6_|@fã˿a,`*D&fz|;~NBƗS7}Zy]C 10jM[qcr@PXdsIhdwo+3Ul|mb}oN)X(eN6*M◅N4f˩}p>ӂwظGr`W.ce}~ؿ|wÎ㫛Xxl KT|ZWZXoDe KU^m %VW=Tӛ^eGUʤ6O CW.GRde%K3kmLoCQZ 8w]qa٪jbt?xJO3A]ժ9QkNʮ`6 7!nDrÄ<ɀ@[h v*<dpxDd2TTHe6Gڢ k'H,DVJܞ^d=^u$nl$Q(ZkU3 pDy7T_GcTHU|_?[*0ߖ1xl/rLtF򋻍6\ED7p77 AfRs\UO?{o=@Dl,Ae)`|oz8|7y |*A|l6\ g1FP\"-*Ƹ~J49 4!XB^KX\aMQc/g_uJd&vT|RQH96 KS#M V{ñ#/) PGxqZ# &"3{|5 㨲H,2:1_|O2'RWu_z"DEb jvDJ:=$bp p|tpu'isI(\bk,c&EbTK# +"f^o3 p|Pb͗dᛷ߿=.Wf%)՝PX^g%ş<ΪteA?ffRKى)ʂY?aKd~ZDqv1c(-b]M9tWTS\Yz _6<ԩ 71-iA%z @0r-քUyjB$W5@va}LW#u;P+A:R. sgBjt0_O_ 7]@Q/s#eApʀyF$1huwƮm;dhjvm $Tw.$aZ1 ׽7ĖJKG.[ ?XWN۵+ '5>TQa觪o8>^E0#YթRu/zҪ|ȳgz#_i%E,~Jg${H;LL._ͦZV1iZ,}U,Y>5ΖDųٱ>FGCcybp4Nv|N֖@о Vr 0i4oc7Nه?7]\e-.vYJbTsn٦Շ`盛ӣLE#;Wp{|A*뇅<.q? y}῿C?dmV{bwyqV]^I4Lc y€In1XZl+g+4rAm_ĶZesbW=Gs!w?[%zڰx's9,(6Nal_ 8XyPKݗ6y&]+xJ5\Gg:6?mՃ*Z'a-[بݶ]yU m-g}<^X_ͫ X)=<9l U-}c̆Nlѣͱ` IJ6UiH3 tҹ4`TydJ*+K@$UAVYk󕬬t,B\F|>Qp}uL>N>|? Xaݧe~wQ.%ݏ,-wHh^J%<6`\* SAQyʽB T GI3\ȢXWG1HQ|$.|+/@¡+&~T1JPP_YB;ꉃhDUZXB`|XJV^% UBE~* lXS.eNDRx]L_J-G*ʒ-IΔ" } ,!8s@VЪGS( 6S ߧG1_5kNRldzr-ln% ^&fHK[c$-Ϫc2h-U]>=ot8<"(;iХ;[UA =4 1 kR ZO[FhZj9iiA4>J(pO3] Ċ Z. ifs-ϟt;J\U& p2ʭL 6yaNTSY5D)Hbs~vX.hz@'H- )%hrI߆(`9ojf,foIҝ'MV cq' ZI| ?-( Y;{ yMtA;#Ls%n^445ՌZ{s Y+5|C6 [$hó/>:0[_JjB6q2fm IKNh 8 ŢYr`Zw m*4dE#E t#fLg$5 p~0rGG?h٦iDUjHKqi}Pg{ƥӈ`)uUHԓK ,x:-,*Gai8ɹ,YB@b.kД^܆v8O##5/OrϼCGť%1g*P<[bߖ+m~H "{mTlx70 Vj}c 2w^}pAkؔG)c\#EI׊x8t{BI /*Ci0F tk /kؑrhH xodžT8hiz$7$ww"y{j00ƈ` ҲF>]V+?Z8˸NrO4vbiEz?q#yg72Q!Q^h*dI%0YEJlAeZM{6d2>u \"6YJFƴoD xa{lb5]#Pt5P$I1eYD-{CmJWhk5sVv? c_>[16tj4Fk= GQ4eTGQ ZeX~:X1kbTkלR;KDϖNZBHjyje'Ζ}J {lL S;콿ͣk #ݞZW?؉e+ƽ ihUc;yS,ݝӧ*(.( (3*cRZc,I5ؒrlBذk0Ӷac N5nJx= h[$7 xqg\W} `_8^0:#R;5,;o)^'',г{= i9G[8ћeDd{Öd豈f]1BNü_`z4{(Ԭw'e <B+qD55Z[E鮧Y.a]iwZF|I MM0Wf}vOR 4/|j-!|4u!7ٴxI;Ź̡rG2$O*t(T|~5Occ,С~*;]ű?+ލʺSe] zwA/t)CoFI6uwXz+sX_@ ;q7F yKK|f(a]N +qX^@ V:xk}MԷ?X$i=6Zv/]PEt`4HsD->?BKlj1@>COK+ky; yMPb(~?ɟg׏'b2SL⠑e_f*̽dk<&ϣk<&;7WyiU'i-)/қB*YVZ*(w\f'⯓MP f2(IہN&?OP]PngUl텯 brZ24`teՈDy,jrǢ&&u%ABn=Q%W+Œ rF:Yt hvVv%Pjz(4#?$74pQIXKOT?Sh@; [K+{a4 SS vfmX98CS`\O,Uż Ry)Ӧsw\\}Nx'ji{{'qf%e71̰n`9A,-W$m5sqqǖuPh3P_S砽Y`h<9o~JXյw0 ҠWRX%+C` !wUiY~g"%Zm…FǍ˓{'=WWx~G~}tW?[)\Ճ3䇹1:YrXP` 7* u-e+T(+e4[`uXIf٣ERERVƀW2"b(湐h]n.RarƖޱvS`=*֌J+.=:>/8*s ȥ+B0zniݛXJں0 1bȑERSk 0 &hJeIB!E&Hɹ/½W ]`7. UҖnQXjnj!{nm"|–NAf?ɹLk+ JOh2yiIoIPzLE!TLd M@yAyKH7|AkQ8rXRԚkt :YA<7D9,| 0g|.?  ~7$v3ꉻ*?{ũe@k<-~|ð?kLȬ^vo"|Mǯ& $w!}O8}?sOo^=X]_#h M6O?HANJ93hK]Mq\i#dHNc#0X frRQ>ãx y>ۯOL§ď'ۯ>8*Z@`dE}t o@4肩R/&ߧGA5(ӠpӣVhҴ) !UxV]e(7+Q,)-  >meeɫ9/4jtK YvWz*|X C&Qp8ɡƣmSn1gɽ09bW`Tv+n26 ˔m;ԓs `H$qBG硸QPĦm&KvQJGLT9e)QP+YSh&MryV{ H-J)KI~42A B] :ʎZHYGVd_h3<6@ G-A#mŸ.ǗDHp,$&Bx _mU?sSђgؑBխ 3'Ĵp,ʀAbH99 929ڑ >x6 8SrEj6p/0&|m v;2^sLz@_Zߴ4vg!\WW߾~=\\Od#˩ dȡg1K372|=Ș l'Diқ.i`vJjS9plJfoM)eFA^~=VTOGIwi?ݎvq!5=XNIP=;(|b}_P4#I8#DtLdrT9~0|־nS;y_0΄(U=,D6hz#h9M\n˴;g 86 Js&bdJJwN%b;A΀PJ~˙1LsF7E@reșF`1!3H~ZDi4=Wζ6 *-X . rfiYdwfӃfiSy\Ɯ_.?@n  Qbń>W[#D >ғD=F@u~KaD5{0}j~zv6GQ 9:@c\ 1* 4b}v? L>lz1m \]]^ŵw]|>[_:lz)NYé)BggBmP?D q9J>S3 L P ~~BΚ9;򓠇gC I5~hiyRh^Z@r(7J>-cۀé6֋y8ΘJѝ2@|г/2|C5TH!))7 Jou;&g嬎óxc4( yojr sf6ws~ݽ7Bwl(/ r+r9z rRm,e1ߗs{9:/UXFzϻrso$ EtLI{vlO1Ǡ݆bEtv{>ߤTϧ݆'[`i/vΏUYb^l˜+;a {.}m)p]^/DBP~ d!hCBYas[om<_~wgZt *Mxy3tfuH gz51g -;lfO 6UUtSD;YO-4A,me_+ma]- +挩LJɢNg=hjW ޏdeF`#;oosk!᫋eUrLT_M;1 ?>ݎ FI|n_%dGӧ˴$@~1O8S빋=GqlpS X2N{ͯDj {~Q[\_LSXRX #ݎ_[ >H]H#E !׌[òayPI6=c^VT׾BUwf?dK5Bqc$oSGK6T YE<NP=6å_/0 Er"dz;vY['wܿ}S:嚿}%[zΒ5(Jn0[Q>k>ewW˚Zquu{|V/M6syӷf3ZiCHzpTZD\u+≸_:P; ֐A^IV80+ [*t6D`9d1d`hbAJ X$dXYo$w80dLjuzZiw@7(ƌ6.gd%S!%hHrv Ap``sժ)y:^ S\eŚVC[=ʂZ稔B+sB)&Ւ*} "2䅬n!BTշ!͔rcqͤ&?<^ќsR4PFя\i6P>B^ #%=hCȲ|#nu .tJf+}<΅}R*vUŪ4MH י%r^Oge}D#tLpR/)8d'ujA5ZI*ٰRvI"Q V"lPŘ6a&? ,!AziUR@7)iØXszwj1 km 4hu`ET6ge"g*N1Ya3gD$WK g^9e%cRY=3&G~fUwVFvȌhx~.,7Io̝ 4ϗo(zit;N1Ē;?Fco|ʤ~K%/qh=Sb؋n #/QFHq49_D/pL EۺR|֝< {,u GcXQ.nE52 8 i)M\g/|1pgݻ5aj^n&DTX)eYZP* )x3yH,h˹ZZyaar]A` B`zs~w$fO񡨻vd?M{ =DhqjE4s,gO7k<\#vX:iILހQxuۛS#bA=ؒAZg'wñGl\~p˿WomUV { ƻoXb1`j/E&;GI_>)PȨX,$J)yTNd{׌jC] I9ͭy/,РA()FF K9'Mx4#^3p^lNr!'lr"gflv&ɐd{|QJo OʢbT@MQ\VrU='0v7\8G*!%+ = VmʻH*3eG…~(`ρxSmWWK?ÎuIwo(dE"S&ȵiȒR#uڞ%GJ%wϒ,P*w`1 ؖ ɬaB7b-;b-d0M܏<11Y霃B-fcO]CE;|~˽C'bo n1ٌŵ#; -8fŃ]B^ܡ0 oG=fXSH:x9^ֆM_`}k]aNׇ wy֚@ZJZylv1my[N̻|f]>8(ud̻\{sI雡X<8?n LBd߰3^l{zŔc7s 0)Z RS¡FxiciOz3p|TJchmR-WVy`[Rʞ1@:YmC+^Z+r}{m,87%mqy0|_BwhɡPF&_bQּ>H| 9byd3=X otӏVOͩLR]fÜ.ga ÍP,9ht;U鉦stΩlNi`5i3 y2e?$Cp$&r%M)Jr\`Fc%+aWk(*5f66ֻM4h,lm On:T|8?׍(9ՒSn,g!J8+vKEH;#qaҪ&YӓUM2Iծj/yS&O4W!!\D)9/cִBntt6ka:"8 #$c`:hOC4 I&H^moh * XH>AAn j ^Ed@F*h XMiTk1KeZ"7W|Qh#Ԙ1o*>&A30}uݳ mh $ 8+ +lY\i, S8#,mӅeT420ϸJ(bWL* `9Mq k¢.3 )ɔ ٲʪY,4 qiF_C1Zy_S|rʘw<>;u7\U/Byk!bz:g_Nu:^͏_3LRce#E"e$NӀc:j9 8* =[NR1hVaH*E!-Sǂ|cfZ>Rx0kp7kh0N b9+AKkOŇK[a",5DH+5H X&K+"=!ʴ@Y>8';$-m^Q1Jsc$ǔR`o{ƖH(+Vv$ЌxXk(e… <i"c+5v{vϛ@-8-+e%g|<U<8rYM4*\</`qn.?spGSNsHsDUZYdѻ@Sf+lcVyj]{AC_+fyefRU.ƃ̊1XѰ7am,;2 ƛauFr3Rl QO,Q? IBq1Zn*!P\O܋-]f*yf@]j*̿8RucӎxM 7 #R[ǻD &DuCTdYAll%e-+l,Vzz2mn_~2v!Cywg bPlSx|8k<$^ܥ .sVjRXoȱnΫDS.~"cZB@#hJT+ Y8^ճ.1ٻwRS<ݡ@4<p%;cquZm;m`4h1qђ@I?y;fFRBO=R#bH,y]]QBqcl9 V ߓ5n)8t "a*k4(1Jr68k\Q>(4jau +eӸ'ZGxpDS[& ZE+\!+\ y*cb  6 d=gn9l {jhfف}nPˉ}9a $FwK\V1`x}{T G هb0浴˛jEJ ,+`0sen@d& 0uȥ oG?]:5am \_Gp1b5.G'bU`4dHͨzIcQ:Tڀdħn_Q=LP䓒uG!OM)YDyl, wk^^L-L4>Ub27.ô`aSX{@B^$Gb͊Z4emv2aNFB,?>*uzhWK؂ o?Oyϓ^wpcOF Q1FI7ZsEAѿx>Vb $VkAoF+KgPڝ߁[$59 6M|a,_1Vȵ2,}zo`W&%쇵ݿ\4j 1p[2TYᷗ\n2s-$LSs؆­WTP`gFhXGe.F5 E܇FiN 0 Ł#"u&m9[x[;RE%#`09Gʉ堥FjDD-hH|^$ȜsiQh jSz@t/AKHQLܕ/_;x 3$w|.|BT"Doz"f.\aJ̍×:Ő b`V H3e0Ɉ~^'YY^8]$n۳ןľ^>ur^>d d?5;z)>6|-`{RG1[34XKeYNY:{v.˨ r;o[|Cl Ʉxw^v݈ U},fe>MERԋ+^o=ZM'P"ŖBz*"k\hIB,SQiK*KGtd{]2hKw7n7F-jI}U\2z?_E\gzɕ_1) "Ye&$lb&ٳؾ=;`{HɖZRnƚcO"Y<]/B'1ǻOQykQaq5!Ԥhg3Tܵc*g9f>h/\^%J~46x/H@P9GĈ[XpRϷ-!S:ӢE2)ɤ]S$ W y;$wd~]I%gTqM U4c|nNXKNZRtqSNJr6k]NpV'Fj+g~VQp^{B\pznUoO9UqP8#*aD5T(1/vi= #3PDOנH bEE9i< N1UƝJzrf# x< [ E$74<\Bɩ5.>ہsR |UA19.@C#)\ٔѵI)N| 2LC ë\28eo/ 9AO7ҕiJۙ\YGs l%AJ.stYwR**K%E~p PTYZUgXngc J)whbPj#V ۸y0K"$FsQUP28H^G}iXd m(kA0jz=tӐTs2`aW U'd b1pP)Jo\H[;0s? iF rM4Wy+zziH!.4Xjɑ JrD 6A?11xط+I|8rv4|ҧI7mAs]߄mũh܁wL ƅB"e(\úB8@"EʡTIIQ pB$a'|i\a RQ\gUv w |}%f&,Q FG"Gpc9EKNpm\.*XY05L!\gu7lgwZdVo5و uks1H}F)["JKZ6?&BxަIʧ#yX$kGo1sAQwnQG=S$vR*u9BFxyx`TK{n9ílw9CENމR"SE5$ fVF+F[XY^yao><噭f?<7vl ݷ_ &K!ZE s9 /,ρ#<%hURk8 3&-~].bgX#9Sc矉]x^yΆ2uw?O(&T{/G<7 8G0(HTbė.A*%"0ξYG11)F{p©g. p.o*1D !&}#'*~dI,50%<5Q*DK-K0iD͈ pRP/ xC2MyTDtR>v1)Ϭd!@Ga6w/P~S2n%m=l%ǝ:xp_Ou [B_S#[GE8~ikW\Do(m0Duj~!L6~݉`O4k1Qm1Th gc f|tYRAǝPj-2Б(+О0'ֆߑ@uz9ڰS ;E ) 3S*ށגy0•*W<:D4sg&J"=!? _3w+*_ltkufU;Of>>cnNj]&ɏC` g]}}]~Z +<dCP5`d%x*8RX.1g$5ԈR!PIST8mDUr}6VnuwtSU>0WKX[!faSdx,h%ZB*P!@Y-}W<k}4Ga|FFKMq^(b.\Z[eN҄?]~ަ#܎ ,L9Y9Ŵ>\m̾bl/X_BT.texH5&=;ߺbm$Fk5N :J.3. p' Ҫ00Rr|<,YPҫVqL $" t]ߎOc @6YTHG go7ԥ@tjû @s@쮘#0\u 1>!RQK+hQƎ*i,̣< -cKTB@ىܾ JWؕrk-RKGTYFP+ ]8i@F%jQ/uzȃ|eUYۿ[mߋBO"(-m|B+{ݳn@䜏3bt6Td8eD)#bsӔOUQ.linߖ $ m=\ #zׅyLCr9C mr9C FszeUj>o>kl=o-ʇRnxƬ!lFi^ U)!nқ#4gdpבQkx{ŋpڲpMGS޼)qޒ dd$mZfd7ync^1z>z& <5=W^7G7QNGoV׿ke5=+ϥ}up^;0aliue.xt/Z1#ڰ-voUjD "qevPd+"Iң Eft.4?0`H9!T˅[Yfmiۇ-IԽT,ɯٔTRs@)RfSѩljG x>m 2UML+;l/>#(u!N ϝZWBΐXKjNS\z2D oڮPgݭJVN3:<;j <=V`SxPSx0L!H'ۘv(?F4KT[Q (ݬ>'jUR)OUܷI}x퓹3ׅa/O> 9j0A lוM'i{S_krDRPaJ95D4g[o?/UɗCWD+=>c\w_{|B[#l "K˺E3֖.݇,f} bƲ\LgN6=1>N:׋ֳnJ1` v5PO-7ќ6%ZyZ 1T=ZXn k 9*`X蒃JG«JPj){~H"l}vI"Nx4͟w-|L4Gg>Uj;&_C0_A;&#s1w82| |S8vM/N3s2Vi'\ȝ̤sfbJz"P3qƛhNNG7$=at >MtP+Th-N:1BxjSg_wt [.)M'mۄHR:5ttgNn}7D;)N8O:q4$v7`*`^烵˾>'wݺB7y"  ̗} ߾ 芓e]ѽ4$ZH|?POr 05=z(='u(kkn,e/CVvS;vIabhTD҉g*}n&&.[\wG}XɜѺaLʀfXnwda>2C &НK{m$(`cTx%2efbۡY> fwNՅW0bZܽzE=W_e 6W#^=qXXN<"cSMidf~!1mo&S|:W.Tޡg핲vXBL K۪_[b-! I EHH& pk hq-4B*<-(0d/ -5D*8)B)PF3 a PTRn^[mJa5Yv^a58TwwYjzbK gu.l(+Vxux+kWhŨ^ZҶ*tUuUjYkYTЖ$~i1mꮙ1*J!wDlذ0aSFQ!t]8H *e<=ܪ|/S,D]r6G3bz84V8۷oUz6/Lf_dh~t09LGӑ x6$ޕ`+ d 4I)@ no(CeDii08Jd_b9Uĸ<9Ù`<]1j+vu# W\C !]UXP(1Z̑QSTA8AX$ 3b̊L3BDё9 p?Jѐ؎jzNXWJ6 Ema"2!Pje ѓDb)Q NQ FXehIK hQ@SG'?Nɔ)E̘,6rh)־0C)[Vo۷7Vyw?u˪1!OE"͑v=tݯ A%^vWjDq/c1dmJ}w#)%GuXșhM kϻ wb116x#! j.^rwBDO) 05<-ݓ{j(Uqг~1 #t)228 /g eDrUǯe_^%5Rb*JX٭\će2e*)!oVcHl&ZG nUQK,C0rȐCbJlP($ۦ ߦgxBWVK?#YJgmU@/{zW8n^BLQDVCJ#ƽ~Aw|އ @Vu P'{6c=e 9)J%hvU;IpCx^¡ Xx} ! };:h!ig)e h2\7%2Nv VrJk~;)E?S %H3bC-v9D\:Gmy<<."Vq8+I ۄ,.vٍH)x#Az`o&ʆ\!>$JgC  A$GƹCy90l晩 i}Z_.o&gA9KW`0܋BG/^ȳ#}{j9HC^ c}ZP  ۇ9gxD %$٬TxR2i:ojޭr}sI7`m4~~}a$I )`Йʌ$MB`Li1%H߀?n.746t3X@v*xqgvs~:񓍖7:uS׹'9rs}~]00B?kbCÛ+<+M_N&VWS8L|݊9 ^,i*r)BDALN?tcC%pn4e=Wn*Nd݌8d* 4JZEE; "1eMZpGvE+~sv' jٚf2]QQMv8fn^Y^csk(\fA?!Hg,J1J;4 K5d8CBI*^Jhٮ$j"7RS7-P_w}}>#*uhB:Q!{!~:[qGG}4{mXs& i2C$v"/V=yv+t7!N6J۩ ,zfo۬l_r+L $ jNueg R̙Thfc,O ֍rmm_ w}%D_FJ7LkSpͅ`ɢ 7^(`\<,/+H_A~ !g7sIou/˹e>f䊖61ng~LfXwj:0?? 滽vh0tTd~u!@jlXOjGn:_!NK A|EA : > |l1}TNVE({9P][T1j8XB>< ,=Uٹ\ ;9O;Y/WBBßN@ A "_Қ]?װr^5>o hrP#mr Z&}>!Jx6~oR-*6X*h'Ѕ Eﰋ|?qHxOIMI8k%:XA`8R> I`OH >,ySWaGuu\9+%TCL Zi7>CKk0=1%3c)}DQJe %Y}Ė F>C+x{SOɂ`>F!5BxdUN0Qҭ<@զ kl:*Vn)qE߫ŝWQizzj\upb5A c[MP.>4A4}lR~z'뙶փAmmh][uAU>n}g1%_V%q@ )A=}z_zjWC[[== ^[-n]bʡ.@}{WDhGJ6met;w0!hVjpz”N6 dӎ;ٴNF!m.n}+Q'rd2wiJܗm N`TZr|"^\Ӂ6-)8ѫWhl!"D<",he=4~IWM}}ȣ_# h4-lk]ikrM2ixgyFA VOn4 ':c(.:ĴUq Փ[1Op0ۥKa Y_q> ?ymDA˅4# * Xhm g%ñτ:.&V)&XG fZ%);>lۯhneXk*!-*^w v8Ol9v~竛]ƻc# +("($w/tgfEo8Sc>גu?o s$;ii8uzufᩲBDlލ{,SzX BL'*ޭ`Jy nuXșhMy[4Rc`*n߾}ԜvO&7dy k<ߝ/ cfFf|,jF~C\=xǏW|yzSw ~*S PG)GHtB%򘧻R_#F9.hQ9 yuq\޾Wޫ%c'])J! j7w)QGԕDD.tR ={kSRhAnޤN V-f~.yW?fVk4 Ӯ5Ůu1D\ Y61j3i(3% `!gn'gJY[Z''8ξ:{M-.tF٨b3Sx6$S'0?hH_\qeaHv1FeJ65^>~:[Y< v-+peo^~N/jӤVZ˜>)aZ<M> mWW <*5jVtڑ׵57֒uZ 6JO3Kv\z2i^zї:9lXyG[z;aA1$Cn7YnpLf1}zsֿ_@nc8ʻ|;אּjp[ {i݉CLΡ>ļe۟xQb(NimBȋ0YfZ,|,λg斖zg9Ytl > 0(&Tɰߙ ڽvuGp(Zpژ㺖T`\wN:hUjGuZҕBCUqؼꂖ)ꛓ`; DFro$J8=:xvgt,!>;L_K{xveӻ̟L,+n$3Md'Y)(;w2󋋵Zh8y#{\q+"H;7\[6vF]||D0%(^rͼ\Xdb>0#RY"f%9Q%s'wSn9%cv)i[UQLQ*UEoRu ] +|ZD7B5r{߉4b7} iV9ϥ |kX'))54Au;(& !8ւ"IA/)yR*OeB`#A B*-*"cR dR!@M(8^ *#u-1+)Q=,.B„FEKהs1aSqG"#v@3! Ƴ`Εкk{Ζ<{oO ޞ7=V^eA>IlQN q}Hu~mIz4g?Yh1iwU+I"ŎJ8{\Dq {!⇥K#m'" 3+䵜/}p_%V;pJWN2:%ʶ}zbR|iso+գrskk vȝ؞P#1u[{#36#ibO#bwT$x|{I.OcpW5?' 1&B3ea^D`zk$3PO@x~&`}@dŸ[Z~A64/fVđkWiK()#g?{pS]7,`T\!}|6~~ mHhVB$,lB%soWn9lJ,;aKp&ZnH!G+~+?嫻,[Ƙ.nCb] U)6RY#b]8JcGt T湧ܣ|x7T,TB* 6&z-:\qcуn{1QMA5u9= 'Rcӱb;A~e[UӮZQq}\pYR^yTqKYJϸS*¦Ve BF N)\.~dD(ktdL܊'/o7( 'ӵt.:#OUAUf2v2%go4ojUOL~{fiәhBgNfZPqڏY|GIC@-i礹~tt\ŘՌ %ͅ97Üi&$AT~&jSv{/",o11`Gpɔ4.H)!%-c CH%%pTxf}u>tP;@JL& (*mRe9ˌ8hEQ?A]irgwZJԾ{w坐Jo.MʷWΖ?'WRŹ;] }LƬK&oq!=77LC9,X\zi\+Ar4v*ڍf<Wҁf=u\r+TP(͞:!&8:ĴW$ A [ש(Q:/3h8Ԟ3\2:p @V9"s9Vb Gm8 5wQ"+R V]s'$E"\_\OǮt}:B3~5yK5p@q$B9*p繗 BM.n ^1D[ RI4 ++R$NGMŀUzfs뢰{,>MA!VV.j$Yh,(X$$o2Y|XC<Ԣ"uMUQJX!H0'SҖz NfmKqRm$WSE!f}c&,<5bܰ{lɌ1RdE씕S^cBGGy]0  ܣ2l{uc@uGip c2ߺ ]PM[-VB+ڞmQ;B18UH)8 IyÞEGHo00؈^JEk*BZ%uR!0x%2(D r*WR#A{n-0ۯ lݹ]eJyG ZBKq^LBrܧ)U*u *ru͙Gpc*k j$w*QN.,XdAgj2# \BzTj/}PڐT@Do~A5BR rRѥtO)JS\DKdJ31y;[)9S:Fv}V Ѵ[Ds[2E p"_Tޔ (Oɪ(K}Vڏdi,X2Q aBq9)Z7rT5$ErCo ;BӈyY{^V h[;L ZДfVem'e., ?8?8J^]Ő1RL{Y ;T+Eb/Pj{3!KzjeGh2BHMiIJmPυnMaǾK/@Tj~j(r뗩unѼ9Qcw,seA6uQYUm ŏ40SN$:Ū *bO\!8meR[ 2]`3}lWޣ);L\-4SuO PV Kas0f[j5!l) vo+w@~8&\ .bL^Ԃ#V0n4JP\f"J33TUʤSEsr_t۱ DffnEJ;<y{yoՐ,!Tjsu(>%۳EtWb1)"~o$- ]C~ӴA>狛.ϊYƻ׷7w9m7=9~uz`-` {i D6PC:˛h,F_9 D~9_24?N>]mL w/ma}my!f`SBq %׺uVarBy ̯}ZG~8å4#כy:I<8,].cH,)oƟD ؿg#?> ]ߌӧ"6(}OlX0+_{{o.:$[ q9ilPdǨLE&·`i0NuÉe51Sѽw:f_nǂ9sc/iϿ&2qOМ*E ‡-[$h(e]q/^c{vv]'bP E(8i$*eXi/ZU rb\:Tܩ\,NTdira\HFj+3k6/s/ZlU1Dw+V(#jC%,4ai'0tpnru*=1\5g9ɍY%R s2PPPg寻EJ0gݸ>Ϲ7?g rN У ^O];歇 0ePzM2^Pcۼ&dY,o>hcFS'nZ~|I_xu!c[=́7&^:䭇 t8QNtmMm+hMm+Cj[D)<0Ri=o7 B<6-"m\;uO?FջiulVs^e럚}cݒ1/SׂkE{]fDJ]͎/| ^ryul7妌bhu)>j/bO?m59Vo=M:k!Dw}W?NPЗWEE_k`Uq}-+ ',,?d/>2ZwW# F vkE~5l&qgWqsepsȵq^-sxqEPs(rм14M}!B&ˠ K^Մy^+ڏk9nHeHx`# +ȈBP-ŋNiˢD;t("V+DX{C6 /{|b(Z?|w۶ ^>íP% 3#h?|G//7/\f,G;]ݷ?x9O_ ˪J[U⃶(-/Lb@&W}2It1> +$6Ƶ,>gŐ~/ǭ4%BeYˣ#y.ƧD%4?X1z(:$=߾}0Ycg*+ >-/l8Deoӳ}^}]a !u*>}.Βp\t 0Z 0~{V֤/=VGV4i +!DmNJ!3D*Ď[hh؅s=ݧ0t'*ڱ >f'/} 0rW/;Tn3X$|m}dmNEV)XDLR+{mL昰VۘJ."yv0gۥQ0f#1U0m,Zư4QF*&*5{Y:CsQfQЁC'H`ݐ\#1h+T][Ա> B3 9ძ̳&9G7z3h$Q6%YCD[F,j|Zo Pw} 60f~Vy3LL+),L [q+ [j/Fq?銪ZH@_O|V64ׅ@5ek6C*EO^y5ju^zr*לI, T8Po+td)/FdqFzɟo/EQzJvאpcc| E"PRW\7oO9 Etq:gM(S$\R<d3TnBO zPӭfD\Uٟ}ǚ@-m"{qoUq*8ۺ. -Og4Tk<2QDAQuFٓϵGD{=][++Qw2D7x0#\'WRgg>pL(g2?i<(4]w~(*u֭tus29KF?4ͫ^L3aB-Zgkz3pֺh[ hcQ 񛊲G ;{mԃ)zEη1R>Ĵf|Y(\ؖXmO%/}|Wz/g G> ?qs9}{gwGy须;-XD$&jupMAe*2NGAZi! aR aVafU 7Z2Qf&+ |1In7J(Ld[1׊keh+ }oxV RZ+uJIE-ںB)a 8l*vr562Tڑ,}cĤK]ְȩi#%m+졄SrHڂ=rщRX8^CʪR`hj-*WzT 5ރ;B9?1{0cs9"Ǘ\+:T&+Imdd62XbՏGRLPE΂Uy|jV@qUjFMRgӄp8vYkr7lS6eWc P)|2:Xg{ !(O+^3KPl[>nw!01t f?w!d(1t 2j^ Q\|6&rJhXS9|Bx|͆C.>GOO bx h"ycupBT-"ZGe'X9< ;Z$;YS4(wV} \&jP{Bd`J7fO4Z {YW}!ſV|`JK4՛0u"}L;24)L4|ɬbNuwhEqbI;מVE\1DDn (a %k+9Kv;ވJdW6;}#5Ez'4a B#?Z#"ajmDbmJ(Wsg1JYpjqigMwj4bNY@Rh2229\RyJP)s2שɹ39̅\qId22yQQ3?_7ɘa_2eG"ILoaE:59fgs K%q<`)0/uyne۟d+*L$"e2yK2qX1.پ[trJJGH|b1 =?3S!OYN§AxƥT+M/c{Ou={%SD/6S| 1kUBװ(\|9Q(j5`ڇ"8pA]\y8욇kt>4"ר(s1fQr.T+#\c .7wAJp-uǢ6a5؏ڄᐨMkF MF`$9%sZJafpoMQHi"8>22!Spxyí甬,zq^4IQ~u%0rxhKn/t4[*-9g2h)'f9ISfGL@~Rc.2d"7{ .cDmzybuM1 r\ Bn VJf1ЃASDڔpT(3ܞzM"˔p+4R/iƅF(3D stp' NQQ9ӹ$vӁ[ =ԗuE i护Y$qqKՉ sŌJY.ԎaF!QI(>O-Z^ˣcw4 cWkD7-npO `Tt_r[cj^R$$\70XK}zO BPWS:_}9ȅ~g X2=SLn5Q: ߗt?~t%%{nJ!ST⛓}_ކhx0Y/-ayMa&[N(fTUƓ}7 I`s?y:Iu-'SŦ6 sZ߫oy ;FzJe&NHhaT'Z0XY#i#kKc'`bMlq8Q~~ 0ꨍmj)QݩQV\sitR37|y,GRW5.Fʤx8JG0JCJ|S)xǩ"]p$C"D׋.n٫'o@J\\>@ ^t!'SHA]ӓ'#s"3㜈yGin=F '3 ?' ALAg 733 +G' Abf]٠ ![t9ϋmM&|F-ss>lt/bH64!b:a@7fny7dP'nMXזX ?oT;?W"ï8Yk{hW%*h-i+W$*yC{{p|-IFvuatmқGZ64䕫:N/,4$s d}>`  Mb{R L)G-g{e{ż& Gb1bZjRlM&˝1d?3{cʎH7-r<3LÿV~v/Ÿi'YOfU |{{gݧk}ORY__:\$E88OTV85s<!>TmR5h)G_Э/`cKF1ࡌDFrIw$ L+i7fumj,!rp~C>pmGV~˰i9"7bu_/VRe'*I ;9SDOm\~5oWE٭nQC m!6w?~[NĖP^71TC[ph#D-5X9K%Qi گ$@rHQ)2~&W g7R^LC$Bۇ:`9BT~л' 7h wR @TG Sa)>v[q'a:SbA=OoTɫUf223wdd~j4S{۷^}UVyr}{MJU8B2Ž;FZ߼} ?x#o?Y5]} ifU~?}f6_=!a܅bNds.d"J@= 71ȎQ 5P"8Ȥ6S,3,prc^7o.ugMԤpA18ǬOx(BP;KT?*mnܭ8Yn9y A!fir䷇Am!$\{O;*5q~Xw3w[ctLmv;oOpQI aY7aYBc_]{RmeCԊ.u8n,A@Q;WS$O6K6@I3[je\#XxE0 #-JW5K4򔖂Ф!zp@CCCR30a2;cUJB3zVx+UIZjwٵ*3`~h&+P ZQug*̹2TPH޳`37=\SRB+_ 7BF{]ӯTe͎y*B Yh}e8v7`w{޺N5MoKWO?8JI= n[0,RZQJ[΅2ύ뚬7)2QG˙M'_6;yH We$@iu[#(HA#K j9)p޼}7MH(hgbWGV =c\n "m|U!\dV0#ʰLs2=dw?|jx̂g\5w_g\GFMl}p0drE*c:GW=,+=;ݜiµ lNC[t?J.9\1~BRF9:2BAkקۛ?߇;0*HQ7 mNrcD=56hm۠ٷe޼ _LwuX f bZp.*@Àj!h>܀nFCntCnFJl*>jH$0RKWIq1C@| ϻ^>7ҖR .1j4 @+hskq$a(sh0rRk*i2^00[I@dgh2 `-K~\ IP!NrȘ" b_oUK9JYեD4$5jf~m/hq]u[_p$x y/2S4'ʃ';5/[-@Fz)J rJeggHte {DxAG[k894d899Ny%x/K aH{i=,1:Piѐ4kXZRNJ1bqrDm?3OXrЋ 6S)R'TFjnn(t?W y˹kccHaDmq GYo#e:5utIRrDʖ |\EINl^(:nxÿЌ2GJj|o3!G9~MeǘKl,oW6CJpiP:(nH5ɼ3gTf)ƇnΘg阯,,Ɛ1wYw^Ü%YaM5xe S(f_ :~ &LH^WwemҦ%:)SIW2d~4"HBiɬW;1:IӮH 5Վ6%Q 7Fn$$,wtr1IN,qqE'רHS[Ӕ` _]2 bY(60ݧ6qIPۨBmw?,iFJ}i|QҒrӫγO_uOONa1>!1R 90#"ݛ0ȸft F'x *)a襖hM#`E閶ѥ $u.@#^z 1i/wr)[T*rt^ s&c%2Qð'a%sM&Մyhada\ZISVm:wX*t,UCcGa~ *FBA0ܰ>A0U⡈\ kOA H&ˇy^a~/1Sv֬|{ p';|i#>#B1x΍rb]0T s4Ư/~%4J#T8"͂u44^XImWߥ7+t*}ER XEn 9S l>C~@>LeE]~ yO"w\.;b??Eh,v3",aEw:']tۻ՗?{ܸ/d+q8[M6SMe@ttLj+4H٢( m%F}/" P!,vZ8*`0c;;y*]Jži 'O*k˖ay[S\י (]5:/ZH'+p}^U4hY+A@`vY:%{@ R1YJ<ChD1L恽38a(mA4U.7M4 Y*(J L9$JXHJLjL5& lPؚ%moKrGGwU,QLj3KaR0*gHj3=Tʕ$i"EF#ʚ$@/)%&hO¸L[9w;1nSD kad2rD([=eX~uR T#1CCaT9Nm(uJwUG֗{ϯ/3)J׃LYR𻝮&iQe n2yz<]cl\S?3by20$k)כr:s\9-ܞwe;^i ٸ)?e S֒rRSݣw@vAĎQGnnQڭ y"#S?j7Z >vAĎQG>?Mj6$䕋hEhLOLL]"s‹ӛ3^zVt>7P3jf_r2N.C_z<&r=vu,pliL[(F"ubG҄gX,.A:Iwjɕż\G^ ׇrvuy!_)e@:6O\>0*.[-64&&ST|π={Y45b 5:qZ0sz2 f|HBNGBD{(1؎H >T*J1 fF8F',F"'3*9I|1Hւ J\%Kf%Sn$VY+e2$㙰iQ1#\ZLR!DŽ6Zs5T&=1}*g}@5@_TQKPmPʙ)I|$a Ts<|C2B4LkE k3ɰ>V0\BwK9P%V0R9>0&)`Y}=,(MSNi[){@ 0"3{U:͊; u]@`=&(^|^T P\0"q†'s| .Y O^ي?u)vZWl)o'ڰ8VD`vݺDގ±S 6¿m͋_[q,~fE㇢q򃔁-ۇPtŭ< dz,O'i g;Oz^d/Ia9ӂ{XJWP,f2 ޲8T:GQ֍9l֤ LOtj?͈I, _,IXF p*Y q)hf7V@F⬰&qk8~RJcs ,ZJ) RFIH|nءnL-tJh]wC)pjś9hQZMt=̈I) jH +'ՙҠ)-c{q A,E`5nӞ/ c xAl~L*D!(zC@C> jp γ7鶲:0$8^HDVDAGX Ks\rU!WbS໇{7Q/lEF\A.y.?b2&9L¯rG/\~,v>E*qq u.FX2!dI9o;X%r/?9WϫAq[m{c,2*M׺U4zt6!dƸaƩ\Zkc*_T,Y.ʙOHcy33SZS2Ӓiw}X@0oGA{)?#,sdNlJ"u9|8}jg0ȫ-So/XWU W!3ЗJ ,x`J AvOu2$!XgQfaD{C5b؁>~o+z,+֥(0Յr8q^Z̀4;P06N@Z(8)eFQBRo`.6wP3fR ̄| `Lw(5Z\ җ||?+ chfh.$GZ0>9#T2D\#}S#L5URPQ+3ػ樓1Ja?h,| G'7}s,`6 OM ֺ1,M)`$H)2픈iȍ[a}H4cxB1}O@Ӱaoo~֏iwwxV0=Рals]J7IUo,5&"SXiߏ߿99FaAccXKꌵGJ":E6sJ:c dD1R%`N۟~ҒJNZy1HelkߌU.=>A񆿍_dOFE,6D rd<WEcVRy*;A,bW$X0DRΔ mM?hAdnaзۍ u?\|d)T<z?#v -9NR r U[u~_!70NbhQ ۛ3a5Lg5ٸb旓qrs H /cWEKN^= ǨR-X%܂ R.GQtkTe:l҇.fU}7UbV*`>i~n7-*}-˒i~w)}t(.\wwˁY,Ɵ>'Mb0ay}=PH)GPOaɬ{(ZjquD?QFr (Og]*}QI9.*I+_^y[78aƴ2dVTs4Ed J!~WNqK Bp C9CCtSF'I[*66Z5׬(uD_L%W$;z'ja‰ZFjrCFz8 q\%Dٻ8$W 0@^Z`X`wv !hєDR[M[nv3*H-Y*fU_A92)zvrm#V18>:41W."=܌D  OJHb[i%Q;,$KA0p C ,L҈O!E"2/f(m)_FQCqD}DEng&3`c)^[Ô,_Ӥ pdXylТ*vj~س>(MPk {7T+K]k"=B+[_Mraٞx=o$r%QkLҜfdK[3iKKHפAAwNW} -yk拮O2_1汐y(VvOЫ :}Bu.8އ M>;D)'IAH )IP@q{!J3}vRSwnsoZ_>^~mcxv7/կI;2޿WU`o%ӥ 4+ 3 /&X$(~OX (6Xms3*7\PgF'fkEyQj$R_J\w5$4I({'JmsqŇZJzJzw-/-wWRj`x!7l잾+/K_J.mؑѪaBH[\Q~{bq}{)Um'㣚;31 kyFuaURlzs Ӭy7ҢlQ E}4`B޸)W!zˢ bqmXe J08>Z`ذV-џnNn sCTgZdӇm9./'B5 D+L̛= _5NJfP%;sܻZPP&P 4R*/tvsT7_6NudXWjŝdA[nކ7fy>w4 g05oX\lKڿ2u(k?g&_[{wqWh_K.]}D/F8$y]ц_/Y%(Lg zp=y| CƱOݯo>-Z4#f #u(bNCQ٠ҾZ:k tQ jjw)چNDkm(;/v&DgNy{3O~YN?vУd:t|Ne|DPe?Mg 93IF*NwT}ߵ-K">P\4i 5gv>)vs?' FZ!5Tؗq>֗p0ڐi&ڕ.P*.Om}, %jUHig$ ~+PΘB%YiVFS_UbE)cV1a/0 uįLdB%Ui0Y puPU8Aۍl8ɶuRkPޠ ⱡdJ:F zSd Z .l7s*6BK -x2Vi[:t{V9tk P7*WoLP$/lM!6L[V2U:4kWJ!$Rtnd;JE E)2(ḥ4Ab)P /@+5*In|;}hay?d .GPI V8r[( g ؐ`('Gˣ u"<~e네BVH}oiSPkMs+mk2k #EXTJDOHs(B'2LԜy{0 K=YJ``ڧ}oQ1"pb:Qk (!bԩGtA }Rd)\YT7b0?\lVQʱ<,RJBo Q^a3!36:Lzav$ 2R g5Ă!2'co9Δ # #TubPUE5GgHpZLI6C/dog05EtQ6F> Tkr1d&"}0e9Spك$W`Á]B,옧=Q\  geljSDGXiwB+`듫r9c@نmR?j4 ؅$꾃rFOĠbz c9] Q],ohVP'󡡴HfPS$`+:Y>&s79wo4c(8pMuoiGaL ]kMQʁ5? >o5$X|0e"QMZGs نyGWmd+ᯠ2\8xZhLp͇\`v{6}_."WwK|^]|FѻfymYDr?6fg6;ٙ: yp@R^ʑ5dO$矻Bܐo}Bof#׼kTJ9)#Ђyكڊ(;eЙYZIXa`A6f kkE睡*P4<OVB 浲͙}D>R2Z:4 I,0_1 PoQN퇫oE=~ᨂ 0oN3OR2Z$dgm](Exqp| q 1X8Bƾh=-|͊.9s eѕ};|g~R=& $*[(RQVÞ\~Sy\dv5XTck0˽U=֊ wkFX1]Zʞwx.^=}T >=\ # ZFm^ҫC-kz4rzJVI%ZĮlZ)=>=X}:9Dߢiep8+g!ۖWTYA&s799~~cRdțFpk!}|׽ TD/{OYo *v\u*, &BC5H]'Y̭{# LI.YjX%]Z=UҮXᘾxJQh,tr…BVx>:J؃R_FwFȷRC,Rw|BH}7DzT*^ŽΫGe {QS^BUZ),%ͨA>ԡ S'`%1do6s쏯Zm2ݭɿ䛵&no96ۡ9ܭ/>t6f|dh"U =^Aīd~=ozծ~=C߻x5E_7udH۷vc. ^w_^n۳IU0jwa=W%>mYZB@ V*(Jejr#^&9fYCdYQ3h<0[TYrZSVaw2JD^ J:tG5@DjK}s,PWڋҡ <EjU[}?[5`x~Ҽ7cT$~kV-s𫟚Zn&.!~N~ݢ^jݗnO(:Xmz`~}K2U?XbQf.jʘiA5YWh[֝E喅"6ةðsFQ\o{M-]4]oTN%y ڝҠм3mJ5vU͙Ql_?5Mla1"y%~Ϳ|9yƳ7F+^xw@8mAֶ?GChmӵm$VI)p"/9>$65,o* vϢ/o0hD/>֟"T;NCΦl:є%tgqGW5B<\xkЊ.ʚUw,HPU6+X 948]#T@[8]k /zSLj,g]WޓfE(]WS:pk֮Yˊmut5#5 ~ww;0QG(}@1g9c<kov&$!=),܏?Q\O@MuM̫WukQ+իqggbb,I(iJ"ZA& .SSd4GNptռlmLqA aPPkvjmU,ut ಴#JkFXQ !%jQkeO=O(MbPkwWC*K8JFj $J_}/GJi|V>ͧ ;!w:%ݹIќzT|O`o2)<#ސI>љ6Qb,yG.+`W>#%L; F2Ӥ٪EM"w48FJR1G?NZR#6&|gLkrP(EAPFރFXK= YŲ ,믠B'tV$TaqD2TʒKWs9Qwl%] %kP<~qiBt_dwt:G*wL@nOfsdPfʈV.oݗ?9D,,zS-bedp!ּ Fzc9vRP@mIwιP1>g*.{͘hD+J)`U<"w; -(**IK+TT uGM{yq-K *?jF82A=m$ 1^p3?w6F祏K}s,{җuG"G#?~HUƸQ!q4aSk, `tH\W`~)^4r#|ZѺR9*g N-ߢ^`ߞѡj"ï8ҁ;β XH͢Ԩ_ʺTsm1jrf[jNCMbݨ[(Z.׾@ѦJ|[6YWp(jkjK_xY(FRYJ$8{vs&_iV"ܯ._/ᨒqaou$o`\XǑLϑG ټ[..A-sfZewBJq(/{]Qncq4mb[-5%Jy[U:< |V>v%AӝvtC+ &Qk*ݍ5k äpd#4I/o vt/կzU ņHfz&K(^;Gƹwtsj+TmWC=d(֢>nE`cF}TaĻnACfo~j<)ܕBeq7@dx:y?c y$*Z{aIg=0XLXgU$@DɋXYPc>YJ*!:gr192$}AFRg1)X4jf!Q\d}zv@39=6Ϣ_W6r⌠s1(EBJ!jh{!TJ@?3*I^bmh[iZ䶀` ɚޑP,<@@} ]T;m<.!TԘ#Q G5|rXͪ{cխgH%z"7QV]fTt!Pu.o=jm5e"ѵV1Xi>Mnque-bք2e-l5%ּ?7z+2q柾aP#7ãf,}K }ЧBvr+{ nw{l;cPJ 49 dl`+mK*'c7{{ÛCmzsf~x3퍆!I4&Vn|sc}t_!t׶G˽6gޚV*ϓJX*7.xY:&3f):1SgŬ"_Km{) ncV|ѢpK(6W1!9?Lڡ6ZM0%o"$\Pt' ,LٿAQXӁfHN }Q;Cgk=gd#&Y,k(GΙQ"ΐzTm_]pJ"0z;-cVQ'0K )ȑQ.~)LFׄ!$"XvCsH2e#&ՌD'4%tO~ZvU\-VNWjQ esLe[&$S}hPANwl#;_M Vּ;_h˺"_"QUjx`wzT0XyY$b /\zV)-μhцh˽ k86ok&zqp= V&NWb56;T] pq-6 BZ4 /^`/gڵ-u"d M.ge5i Q"o4 r֨P'Lj@`qDESFTޜ*$#K?YĹ+b^!~Rh`)b ?J瓱 ~s":D_xw@jxhjq˳`s OL(ZΩg26kq[m66` bZVoHze~Gcb^Bʝ']t>:>+ۋk_YcޮэhW^2f*k>T簝v|`{)|dTqFc* VL5KiFΕl8T2oxMFGό}FG dAk2!W$oC&WwaI?q %g1rL`k՜M{O' 1fvVzJ;'K6#ێRǘe]fBJBzhAK2"Du*[Ԫؠ)j~dgƝ !mÃJm͢N2v@C[1KTc5jn%U S fm:W}hPA9 .CsoǛy:7'bRc<-Bt_{|!VƝ ;deԒ?|J~{Xw?|?^cg!Nh)t.00 ǿ׷>w~;mo}FT9=DŭF)ՐeKod Rѹ礙RWZۏ!Y@--䐹; b߿^ϑ‚]Js.?-]'\#|,?Uau洜c, A*P pM?*h5Y&@ӏI}s"v]A/4⥊t7r %9`RD]qr('_BxY3¹iqmvsRrON)㓯a OR"Sgc9/LY #OXf%&_vOZy6^VjϝNw*7"Ū|KL¨ 2Q̝ݤɗMEĀkF87ʋ3mw#Quxc}cx%QN ocԨG.jkNü䂖/Wƺwg9*wX|b_9V=lNG}>C5WwD^Z qEXHK z6˙mY6;iܦl6Gu]iX@rljs "%Udʼn4S6 0iah#km#G/;\+dm i=,>mۑ$߷ؒ,f3@KvX"૗D9T.s`^H/ 7D+%q(mZ|\iM 1{)^@mπ4~܏XI8{9Z/%KL*N`?MGTK^˞=iB>Ҩ#oI.-tlR飫]s2Z~Nb>EGTwKu^Ke]KTۋB0A0+M,Q # q~I# $8 6ׅ@Zè*0&  p9pKBRIdq8T6z"8`NJA'.q3<3d3=Fz8'Qj8՘R=BTr!1N쭈LyBGG&) 2"hU/UP 4`9.iEQޤ :LHU">>dHE+,73eMr{(rIФ"xPĉHQ4ѱ|U_QYx@k&m@WP͙]D $y )\3c H^z4(y䦬+1LN@ D. )b?yɘtj.㻤yDI\ed+ChRaGWE\rK'%GϏ4ՃX "Kubȕ>0)=ZnV/;.ak`dYj`4BRc 622$y<`Ώ2s> ?)6]̢)hّZ>С|GҲ#3W+}V Lr˔q~ {pcE55yV TM@s-'"C?: LgL:1|䒍>&ͬ65ʂ@1$چ(<=>5|Թts%:tye sSjΛ= B)_bAQVL!&@DQF+TzlR`I$禜k)k'NPJERUnHäD$QsVb2 .D04{6jA2 GdD{e F8JMPZ['S!CB <5TI1jJ=CIZBFǝ%R$(Y!$8kAZ̯70|V2os,/orl~4?iuSgE,ĸҧ2F5t@9q_"yCuc p 'ePb*uPl- RgYy <~FF1|D{Da,qF023$5q.Wf[؛ ͨZ 1x"DGvQTǨCGKFTD1~v^T!eë W5@@:`RN 8`4fύ+^׉^: SuBIz9h ']WQJJgh?9ݞd.73Q5"_L}64g|5~>j$!yE F apk95 l Hp8z*8'Oivֶ,'/:ͳۛI.*<vmhh9Y|P֌j-y- 9AwK+Si43R-@)R)q«+Mؘc%z(HvɯU2T"e8v.VQ:䯳pc6KLʎ+,z|o~!bf/骑6俬 SF/MےJ#0Mrz$ƨF҈Ӡ UnYClj21r25Ag񝾷?݌h f41f p\b @#֠Nv2)ܘgPA`|qܖW}LRZN[{\ɖ|L@ /qe[raߓs"jg}Rx|ʸTjz\- r2푥=\[G+mJCxPT…Ęanu ksWvl|{h/A13j恋z60(zJa ]N-9T8\﬈E)Qe~H"B=k}fo'pU|lb|W30Y̿5/,YsuMҼdϷ0XꂢFHl=8Hge{k̵>NQD~-j`CcrJBT.zxëO9~^A&˧dMKiݩ3>s&JnϡTzn~0hZ8 Xz pY)ѹ9ܱ)]D9co!Rh~}@Sb}jWiNw˦JC1YjS͸!}L.N5]k#e%5!'KmR|6nU!^_}v'yoá e~z9hnxwu{CrChk ~=;< ?M^j<MD7Yr^JF ;}]K-yJb 1&χry6&ZeSWGMQjePb:ݎ}N;wD+ncXOnU6%Gȏ.jePb:ݎN ?wPưDTF:~|]l)t rrKMQ¤sxT߷z.oPPSsͲ65fً_O֟h2_fH<ۛ[mR·5B֝#.q~iٻ[no3z{v'#`nVmyE_bP3gPMW"?| |O!s)z5 2]0:õ{.:[ k[PnRr x`ӥX*PN_R:`5TJ [bɨ宾F)12 S]~zleHCXL3(M_6jyȦư`\hC̗IcI2"Q ǶUoUdgF!a^”3)g;^سfZ^rZ1| ?* 3']sz';PK:$?ҟ P)TjC(mN.цB=ꨉ6ՂA'"g|-}"N@j;EEVbe@$FCA@.|1:i틪Ik_aVM)עSR_sl(eB&C1IJdYF$-F f X*`lͳ"bO7v+SE{T(ur>!/!%hdhIʍжю%iTQ|L, CL"E  &62nʐ.ieՌeD9y6n])%0!켭)2ho9H_̲01-c7/rSJ`:DK'g8ߜO _=] )ztJ%!$\T:o}B &\IY\:"uXMYR hƇB;%P|CX2TsY!2v0=1l`rJ4jEX &Z.1*X`0;l'HF):t 94(z6zk_q_O!A(A2[O[8=K?j߇Wiiݯy?(OQdZ]HXs@>mvq7\F:lDVjXd3*L92QRk+٪3'%ƌZX_$z4\3-$x+d=@=TI|RXL&ENFtbd1Q7(SP@g1ɉ̬% *h5eECNKB1eBE0^^b|Ŵ/:E9;@ at@{tp6`z?PWeз>e`Ol}>vv'r 0qp(23l)`G() _JR#r5ULM] Zxw~+oPLʋ>LvC: t=S6' _',>)o'c 4JA1F?h7ss'+K\4x%Ƃ{3h9XEOI}~"jq|KUz ^~ʞxcyYmDIe_]i85Z 95ꨛ=Zvw1zNtۤ:9l J6.~l>^,n0^z ze^>|K Z]Yk{}ꦈ=O_k0|1c`Rnƣb0˸j'WNWYLJç jHgLaBF)ƣ~^\?UcFW{1 v]11cz#Ma)HGJ]X-CH?is;2 !$aI' B4$d!A MrO'%#'ȼdbd&&npNS<|c0q^ɼ`}]}T[<<|0܏̼Wh%hɤ[[䫝xF%#Vrl361n%o0'0 N1̦^c0oe#3oy AđtoPo)Fit..fdQףkIbu0T8w 2뎶`O1_2bV[Gk82>6`4g^c0OlGV2$FPoAcu)+ u9OI/Aʼ(O5l7zQHru77(vJd)VՋm#AF0".:z]XXF Ta=¸ȍӚHa_|F ;:OM^IW>f3"޴NySvtY%}Ejr!ie67pPCuI I'^**P+XJI RTuCFJv w4_^Zt 4d2}nPwy񟃛[Fݕk܋tMVr`Mnq'{[gImH@"~k!/'D>pzy*쟲>^aw'׭.A9罎&~KrgO'אJj`뤐 K?q%?2]E@p]hAeЁ6SxsũWR䈰YK+'+ѻy}͇cC'0]Yiñ'̀Uz/(ѴEuW^Xj$e{3zqA@%Oj,=xFA'@OtLn螨:ٙ]Q}1F9n)vMFɤjv*e41@)L7b.Ȝ|+ uS yFsrN/r~",qIajASZ'Q! .QȃQZU:,~=(H'F{D)JBt hxҊHT)(_2QUc-Z+Q#c\kJ#1d"*N)~pϢ2uFpQYlp\jLcq; )@K# G%zƣ P< hs .ց AFP>H\G"D_BK%c%xp$8@uPth!5tPj˥qֆ @RKCJn6^K.91J*XN ͥM%ܗ@VYREhDph JbhA[H ZXsmזTXQG[(1j,&8<¥O 50O)A/fΔZZ y%5]JP١SAJnɹ@ W!J e\KMLj7)=A@E4tE> B5$ c2Bj{4UM:2P r H'j?OHdQԌjT&nHyԜ*NJd.;Jԅ@wa'vtxJE=z*BN49\%l- u1O% z~Up=cğtJCT5uSe/seOjlY\In,f]K/}88Yfucb(1D%`'c58cNTB=N}gp~ 34j}2wr=zh4S^X_yw}Buѯ8޽DYk%kg{E,4γ~шwxu8FMH+yJv֍1Zhrg_lݶwm:,z+7sUe^)e8ujÔamүc]#,*sf ټGӵ ̃TOjA'dBDT@#US-pDyȞyxw!=C0sl hA \!b!R,+*S<`]Z 2][5qw|0Z7*٬/%\e%6pLj>GYǟ]2tWfs:0%J:8}qh8m5SL`H$}<ӥQPq';e;= HIqOQ d(=H'H>\nJ$a$GMY?LPḒ^J_X1s65\HB˜2j4^0i@Ok*_ʯb5eЧa߇]qT]e^0gqƚ%SJ_,hADQi9]v*Ŝ7C!6_b5Yxǻ !Tpr}屆yEo j0ԩ }U:͓l ם<韐B^׳ur4PM!#B1 MoAܐO\ Oύfr6M1|*M,MRtxJ8W_?pFa:8_~vLϮ'ؽ_a/ q _1|e×1jM'wIs-NdJA`Ac2D`e{BQ@-e$DB.j(4s:X%8ԄV56ɕ_Q˜r4,e#M?ӔEl 9S-d2ׅnOu^t%i @(*I/KjN+}Kn(eGzoўS}("{v~R-N WGI}c*Yh/2i~ΰڬ߬\!3 ѩ 4]ҥv`EN3Q:CS.ai~-47egYؽvC0|yPqJ7BVFe hRTP"S;k`-_¨2aTc ̓c8A ]DJI)I.T84km M&wZVju Dw:Fa b6jvqr=^*<{]^H}8V. ML;1Ww@3ˢ(7рF%H^%FJ7hR/ bbj!t帢g'`wx1"۽2ɯ@Qer܋MS}/w’+kƊJ)?vsC(\h11T1͝dczK{"(jznnxwkFwj1bvAB:@]ۛS~"rK)5qsBز2XMd1|+ܯnF#|ͤqQjLfi*Ձx3!NOCy zf@#Pr(2kA o?h4FJ"+מ5V5: WTP [WkdޥwdL #ޠ=[^G9%]4XY7U|U7=]uӟ_nS_+^bdnӃ]}>-QG66\n>78 ?#ql_]|hNڀZ-u@1zA ` ɘ8(N3+)bOkqCփP*0l۳VR}GK@5u{-[_m.5BSKKR"{>+@.TO?q35iFGFk ч@#{5He 5/+A\.Bw-jjsRGi+Z]7;=WIeퟯݢ̱M_f"UcI$Luzlҍ}1ky(_-[ΌDwh%MR}s œw60 F]QB.1yb v 3oazqfo;y@ [na*፻a hjr 9!>\@qsdԭo6 D(i\%iodԂr3 EZ C$98MQ݌V'tSd͢ޥR|KLk r3kv|Yٴ{k Rˆ3K[,KS>]-&l~4Z5Nbu$i/wՋDa0M}pbືX:@ѳJ*NFƔ:v3ڭ-9C;hY_J['ڭh1M8_h7J+nm1:mQGrAiTڭ}NLCtcLeH06l)2/y__N ;Xf AFZo#@ e P`ۭCwYfX5cx7Ml48/84B`#*SbAZ=3B(K}QsO\ iK8ҋ u_˻l,)|Q<=droclg!ixkl7NYf.cR7I_J`s%2 dD_~`fѧۛT؄Dr$$4 5!X-T5Jֹq@*YG*lB*h5J8QH*< @YRL+M܄SW<ء9d 388D X$ݱÃF`c(A+ l 2K` IJis;ܑT/o)*U׌6BWRsnlEZ O@ɽZ?)]SR3_;[d:)l:Weߚָ c3)y?;2/~{WǦT6.k-oշ}&xfm&"~66ւ8s?'b,9ow{ލW~v'ARaXa$*˩eaՓ5_&-D&kjؓ tT!R, a~<4 x0K AVF V?b,>T[\"DKVO?Shs$L9OV!MWa TnS8LTKTu2PO QBb_⡔~6#rNhZ[?,z>.Tƈj2JP8#X#V94SXTˌʌ` 4XQg\E_qkn q BJ@,FP\&f]_X P6[0iAf9OI9%Inv߰e`&JLVAa>x|r?s_v|>8?0"P0Isl2X}`D? 3OVɝKһ:Ffy_ D)كj},r%<%H'Z0X\Rf-7/9ZX:{iP T[Vj0"V 4*=;Wag;QS !1 0)xL`DZ !GRVZ;wԂu20:φeT 3nܓYbk αY *w/}Xm%+E*ߥ`rX? fG lY"?zp>ͷS33!r>sno=ZvG'!~gGp5YcWZ]!M%W.8%fӰ^ًY :P|#i.t0 a`3p2Sk,#幥kJ5%5@T(umcK}!@4a [Q.z?Eŷ+C'iqk`_<=3fsj}>oEƏ\waUE&Ὕg(>u3ys7DjjoϷणlHr_Ĉ[g9b.=y7B065n*F!ål z{ m_5oy0h5 ;x҅])}&SDֽ8QG"g Ά%׸8 L=q7|aYiP瘁 gx CTI{~B#D+ +cj}α9ZCvSP}49Ƹ=(5;f]2ټ|zAX*=l^yCA:bAM&1ߢO@9DRG8}©%lr>]M'fGq|q7NilڢqXAeQQdzs*&6V @K ^R8ˊJ JQ 2{Q# Q1Qb~t0(Aq)TYwzpX|&J ЙJS;j  ~/Nny(1cdǘpq<:/?o.X`Nӄ&ybn2,Ov|!U{kW!l{19-tb:^Gd_ r* K WO^#~bN8F^، =t A#f/DNa#Jtց"FzB9yb`@),BGfa_x{G(\8F9S|+O-"[;W>Q"]];1tUqƐwF.\ys^|}~8BwG{L[>9h7 Q) !qN֦(0>`gqx%h%03ۈũGM^5`1DP0'p8̚RZno _k%(J0ݑ- EB u:8QB#ltzS ŵDH\nJxO;n:.ۯB,2uUR'3ca~sbz>O{|)* je6a2L`ևv-ܫuI5S{8J*CrĻ˱rMq;(1< [ J,ٿ.FSٿD9i;ؚ>RקlS kJx]r$Btɶi@K~ĜZIbF{cEm#z#f`J6ԭ]֮rՖ#UD+Pϧx͂Dl7d/;mVݞ"<4ZM2$k6rZMP!MyW*tX5{zJY^\orV[g 8*juJ2"*)z,Nn͹$}h.j+Xe\mǐS>#Eq,SYr'Z[X:5sNK)x8:•cRw*x`e[ W+^ fNՠyN}' * xs5opf^uĊw tCSrRf@7) CX#:AD^0r!םZKDl`%$&Rgp#Uɹĺg ={M|(Y"B>P0;w̢ܲ.|d7ͭd<{+WUOҡQeVBSPRJr̍&q*S'1oDn .7DP`r\Nw|PC:{?0_,4/׶ѯ`@C_m p^G B 0r "J]!dy! ,qmqFqm|$䤆7Qê"ZS7*`sXdLgUpay21v:-RSaUz8kI!Ejm, :"K-)e+剾N+2rRWI a0{Bustl;rRl4Fk8̟ 37<3f0j"Sg`K Ӷ<9AOv53E)Dg2\~u]i3Q>5#oyNwJq>^e܍`z䳄[שּ{kR4ͧ0f Z!|Iu/+z: dl'VU%8zOu5p^XȡTTRqQ^J r;"Yq]F 21) EIp!-e*cHdrMjHNEaآI!5,4bޜ~I}!^;yRj(nu~^;M/r'V6/GCVys`NG S̰s:$H ڕ*:B;‘nM1ZsU*=AK<qL *[BFhMQw\h,\hӝ]b-N4BfbpҌwYWwiÅmsEK9WO^pg4z3&L`MD\ZI.6p4[, 63z1鮦"MeA](APOwy ]X*^'|d+QđRDT HMS/1FA[;nQ;=e1$I/"Q)N rjq>~%<8oMpU BDf­#z"jy)78FĆ2Nfb$I"lΎpɋs֑2L,YktNt֭%pDjѮZ ԖaPQ);]^{wx?tCwx?t0YSߨƒT=1ukU*mW,5C$~Noj]Z78>Ao?iooA }oh5" R?X}}'nr A:x޿khةyO8ߙP4~-?{OT8? Pq%"Xqg59Y,n[$uXmH<"AKȑHH@ _z|$0b,›fߴ)<_0,{+BvzSX1" ^9W&Ұ@< .z'O?a6WOdؠC{3G ja-Ks̓5C0~g\I\!bpx@%H( Ʌѵ`s6"1^B*'>\/ O`v!{ds5f׋r*H*0uSZ^e<ϟsB߼J-=^쬷2X@d.m\D8sN (dTW̫%JƥIQQ&zq8 "qe6;UALFϘ}|]?nqzZVS[iM65vL}O:ʑzνg)M&z99Ώ6<oIpQ-)qKڢl!VXR u2BK(op`&юu[IBw#^iv.:c^`vu1;LHJĒ낖AТnJ(QtxXkEjƼp7ҧCڦѦF];`ml*næU޵hZQQU 8F+k-j{#mYƢIk^ 84P+;W8)%JO; `sA(]%ʁD u`%|8!PgZe'0JZD=goDP/;|)3d[3ݎK_jQ=~{p^4Z쥗; [ ad襬^zc=^{Rkz5C/t^zcݸ>KoΥV`F\)"0Aƣ~>|xF&vVQOxKB<顼B$%ĚB).guwf,6RL/egݹ>42U~ ?$tz:z8Tfo,"0ZUNa<k&8`R%sL>:D( D"0L+(.aQ]#y+sظ48%U`rjEa"ۭۭyZNC^sg9Ռ.OSLnݤƍˏ73f{y -]6XF4_"O!MA(v5_o#̩R~sY aqB-k*B#nuzNXE]DL 4f3eϾjًY"@tFo:Wo[ i$֘F5]ykII9ܣ2݀ DXA׌CRE$ZI5'f}Y^NUh]xD/3xm9 Zoxu̗YQj' LZ'bs=N`M*uD2<鐼h4RϺ?s'$Y 'Vpm-81 sRlEϗVfaAH h[n+nP5Pھu=RƵ{km Drg%D]y޴Vbuͭ 5jf@cjvFS۪jjAtj$R$б?bO`?y_Cn~xgje3-(B?eA<>&G&Tc@R9$P<{ۂU 1J{{lQswҖB?0PJag@Z7LUcxn:4xB Z15 7c:-XoL֯(yQ-6\0tx-7/%0M@ͱdj]Ʀ'fe[esق4lgO z814Qn[r3G0ztvN1 #e[dJjMKCײ(* 2?sb qq3zYqF[{Α9Iwɼ`qT|v[$G_R0{'nJDqQF`&"÷P#wvk@@_@g .F\G-I2c֙q"fܙ69ji"?o4^x@YmEcJƙ=R.G4Oh9ڗ0[ATl/"_D*Gڵny6U-3 '3APR$;+׏9;!Q3V>Ǡ׍,)*+a2Mƽ^״x X.`0j/`p_GhкX Ȩa&8>xx_y!0 әekH Zccο{Oԉ,8#tk1/[}tE܍hܢFx+Ys o} > {K`[%qr]S`h`6gȰ`qQ)*$WħJ&-Xa`my7ڈD A簽I/`q ׯ;mLD囨C`qNfLI8%25G=]oSf~^kF39e.Hœt|v-@ `v2/!\s3Ÿ*;*ǘ-Yc4s25!.w$2|~4걆N*= t BUHox̉ow 0VDS(^CiD]eS/EZ9C`i-$AdݱUE p?ď,;Bq;u8+_+Xg^[mQ%3;̇OS|OK_,ћ\ĵ"*X\NK !xI!혞A'B|R`);mLR1 I +6BNEH !bu K!dKQ#BG^A6 ա)i%PF&I`׎qtB ar&5n5<R4Wa?~c3K=3yVQ>I`9=U_?Y2&$<<ڙb|~m?k;,sv|~FIFIk\CO#j Hojl65fb*g6g*J'nB8t}1b3#o4զ]`#s_x@FtәsWjpMJfAŽdܝ)Ws]ɯL9lHdG4{הXM%JdRL0Tx j% xM(dd|5'%I]*Lm`dS֫%5 ͣN AiF}@A.u:k8(3=&װrpj](\@ךP0Hg{/85_/s8ht-\r!,JtaE͎=v렮E18@^bx߼b6>֬ d[`ժmoH-]UAeUHڵVmoEu|^*ʱ ZlmZ_վP֍U#B}$cYqRXFIʒWtYNP \YN"`m%*U9 X*om]hYNơгvF? ahYj'G~}ViUs_v=NV F;(,%Yͨײ7]Ӻ^j݀7*.^j$k/^SQ^J襗;Ko*h.*q2qnO>ќ)qtlsXezHl2sJhpN1b ŇZNyK?C'Ez<^N&ww:ݼ]rV`XtEKK/H(zIo,%d{;$ +.O.0NX)IEӝ-ʢ1}Rf[y͋dV8oz?ӷaot=aݿ[f>~ˇ/N9~~h?6u24犳>{5R'՗stEӁ~ r`w`30y_l籘d˖Z٢FbR΍vF/ ^kFRR#@6Cc4Jh|'X?=8z~,Uw(.u{nVnj f1A˅D9QmwywjV{{[+fQg`6yt' a:NìͭWIKu)\OD$ 9eWe)Ve{HP#1jӽ@3gn~)Y0HDz!88;c!羓:NtH(@ktQ3H-6:ڸ uŏldjG#׌z̡8 x3̅UU"Lԅ6.$!mPKPY+n`3ϭS$pNҭ%(yg,u)J% H\!beLfVX2+Õauq&cjm k8Us F] k=I_f X[Skm[lʍ&vOby1)4HwS2ANU=My)9xK^o8ۺydۡf+֤Ω NBp5; 9K _H P P`]ULH"8G qzq[ q_E'nvΊ8=Ń Ax"13*Yz0[t6].*b8QƮv8ԌČC (JMM WP*#s2yKq.)9&!O#S4}t✀=vcnjk!J\?!ȍ% j=8腱< QG}en)6;tL*pn$zᰍKo.cL%9Bd`8IJ\[.PCmf[>gT=';w A:48 T;뷸$ԗm1[ץqu(MԓbmdɍDWz\n@ U 0`vWPlAMS'Wj+"rq.x|ÐiNZOh/_'. ~P)-srvnjMo@LF Ɇ[ ì64{T<|T̔u5*2Qi51RBs2mt&Ik_/u}cVΓvu^0'Z|H=m>4Zb^nI:HpB]Gcn/`wCL I& g;e ZV}Ekhpl %m1N *8&?ΰݠ&1NAMIc XB;w~bw0>DiOFɐHL*d "Z`Jh*I)Aq!y'75BK.|nO-B&P"r c cE}_XjrEoRw$BGM.u#hyad>8y"ddӔ$6?h n}=3E;RlAnOmXAh)2%r}wIX$}/`89Rt}9ڱ$+Jf[_ #}Dxrj=~"m>"/w0"/ao!ƌ>|CeC)m{w/ G=Ucxva1jhYjqj76<fucscFpEJpG^ӐrQdy!:>E =x1ڨBWUzuZ0RzoZ6;K^*ر3,SJſu|ď5)2Y[8ZsS.{%ϳ݃Kz_Ӫ m&na|H^|qto뻿ښ?_??} tu'y5j2?F֕+YW.ڑx??}kUبtOY~]\,D.N E JMG3`W7"p p5OmzޢBy%C_m^YvRnj=]b{BՕEkCw3)h5-b/'gάMrQ[Ȅ :=&S<[c^%mҬ&/%HLGwKf6~aD0q[ܖMEWQgd6"ٯlHGcv>k@aluOߵZ8/3,TW';<2~|޿z!PiG{-Ybl@:&/ ד>Z =lKMMBҫn`@y͈g։4:Io4RQ"B $L"tXm@&h t%~ˍG)aTF 63T1s:qf0p 8F+>|qw8Qʾwܸc$I:4,Z_^u>E2-ZHR,ωPUCY]pI$%k!2F_?>9xѹc\s^pA5  Pgˊ [A*OYL#CӉy4Yw~84M8C`M;Vc@`o'=Jb̗CA$Ro99ijWBK blޭ 5MMԦ Mٞn=ֹ Eo? ۯvi4S/Asn_߆n6 CR)~Z)O@b w(e4e \e$q"O{ 4 Ԭv!ZE[׬[T荶tvN1Յ}xF@2EuwmmXy`y} ntvf^fdmJrӃK$t),͖$#|wtZdT *&zџ~ޡVlB`n$ַ0&jJAWJQ9mB*uT3n5"tT-BuFɺ$ٵPTʜU9W#y91hQ\b4X ]( -VөL:>dt2'k;n+am=kS1*A>DkB.DlJG= tRzhQg-swkB.DlGލ*ZRub:n=ztݒ;ݚ 76%CG`1oN͌<=-~upώ"SjuZݛͫo BNA>CX}V nuyI?JJqiSZkW*%!$d-R;qʧ/X @)tJ.CBJ3)H|~㤉.P.{2LwB'r٭@X7;dhl" YR9%*:Jَ!?ں%AD1"[`5Ƃ xǛ Ukd(s"km!$J"׵P[toy= pӌ_WzoC]W@At48 L)+6CB K&!P#!tX`[$ W:ٴbrX05P1e5g0g sͪ2cU"WൟhWSU{}{ <"Q[XŹ`*nj' u͸4fع* $#@ 0tL 1q@9rZRZh5FuFM*;eeq*u_ָ @ a|P Ad3)R0,s( ՖrLfylJ1fsS#@GpHRP'9'CNŔL[LRQp/bwf%ɫ݇O[k|.gV0Blϳ:yyumoG|$'£[e$}pPؙ>25QnnAv+D+X_qA2q0BHyB|䅲0 Ųmc#1>A(al~/$Bx"S8'Q5=vv l8!HdB1 [7I䠩ͩNB/Çyx7fC"{SZqgv=^QF}0ۥ P͔ҷN¦ SP"b>]y }O/]}Anݹ)(D٪:F[3wLC'ugEGGW7`i͠QQ xf:8j#Z\u6{@A ɸז~;5 \*JRsj`u#,&_rrO}*F)W1G &GR9;RuBΎ 8;@Ύ&,M48 rswコ!xTN;b[qg1z&,M4ɦJ>C %[BL'1ޭ%wμ[ō@ք\ֶ)A0bas|k!O8ɷna`0ǻͥO$KMw⦏+0G'̓~NrqLٙVG夝 ^W۔M6_܈LJj~{s .:ۦhG-n /_L c;~IKP2 g~opbA3M %R r,3,C29VnؑO~%#Mh XQJ >Jɤ\j<t>x|(]y|G*x1xR:f/h(KV8!@ Dsy0rIJ5Hh1˱ N`i$͍.t3̺P_ c͘TJR"w贛ڬf{:e6ǿm` mj@Ʊwcp ĂL0DwD5#z8D %az{I1A|-)&h4zt?]gd(3cdwJZ\Js[e"I۴:QAVQ 0rFuns3c%͘*i(S A)$"7} 1k}nu_X>䳃L~<Rk;e| H*Cz]8Cr;!~! f DqL'0!=N,1ڴϧW=3$y;zLb_}=)*2Xf_\QF껣+] yëm&Ass[Ej _8-^Dxp!8C#\|@T<\U$6.C![d?I ro.?qXlꗢ0!XR~\K`og.,zcZӔ{ 9M&Re S>ֺ!ÍU kP&dAb)P!zH@_s mv\R$"8BS(Pi#eJ^udm ԡ<$* ]FY!Vf# nQCMfAB8+NEP K;W( hl܌dĽq>F"Z;]b*FBd9.7f0nBDi@HȜABfF7/pJ>S4hBD,ټ߭ %.H(?M$aLga$مճ p08, %0 n #e Iy=PJ9 p#L; BQ}p6A%тCg<iM<-Q7Wv8W80tfVO'ՄN,PlCvOl= RĎxV-T@ z #HbtXRZ$C"5-UeW(S^M:ԵwmH; ;x?IR`'JH khĥpmо,*ܵݫ; Yәs?9],CtnaoES b^-AroSK [ޭy;]%&z [{P"%,s+M$be((Lo)VVgcI\cM)Ȝ(D)Jy7\3vI@?,w˃%ug4^1u q\==QSP={\2FgUʊD zE:1$$ECRJ6]8 mtDfbԽPJ0wY4`Rm2Sro-,E@ߧFF;Fڳ3>_G;[:QϞvW8ES^e$?viu&%4aU?WgẈ^cMz z5J4Pp&v;~g*!~5+&K N>_o^C>PĀ65 j5 nuyMZ[\\3# aX4s{yZhK^ڃ .P 01*45r RoFP \4[psu3^Dr H-NO 9'`7s`RE`^zz@׏-:PD*%[ǕYeQI|{TΩZ7ɣ~ ;ӭF:s_+\Aot Ĝ y6R|==~/f~@'Kqsolu ꩚V|ߢ(Fnx{K_;@ldٓvd\)کwQSѽmSxaS9hv:Jd]0>V1X%#plHf4*ך1|F؟yZu|XevH g:NpРĮ!3Kaa!G7[0*拢d9vqoskO3g?{g7Oӧ1Y , EOPe3!r[X1Ĉr%D }>܂oG}㢖U(EM]pAh75:AL(oO`D! OW-F*erd1A$Sp3 "hY CdN<# P@ޅnS6E`)x)y<>{ڽ_g1+(n7Gyi7mBYq;_9BB"9A/D*:%@i)>碑Va?{䶍|n*yf6q%|ϱ.zwuFCQ Jbn4ݍn7L |jrt\w@%W뀾}{4{g$n{FS˶=gj_*i5g 9j2\'V>(k"V[tB/*Z Ri`Ǖ@&0Wg0` 0ry>荾/G]KJsX?)yP+6G+ 0g_$O$I%9d$2F|;^}]Pw4xMѓՂl!]Ef|.^cS<[nydeƦѯA Um5q TްD. qcXf tG}+qmGKACUb0h&6]4R YRJFprbH.DVzȢI!%'IB x#? $1!&"$4!!q$0H,(;(t!t 0'- "0]C >aK~ȴ]  Q(Q~0JDc 0NDPA ` 1 uyw> ~`֐"5t97$zud~Ui}Vϰ>OI1񧻯nLq;QwUF%P?ʰ-EB1cŐD&8`JqEQԘ p_%bqB5P9OƧv*Z.+,n"}@ 6ZC2qT\%Wj1I%<_=G|)7:KCAv#.FXXP јωFD$!L_g)(]Fumͩ~rEcv"Lm ::g3,ɧBM23-UߍGϋοz<0}~̳ϏQ6_'oY=Y[ 9r+Q@訃:,뵶 Ջ◉ڮ}e|T!)M?ݺ(qJwk p~JÏ\\,оZ5?ZߓSQ ym@:5+c@*?vW5 v{Esm*9ksՔ g5&Bcdr%#Zj^N5X 6?9\},.ƐaW$E95|QajDQZF[WH{%KKcZeލ~ZlՍ֡9\GzggBx? f?ňSɱ2YZ2sj7+ԡv}G֝1x*s즵[Rܸ͆֒+U@ ڥv|G%)BPўVʸv !7.˔BÞe-2-N隣y9Z_kTLZ/_to_}= F GQ1I~,R͉CiL5n,4uoՏTK8]鷔i&Dj]֥H51OJH~NK).HiT냔ZJ!5RHS) )=EcjK-I)ZJWNSXZ06HiԬ~NK)!RJ 6Ҕj R*̤29ԥRa&) g)?nkǙdr^`xEDsmsc ]h(kf?JSb8kf6x:J?̖ϛCpoM;+f =Ygz=Aeԩl+%k&'_24dwMs:+jfk칙Gopn)4a;ꛀvWmex $kSFn,w GwR"0`c׳m-*(Xe󦖖IFힶ >l-Xעw&P;SMmJ1(>q q#I֚O1DsRk?>K0'B:bO'; /UOtYْۘ'E_yE˦`YSCq*1)ˈ3GWM\kFWm] H{\?q7@k$j ]^Iq=*3q}Fi8 ˽m{\8mNޠ0</wtwϿ,Gq'mKp5Yn6WnȾA>`5 o- Qb2JGK?ۣ'g2@PWս9{k }pU8s9 8W|ڶP6lO@!c :q  l>8{ˮp5D搚Ҿ r4_C3?D͸^t 5QZX|>[.]#p6Rǹf-[Ka\7ۋq~%hz1ڃwl{p4gO I#g\0p>bQ|mH㦔XÙ0׎Ex3x .38N+m>sa/WԤko(Q9iS[Pcg]*^h\%wFi\gy 2i,0Բ0@cP2/q-6vSU*4 hJRĝS2nF_&w\XYBc\)#ѧصU׌А3ْ?8^'ϳh^u_h)DOe!Wweo:G^rPVB@yH)AZ.kO:QWkN6S0󐘳/obF)R aGlsA;CAИ1{CZLDgk(a[)SxJlH[n)eO|)gT 0u>)O+%!L0H‰m<$DAϨuPjs9](+,N Z!jH n>`W^İkOc)udJ\\Ba.~%J7ϛThϣ`5XO. V?*=:mY?FyܜP i.'Qj`e$_櫘pԟEʼnI_ZסFح)|Eu[ϿZS$Rpk)%knqД~#Qb5"uAvXigK؂"l*@[wG{- $5gM"؃fmkOoڎyzu1$塇B@#j.X95FsCV-,# Zϖo^4Yq&752Wq :c^`@]gċpCAX']m.RVy"@:渇"ǡOqɦ|jYy5 ঋP@ OyQWK LM&<-ِgDK?L}/ 991kCsˉf,b1qXP/UpqLC|0~BI0 Px#H/ ʐ @@}W{%&㛰VtGD`0F1q'6Uq{TewYl4N9{&cP].!2b;Qr6!v nFt5z1eY֏ XŁtM 5$1Z'=$ ]M2bĹ}W`-騅'b o19ܩ8{Q P=C2 wFlOw1F2lWDBFG&\i[:)Z !#Ü֕zp1JIchj_/Q'1Xo,ٳrs,$gr#2itoY@_*P 3R%גo6/BC4bRBȾxF1'n+olW"pvxk7 4;ɽk`d EYD(Zd cӈ-< w$]}8Vg#[W28+p8pCIo26AtiIip=<5OZ4U ,?8E [E,VY=D$SEK}^ՃX]%Q+к͔_'8Cun#!^!NJ"][o#DZ+_rrߪ/p`ɁqxᵓDF-QD栗ek3C,i5nU#x2ey.VZ9/\uCpo i.Qq7pL[dz;VSQ΄xܑĴ]itYi\QR9=3'kY-\[/ZlO$}JqېDޯPH؎L/]%x|`kQ7Wctkyzj̉CBY +N4ָ`_rS ܥh8mը#k?qs61lٝ'_Co/w.)vm%32e&1Q8BA"虃⒀猺PY!\eJG~lXǒ9: ͡,9ScܡzͨNҚ_oFhi.gy ҙغΕ8}%Hi k̏bv L3Ó] A;.]΄V;\췄ɄxsG1 >^6@K.(3)K;0%;|JKzGjÌ5W9B8pFJߗ?Z%5B5{ s dxڍ TC{}XhВDEiQ"K3ɦ'D/Z2n@$!4D;#)_%2$9ur3f.fd1$ǭ _'m bp; Ҹ;LP{*ƗYbɒ9 *5-CX8YM{&Am|Td`埽CX=7=z;&( Za7` fɶQDU \Q]]Fx9?)/9 gϤN?y80Ί` ~ Y` >)QaQ>N}!ypڷDM)%|QGDxD+0[|R G%~ <=*B | @-x#C!с 4'UB`KDA8/D_ Jk*q[A?өR:[;qB)p}m1zڱzfćSқh/ӤǛk_8xoKaj;Lm{mܶy@ {ʄfyu/5u86<RJ֫?|J8]<<(ף˞^ՉN{( w'$'rx2go|Kz=]ʓo-FHц Y>JcS<{$2h  *Y(]2RM>LTM>Lr5yY+j' =>Hu`C'$VMN4ÿޥaVfkU U'R N@G>r}&m@<\֐qU9+H`ZHzIm=hQ@Y!fU!!z)%R Y l،tKCaj=L лFjRt gڠFad$ =Au ,,EfI:fI:Q%%_u+x ScÔBB@z@)!;EQkd:Ȳh\|Aʼn'^^ZW9ߪqU{%)V8=n9-rp+S4hC\BCS:${ -(࿢@ᦰgvq[<])"iq !jj-48S '4 |pb'"UpCn-+a{r FS/("3j-< ՚\I Ө *.L$e9'A=yO r 8 `I NMz#V@3)gOP$0g6?1zUX~'4) t/6߽|lJ"N/'C|_p}?%3(?vq̂O77'_ףǫ=h'цTK޾ݻ=32H!D}7hA 4V$;A `7xׅs_LVƋ]lH5o4rSbEڶ}!ʦͺE} D!C|5bD[  >W7] .cUp w⊂*>AbsYIavYLYYu=pLۻdU+0Ø.lĸ(4˻t mH9X+nK(P^{htQ`t+K%$ѽ=ْa](Wy PAS# ˳J3ؕ g̱zEIeA, m-c,d<]brc!9+b;ו5 mU \aEB2"Xk:H@S"ۿoQjߤ#fQ!qAAk_"醭qv `W0]NbhwUhswi|"{:yơ4׎1> ouѥ\ =gҴkԁҀsbuqEZ{VYyknNLalKJk`P>N6; ϵ_2rS+6R'7OJӱs=c |:%c]izpw嗕}'UV}r\HqR0sE +,r)5zbֶp~ăŐB)U%} D+bAqoǘWx0'S2 !6 |`^Kb*k:O<1IGgjwE Kxf9c:o*ZEM Ž'a$FeJ1$2x3ᙬ4zz+;0 j%KM%NP%)UЀY+J8.Tb:N@}(BO{)ᷚv}dܷ&7 kJD(ݮ;"<`ゅ]/'mr^:uJ@kJzѕ9MYh/3ݺSde|>:Z|^E@{B[0}S^-g3(=Ih><]^^C/@7Odf4 XK3s18Ju5&9.iK黻QoN:qv|T V:8-8{A jIL4fTc͝7~G^Ft',{19y{O)~UXLUU9r7~,dn3ʚPT훀;DSӜ*غ<' U9mDYz/oM9Z i.CUOl14dSs4{(fh)1 *4r6b[_.ce4&J`b 2ꌥN>~HA)+`"|ND=*'3{]q5Iȁhҥj洛J=jN96n{6|ڭzzUՇ?ח8(*^6}[pm}TEUY޵5q#fwGT!x_Φ7 3,ź84HJ3$s!*8&9ЍntTyD6R gK[Z^B&_\p)*XR1,[ a -'Tp@eTAÍ$70 ߀+u&Fx(c c=gg)]oyeKq:6b 0}6IӧBIa+Bo%dU5仇y|z&ͨ8 1ͬ2p[փTPʣgHݚ!;fE%1R_GQb8vvqLEc\V/nŸ/D'SU!7^2 ÑЬFX+{\>hTzD P,Dj*jr3'lYœ4>^=V"EΆec˜mJER~XWdI)ɶ;Sq' /x9J5˾f1"|k]?yAС$EɿE]6Su#jzXBOJjpeVȌK7~~|a#+LYD׀Z QcOFeJ?5hK`w`!ZWiBCg2vZ5ܽ/{a یW4U_-_k_\KyK /[`S!j5P+u.%7/{p2ZFAn:O55k[$ګ8;q\U&EN @1㨩Af` HhX䓟LgsoտF/HƫU(~ʭ!2AMX!"10fb.׿pF !FLQ^*>,^EgD'u7l ʷ5,(M2\\&$7N3/r4sׇٛwA ͔pI>믔}Utci)} X Ba] Kǩ˺-96\wރvP*m|صE+LJ@HiSpU B{ m$UA:-y\ZFWL/~88Gvcфɸ8]ZzfN=J7=,Z.>Xq9v̏V''Ϳ=8baALj9P&n-+IQC к'SW75eBJ{@t,j-1w0$KDVVKu3qh'POZWuoߋlU|/Jg,¯#Z\:?YB1]f/D!dϏkr&+'gr>QngZ gtl՜K&up0 n_E|hݐsB hNY@uZZ ɼyiAg&E$Hb |n/&'%<[o܇j (~{{~ąf?~oQWh >؛13Yg?H_ߌGj2}X0Nj1~NgDȷʿӇ_^M/s[~BȊR*_n.f64gP\K}Շ3\ڥ>Wj.Ϧ쏧O>ZMd׷a_)"YWfr>{'t>f۟m,U-N-zijq)7^b.Ϛ%hˁ[/3[!A0e$ڣϯ-[<تD*qX8nUGp1|EHbEF2_}]B%vu+Y"}ZqYPMZ AgJjWALw IkᦄP=!RPH(2=U=Tv&g R:Qp \+ 4/@0WVh01=`leJ:u/4IøX 3"k[%+ c+ڣF |Sݤ<^z{pa찅kuH'|q}FcZЗV.9\3.&Wfʉ|\sm!d Pd]yᒶ@MC! C d'Qf\P14N~@{]:04..Xơ;D @W#½[P᪅Q;5 +lP8y ')TNHF״Pb6UdHFtrEc}*NXz(}FFW4N$ 3#tKUd=۫y#~1~^Ю~~zv2*!2kMy3=]5|gy.Lr]~ޗ6ȀbD!%m(9c{Y4M |̬1's_+W#\.|{oIS-:PmL0N܄ 5D|7Ah^`dr7ńxho<8L_3eg11ĿXޙ@DbOЛ^FZ !jJ]#yCcQ^|O2qnćF\^l[w9jXPd& M,&*9mu}M; +奫|D-gyPk5w CS;[q#Y/{]}0a<;NęK %Gf2jI{S5 |#,~U]]U]]`kIfYcdѱ0gUձ&vDƉ$ |JSL8juwjR^k8nreWJE V`hms=T R = ]X1V\M1FWČubRgVT\1gr AI>Efb0,A%X>R~&W.se̊zHeMuhJOn ̬ph=O\6o4!޿]b^[0ca<${?yMasgmOlW5;[ n.<;vxFC,qڬA5aݬ1=C;5cgm?Κ\Z FJqT_Vp8- N3V}(R adU@)Jr|w\l1¿_#㕸l+r*'SJm24~GnfEU$0ZDﳸdFF/Kr[h6xf[hˈ\/O`<}<h\OT}8RdέyA>X7ܯS/Uk5ەz8͒m}TKTԬPI$ VkͨP)c>%9nV`EQR!̼TE2ac.JհrRѺdx{#a53֚DGհKSIpw~8` R a)$jq"gq=}mPc#H$JEJ$JY .#ΚhuJ?2U7\7njᗱ囎q3?_[ I i9K NЃRKŒ]I36 6ћNٱ,k`(£ASdMw[XS zJ/r5d U"Ky,d(JFXQ|6NW8ǕwΌ*j B{~XA͆<=}$M[ḱ}l_bNt. 弐xyV-8\ܹ>l\B (@VW\ALDR ZڷDZS)ۼLuN }o9㧪uҦ'bb~T̘Ũό&]/(y;L5śa:8HX$<hr<|ijpU& A|ɭX\K781^Q_vn\7!O Rf/WZ[xT֫I3* 1`R B˳{vf/DR m9 Ќa6 eOjeyqOٷFaφl7Ti,qC=.+jNV6X)-;yǙ>SÃ]rvM Ftk.[XPݒtsԿ3Vu[UZPےv.CBU~fڹ 5o_<g0%LN~n>=CH|^WU + >>ʝ1\@ a.AVԌ<@vy#'l*s5c}qߞ Lg';[%n:~%_r|nm|` #N#Ib(Fki-)֔#3 2f ƉD"!οOAtޏF0;`LQn:YMM@َ> f_3;٣=k/{l,kK?|~!a[*djRhLyL$)q4F'90/9RL@&V1,eTJ>>`5J2 '1q'>w7s/yԫ]w{,bW|U$sdbX.f;S-J]B8V\('30\y8ly3Z9l3n.bUVKL YX,$<3k Ŗ;bQ*^qeSM1Ag L:vE腜AZ;=_e=3(>u7mF5MR&Z8eԀ:b(pAG  G/{7RuRtKOiHʻݕu4~Z4 u(XJ'dH J"DDA]U)#c BpZPADp&%R(da6A"Z -M0%QHwW۫ʇ5 g,)3Rfp;Yh?/#ISit ̷lϛ'7~UH̢{4\J?|N Ͻ'Oyz"Q>xW:~~whe >^U7n?ӏ?^gsl~ p'p)_|tW 5S>Cfa :q㠰+"JȰ$g' ,&aSd^/ :ݢ#:7z$a輠aL K*?kUrHh'h#X$ASdXxJcEM6&҂g4X$Y[1E҄j^Bq )' ~β.p8^T%uG]XzS_n> mCHǻ3 sb»qP0_)Xӓ,>βl& yX*)cߛ¹/;A]O>䣽>ov淠+3EÛ>u>N斢2*N-vVP ~ث&\zNwWW3pp kT$20&Vsf"([=NM;''7+Od@L :^=f@ƨ S2~6?&xdIx \}ns~^d湳64t9p&[:xPju*M T6r+#dIw4Ҙ'kqYVQ#DqF@F:mX&]PwW [ HPr,/: 0e@!}BPv.CUy>*) 15˱oug"X)ўs덧_> t%,x6flž3%F xW@'-ţT?N27=ٙ;<V n{V;j?k-xz?;KHU7qƣ"O ӣk#Ȃ'-8MJ5mhGN Xۈfnjָɲ]nthguMAg097bܗqӯ%_]mX?Yv`n//5]g/Z\*~YUamegf2vsJCs6V^nqFnI)JUq n= hhn, ~GwsK齡ްZmV&֣fg}#R'& 0i>Q'&*](@xWPkȼmڃEGʏ]`ӄcՐiW?désS'a t.5?1[U k7Mh*KrZ]jxuN0^<0}t`Nǽσ'P{Ǡ~#rg439.3ʾ>}r',s+ĚEr\tpEIZN KTz hykΞ[xKΞBNuVi(BNɒ3׏ED=xz|?W2"Z"SjQljP3<|jYTQU^qIZ`)P) [&!׽ImcŨ`hZ: ""q;*m0QIe9Z9d-{EgK$X pYTJo)q:y 12@8n"x%H'x%G0_dd[׿>5*"&lŒ ΄MVvoY(oCZe%ˈR2 DgH4SKgXSkR_;Efk lOm~ 3 >4ֵ֭4o»S\Q9^Y??6E@5?z˨x`I<^lQjHޜZGG5r }nB!P㵠N[PlE*-(/q-\{# {n֦OxS0ϨR#En\(NAc5L+1nA@cfPLh$2rɰ葻"Zl\0,2s^\Zxj pH}Zz OaQ|˝sv/@{ d8 /y9q,{qpoi')Z?7)Y~r=[Ql`vc~r; 9fY28>qnxǔ-D|:Y mNVM~Wo{sw| 6`7xٱXsa|0JG^i/%e#mw\>KLo{yz˼k[^f=!cw0cluOyv}!"KPN9`RNyF=~>:_|>1M-lyd%[E8jjn+AھLJ pHsްv@u# .uwPpD8(8dC۸t6FFhq3oD ]5ITDTq\쬷jLFs#"%t$!$ 9}:ǫ@!FRAtp+4:pC A`BaQ8j\yj(Ɣ2=#XI4ZszWjZ՛*. ڝj`Y-1$(Ƨb3)D>ʏq@tKSNl\ǽ#i@a7{0L1K7ZD-OuJSs%'zY]m=˹Y[rWoHT&شV>}6xvp?o doɷ<~~^{`>M*7wsf%*Dz{?t;?׋j|5>Boߙ2߹b-<<@ZjRbӗnbK}Xos A%F"pwYW֮qdR@ E#Z*!m!! pf TKjMQ7ݟKHԤD\'ocd7fNHyg HF\7łC7hC&F2; يLI&@ZvV\wg5;S-DHJ2-W@f늁󩔐hH_e-Wh 6M!H[?; 9,σHa{m wjXhM 64BZ;J OPM33)H)b&qтA0CqAo-h0)'Fyl{6%UIy(]3,RB kX+ +ɝ( ڍ8Ր729 " t&yNbVؿG_B s]C{hP*xw-UTjWYr).`-?'i>x.= ;7ٿ=dk{}E NE..|{|&3B2VL>Ki,slFScc9w7as Ya=#qbPY(TId~ UǨ 86/0|H(BHG V:R`)Q::K%Qg9 E%0+JN-B8x""db4d]P,b4F^Qnu:nyYYZ#iI AiDF8H\@ Tԫkt:R>WWսo@A$`kB` % "x?dü@Nv ߥ߷^`&B߇|vlaΥ$hF 87l>^ߺ n"ćeݮԴ޻XKWe}g+ Zʝ^iw1UdꭋAnmICˆkZsVe¢_=<іf'UͼW3?qf59hms[p#ᦂ_^WVb|jz_*{,n}y1N{@ הG<Aa+ esk,95 c'xstn7.l־'=FH"(zJ;B5׃;\CʤZʑ PU"b"7 Eӷ4qUk.o̭0z+3Oh._]i{qiZ}DMr=8Xw![.?ެ/4OV]4><26R6D> 4}S W׻aDKPlF" jюłkE<^'=j3G\ 9J@WW,cg.jee խ:+oWe<.Q;ݔ=zrEdS6 ܣ͊Zy~o٢O9ƹ~8 aEE!(l ւT Tޥ=%!/^'<ކ<;T_;*!PYQ/`6U,- #uQ.WHܰ?FT}hc8<1~׏aq?f߽ۋ8*r ~iS)aIŜcɭn/n}k;KM,_z05WhD]bB$tC`kp@]Q;pUV6М}_jgz~Pjw J 7~@8ZW(=ubdQǛG E֯a>a!zIڔRYOmu3:'jA4'S>u&\@"C ⳑ/gy Z9jlWpd<yIJ u[жc#iDxE!h#Ĥ,rxV)&w<I)-F谑nFJVf %u7*0"Ziwl>aKEy)C8?u E\aKOUz9l `'4 R4OR!瑂5Xl%A,UF'2*b_Ag"gJCO[ߖa:t)n]seZ%b Fr:2APyo4:6#ԠIoe 2TbiP,ҳvhmvMz !\rzMÊەW񟮗w./|DZH f~MϮˇ~q4ſƻ Xz図| 8s_ker"dI?L_&/oGxWbZ8> "LyH+ҠG,텝]nK 2 ϗ̿=L$Gfai5kDX5+"j& $ (L{H3&5#v{WHn/ i `px`7weo=m$y`-YnKj~Qˣ`-T,ŧ$.3yǸ&[EΕ=%Ĉ;v?/b?n{f=ioe7a&6cL˴aMIbl)*aTtɹ~pM[k-l38O:xU\tуHWGiE:CmvGv7z2 ޿_^A;y=k#ƒc3:G`+̃[AKz*ZR؟ 8fͳUwhShF|°>ؽcW_l?a (\ 81N~o*iWJ[)IKMr DI.N*Q`53 VH`0hƀs\*J+\ͨzhU\:2rrSsۯWii:MOY^Z=r 7mBE?ǧ_#aIz~I.?<=*(.f7wq<{xwZ%2D&r7t6_mQcףw ez6~U]^tgc9DցD}8TCYaNK*0/Q0빽^ޭ@\qnd,!BO 8J8\˿~ߞn#Jv.b&aopͤ8eoÿq&$i6AwvbӜ\^R£nfue3mHYa5ԌRLƸe  gF`fUf>Kq:)Ene|jJBϢ/M>x*8-w]' AzAuh\U*:OB J >X@D%Tǐ_1 dIq^S .zr*4Y槳 xxWitξM`qfG&B}^`m6xk'CI -D~T1cu,zNLHg{pExǙ--3Dix@jG#OR znsC v^ryZSE4643D<7sӍ8C+-UO)WFKBN`uc.'v  Sc0G oS &b1cTsmhv`fK+ 81{0 Hi)-@d fJ5>e+(FYm-.KpgſBol İĂ„TzaB D!>4DcWKӆGHO`!=zM$ Go͍#G_/e^M T4Y4xbm+\i:uiE\` 6VY)iC@Dh Dfm.tMpeVa ;5"1" C\p'†s548K ÌO)OB+5*266*Z`T% M@aپ֫^IX#ml@511kjq0132Ⱆc,qylIErV֊ [FbVb qoDcwy5$NJbQGROX{^*pK۲ioEY,E^%SkK4aIrYZ-ESH#0&=Q1҈N j=q OF;b@lJƈ(s>5$w r d\nܖ|1!&8>XuJI (Ĝ hȬۋL1nS:vSw0n4W0ؿX\@$RW {j 7iP"cҰ@I0 g&r hȃFif;ȡI >!eHD51E'uj,5{&*N{@duFyn/"N x';ܝF"çOR>ǡ+bB K7|)HWd9뉠0ة9\{D" ѱ3|3A@AI߉ ?OO*щOp4k Jٽzx,1ߎM-JFic(U]d鱗|t/RX?F밨, 6g$e S-rDdzW3gs̫d]Ecvll¬9ɔZ`49!"ٸP1_iXΨCp 1F?Xvt9N nYθiEoG?#z}8!|Q[0E_Pq+OLj~)CSƛ4MA` MMPEJQϣ[Kؚw?<3үw=:τUY}2'JĩHJ3KX.pHL 뜻|#>vYRv"Jٳ'ş]jiX1/'g }xZ8SP!4fĠdW.vNT1w.2?tv e4مw6 \1O6noAԟEbKh#rc&KXGݚbb:M3s%yѭ'qtkBq=ئ昗jnAjΗNR +@ZvJ y-(N*Ne?WoOWPHs`@,ۛۯgKZ76y5؂c(>/rvF I+;\iDߢƆxL jh,(5Nw JEGmES}`;櫦n)*|_8'[iś-|F ZMx@+$ksל4'|+i+e=P0yt: (D2@g,,3SaaeUy0Nj[';0x-(S4RԷVWxij+hnW9%UvsjKG99w$)hOmJ}^X= ٺzSKTюy#][bJpps &OkP!A4(=Ҳe_E˾}U2=;A S`42 l$7^"+NL(ٛVԈikցMDf"F6c bV Jhs5q(l! cq+ WbO0y*?ö66`r(R_K-D:(ӊBY$Z)3,`J/| 50z=+K.4X:qIGZӛbHw|}n/0Ƥw!g!4D| x hLҬ(:e i"S-rɌ{7b4ft<@J^fWuRu6DpezXj"`cF)ZWmB;F7ŶwmI5i\9i,ӛq`p+~_~&Gg5ANJ2-_;7D0p6 (;)#+p)%l܄[4 '>b78&%wmhGm%\tNNIvt<|6OeI%'>| )z٠HQRLH,~v Ւ 07~gMb1_ 7#/4'o +# az`Y [mlycl]<}oTu0ڷ.UL2jOxlyGh5f ljpLƑST_PȜэU(>[)aR84"yL1!# 栕.Qr0h` 74{l1G}1 Y= @`lJe=9.)qd./w/ Uň-CV.wI@3,]lKZZ$U#]FI76IEҬXIa"Ol5JEũFCy޸~'@*l]eYcRI $WU@c&*,$028ګO Wp(p~9Lbs aCޚ;K;2y 48 `Fĝ N\UyWa2t?ubƆ65X0:7vv6$cW.־&^˶^9m`Enh鞵v6TnCfG$)lctGCN,Վ?Y~uNMu2@3J\tZN\ĻBd-Ahxu=fт6[Ɯ"OFS`ᚓ)S P8DŪ<-+X˂SyDU:mB(t0BSFw=&aĺn?=Fq<]ic>)O2@fjZLO'$>`nƟ0ΥEVP^8ڣǹc )i\wLZBP&}jċ^O/H}0/(KwF|hovVvUsH5T d҃Ǘ^ZRcW#f~ L~3-Xwn^oiUx RN4L& eOJq?|JDAm$7zueث,kay;Hӗ2^_ŧ|,߀O oOS%q-?wgi^ߙ IsBSb:[Jכ,#>i-}/FP6~X.d܏&q/'C0CJ$K2:AdFUa82̈Q? {O CQZ1X=Q|,}9BC\EgY:j$Bƈ(jEItq1N2Eڦ3ߊDv'o-Bt+dGn%)_pm$+rjY@Vrdҟ> m"ڈXG뇗0ƣd3m6)lnPva;=/ۤ' rCyKl(<0{uww `!7ΕNa\S2JJUHAO=^u2/e$-dTrZs +mh.Qc|ӭ}fбP\ٲge:T\>[U}ޗn *`ː;öICĖbClwH )w8(jɟ{T;7z5_ Vτ @(ks}FVNgQb&gy+8MXnqTLN#n!YtL\OS육R+GX[]16rMœjzX,UOn^fL78x=mM\Mkcoq hizpmcEB3vNQx w1٫]I!ǻ}Uy9 re7gۓסqN3l*?[2Lv8+J89Ǔ?;bRB>g;Lrўu"/mY{٫k}3odR:D%A\ `GqhHsa PՊPC}IcJFy:R4B|h8%5de(ö+lH],Q?lc*Юm賧mFB L)RĐtz>N;X rP)5ed>֡1*uJE J 4X6k#,wiÞ @MaI/s eb)D΃ 088_cPFI5zqciE\ Dkq܍޲D/cL]e:<ມtP),oݝ4?8'*: pZ/ܰWqQuOWđ>ls.WT.X?Z?½~aGxeIQ+4jӯm2=5eeeeuy(  c D\h"$"!}:cFK[da]嫱>]jCݖl$-F1 M+mƷh#w{lX0.jpt@EH5EMnrE ה <$ HD(rpDꌝuavϝWFK% ?J W+0 GESG1Q(0J)}$CFX(CeH0Vb pQaIr7Oj +j#刡pAI|G0Ga(Ecq}| c'0咁a>U`#2F`뤶)m:H~NG#WQ%* dqHG$9 c QG)GXN*(ԇ #w]Di]2׸ɭɩZ?rF&{yYz=I.&[ƷJ6}Mӎ{Q!gm$zYϟtf\eM)[_X3衝(I&qe220p&d|Lcv5O$M0%=r8NJ4bY#ۖHb2^T`ƣL4ZRgئ{trփWWvtpGzyGLdzyKY]uZiA3S~AFq{нJe},.Q($wّ^'BOתsV{^C,zwJcvTf}%_T [_ydk;Ki* we?N伱d.{8<1ӊ:STy"73ODTxS"yO!Le7\*LnO5ü<9MVq[OF-)hJrT+^AZ#ZpclKoޒ`cv>K_>Y4+c񬻼sA6q][y80Ilz Fi!E3,:4-qj;N`euz9Xx$!߸VB١+FHVѩFm&=Ín<g:$*2E19xbuvSVѩFmv!T47Uuꐐo\D)F9R h~K#)KÊ29˂dcBq&=( tGv=vTr D7DgхdY淽ӫ8!h+::aEexV! [6l`N%] g}Er4^m/u-тvld[o\`^-iJ)Æיl+V9)oڬNaT>R ,az,Yv^AjLHh mJ[5qSĪь;a!'ԀxmuVgmLaTD9Ǫ R96ܤCv Ģ m7k*܉+\XNoke+qh)e[prߚKȡT&^bhnVcGnPvidv$VwѵDȺHӒ#eF*x`mEFB9 bZPKٲ%wݹ_N=Y/U+BO#_OON ?hkNs 6swqt篏G>?~8'*oB~T8k`2(-ݸeω "yeffRY|d: b6>1Eӗ? KvJݹxFN:b*Pa򚜉yh~tWyR$XV)Yi=:d*Qw 7ѺeLKtQNi?f58],)rq6ڪ60vjDƌjp,el c; ۊrm^ǶY`aɗbP Ga$Nt fZEu1p*PԚZT-0! Ȗ:J q5q *#qE6ͷvtquIWwI>V[um0}Oǜᄏ @DWVmאxi*``KQH"8"1p[-o Odx9N\M8㠕qcˣt20tk9kvPy,jV %$#ևq5 ^`c}jG B\AcъTGD e$%}rA2MQFi1S1% 0Gow>ܝGډzq]^u=O*hS6m3 ٯʞ{ĺX8X65( J6%ZMDCd)8l_t~ g 8UIVJgiM'Aΐ,3 f.ҷJteMvd-x^Rw{^,;zw%%^LNp6XlXl+tOS{򱳔ʦLHò t yz;W;^V(FOɄs , c|H,^e\j"A m^Je.8<1 x:ڲDrLi lEz) P8$l4w/֓<UpY@%ΑAkL "_OQU[c#d#oHNS,jr"q+ >FAQL蘜'hI4̬2R+{(V=J7#kup͋ "~B㲀V+vl`6eQ w6pAbZY4$svFmgehrl+׸;[&!)uؼˍ,1Խu 0>髍,loj " FzmH0R1e)HF|\=![ hZg|5QJs?VaۡL2&w7^wac7CPCir/no6eϘ@;5&+͝0ֹG_%h+=&)R?QA&ߓi+kTXO7kӍ`$\TJ4*{={7=Tp}V#I~X'ndNkkWm\˘q9=[%8x,΂hy 4H~v~'Mڋ?f.$\~IpMLr@*OK3DdhciF\ GKNGg׼r2yN%Ӣװ~U|*̊N vrmE/ je9w"u4m2"Iۺ֧g)bX_.,ֽtSkXFyjW1dkj \ .j .D/_uҙPp k1J <* ׸/ Yz5Zķk"o@ ̨31 IsEsmމQﭻ3Ap&9sc^[;k }XT޴ H-4RmbַR}>d$*ɲG?J6I\)񊸗gLI*c֟xFX#Y=O4€('n|5VI\mEk5輙 J2H!UW°qzLeGOCf&pB 0$\;,nγ\ }ؙFiƻȜh0KQjج[cLa41p:B1nk/.y;բ tm]^=HRijļ%= UcCvGʛx֏\491%Tٖ~R-}8yZ(Zj9Dss[2(@BЁ<¿{ 񁓝x| ޝj &L :c^6m K9?~"TTm؝jԲ|1xJlMkyc%m`UL[z( y”ߩ׵*3V˃ԾDBOn[՗LnC CS-5fvAtjQyWNvlV_.S wLeh _s= N.4\ 3eR֡E͝uX2}s@rN 3 T#B-t XjܜUЙF@ @%A3Z]R.3ӟ4\ ނ_^ `zaueW&f-k] Rn¯Q-B/@ZC/zݕvvKq3C?H ܟ ̗ QW0گ9ņjʐ}1hLB}Jέ1E,JC'P]x x>Dz0!;jW}%bvZ ́]f'sr%ŽՅvZ ͙;[Zv ~%!ө$Sfo!Ofy”ߧ״ajywh^$~[}L6B9D0zv'RU :(n &]}~>>=9B7S9GMѝ>Z1ewΉʛ&?*ګ*,Y/6剽 {nѧlpndsyfGK+f3m7 ̢ ip~~ bn_k9>Dӗ>YH:4|!Kw~hLY̕~|hn)BFţ4rRZT1%iVj=V11F3oc?WJ;O~$[ mSI }:-D /},NwZ,i?f58],)rq6ϦVc'jLDKlިR2/eYN. [?KnI9,n__wǫ(|<_n7w.?ܸۯIc j#?HnpX쑳69!qy,||[f=ِ/{]~j W_c,PjЎY;H{)*/Bo855x1錥kbwtd*Qw 7ѺeLKtQN,a厥O*  y(U1$PgT+NM9(M#PqJlj.}("B$]˃R!P:Znb&Jam} cH>4 7ڙO^02 'WR߯_~VǾ/-3 %L“ K UxDŽHLV`Pi[|l@6NvGxϼ3v&=D53R!0" &o2'o-]żA (h;Lmr[}<-JyB,@UPM˩wUDu$zZm 6m4Kh NpUR#@$G$vjs 'aγXT p*rE K39eZAJ_L5g㟷Ed|zRt/RzRy] 8텔ETK /k\Ji!n/on]ϣeY)fC5y\`b9{IF-]K5#TtՁ E1 {OB;yxç6lt:* aBrPV!)A* )85$ށ@)Ը;(,b ɕWQqPd)J%T@R8>՛*$=i.zŷ`[Ytj 50ٲ`r՜El-0[li|5 _s|p}[x"ez)V7$;/v:J^C<?|. ME"k#h*6Fa+NuMc/z `H;8\S}TP_R#L$@>Igk`C8Nz8Yv %C,"E?.DKH(( jjML9 3 f$I-EZܨnCST J n*M$i !4v&[$ Jar 8;g g)IsbU?dj{|{91S)=#il5LoNeIߕȔdJyR跡4&IbD,9C Y75DǀD։uZ({bao) I}-p:'hQg)3. +om_ϝ[[-pnskk]nmʖp:}{F SPI%D`J攑180:Mm LtҀ⡚׾OUj˝?~X }ӼZb uS%Cg+q,([Ӽ;_}ɫ+^]0$YdR=xz 9af M{>ʆ"1pk>9ɹ!??ªsP'1TJ\Ob']jRB0miB?4<Ϟ9m6k7I<;/MXw5:Zò]++`+`CeZ!K+pSJCc<7P;iշ[T q)|~ÑJ۹8wfm:gOfcrO&*q]r]4?Zͺ3z{0,O3SO0j4a csǗ}dm@۬dY.8 68<٪sQpK8~ȉkroɂ;Uװ\~v9|NJg8w#%0h!;9d+")[xJN,:GL@o$}҇w;SltW {٣=~9kZ9׫_2pPtb L8R929RrӚiY7pAP]B-IՄ ~[ >oc!>j9X2y`GهSMuQ[i*ˍs:޿e됓3ClLOED~cZx,Z1zřNo?x= q1vvUm93׫an샦) c;>r&,VuSيؿWY y]nKS.rsM]!3# T>i% h#:eD;vڭ)BD3hY_Sޝvk1ڐ?6bZ[ I'nMy":MQGҋδ[Nڭ 2%@:,Z[E!x"0՞S]q}ZlrN>'ۊΧ p bާ . v̍Elȕ +WvAj{KfL|a}~Ǩǻ$S1s|'x/% ]fу}iS5O%7b;Mgm2u=g,(ݳzk[>_YtB F]v˜Α,>5Vn3 AG~&|'?fBevz[ 9uMeaUîL#2B" &9'54 ηj2Z:^Q*.۷~3*z_v ͝-_)_haH-j\܉خ# e,╂U'w Bўye~N)ƂH'8CW֯YQiQ#:QW/E 唎{pScNKu2B߶J+_-.^_k g6~vhk1g3S"DNJs°AXDc-˙b iPh8"1B5u<0JŋTQ:xV'n$Ċ@Ő#Ŏ2h77YVEnL&i-d~ @1U?$SFŋIye-νSnƃw>}NpfSV? vA%N%TRKCRBOK޼%э@'}OYDf#<})")@f L~9pO(iKbp%yfa>clwǚ9"'sM.?fX[^՛E(̚>z@_*0hv'f|1w]m3_txOf:~t?>]:ך)gB4"$BF($T C8\@B) ^!,;`#A&!%xiX>;  [> ⽽Yh+cZ@'W쳲Ho7k&_ ,r  Heu0:d%z`2z/gK/"pp$N M#@ZV@@4#!fT s3Fjdb$0Ib)5|VaW1J$WydZPbMZJ:[iÅ;S4ڠ ЄM7N@ͼ|ټ Cds;F/gd1z%>!>rh(x-Y~ (eS+Q4jlmvWd4 >̳>Uq:I thH<rAW?͟\~aU9`d#KB|U>x%{&e -}J„!v376wRua;/)~BkG9Kh AwYޏjRMnO42jhivQܺIZ ~;xYo65rXfH{P)@D2R*6XSN`$va{'(z~QlP`ѹ!^^C 'V̝CΡqsք\ կe(sk] @ VKqh ȄbHo:8la5`g-jh_Ž"fϨ4?Eb4 FmTOrkSRHMj")g}0ØM(&>R.\Z@_ eH.%WQRi!53OoRVU??\j}?pu? v{hFqn\U|&g.{Jծf5[6ދi՜p{/']ߏLg:.~|`U߾^QS=ުQJZ\W9G cYyM*>s};mq}ݹjFX;rHDHY¿m: )n a~=ɹ\=#hug Bҁ{4(-aDP]Xk6|3wtƖ@1Jj%+=0" SE< 8W:[aW,y)Bc>POh RfBVd+R:+?_ܾOGXl]Z%jn;W}GcZ{%B.?g/bIK1Ƴh}5y9'O%>1G}$7dS'g/md<T /bپډ|Xf$e]Mƽ1:(6V8boGv1ВqMުybuoMIQ4IvH;(ċ%u8}ug$ԚyTUr⫯1Δ%:? ;7 "`v5})D"mBkttVg?+H-&6Ҏ(t7Mo'z燳5Ga M /(uTE(—q#p5a9X =H2}ru}]!'.<zR=PцpG޴nߊ8ЗiGL*Lu 9Žd?ajr{qڥ&.M~ vnlV-/OcPS9A wW(!_~ǀ1?= [RKuQVcHR?!睿k3>5;*/i}ZW䜕3z>=x(\zbLIf}9U(OсڂjUܟ5Ω=\'TofcA"n{ NtJ]sCHv;߾35;)wI!T̫]ۣdsȆ( x_ZVk0ƀ(`s-:5q#@(sf8 fm~yu=?"8pE D ˲$Egh@hRʷ<╜7(^ik;j~k\)ٷ#6&T N=g48"imF[4ѯMjAԆBOע Eb'o:I&Û:\͆@K+5SRUo =`ş±,<0BZ ,"jH|k٤y v i{[RN/TXaQC&M'{Czå\:_yî|ef[VIDDI%IIdgUX%uQH'ЬPrgKd$d=vL4-qrLgᶯd5q{V2VWUۣ 檅u{U32ykќ}f&as qn;nQ<y3+߉To|l<"! q6<\ܺjk"9ҋz0L ^60X^R5n=v[io~6@&]x?OzɊI,ʉ Bl,XSk5D Cӗ&_2QAknYGV\>5T +An2GezŹõ :/%&'̺~-Er},bZj{)]| %vsoJOp6h29.ql#U Wn{&;dI ^n`mPIsgr634Qיt G$KﰽO3]N,m͆z0p%Ġ\ 4Zk+`5fg1)< [FՒo+"?!7 (Lښr- Ќu+X(ɂ,%[ZC[+vKC蒆ғ2R| ;Ks*:"1zO !tt-efp㏍5``fzVXTC=\x%%m*4TKf֌vȐ0E: %"R.z0jWz@^D'ХDE-f49Z3-g-kTɧjdUIMz+xlV-O^CCﯓ7Uٕs/Rez5)W՗ mXב] OS@x\ͨuf&Myf*jj;/g~ +~RG;V2UkUs;sfjJΨb#:mhh OJ d]ٿ(ʼ2=ק O,u{fU`6OҎUp ui@+v񦴳r2Rh'dYӛHT4ZBf 4@@ _1 }pyH/" іL#G"yE2t ҁ)tQMIF k*7@sQx>¶ WK~Z #AmY!(qnD%+QBa,yRzk%œhHfoMUMNT.x4owBd6u2.q9+_Rߗ ̥*?lN*Uz)j6NjBsؖl s̯{/D]>GD 4%Z pR-n L B!6VB!t&;&z֡ (LUA,tcd ވ@"ӑ*{5)ĻTǡPMY<7ECB@bїpjzܔ62mV7J~)\`$j(D9;tPo}!ٹ:0K[k*֎R |-V7r+Fh ֲҞK4 ´T/L BP$=|)>-ÁLpyk ,D$ $UM*}RǰM,e\`薦GJQ匪ʤ<* H9hXv>xvo1YGT=:Qq_M Fr|16%$_ 7I)7jD7-DC|vh&4EO ?}p@**φW`\j|Sl }l} =Z잼²7^(> p~k5 :'"u yQKz틝R,&5g W:=rϲ-Z3ƽ,V5%ԲpMs;K %̨c=-ocX+ŋἐFzVZ[b&UHAr,E5?qZN#଑mMoroNg$,F%!3֠<;&+aکԼfXmַWjޗR6 ^+ס*7dPTUʯuD&̦r[b\W#~4QHdUIxh:|)*j[ݏoH7_"4#j<*L&Ioz<jھN/1)j|BQo RCo[{vϮ~ڎVD+L#ekhj$FD :%C )5g";=h%A{(,biNjO<*.^,(v $amƀZ@Ɨ^jLF傧ocD,-C%1.lO%$:|ɚ҆Rn0R;dg%ceҾb)Q1%i颖AҊC;Hp%YjaU"XS9~NL FDr3ElNݶpf Ҫ3ڃ:Zu_|?@L8rqјk9dcj)Shd4rUI]}=pwGַ)E[=A#Ǵ/\]K1)xDGK&O|LO:yf5Wm.qj`v'pM .cH% <(%%T^oмQ,0= 6j%J]!H ּoY;ĺޡd}D#{<6myRB?mw.?,҆EQϬ DJrrK% di]_φ,=J^|Ivx#|w7f<VR$8'HDP;x͔NH.Y $$-Ff nmY]T\'( @JQQ8nJ)s5& !:9yDB㦩a2$45rShL͓Q]woq.g鹘J"Ύe3v59R,]^R>E1Xf6H`moR)oT!Uv\K,ug&z[FRD7n8%wZHm?Fhtx@c$ܻIt-Ќ2V:!BtuҲ\0oRרU(wp e@Up-=z 2Z,w|;*V\_oΕW;u5]EUnя5v^4tvt9ot"Yb*}[`4Z;Ra]hEZYJ(mlC~<0CT3_N>&`:|\zDM!eXCxV)oujuQ S8TZg N%$A9UXaCY'ߏ;S7~]x'QIPwvOlՎmgqFVռ-r]\m>Ⴉ;Å}Z)pmn4ET=CkcLJfSߝ.u!YQFunKe"_J6sڙnѝ8_ c0rmy'"""\.8/)0ࡔA[!@[ApT:8Ri&\`sWA;',O\/ E;rn]OJ4fwvfi 6~¬ԭX!jh:pԬ!UY)U뵀keJ`ƕ )W:Rpn{YecB*XZCE.iTcbnJQz 2XF$ef+$ <]3RR%eHdGNEHFBkT#w}H~ku) hCM.E%Q,FtƖ^ Tj/(HB1mBEa4mС{>e}ӉsN^8np7 JRUzb?|=4k$rz A¤ygX^y-mJA;/ L鮂9.6.jM ):b)c<댗z4¡ i ++?Հ q?P{;nT`BHɚ;!W^^б@@IE/g#]28mwE'%●ڑ,UF0o ㆽ ˋvyXd-T8nci%Cc\JV9;fbOLB{q\ bw{Ʋvr\:cblL.NE\?֮ŢJUh|qOKfz:Rs4=8oqAJqc7=yYOz>ӧ&^+z'O4R䌜wmRv97ZS;+Y\W}ttM.ꬂK8N*0 ^ xEHS9`ZʯYb*3$sB IA$TV34&) ,k|.b~g-ژ}Y7;|J_1]5ݵ^_O?]כa{Q2$44(;y~$x6 bxb)< {_#Zl{4ǿZXbAw,f^=/q?Br:; yV3DzmjQtuLJwBz4r}Ғ|"D=}v;VCȎReyoZmΰۼi?b-qk7qbPx> J*)ěLРw ,0xGn] &Bk(!YiʠXQP 7ѲDҔnC],L;&Sì tb(S~mZgTno`pVt6SI\J*ڨJ*%x_>¥s! d ɂvq9L\,L&? 1 e"lDA8!:nF˥/L&@HLkLrL2x"-4Aر%vgva?<ݞ,2j1I|0N]ׄq|N}e傕Tߦ$BIY}Rr*yڹTFsHɅh>D><~‰՟!MLl 5lP |lbEgH&a9aG?1 X5tgީ27^eBזt/rk=V#PmYW|t\jYgt`1t7܇+q܎wЧ : fM,'^(vPSHr099\|-4'*8SߚcD o G_:mh瓼Ot*)qCuDU$s]}V2$ % ?Ii1رG?{7Fz,i޹,jbRLߞ/z4~Asq9FSkEb;ZPTy9[p? X\h%:"}S_@LhqgJdA Ί UO yw`ܳ#ypИzP= )*0IHUP됆}˓3k?"&*,2o~ 1)9R95K\TOz{Nn^)I P6QAl< & ˚J ѳiGwzy7QKYv\[]tY ^pN,w<95Q)3nǮZbWn])HoW~f 4GM]R2Es*x^"P^x"B^H7Ypә14)?Q,2p2Kk'sjH694FG y° uvSo?Ǔpԕ췣+ (VuڱvkGuˋK{@0noiE]-Q^޳?oZGIxy|N>rz/_%Ls\ޭr`]TXNĢᱢ X: p4Dgx[utq'X 싹FOL'%iR7ry)zc|v)܂=~vQ̎.>w:Vk/Mp91eV'f7mIm|9N֜G7_ 4Z'󿞜/Q81;d( -vz*oI2hT ф[^k@s󎁮"p=L.vcw{h+Q4z+U{2m)C;6=O7<}c*UuPjQr=JWmj7m tsՖV2hC;NĺoX@-Ynl깜g/n .D2hc:}Kt{t68lOx6|peSv+ߣ;QA]kvv}Zѭݖm &lрʰݭmTmΰ~-WI=$f3 x@fP&57 uE^[m]  `;# T3@+2*D>ED{ah10tWBT3z^E`R&c,ZYl~z7rgܯ7&M_a; '(4A6 7iG%]F+VJ(Jkeη`#`hIDTFpPG ɟc݌tM$nPҞy"I%аBR_ppzzf1 s._4&Ft`eIucAy/\ITQ)m'<Ǚt]r%su;J[lt2tޟzA$o|)hDžOAs-.46SJ ; N ۅܨH壽Q3.ʺ#0.mӝaJhd,h QqjZA;9i| jw\h "8(`(QRHU}E<_Om\SR{¾D,znTha9lW9wv8lc'B%WUgWgi)< J/͋+ [n b'1>:%rFZsvl}G-(PcʢIFD꛱P8BS m +AjXgnX_^%Kw Qڷ2:{-8 HUEZ+ޔN;뉒=*;Rbԏ dm:~MEӈk pgZ큾=&!7c;!T ̦Lb"fuL|ļւ2jbWAnnEMu)T]UXrAƵ 2KET`Nb#@I2oj E2 lE/պZ,#}&:)0]Gm c:@9紐;17pV;Tl$PTJ[Y%ek2b.4ŐrQL<92_.?ɋYES#o]^JtͲtˬtCuˡr/LZ_Ӯu<-l_oٳ^Us9_er>"iG!+);٭6l(mYH*Pc9ÂuBדU>1m<fc]薺*@vG5*24ۘH9SrʒPFYb12O2aᲩ Jtdrt, C1gAzPNМrݧ H7.G[vSSkV߸ IhAVdgH@ňTkԐ%-Jtx" %!~6eMKeVSbMb-z!rMA</q)3Qt17iNr|ؚAp~3,61t0kdc̔LusKXW[X}:^̇‹v [_%Ѷ|vqei #|!1+yx>=\XCkMDjlRuKmڸ`Sg]+PݍOqk7pnjn.h4qlYI-jj䈞zm:D4+eJ"Y͒@"HB<ʿ/zc 8k7_\\56`c6,ET̜ʜ:hmM 7BuVI=ZGBs%pn [Z*+ Rv?([M v-8'yE'J9% .K%[Zmq]\]j! PV\ejz}:WjEl=.;6.;츣xer 8ެZ|w$dء> U ^?Y0LNϧ(~ @8fܵjlIW?tzP;LJΏRf[x (-6.(L@jR\AA2?jRaw{{NaCAqYU$>:֌=a|GƩ㽍qy{]. hF^^/O<_+^=-(8-g3zW?ǝennuͿFx_^R){/n^͗I͆\-b'_}CK߃¬ zKG橇Enȉqq$ Dϛې1 MW "S:b)s'*"[I^vs;CnRqMʅD"rl!)a8ÛCۥ$pFhts‰ L.SR%6˩lRhV}d`sYk0Pz! 5Kv4!M*`LQ&q-W'|7L{i5ɩkէZwW~#|7ާ|>l2S7Se^Yy*5E_]bE"ڍK,Su|x8~ $l_;b} i0zr"i1h߅m˥yose6ۛ 7^"AR{d_WWϜ"y&y5dJ!aMmc9q+sFo0HAtGUG& H槉!1*Rͩ"bw锕QMYlbj!qo6¹YAa3U5ՒS1k)9GE4K-}&ՂSqki zb9R,˴:jx2-8Y% 9/TKJ;i1j)#eZTAh)#eZZS-8:%5rQ\$< 9/TE阵 "i`DVg̡//.^Lz*jUc#qyb/Fr}^a8]/S5%d!Wy9T_ʿ7O5.Q|ՇэMwKUB8L`Vh4+Di=A,1TYk͐9˰0$K6g'PesV@'9LGi PiLpI=o,#a3Hv H1;\Dvõ5+5']=XVVkՃ|0+vwSOPڳ2> Y#=xY}2$?|Q@ *;Dٻ,(XǰvCE̵r֊66AP k*eO!ᄏv]! zgF\DL*|JruȗY~sVGx0婖BrU TaX/8׊.˴/yCicRl5:D͔?ÍX]Ԛw|&plqp==ϥ칇 + Xp`X mpL[l6Q%jKVM𐊿tMI2cv\(=pѳOs5$R=pZ;a€l耂rʊC)H!T^oлX5DWTM{;|X8M*Nmaj#i L|O:IILxQI\F+}Q烾^o~NnPH& Sa413Y(BhDƧ{͹GmԌtPK[B[Q @2bcљF8 FjQ 1 Lc62e22XmB!B–FCK2)<'AR9;[\u |sf跦>[UY(`j׷zM qd8?q%xv&k6(.fZL|nRJȵg,ږtwuUJOܼ'+0:C _>T{y<j>a:/Nx.G׊ǥi$*KϣCBS$g҇>JwCfB1wƸPq}w&Ւ)+qaS1M 1W{ [/M[偳^xxh isIC1&=eWӸ!儱mW;.vpmjZ]q+G*n[E43v=J9c[PIwWIX[iļeJ5/Y4x]˧WR%Z nnYFdR7"UQa(kX6&}q.F}U[ީЦmkY8;_*] ;qŷ&*jpP{ܹPep.l &`ZmdN7+þ{ɐ kzu~s#|C;goL5#BR)~ݻm^E膕:FmNdXKTKԙc&|G~;4?Nnmx:|3w݌;k7Qv57KU?7<?jnEnz0V5a͟uV<O'D "OՇqݫCtfPzҐv)!~^ne=9hݺʠDuuۣ1n& [7WN:{֍f5hݺʠDuug);noZ>4o *€Llfr%OXLeFLX&EjWbǝL eg(9vse6:' B7ZpqqZQ4. pžI5,ǨXi)Nb8# 2-B[JI6YBtT.?OeT!p- 6\얢vc!Qم}A\閐H> ٜ%H$}ϒ-ڂAsHt.`y@^˼`5ٜ ,Q9\{;z~IiwT{cx8+`MJQULeNv@[?}lB[o'qzz^yru)>T韡)zL jI\tdǵV~x$*3ӵP7P̍SHPė]6,yꢓ۴D$:t 夏立-! 8~0e0 [B4;bY~j0hDevH,Sq%GBȐb.$ۯ[NS ]UV%E' wJADЌZn#{mB'烂 h-6 " TEҊ(@>n5` 40 PQ)#5(DYh"uI<"6 g*JW ' Q6j LYlF]D8dul\izyi-gV8}3%UL`ͪ!G $[T1}ĺӪ*¢.ӥc-YJr$ mX+Q9o?T&5ZZY+hǿR .WP_1tPf*WkdQvWZBO݋ZS2Lm6I] ]K0- ]0ў@bIeou׆5K+&A:@ԡDkmX+c8S9nX%#dxQg*/D E(jPW&-mA6E/8hK1?CŰbHU&UʕBah;M f pM F-S%\3^k "KߧB! @ +^J+i 0aXʲ $$&q.0I8y52:&eVm/zl).e;J~[5ta@aKBnnj.2V8UrԪBE:B%=\H]<+W 4 Ε:F,VVdT !tZx #6x@޵5迒}$6S[*^c%Hr_ز#9luKV2;eM@\L\]/R|0#1BmLC(ktpdJͨKh-!boC?Kg{XW^uуbVw pú:5Q;֩COJTIsS;`S,Рh@*^BX)8 q&\ `GP?X_Ζ=WBcҙߠ|NOkeRIW H\ZCE49L ]jO Uh5༽Gz zt BࢇF??#'qIоDJimH֞"Rse(YT%NSR\K+t/KOރ: sp85A>[_;cfxXaq|}> QNPV'#̼Ԃ q6o_x(B|Lf*Me*Z^ -ƶ-;^C*X)'^#KI^,MmͮzKT>cKW>R%J@O]\}{x#U>(+蕏Ugw[^>Aksk~ԯ{6Kcz}nzߚ.zwO= i eSsŶWlNz|̿NV]ڦ^RmHh<*@iNR6h]g$׎ c;zV~sd[vڌz|M쟖F~Z4*7b7p.o Z)4?A= %KYP2 ebזz 'D=cvWG>d3;wubpF`snpU;θ?SU#Ԅ[6} 4LYvϫ*ZxRUQb($G#W,@2ףYz-rbo2CꇪN& (MMx "W&&UepVALA(p҂/Q^W/'EϋSrz>cP|ƹVRmvFAݎ! l~e Z^#A˥i֔o˻fIb~eeRgt X^.5lNoXvSgwK؊ӽm2Zec' ^5y@֚DyMLMQzSaFիWи2漂ˬ+豮t#tNSx/++l>G¹QN*5}Q91Eń;IՈ;=X1se%ʼn$ "^`s9aI{G3;K3vg >E U#b=~V}-Spҥ/ }-=7&"\&*aY{\ !r w$˼UWHf@{Tptvn7=|lw_~D~2ZJ\tҞ<}q.(9^h(.iA([%.⹜KLC0pGW._N:[q;,ghkWWǽA7^ Gَ}bxpgm$g=g$znpay9'kum7rGUejL(>ޱy{t޿f\yB>lYOaFj gj'n?O -g01ceV/V9 }Ki%&]#DV?b(L͞ Sk&պj h_8 kpX!wYVr@R§5ǐIlĸ$zG*x4)#[t)dH+6 H<$1u/F Ti{\[}ݭWkX-K~~Wmgevb"1^i.zW&˔}!8;]SJ_; ebn)r\2> Y!J,wY Aka4y.PtkwY!.@ěC_}ݬi0 UݻX`)Ys|V5B +{. -+"O)ytL II6kio堠%sg~43oTIuENIrbCbMѧ {(}tΌgDˠ:| fNRA5 *jQ=1h\E;&͍;d0>Wm.pP&Umc2p <{i(*#k5&UYo ౔$oէ+7<;O!b\p+"X~Ԅux쾉*I&ET 4/ā;<hۣg@Q[-YE^Pɚ/WIT{"M\PtP0P20BREB*J)tVy.Jݭ~1G\sYйt}JKZbuW}gDc aPpbFL]҂IQZkIsG|$PkL>(.ѿ%2"0 rR.aN7u[`$"e"W?3!uQPs JPQpua(@i9_p3h :9P$J)s^Bf|Wp<#8o s) mbxcV@2AT"S=> 1}[%9srIV8h# VEtѐIU{9]!u;8)r ))+qt;vtf~1缞Iq}Wu^EB*O,i'L|hiųu`2$+3ECQ;hh W8v!ɲzT }6я-"*sR9fePi0xRUk{Z ܀nsO zgHV&S?L'H'W/ x:=Z>Es&3L%R,ř3N]TXG)| ;aH=Y4ZRYX c>c`Nݪʅ}o8%!.1p͜U +@!ҽx'^q;B$5pP\F 7v$}U̳s#|&J0+\&_*!7*zf˧&hxz D}-I P]WZD ڶy! 371gVCkFu/^ ] M&F_L)Rrd1 #kpw͒rؾ'>}4cqvdF\#Yiow15 &$ګ0m+w_VLi*`5U*ք$EXU$}@Ja^-hQ  Zc`t@+CǾ& ED1>W~0͘\wc[o7d0i&%K Y"3p)!`Ax(ebiV([HcIiQm%2dTP0ʛ3Cj-{ˇEȬG0}gzf *Q$2ͣhƮv]nاa8'8+2Rs D,3!P*HolwM9V Gv,vĄU^V;ͱt_F"qȱ>=6{?DҐX5"%1ǽfQ:よ<{'V!͇,}α/o_)̓U` kwDK5ݦ&7;etҍ`]8z|7 f];éߐN#JPy ИcJ{H_!ifI64()Qe:.ZY%m-"_ܑH wt]Uăk=-7> އ#Ygd8z0*)񰮲~vy>(th)jw'ݞww@eAd/S:KKBqÓ98Aiොz>㬖lRYOov+ZQx'5kEI5ܲYxY鉙#0XiGǒA(w~}.\K]] z n7BRHL/'#2BSԤjYD.1-\ h @k,_hVL1Az-B{1mɋV/ଳ^ {Em_Sh)Wo #{b(WYH]%awKxScU..gy3H]Aب[NUUR y l^Ѥb.-1Q *y!f?Ǭ;O_3`ְ7O,m,)OIcwʭb}b#N,b„CGV=Hntcu7Y6B[f hSE[ R] ְ|7u#%8Xz܈ tNwaL0"k_3 JKgY0X{,tgsC\VGҝ9Ʒ?&K@[!c?v$5Ac(ѳb1b. ùr'Z> {3ER Yλd]Hީ]o6 ^&uPؕ[|Xg]ܳŚ2}:m"#`v$lXp}?c2rڽv-[λ9''N|,%XeQVJ"&dzpx.a#z=MJŁ{~e;!c05v}t] )̶gAs +>|_e%myb;hsţ'* !zeTM!z7^ LKiuxDYAA>0]^6a]6:~*AzkpŐ׈ŶCnT>ZMNK.YwyY1l 1Jl`D'+cXUp M~Htg ia09Q{e#2W%؄Зl&\b"VyVY}x}ϽC,d29cG5=_^46aQm~VRCrP_ +)j-vt]_ǣ%|s7|2,ϏcTٲUBDGeMYXhl8\])aeϯ*ewUm -òy-Q>'i/Kp`x[K<8ߠԉv-b0*m5ID/d{{ޤZRhT{@{ptfX|vowY!u So@a2kleHQ"}ub_ԍ/zᰇJÎfWNZ%9)~I/gK{1sWэ`gBi-~K0δsz뮙J/;\_ih,*q:k^IlLg6E"8U'+-Z7[}%ѕh4_z\VKogX]\D`0k, `[,ô%feJh%}~7@mv.,B'ye> '2 J/~d*OC$W |T7įTT 6Dգ@i)\AssGwck`\rαхIPgh6Z 6E\t9||]![a }NJ񵷾L䱃E(3`nFae=`+,?App[L74x%.L^`eབྷhoDB?Af}8ih'T .h94li(fk[]IW& T`ZF"3rV6@ hak3na;ss6[LrrV^,d]Њ9fx4e.ENpp1};,˻Y{\KXO:ݬ~+n6tkS*-?sI?J l 6"÷D\H^Lx u慸V{T`$*zKG$꿶F{GF1dWx4# dwKIWص.ʢ"+e 9Toi9c8f9>ݯh:(Z&.bldɔ#mTtF2LrK.Vi x+BCGO}!K*5B+},X5̋lASsW8`yh-k #j};Nv>z:;+垳F*)K*gTӦ!!ɚ X8$a3}&'hÁ(^R|2دm!@R/e /r +c[ B*+%e-/AH#g^z:)V =¨Ri EdBIj?{ 7xɃQ<%^dN놻2E)ANi, Daa%G7u h)%1Ok!uLjc Yhg}PK563sv;j!,K57l.U .pFX(,&#|y]kcܚ|>{]LUEHŲD œ*Y B+9#OK[4z|QMZA3"ۑk+`_^7 *ZB}nVEp{'6)Gm0+P VΧdzK %&Rq X.!azu6p@I/tV7+!/ϔ .N##_?QiQr6XoyQ/&ѽ`ZVx;nu}a\ϵ *޽;'T`I0+D mpPKn0;Z*rIo8zqg{#p>D4=_%Gӂӷ|X!d7W>cz`Ȣ2a9B l|:vЈa醧cTFHEoe s:FNa֨dG`xrwTɷCx%|TrvSJkԲrtjE '@i,1&:I)k*4fm߬d٧U}(AK3A)n㟓?Fh'kHΗb}xa&os}ck 2Jܻh~P@>4+ܝ^}u ~'0zbph5woӭq;ȕF%sڳP*'jd dZꇱ&}nzcn`ώ\I6pӁ쀺% )V3FyfXw^pCI|~Hm$6`SBL"eűhꌑkT\ʥRs*רg%󝁞_4OnoqmF4l>jS (qv˽4-7^FO!]hb~_渫!%h.o?/R8Kz[}";jx>x1♺ܥ~d]h"_kr!1텚?&>o.@Z-Tu b[S.@Y} Ů|[u #˳(n;l۰p6F!"R(5~%)HP;ą(Be >TYU-UIhNs?PދBcC gTZH8ଗJD# ˞rFhq%uᜧNDu֮dHr0B XĥqU4NrTQ"aE:evETf[*a"[ZsW˵g *BIp3AD*]~8%WLh8ҚHd4MX!Jb']-6qk2(.:?^}οSJ)MhSh1Ɋm-[#I ['Do(2 Ji9T)8RQyӫ̆[]RcA@"N4pj 3wiY9E(:JfS@&2FDrAPjKAR|E\T\TC2 jv6BF U2)KbAXcņhDB0ISJ!)o٠: 4A}ň&ɷvӃ6Zk˧{vV]luۇP1E=~ͮ<|r{釩˾u&<|abr][hnM{$n3׽WM aL)0GC=I|0Wǃ!RIF(wdϋ6VL9BTxA:(KE2D+NyCMTڝ8.,a \ #Q}cƀvIKbId*5Kҍ5\bî|ݵZHԸ/aJ+,az }gxD% $| k$p6C 7,Q8eJ7ƣCzrKػuPEJחp`P:͂ҍ[0˕%-ǫufḰu9I[EģW,C  e3ܹxC Y53 /FLSMXsէ*e5W0i+\,I;- 36Db2=b^^)szQ#Z%Uwz:+xMEm'?lW:a#}ja%>̨v 8NjXc<lXk폐Ebk4V qUJa[2wû*s?z_5?W4p2zյIp}֥F꫗(k\tlmƻu=׺ 8Š`We*fo :(sL}bdB`Cxqk9뒸 9fb~!W =8cN$OF NW=kc,bl<>ɴP>y67y/:?}S_K?sOhѪ(SE=5/9i5luNy[吣'}LQ# !Ct!㮫;v ڭ.BSe/5V?!P5WZvCS.h<NwnG"|w{ڭ~@B^9Dk` w ;z1dM^i-05uO$W )7A{Q|{3WHD+W;?co_85>alP +UnT-Zsu {uͺpqr*!<.lZ=^\ i ͙i/f ˟n]h*D5Lul kO ZjqUɠ2L0JZ5-$,%X!w;wbb/NJ5ʱP6]u$ TJ]|Ԗ|SŻv^Sb"1V@Q *ڑj8huBA)LԮ[j́yErA,Ԉ8 j]\O}Gm^O jy)%:Ooj7(wGnuy(ݎg{yW nM CŹpKP]nuy(ݎE^7nM C26k:-y_[;D`C))$Xw)d =w9XWrR4BJQP) [CBmA O٩xaROVͧۧ)DS.$"L8/tX+>#˹_=jj_Wküg-;toGXUdU0GA U?' 8C>S ?C?[nͤ4z?Qm Oړw9{{},X>py&o\%'q+$ T-*H{שڀ݃% w$l0-@^Zk6<供'&bʥ )e LJ:Rw.S q8UmmbdOfp9F+y" J[A l)壥Hh3,e[v»ש֔}F)a(zB'T_S4?6JTxoԝ@0z9apN"{K TR=-g]z(< ԟKYaRҌj ;Q rt PQONTPRCzj?=?HNvar/home/core/zuul-output/logs/kubelet.log0000644000000000000000005437122615140077033017707 0ustar rootrootFeb 02 09:07:15 crc systemd[1]: Starting Kubernetes Kubelet... Feb 02 09:07:15 crc restorecon[4670]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:15 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 02 09:07:16 crc restorecon[4670]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 02 09:07:16 crc restorecon[4670]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Feb 02 09:07:17 crc kubenswrapper[4764]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 02 09:07:17 crc kubenswrapper[4764]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Feb 02 09:07:17 crc kubenswrapper[4764]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 02 09:07:17 crc kubenswrapper[4764]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 02 09:07:17 crc kubenswrapper[4764]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Feb 02 09:07:17 crc kubenswrapper[4764]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.607616 4764 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.610755 4764 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.610775 4764 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.610779 4764 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.610783 4764 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.610787 4764 feature_gate.go:330] unrecognized feature gate: PinnedImages Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.610791 4764 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.610802 4764 feature_gate.go:330] unrecognized feature gate: SignatureStores Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.610808 4764 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.610815 4764 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.610821 4764 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.610825 4764 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.610829 4764 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.610834 4764 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.610840 4764 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.610844 4764 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.610849 4764 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.610853 4764 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.610858 4764 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.610862 4764 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.610866 4764 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.610870 4764 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.610874 4764 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.610878 4764 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.610882 4764 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.610885 4764 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.610889 4764 feature_gate.go:330] unrecognized feature gate: Example Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.610892 4764 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.610896 4764 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.610900 4764 feature_gate.go:330] unrecognized feature gate: OVNObservability Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.610904 4764 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.610910 4764 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.610914 4764 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.610918 4764 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.610922 4764 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.610925 4764 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.610929 4764 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.610946 4764 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.610950 4764 feature_gate.go:330] unrecognized feature gate: GatewayAPI Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.610953 4764 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.610957 4764 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.610961 4764 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.610965 4764 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.610969 4764 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.610972 4764 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.610976 4764 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.610980 4764 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.610983 4764 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.610986 4764 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.610990 4764 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.610993 4764 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.610997 4764 feature_gate.go:330] unrecognized feature gate: InsightsConfig Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.611000 4764 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.611003 4764 feature_gate.go:330] unrecognized feature gate: PlatformOperators Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.611007 4764 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.611012 4764 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.611016 4764 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.611021 4764 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.611025 4764 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.611030 4764 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.611034 4764 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.611039 4764 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.611043 4764 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.611048 4764 feature_gate.go:330] unrecognized feature gate: NewOLM Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.611052 4764 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.611057 4764 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.611061 4764 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.611077 4764 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.611081 4764 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.611085 4764 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.611088 4764 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.611092 4764 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.611173 4764 flags.go:64] FLAG: --address="0.0.0.0" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.611183 4764 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.611191 4764 flags.go:64] FLAG: --anonymous-auth="true" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.611197 4764 flags.go:64] FLAG: --application-metrics-count-limit="100" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.611202 4764 flags.go:64] FLAG: --authentication-token-webhook="false" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.611207 4764 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.611212 4764 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.611217 4764 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.611222 4764 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.611227 4764 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.611233 4764 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.611239 4764 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.611245 4764 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.611250 4764 flags.go:64] FLAG: --cgroup-root="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.611257 4764 flags.go:64] FLAG: --cgroups-per-qos="true" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.611263 4764 flags.go:64] FLAG: --client-ca-file="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.611268 4764 flags.go:64] FLAG: --cloud-config="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.611272 4764 flags.go:64] FLAG: --cloud-provider="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.611278 4764 flags.go:64] FLAG: --cluster-dns="[]" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.611289 4764 flags.go:64] FLAG: --cluster-domain="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.611294 4764 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.611299 4764 flags.go:64] FLAG: --config-dir="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.611303 4764 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.612351 4764 flags.go:64] FLAG: --container-log-max-files="5" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.612369 4764 flags.go:64] FLAG: --container-log-max-size="10Mi" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.612375 4764 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.612381 4764 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.612387 4764 flags.go:64] FLAG: --containerd-namespace="k8s.io" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.612393 4764 flags.go:64] FLAG: --contention-profiling="false" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.612398 4764 flags.go:64] FLAG: --cpu-cfs-quota="true" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.612404 4764 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.612413 4764 flags.go:64] FLAG: --cpu-manager-policy="none" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.612418 4764 flags.go:64] FLAG: --cpu-manager-policy-options="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.612426 4764 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.612435 4764 flags.go:64] FLAG: --enable-controller-attach-detach="true" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.612441 4764 flags.go:64] FLAG: --enable-debugging-handlers="true" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.612446 4764 flags.go:64] FLAG: --enable-load-reader="false" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.612451 4764 flags.go:64] FLAG: --enable-server="true" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.612457 4764 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.612469 4764 flags.go:64] FLAG: --event-burst="100" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.612475 4764 flags.go:64] FLAG: --event-qps="50" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.612480 4764 flags.go:64] FLAG: --event-storage-age-limit="default=0" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.612486 4764 flags.go:64] FLAG: --event-storage-event-limit="default=0" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.612491 4764 flags.go:64] FLAG: --eviction-hard="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.612498 4764 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.612503 4764 flags.go:64] FLAG: --eviction-minimum-reclaim="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.612508 4764 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.612513 4764 flags.go:64] FLAG: --eviction-soft="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.612518 4764 flags.go:64] FLAG: --eviction-soft-grace-period="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.612523 4764 flags.go:64] FLAG: --exit-on-lock-contention="false" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.612528 4764 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.612534 4764 flags.go:64] FLAG: --experimental-mounter-path="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.612539 4764 flags.go:64] FLAG: --fail-cgroupv1="false" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.612544 4764 flags.go:64] FLAG: --fail-swap-on="true" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.612549 4764 flags.go:64] FLAG: --feature-gates="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.612556 4764 flags.go:64] FLAG: --file-check-frequency="20s" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.612562 4764 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.612568 4764 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.612574 4764 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.612579 4764 flags.go:64] FLAG: --healthz-port="10248" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.612585 4764 flags.go:64] FLAG: --help="false" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.612591 4764 flags.go:64] FLAG: --hostname-override="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.612596 4764 flags.go:64] FLAG: --housekeeping-interval="10s" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.612602 4764 flags.go:64] FLAG: --http-check-frequency="20s" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.612607 4764 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.612613 4764 flags.go:64] FLAG: --image-credential-provider-config="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.612619 4764 flags.go:64] FLAG: --image-gc-high-threshold="85" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.612625 4764 flags.go:64] FLAG: --image-gc-low-threshold="80" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.612630 4764 flags.go:64] FLAG: --image-service-endpoint="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.612635 4764 flags.go:64] FLAG: --kernel-memcg-notification="false" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.612641 4764 flags.go:64] FLAG: --kube-api-burst="100" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.612647 4764 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.612652 4764 flags.go:64] FLAG: --kube-api-qps="50" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.612659 4764 flags.go:64] FLAG: --kube-reserved="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.612665 4764 flags.go:64] FLAG: --kube-reserved-cgroup="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.612671 4764 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.612678 4764 flags.go:64] FLAG: --kubelet-cgroups="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.612684 4764 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.612691 4764 flags.go:64] FLAG: --lock-file="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.612696 4764 flags.go:64] FLAG: --log-cadvisor-usage="false" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.612702 4764 flags.go:64] FLAG: --log-flush-frequency="5s" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.612708 4764 flags.go:64] FLAG: --log-json-info-buffer-size="0" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.612717 4764 flags.go:64] FLAG: --log-json-split-stream="false" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.612723 4764 flags.go:64] FLAG: --log-text-info-buffer-size="0" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.612729 4764 flags.go:64] FLAG: --log-text-split-stream="false" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.612734 4764 flags.go:64] FLAG: --logging-format="text" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.612739 4764 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.612745 4764 flags.go:64] FLAG: --make-iptables-util-chains="true" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.612750 4764 flags.go:64] FLAG: --manifest-url="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.612756 4764 flags.go:64] FLAG: --manifest-url-header="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.612763 4764 flags.go:64] FLAG: --max-housekeeping-interval="15s" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.612768 4764 flags.go:64] FLAG: --max-open-files="1000000" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.612776 4764 flags.go:64] FLAG: --max-pods="110" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.612781 4764 flags.go:64] FLAG: --maximum-dead-containers="-1" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.612787 4764 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.612792 4764 flags.go:64] FLAG: --memory-manager-policy="None" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.612797 4764 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.612803 4764 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.612809 4764 flags.go:64] FLAG: --node-ip="192.168.126.11" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.612815 4764 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.612829 4764 flags.go:64] FLAG: --node-status-max-images="50" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.612835 4764 flags.go:64] FLAG: --node-status-update-frequency="10s" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.612841 4764 flags.go:64] FLAG: --oom-score-adj="-999" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.612848 4764 flags.go:64] FLAG: --pod-cidr="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.612855 4764 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.612865 4764 flags.go:64] FLAG: --pod-manifest-path="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.612871 4764 flags.go:64] FLAG: --pod-max-pids="-1" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.612877 4764 flags.go:64] FLAG: --pods-per-core="0" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.612882 4764 flags.go:64] FLAG: --port="10250" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.612888 4764 flags.go:64] FLAG: --protect-kernel-defaults="false" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.612894 4764 flags.go:64] FLAG: --provider-id="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.612899 4764 flags.go:64] FLAG: --qos-reserved="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.612905 4764 flags.go:64] FLAG: --read-only-port="10255" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.612910 4764 flags.go:64] FLAG: --register-node="true" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.612916 4764 flags.go:64] FLAG: --register-schedulable="true" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.612922 4764 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.612953 4764 flags.go:64] FLAG: --registry-burst="10" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.612959 4764 flags.go:64] FLAG: --registry-qps="5" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.612965 4764 flags.go:64] FLAG: --reserved-cpus="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.612971 4764 flags.go:64] FLAG: --reserved-memory="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.612979 4764 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.612985 4764 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.612991 4764 flags.go:64] FLAG: --rotate-certificates="false" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.612996 4764 flags.go:64] FLAG: --rotate-server-certificates="false" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.613001 4764 flags.go:64] FLAG: --runonce="false" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.613012 4764 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.613018 4764 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.613024 4764 flags.go:64] FLAG: --seccomp-default="false" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.613029 4764 flags.go:64] FLAG: --serialize-image-pulls="true" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.613035 4764 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.613041 4764 flags.go:64] FLAG: --storage-driver-db="cadvisor" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.613047 4764 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.613053 4764 flags.go:64] FLAG: --storage-driver-password="root" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.613058 4764 flags.go:64] FLAG: --storage-driver-secure="false" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.613064 4764 flags.go:64] FLAG: --storage-driver-table="stats" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.613069 4764 flags.go:64] FLAG: --storage-driver-user="root" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.613075 4764 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.613081 4764 flags.go:64] FLAG: --sync-frequency="1m0s" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.613087 4764 flags.go:64] FLAG: --system-cgroups="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.613094 4764 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.613103 4764 flags.go:64] FLAG: --system-reserved-cgroup="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.613107 4764 flags.go:64] FLAG: --tls-cert-file="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.613112 4764 flags.go:64] FLAG: --tls-cipher-suites="[]" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.613119 4764 flags.go:64] FLAG: --tls-min-version="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.613123 4764 flags.go:64] FLAG: --tls-private-key-file="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.613127 4764 flags.go:64] FLAG: --topology-manager-policy="none" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.613131 4764 flags.go:64] FLAG: --topology-manager-policy-options="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.613136 4764 flags.go:64] FLAG: --topology-manager-scope="container" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.613140 4764 flags.go:64] FLAG: --v="2" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.613147 4764 flags.go:64] FLAG: --version="false" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.613153 4764 flags.go:64] FLAG: --vmodule="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.613159 4764 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.613163 4764 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.613285 4764 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.613291 4764 feature_gate.go:330] unrecognized feature gate: PlatformOperators Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.613295 4764 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.613300 4764 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.613306 4764 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.613310 4764 feature_gate.go:330] unrecognized feature gate: NewOLM Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.613314 4764 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.613318 4764 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.613321 4764 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.613325 4764 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.613329 4764 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.613334 4764 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.613339 4764 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.613345 4764 feature_gate.go:330] unrecognized feature gate: GatewayAPI Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.613350 4764 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.613354 4764 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.613358 4764 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.613363 4764 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.613368 4764 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.613372 4764 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.613376 4764 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.613380 4764 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.613385 4764 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.613389 4764 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.613394 4764 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.613398 4764 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.613402 4764 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.613406 4764 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.613410 4764 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.613413 4764 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.613417 4764 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.613420 4764 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.613424 4764 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.613427 4764 feature_gate.go:330] unrecognized feature gate: PinnedImages Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.613432 4764 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.613436 4764 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.613441 4764 feature_gate.go:330] unrecognized feature gate: Example Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.613445 4764 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.613448 4764 feature_gate.go:330] unrecognized feature gate: InsightsConfig Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.613452 4764 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.613456 4764 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.613459 4764 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.613465 4764 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.613469 4764 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.613473 4764 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.613478 4764 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.613481 4764 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.613485 4764 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.613488 4764 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.613492 4764 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.613496 4764 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.613499 4764 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.613503 4764 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.613507 4764 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.613510 4764 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.613514 4764 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.613518 4764 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.613523 4764 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.613527 4764 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.613531 4764 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.613535 4764 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.613539 4764 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.613542 4764 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.613546 4764 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.613550 4764 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.613553 4764 feature_gate.go:330] unrecognized feature gate: SignatureStores Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.613557 4764 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.613560 4764 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.613564 4764 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.613568 4764 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.613571 4764 feature_gate.go:330] unrecognized feature gate: OVNObservability Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.614429 4764 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.625456 4764 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.625507 4764 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.625586 4764 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.625596 4764 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.625600 4764 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.625605 4764 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.625609 4764 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.625613 4764 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.625617 4764 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.625620 4764 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.625624 4764 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.625628 4764 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.625631 4764 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.625636 4764 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.625642 4764 feature_gate.go:330] unrecognized feature gate: NewOLM Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.625647 4764 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.625651 4764 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.625658 4764 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.625664 4764 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.625670 4764 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.625674 4764 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.625680 4764 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.625685 4764 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.625689 4764 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.625693 4764 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.625697 4764 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.625702 4764 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.625706 4764 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.625709 4764 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.625713 4764 feature_gate.go:330] unrecognized feature gate: GatewayAPI Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.625717 4764 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.625722 4764 feature_gate.go:330] unrecognized feature gate: PlatformOperators Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.625726 4764 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.625730 4764 feature_gate.go:330] unrecognized feature gate: SignatureStores Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.625733 4764 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.625737 4764 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.625741 4764 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.625745 4764 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.625748 4764 feature_gate.go:330] unrecognized feature gate: PinnedImages Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.625753 4764 feature_gate.go:330] unrecognized feature gate: Example Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.625758 4764 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.625763 4764 feature_gate.go:330] unrecognized feature gate: InsightsConfig Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.625768 4764 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.625772 4764 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.625778 4764 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.625785 4764 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.625790 4764 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.625795 4764 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.625799 4764 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.625803 4764 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.625807 4764 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.625811 4764 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.625815 4764 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.625818 4764 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.625822 4764 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.625825 4764 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.625829 4764 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.625833 4764 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.625837 4764 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.625841 4764 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.625845 4764 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.625848 4764 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.625852 4764 feature_gate.go:330] unrecognized feature gate: OVNObservability Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.625856 4764 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.625859 4764 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.625863 4764 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.625866 4764 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.625870 4764 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.625874 4764 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.625877 4764 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.625880 4764 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.625886 4764 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.625890 4764 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.625898 4764 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.626072 4764 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.626082 4764 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.626088 4764 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.626092 4764 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.626096 4764 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.626101 4764 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.626106 4764 feature_gate.go:330] unrecognized feature gate: SignatureStores Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.626110 4764 feature_gate.go:330] unrecognized feature gate: GatewayAPI Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.626113 4764 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.626118 4764 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.626122 4764 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.626125 4764 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.626128 4764 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.626132 4764 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.626136 4764 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.626140 4764 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.626143 4764 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.626148 4764 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.626152 4764 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.626156 4764 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.626160 4764 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.626164 4764 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.626168 4764 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.626172 4764 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.626177 4764 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.626181 4764 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.626184 4764 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.626189 4764 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.626192 4764 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.626196 4764 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.626200 4764 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.626203 4764 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.626207 4764 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.626210 4764 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.626214 4764 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.626218 4764 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.626221 4764 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.626231 4764 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.626235 4764 feature_gate.go:330] unrecognized feature gate: OVNObservability Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.626238 4764 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.626242 4764 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.626246 4764 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.626249 4764 feature_gate.go:330] unrecognized feature gate: InsightsConfig Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.626253 4764 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.626256 4764 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.626259 4764 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.626263 4764 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.626267 4764 feature_gate.go:330] unrecognized feature gate: Example Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.626270 4764 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.626273 4764 feature_gate.go:330] unrecognized feature gate: PlatformOperators Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.626277 4764 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.626280 4764 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.626283 4764 feature_gate.go:330] unrecognized feature gate: NewOLM Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.626287 4764 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.626290 4764 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.626294 4764 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.626297 4764 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.626301 4764 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.626304 4764 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.626310 4764 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.626314 4764 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.626319 4764 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.626322 4764 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.626326 4764 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.626329 4764 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.626332 4764 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.626336 4764 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.626340 4764 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.626344 4764 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.626348 4764 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.626352 4764 feature_gate.go:330] unrecognized feature gate: PinnedImages Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.626359 4764 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.627625 4764 server.go:940] "Client rotation is on, will bootstrap in background" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.631539 4764 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.631642 4764 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.633330 4764 server.go:997] "Starting client certificate rotation" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.633364 4764 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.636612 4764 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2026-01-09 15:57:43.305565718 +0000 UTC Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.636717 4764 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.658911 4764 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.662718 4764 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Feb 02 09:07:17 crc kubenswrapper[4764]: E0202 09:07:17.663045 4764 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.217:6443: connect: connection refused" logger="UnhandledError" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.681134 4764 log.go:25] "Validated CRI v1 runtime API" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.715246 4764 log.go:25] "Validated CRI v1 image API" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.720133 4764 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.728567 4764 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2026-02-02-09-00-35-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.728605 4764 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:41 fsType:tmpfs blockSize:0}] Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.739852 4764 manager.go:217] Machine: {Timestamp:2026-02-02 09:07:17.737921039 +0000 UTC m=+0.671645137 CPUVendorID:AuthenticAMD NumCores:8 NumPhysicalCores:1 NumSockets:8 CpuFrequency:2800000 MemoryCapacity:25199476736 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:ca92679c-8580-43c3-be51-62441e93d22b BootID:31973a70-ce17-4811-9a14-f5476c6979db Filesystems:[{Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:12599738368 Type:vfs Inodes:3076108 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:5039898624 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:12599738368 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:2519945216 Type:vfs Inodes:615221 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:41 Capacity:1073741824 Type:vfs Inodes:3076108 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:429496729600 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:c6:f3:a4 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:c6:f3:a4 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:0a:cf:b9 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:ef:cb:3c Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:82:d2:b1 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:de:1f:fb Speed:-1 Mtu:1496} {Name:ens7.23 MacAddress:52:54:00:49:89:1a Speed:-1 Mtu:1496} {Name:eth10 MacAddress:8a:d1:b4:d7:0a:7f Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:86:e2:f4:39:fa:ec Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:25199476736 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.740106 4764 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.740318 4764 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.742051 4764 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.742260 4764 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.742300 4764 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.742520 4764 topology_manager.go:138] "Creating topology manager with none policy" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.742530 4764 container_manager_linux.go:303] "Creating device plugin manager" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.743112 4764 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.743149 4764 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.743333 4764 state_mem.go:36] "Initialized new in-memory state store" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.743426 4764 server.go:1245] "Using root directory" path="/var/lib/kubelet" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.747403 4764 kubelet.go:418] "Attempting to sync node with API server" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.747430 4764 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.747455 4764 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.747468 4764 kubelet.go:324] "Adding apiserver pod source" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.747481 4764 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.753103 4764 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.755065 4764 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.755550 4764 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.217:6443: connect: connection refused Feb 02 09:07:17 crc kubenswrapper[4764]: E0202 09:07:17.755687 4764 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.217:6443: connect: connection refused" logger="UnhandledError" Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.755691 4764 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.217:6443: connect: connection refused Feb 02 09:07:17 crc kubenswrapper[4764]: E0202 09:07:17.755769 4764 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.217:6443: connect: connection refused" logger="UnhandledError" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.756688 4764 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.758046 4764 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.758073 4764 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.758084 4764 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.758093 4764 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.758106 4764 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.758114 4764 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.758123 4764 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.758136 4764 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.758146 4764 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.758155 4764 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.758170 4764 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.758177 4764 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.759032 4764 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.759537 4764 server.go:1280] "Started kubelet" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.759885 4764 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.759892 4764 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.760532 4764 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.760887 4764 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.217:6443: connect: connection refused Feb 02 09:07:17 crc systemd[1]: Started Kubernetes Kubelet. Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.762961 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.762993 4764 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.763330 4764 volume_manager.go:287] "The desired_state_of_world populator starts" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.763354 4764 volume_manager.go:289] "Starting Kubelet Volume Manager" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.763465 4764 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.763674 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-19 02:50:54.767706889 +0000 UTC Feb 02 09:07:17 crc kubenswrapper[4764]: E0202 09:07:17.763844 4764 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.764230 4764 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.217:6443: connect: connection refused Feb 02 09:07:17 crc kubenswrapper[4764]: E0202 09:07:17.764282 4764 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.217:6443: connect: connection refused" logger="UnhandledError" Feb 02 09:07:17 crc kubenswrapper[4764]: E0202 09:07:17.765336 4764 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.217:6443: connect: connection refused" interval="200ms" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.765491 4764 factory.go:55] Registering systemd factory Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.765514 4764 factory.go:221] Registration of the systemd container factory successfully Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.765805 4764 server.go:460] "Adding debug handlers to kubelet server" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.766557 4764 factory.go:153] Registering CRI-O factory Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.766576 4764 factory.go:221] Registration of the crio container factory successfully Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.766652 4764 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.766678 4764 factory.go:103] Registering Raw factory Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.766694 4764 manager.go:1196] Started watching for new ooms in manager Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.767248 4764 manager.go:319] Starting recovery of all containers Feb 02 09:07:17 crc kubenswrapper[4764]: E0202 09:07:17.768525 4764 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.217:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.189062c1bad78a21 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-02 09:07:17.759502881 +0000 UTC m=+0.693226969,LastTimestamp:2026-02-02 09:07:17.759502881 +0000 UTC m=+0.693226969,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.778876 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.780562 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.780674 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.780756 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.780823 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.780886 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.780987 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.781068 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.781161 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.781247 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.781332 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.781413 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.781507 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.781588 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.781675 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.781756 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.781837 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.783153 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.783175 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.783191 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.783204 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.783216 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.783228 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.784067 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.784095 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.784472 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.784572 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.784659 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.784747 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.784829 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.784908 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.785913 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.786025 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.786115 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.786217 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.786313 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.786394 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.786472 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.786554 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.786640 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.786725 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.786800 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.786879 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.786980 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.787085 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.787170 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.787252 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.787339 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.787422 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.787510 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.787590 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.787704 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.788117 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.788538 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.788638 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.788723 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.788811 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.788892 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.788993 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.789087 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.789166 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.789233 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.789309 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.789382 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.789463 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.789535 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.789834 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.789922 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.790018 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.790104 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.790229 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.790312 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.790390 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.790457 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.790525 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.790609 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.790686 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.790757 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.790830 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.793199 4764 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.793360 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.793764 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.793856 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.793956 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.794053 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.794143 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.794238 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.794372 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.794461 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.794546 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.794634 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.794726 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.794806 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.794892 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.795053 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.795145 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.795238 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.795331 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.795429 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.795521 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.793962 4764 manager.go:324] Recovery completed Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.795601 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.795957 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.796045 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.796118 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.796193 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.796272 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.796359 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.796440 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.796517 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.796586 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.796661 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.796738 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.796824 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.796908 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.797009 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.797099 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.797175 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.797352 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.797718 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.797797 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.797922 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.798020 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.798117 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.798189 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.798264 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.798343 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.798607 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.798679 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.798760 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.798835 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.799134 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.799271 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.799834 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.799959 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.800085 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.800174 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.800257 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.800346 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.800429 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.800504 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.800577 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.800658 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.800726 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.800798 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.800874 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.800966 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.801043 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.801114 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.801183 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.801261 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.801330 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.801395 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.801461 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.801532 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.801627 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.801699 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.801764 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.801828 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.801893 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.801985 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.802063 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.802128 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.802195 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.802461 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.802544 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.802620 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.802691 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.802757 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.802828 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.802893 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.802997 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.803076 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.803153 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.803230 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.803307 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.803391 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.803485 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.803545 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.803647 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.803711 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.803766 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.803830 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.803886 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.804023 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.804090 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.804145 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.804210 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.804265 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.804318 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.804381 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.804461 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.804536 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.804597 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.804650 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.804703 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.804758 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.804813 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.804873 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.804956 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.805024 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.805080 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.805136 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.805201 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.805266 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.805328 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.805380 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.805471 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.805531 4764 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.805591 4764 reconstruct.go:97] "Volume reconstruction finished" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.805641 4764 reconciler.go:26] "Reconciler: start to sync state" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.807452 4764 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.809110 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.809355 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.809445 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.811150 4764 cpu_manager.go:225] "Starting CPU manager" policy="none" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.811177 4764 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.811204 4764 state_mem.go:36] "Initialized new in-memory state store" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.821484 4764 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.823320 4764 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.824213 4764 status_manager.go:217] "Starting to sync pod status with apiserver" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.824311 4764 kubelet.go:2335] "Starting kubelet main sync loop" Feb 02 09:07:17 crc kubenswrapper[4764]: E0202 09:07:17.824413 4764 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.827317 4764 policy_none.go:49] "None policy: Start" Feb 02 09:07:17 crc kubenswrapper[4764]: W0202 09:07:17.828012 4764 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.217:6443: connect: connection refused Feb 02 09:07:17 crc kubenswrapper[4764]: E0202 09:07:17.828091 4764 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.217:6443: connect: connection refused" logger="UnhandledError" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.829005 4764 memory_manager.go:170] "Starting memorymanager" policy="None" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.829113 4764 state_mem.go:35] "Initializing new in-memory state store" Feb 02 09:07:17 crc kubenswrapper[4764]: E0202 09:07:17.839110 4764 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.217:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.189062c1bad78a21 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-02 09:07:17.759502881 +0000 UTC m=+0.693226969,LastTimestamp:2026-02-02 09:07:17.759502881 +0000 UTC m=+0.693226969,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 02 09:07:17 crc kubenswrapper[4764]: E0202 09:07:17.864154 4764 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.885170 4764 manager.go:334] "Starting Device Plugin manager" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.885219 4764 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.885231 4764 server.go:79] "Starting device plugin registration server" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.885687 4764 eviction_manager.go:189] "Eviction manager: starting control loop" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.885729 4764 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.885983 4764 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.886116 4764 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.886127 4764 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Feb 02 09:07:17 crc kubenswrapper[4764]: E0202 09:07:17.897798 4764 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.924674 4764 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.924825 4764 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.926060 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.926101 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.926114 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.926307 4764 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.926559 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.926656 4764 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.927456 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.927498 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.927511 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.927656 4764 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.927827 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.927882 4764 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.928084 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.928119 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.928128 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.928493 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.928529 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.928538 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.928672 4764 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.928776 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.928811 4764 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.929170 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.929211 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.929223 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.930444 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.930477 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.930479 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.930535 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.930497 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.930568 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.930664 4764 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.930809 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.930861 4764 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.931245 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.931264 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.931274 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.931407 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.931426 4764 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.932285 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.932327 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.932339 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.933282 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.933323 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.933347 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:17 crc kubenswrapper[4764]: E0202 09:07:17.966144 4764 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.217:6443: connect: connection refused" interval="400ms" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.986257 4764 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.987440 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.987588 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.987673 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:17 crc kubenswrapper[4764]: I0202 09:07:17.987805 4764 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 02 09:07:17 crc kubenswrapper[4764]: E0202 09:07:17.988398 4764 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.217:6443: connect: connection refused" node="crc" Feb 02 09:07:18 crc kubenswrapper[4764]: I0202 09:07:18.008304 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 02 09:07:18 crc kubenswrapper[4764]: I0202 09:07:18.008352 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 02 09:07:18 crc kubenswrapper[4764]: I0202 09:07:18.008376 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 02 09:07:18 crc kubenswrapper[4764]: I0202 09:07:18.008427 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 02 09:07:18 crc kubenswrapper[4764]: I0202 09:07:18.008480 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 02 09:07:18 crc kubenswrapper[4764]: I0202 09:07:18.008513 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 02 09:07:18 crc kubenswrapper[4764]: I0202 09:07:18.008532 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 02 09:07:18 crc kubenswrapper[4764]: I0202 09:07:18.008562 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 02 09:07:18 crc kubenswrapper[4764]: I0202 09:07:18.008588 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 02 09:07:18 crc kubenswrapper[4764]: I0202 09:07:18.008627 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 02 09:07:18 crc kubenswrapper[4764]: I0202 09:07:18.008649 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 02 09:07:18 crc kubenswrapper[4764]: I0202 09:07:18.008691 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 02 09:07:18 crc kubenswrapper[4764]: I0202 09:07:18.008733 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 02 09:07:18 crc kubenswrapper[4764]: I0202 09:07:18.008763 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 02 09:07:18 crc kubenswrapper[4764]: I0202 09:07:18.008786 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 02 09:07:18 crc kubenswrapper[4764]: I0202 09:07:18.109876 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 02 09:07:18 crc kubenswrapper[4764]: I0202 09:07:18.109924 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 02 09:07:18 crc kubenswrapper[4764]: I0202 09:07:18.109952 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 02 09:07:18 crc kubenswrapper[4764]: I0202 09:07:18.109970 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 02 09:07:18 crc kubenswrapper[4764]: I0202 09:07:18.109992 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 02 09:07:18 crc kubenswrapper[4764]: I0202 09:07:18.110011 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 02 09:07:18 crc kubenswrapper[4764]: I0202 09:07:18.110026 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 02 09:07:18 crc kubenswrapper[4764]: I0202 09:07:18.110042 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 02 09:07:18 crc kubenswrapper[4764]: I0202 09:07:18.110060 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 02 09:07:18 crc kubenswrapper[4764]: I0202 09:07:18.110076 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 02 09:07:18 crc kubenswrapper[4764]: I0202 09:07:18.110092 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 02 09:07:18 crc kubenswrapper[4764]: I0202 09:07:18.110111 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 02 09:07:18 crc kubenswrapper[4764]: I0202 09:07:18.110127 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 02 09:07:18 crc kubenswrapper[4764]: I0202 09:07:18.110142 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 02 09:07:18 crc kubenswrapper[4764]: I0202 09:07:18.110166 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 02 09:07:18 crc kubenswrapper[4764]: I0202 09:07:18.110570 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 02 09:07:18 crc kubenswrapper[4764]: I0202 09:07:18.110628 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 02 09:07:18 crc kubenswrapper[4764]: I0202 09:07:18.110647 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 02 09:07:18 crc kubenswrapper[4764]: I0202 09:07:18.110677 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 02 09:07:18 crc kubenswrapper[4764]: I0202 09:07:18.110707 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 02 09:07:18 crc kubenswrapper[4764]: I0202 09:07:18.110732 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 02 09:07:18 crc kubenswrapper[4764]: I0202 09:07:18.110757 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 02 09:07:18 crc kubenswrapper[4764]: I0202 09:07:18.110779 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 02 09:07:18 crc kubenswrapper[4764]: I0202 09:07:18.110783 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 02 09:07:18 crc kubenswrapper[4764]: I0202 09:07:18.110824 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 02 09:07:18 crc kubenswrapper[4764]: I0202 09:07:18.110804 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 02 09:07:18 crc kubenswrapper[4764]: I0202 09:07:18.110854 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 02 09:07:18 crc kubenswrapper[4764]: I0202 09:07:18.110882 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 02 09:07:18 crc kubenswrapper[4764]: I0202 09:07:18.110901 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 02 09:07:18 crc kubenswrapper[4764]: I0202 09:07:18.110910 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 02 09:07:18 crc kubenswrapper[4764]: I0202 09:07:18.189116 4764 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 09:07:18 crc kubenswrapper[4764]: I0202 09:07:18.190427 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:18 crc kubenswrapper[4764]: I0202 09:07:18.190467 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:18 crc kubenswrapper[4764]: I0202 09:07:18.190480 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:18 crc kubenswrapper[4764]: I0202 09:07:18.190509 4764 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 02 09:07:18 crc kubenswrapper[4764]: E0202 09:07:18.190895 4764 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.217:6443: connect: connection refused" node="crc" Feb 02 09:07:18 crc kubenswrapper[4764]: I0202 09:07:18.263301 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Feb 02 09:07:18 crc kubenswrapper[4764]: I0202 09:07:18.290781 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 02 09:07:18 crc kubenswrapper[4764]: W0202 09:07:18.301542 4764 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-49b32e6d456b3590d318c8ff087ff6583e1e26afc831d6e68f6ab2bf6dd34d02 WatchSource:0}: Error finding container 49b32e6d456b3590d318c8ff087ff6583e1e26afc831d6e68f6ab2bf6dd34d02: Status 404 returned error can't find the container with id 49b32e6d456b3590d318c8ff087ff6583e1e26afc831d6e68f6ab2bf6dd34d02 Feb 02 09:07:18 crc kubenswrapper[4764]: I0202 09:07:18.306911 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 02 09:07:18 crc kubenswrapper[4764]: I0202 09:07:18.321983 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 02 09:07:18 crc kubenswrapper[4764]: W0202 09:07:18.325020 4764 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-a0fb8d7faa9c9906f5977401249ca564d82a0ec26bc3cf673a7aa5ce4b70a283 WatchSource:0}: Error finding container a0fb8d7faa9c9906f5977401249ca564d82a0ec26bc3cf673a7aa5ce4b70a283: Status 404 returned error can't find the container with id a0fb8d7faa9c9906f5977401249ca564d82a0ec26bc3cf673a7aa5ce4b70a283 Feb 02 09:07:18 crc kubenswrapper[4764]: I0202 09:07:18.327858 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 02 09:07:18 crc kubenswrapper[4764]: W0202 09:07:18.329567 4764 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-63aad3c7a5e012411fcf7e81e9a5a8325238cd78bdd4bfb4a9a197c08a3d958a WatchSource:0}: Error finding container 63aad3c7a5e012411fcf7e81e9a5a8325238cd78bdd4bfb4a9a197c08a3d958a: Status 404 returned error can't find the container with id 63aad3c7a5e012411fcf7e81e9a5a8325238cd78bdd4bfb4a9a197c08a3d958a Feb 02 09:07:18 crc kubenswrapper[4764]: W0202 09:07:18.341403 4764 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-8382d901059228c5096d39a16fed65cea11a1b1a091aa1d9845d94c0e2455714 WatchSource:0}: Error finding container 8382d901059228c5096d39a16fed65cea11a1b1a091aa1d9845d94c0e2455714: Status 404 returned error can't find the container with id 8382d901059228c5096d39a16fed65cea11a1b1a091aa1d9845d94c0e2455714 Feb 02 09:07:18 crc kubenswrapper[4764]: W0202 09:07:18.342394 4764 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-88e6c6f0a413c15e2498e2d0ea14c69751000fac5e5a0e7bc81801396e932bac WatchSource:0}: Error finding container 88e6c6f0a413c15e2498e2d0ea14c69751000fac5e5a0e7bc81801396e932bac: Status 404 returned error can't find the container with id 88e6c6f0a413c15e2498e2d0ea14c69751000fac5e5a0e7bc81801396e932bac Feb 02 09:07:18 crc kubenswrapper[4764]: E0202 09:07:18.367681 4764 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.217:6443: connect: connection refused" interval="800ms" Feb 02 09:07:18 crc kubenswrapper[4764]: I0202 09:07:18.591628 4764 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 09:07:18 crc kubenswrapper[4764]: I0202 09:07:18.592985 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:18 crc kubenswrapper[4764]: I0202 09:07:18.593019 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:18 crc kubenswrapper[4764]: I0202 09:07:18.593030 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:18 crc kubenswrapper[4764]: I0202 09:07:18.593055 4764 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 02 09:07:18 crc kubenswrapper[4764]: E0202 09:07:18.593511 4764 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.217:6443: connect: connection refused" node="crc" Feb 02 09:07:18 crc kubenswrapper[4764]: W0202 09:07:18.612105 4764 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.217:6443: connect: connection refused Feb 02 09:07:18 crc kubenswrapper[4764]: E0202 09:07:18.612175 4764 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.217:6443: connect: connection refused" logger="UnhandledError" Feb 02 09:07:18 crc kubenswrapper[4764]: I0202 09:07:18.761548 4764 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.217:6443: connect: connection refused Feb 02 09:07:18 crc kubenswrapper[4764]: I0202 09:07:18.763771 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-06 06:38:17.583835315 +0000 UTC Feb 02 09:07:18 crc kubenswrapper[4764]: W0202 09:07:18.793458 4764 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.217:6443: connect: connection refused Feb 02 09:07:18 crc kubenswrapper[4764]: E0202 09:07:18.793549 4764 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.217:6443: connect: connection refused" logger="UnhandledError" Feb 02 09:07:18 crc kubenswrapper[4764]: I0202 09:07:18.829202 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"a0fb8d7faa9c9906f5977401249ca564d82a0ec26bc3cf673a7aa5ce4b70a283"} Feb 02 09:07:18 crc kubenswrapper[4764]: I0202 09:07:18.830026 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"49b32e6d456b3590d318c8ff087ff6583e1e26afc831d6e68f6ab2bf6dd34d02"} Feb 02 09:07:18 crc kubenswrapper[4764]: I0202 09:07:18.830906 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"88e6c6f0a413c15e2498e2d0ea14c69751000fac5e5a0e7bc81801396e932bac"} Feb 02 09:07:18 crc kubenswrapper[4764]: I0202 09:07:18.832576 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"8382d901059228c5096d39a16fed65cea11a1b1a091aa1d9845d94c0e2455714"} Feb 02 09:07:18 crc kubenswrapper[4764]: I0202 09:07:18.833202 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"63aad3c7a5e012411fcf7e81e9a5a8325238cd78bdd4bfb4a9a197c08a3d958a"} Feb 02 09:07:18 crc kubenswrapper[4764]: W0202 09:07:18.961183 4764 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.217:6443: connect: connection refused Feb 02 09:07:18 crc kubenswrapper[4764]: E0202 09:07:18.961277 4764 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.217:6443: connect: connection refused" logger="UnhandledError" Feb 02 09:07:19 crc kubenswrapper[4764]: E0202 09:07:19.168623 4764 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.217:6443: connect: connection refused" interval="1.6s" Feb 02 09:07:19 crc kubenswrapper[4764]: W0202 09:07:19.285164 4764 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.217:6443: connect: connection refused Feb 02 09:07:19 crc kubenswrapper[4764]: E0202 09:07:19.285239 4764 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.217:6443: connect: connection refused" logger="UnhandledError" Feb 02 09:07:19 crc kubenswrapper[4764]: I0202 09:07:19.394523 4764 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 09:07:19 crc kubenswrapper[4764]: I0202 09:07:19.396104 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:19 crc kubenswrapper[4764]: I0202 09:07:19.396138 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:19 crc kubenswrapper[4764]: I0202 09:07:19.396148 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:19 crc kubenswrapper[4764]: I0202 09:07:19.396168 4764 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 02 09:07:19 crc kubenswrapper[4764]: E0202 09:07:19.396681 4764 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.217:6443: connect: connection refused" node="crc" Feb 02 09:07:19 crc kubenswrapper[4764]: I0202 09:07:19.675504 4764 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Feb 02 09:07:19 crc kubenswrapper[4764]: E0202 09:07:19.676723 4764 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.217:6443: connect: connection refused" logger="UnhandledError" Feb 02 09:07:19 crc kubenswrapper[4764]: I0202 09:07:19.761926 4764 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.217:6443: connect: connection refused Feb 02 09:07:19 crc kubenswrapper[4764]: I0202 09:07:19.764038 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-19 20:16:57.959200945 +0000 UTC Feb 02 09:07:19 crc kubenswrapper[4764]: I0202 09:07:19.838601 4764 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="85e2f75c56d63cacc93b9195c1e8bf46eb7a7234c45a024bd95a460db90edb6e" exitCode=0 Feb 02 09:07:19 crc kubenswrapper[4764]: I0202 09:07:19.838760 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"85e2f75c56d63cacc93b9195c1e8bf46eb7a7234c45a024bd95a460db90edb6e"} Feb 02 09:07:19 crc kubenswrapper[4764]: I0202 09:07:19.838838 4764 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 09:07:19 crc kubenswrapper[4764]: I0202 09:07:19.840433 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:19 crc kubenswrapper[4764]: I0202 09:07:19.840472 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:19 crc kubenswrapper[4764]: I0202 09:07:19.840485 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:19 crc kubenswrapper[4764]: I0202 09:07:19.842465 4764 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="db5a54d3a2117252a1ff50884683f99185874d2904c0960161642280f614679d" exitCode=0 Feb 02 09:07:19 crc kubenswrapper[4764]: I0202 09:07:19.842522 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"db5a54d3a2117252a1ff50884683f99185874d2904c0960161642280f614679d"} Feb 02 09:07:19 crc kubenswrapper[4764]: I0202 09:07:19.842606 4764 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 09:07:19 crc kubenswrapper[4764]: I0202 09:07:19.843886 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:19 crc kubenswrapper[4764]: I0202 09:07:19.843920 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:19 crc kubenswrapper[4764]: I0202 09:07:19.843942 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:19 crc kubenswrapper[4764]: I0202 09:07:19.845622 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"10f84be2e08217a2ac3edde858d78b4410376ed9bae3bcabe2a7f9b2fbe2f69c"} Feb 02 09:07:19 crc kubenswrapper[4764]: I0202 09:07:19.845659 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"f6f064bdc66470aace3180d4988244026b54009a63cff199e41c5d02a1d18479"} Feb 02 09:07:19 crc kubenswrapper[4764]: I0202 09:07:19.845669 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"60258f84a1d8dcfe48b7e9688ed50f077e201971f20f3f6663fbf1ba58a4611a"} Feb 02 09:07:19 crc kubenswrapper[4764]: I0202 09:07:19.848542 4764 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="dfb06a5ec9262cf39b108a66014d15e325cdbd5441a2d9e7a811dd57265c99a6" exitCode=0 Feb 02 09:07:19 crc kubenswrapper[4764]: I0202 09:07:19.848702 4764 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 09:07:19 crc kubenswrapper[4764]: I0202 09:07:19.848689 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"dfb06a5ec9262cf39b108a66014d15e325cdbd5441a2d9e7a811dd57265c99a6"} Feb 02 09:07:19 crc kubenswrapper[4764]: I0202 09:07:19.850016 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:19 crc kubenswrapper[4764]: I0202 09:07:19.850049 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:19 crc kubenswrapper[4764]: I0202 09:07:19.850064 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:19 crc kubenswrapper[4764]: I0202 09:07:19.852248 4764 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="e229f484d19cbdabedf53d90a822a4434db9e00b1301538e885e26acc19d25f8" exitCode=0 Feb 02 09:07:19 crc kubenswrapper[4764]: I0202 09:07:19.852297 4764 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 09:07:19 crc kubenswrapper[4764]: I0202 09:07:19.852297 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"e229f484d19cbdabedf53d90a822a4434db9e00b1301538e885e26acc19d25f8"} Feb 02 09:07:19 crc kubenswrapper[4764]: I0202 09:07:19.853560 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:19 crc kubenswrapper[4764]: I0202 09:07:19.853603 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:19 crc kubenswrapper[4764]: I0202 09:07:19.853622 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:19 crc kubenswrapper[4764]: I0202 09:07:19.856910 4764 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 09:07:19 crc kubenswrapper[4764]: I0202 09:07:19.857701 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:19 crc kubenswrapper[4764]: I0202 09:07:19.857726 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:19 crc kubenswrapper[4764]: I0202 09:07:19.857738 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:20 crc kubenswrapper[4764]: W0202 09:07:20.465581 4764 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.217:6443: connect: connection refused Feb 02 09:07:20 crc kubenswrapper[4764]: E0202 09:07:20.465670 4764 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.217:6443: connect: connection refused" logger="UnhandledError" Feb 02 09:07:20 crc kubenswrapper[4764]: W0202 09:07:20.590434 4764 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.217:6443: connect: connection refused Feb 02 09:07:20 crc kubenswrapper[4764]: E0202 09:07:20.590530 4764 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.217:6443: connect: connection refused" logger="UnhandledError" Feb 02 09:07:20 crc kubenswrapper[4764]: I0202 09:07:20.762710 4764 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.217:6443: connect: connection refused Feb 02 09:07:20 crc kubenswrapper[4764]: I0202 09:07:20.764800 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-10 10:02:59.274385962 +0000 UTC Feb 02 09:07:20 crc kubenswrapper[4764]: E0202 09:07:20.770242 4764 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.217:6443: connect: connection refused" interval="3.2s" Feb 02 09:07:20 crc kubenswrapper[4764]: I0202 09:07:20.858405 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"8fad17b01545eb20785f03e3c61bd5c982f6217e38176d7a3b23ac194298644a"} Feb 02 09:07:20 crc kubenswrapper[4764]: I0202 09:07:20.858462 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"429ead0993f5e648f6256d249ee12aea03514d3108618840552b672fb419cd74"} Feb 02 09:07:20 crc kubenswrapper[4764]: I0202 09:07:20.858478 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"081d7f76d937ff71e35945f9b1d893888bb649d15bd84312fd4b61c9520ea71f"} Feb 02 09:07:20 crc kubenswrapper[4764]: I0202 09:07:20.858515 4764 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 09:07:20 crc kubenswrapper[4764]: I0202 09:07:20.859880 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:20 crc kubenswrapper[4764]: I0202 09:07:20.859954 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:20 crc kubenswrapper[4764]: I0202 09:07:20.859971 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:20 crc kubenswrapper[4764]: I0202 09:07:20.861462 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"664e0519c2a2c60904dfcc8deebaf5d638dac0bedb13d870c0aaea5b4f5428f9"} Feb 02 09:07:20 crc kubenswrapper[4764]: I0202 09:07:20.861515 4764 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 09:07:20 crc kubenswrapper[4764]: I0202 09:07:20.862576 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:20 crc kubenswrapper[4764]: I0202 09:07:20.862617 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:20 crc kubenswrapper[4764]: I0202 09:07:20.862631 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:20 crc kubenswrapper[4764]: I0202 09:07:20.863554 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"0a69d25d352d7ceac5c735bc1fcf446a1fea6dc22c65a5296cfab828cea17377"} Feb 02 09:07:20 crc kubenswrapper[4764]: I0202 09:07:20.863700 4764 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 09:07:20 crc kubenswrapper[4764]: I0202 09:07:20.864649 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:20 crc kubenswrapper[4764]: I0202 09:07:20.864683 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:20 crc kubenswrapper[4764]: I0202 09:07:20.864696 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:20 crc kubenswrapper[4764]: I0202 09:07:20.865501 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"d468715c2da33996561c40043216233fd12e21539d0c58dd1f3b892fd9f91a9d"} Feb 02 09:07:20 crc kubenswrapper[4764]: I0202 09:07:20.865527 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"9b184f7c936dc4046f74c0018269cbc2c4b7ceddb650b4e871d3f8a707e0df88"} Feb 02 09:07:20 crc kubenswrapper[4764]: I0202 09:07:20.865539 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"9e78b6d08e4d616cefbb73e3fd43b3c6886dcd8aa997c7b4948bf1ff63401293"} Feb 02 09:07:20 crc kubenswrapper[4764]: I0202 09:07:20.867048 4764 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="8a9cc673be8ff245dba81278e8cd34cfe45ed36ff0b8bd5b30a0ddf333ec6371" exitCode=0 Feb 02 09:07:20 crc kubenswrapper[4764]: I0202 09:07:20.867078 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"8a9cc673be8ff245dba81278e8cd34cfe45ed36ff0b8bd5b30a0ddf333ec6371"} Feb 02 09:07:20 crc kubenswrapper[4764]: I0202 09:07:20.867189 4764 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 09:07:20 crc kubenswrapper[4764]: I0202 09:07:20.867863 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:20 crc kubenswrapper[4764]: I0202 09:07:20.867891 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:20 crc kubenswrapper[4764]: I0202 09:07:20.867900 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:20 crc kubenswrapper[4764]: I0202 09:07:20.997355 4764 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 09:07:20 crc kubenswrapper[4764]: I0202 09:07:20.998538 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:20 crc kubenswrapper[4764]: I0202 09:07:20.998577 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:20 crc kubenswrapper[4764]: I0202 09:07:20.998586 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:20 crc kubenswrapper[4764]: I0202 09:07:20.998608 4764 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 02 09:07:20 crc kubenswrapper[4764]: E0202 09:07:20.999094 4764 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.217:6443: connect: connection refused" node="crc" Feb 02 09:07:21 crc kubenswrapper[4764]: W0202 09:07:21.211815 4764 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.217:6443: connect: connection refused Feb 02 09:07:21 crc kubenswrapper[4764]: E0202 09:07:21.211881 4764 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.217:6443: connect: connection refused" logger="UnhandledError" Feb 02 09:07:21 crc kubenswrapper[4764]: I0202 09:07:21.761737 4764 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.217:6443: connect: connection refused Feb 02 09:07:21 crc kubenswrapper[4764]: I0202 09:07:21.764904 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-17 21:12:46.25861936 +0000 UTC Feb 02 09:07:21 crc kubenswrapper[4764]: I0202 09:07:21.871802 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"b25bc9395831524c53f9e565b4efbe74f21792efafeb1dd507161bc91aa0ecdf"} Feb 02 09:07:21 crc kubenswrapper[4764]: I0202 09:07:21.871851 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"b44e17b8ca01f549bc10081dd5f109cb1b9fff364bf5bd4fd90d9f0e3eeb04db"} Feb 02 09:07:21 crc kubenswrapper[4764]: I0202 09:07:21.871862 4764 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 09:07:21 crc kubenswrapper[4764]: I0202 09:07:21.872892 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:21 crc kubenswrapper[4764]: I0202 09:07:21.872953 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:21 crc kubenswrapper[4764]: I0202 09:07:21.872967 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:21 crc kubenswrapper[4764]: I0202 09:07:21.874748 4764 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="bc7e49c7a5ae66ea939060192fa6a346dc0b0b9aa9750b29879d3c426cc0d276" exitCode=0 Feb 02 09:07:21 crc kubenswrapper[4764]: I0202 09:07:21.874854 4764 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 09:07:21 crc kubenswrapper[4764]: I0202 09:07:21.874863 4764 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 09:07:21 crc kubenswrapper[4764]: I0202 09:07:21.874870 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"bc7e49c7a5ae66ea939060192fa6a346dc0b0b9aa9750b29879d3c426cc0d276"} Feb 02 09:07:21 crc kubenswrapper[4764]: I0202 09:07:21.874901 4764 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 09:07:21 crc kubenswrapper[4764]: I0202 09:07:21.875002 4764 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 09:07:21 crc kubenswrapper[4764]: I0202 09:07:21.875001 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 02 09:07:21 crc kubenswrapper[4764]: I0202 09:07:21.879292 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:21 crc kubenswrapper[4764]: I0202 09:07:21.879337 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:21 crc kubenswrapper[4764]: I0202 09:07:21.879348 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:21 crc kubenswrapper[4764]: I0202 09:07:21.879370 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:21 crc kubenswrapper[4764]: I0202 09:07:21.879392 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:21 crc kubenswrapper[4764]: I0202 09:07:21.879400 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:21 crc kubenswrapper[4764]: I0202 09:07:21.879421 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:21 crc kubenswrapper[4764]: I0202 09:07:21.879451 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:21 crc kubenswrapper[4764]: I0202 09:07:21.879461 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:21 crc kubenswrapper[4764]: I0202 09:07:21.879803 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:21 crc kubenswrapper[4764]: I0202 09:07:21.879895 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:21 crc kubenswrapper[4764]: I0202 09:07:21.879989 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:22 crc kubenswrapper[4764]: W0202 09:07:22.061155 4764 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.217:6443: connect: connection refused Feb 02 09:07:22 crc kubenswrapper[4764]: E0202 09:07:22.061231 4764 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.217:6443: connect: connection refused" logger="UnhandledError" Feb 02 09:07:22 crc kubenswrapper[4764]: I0202 09:07:22.765066 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-09 22:52:08.813378107 +0000 UTC Feb 02 09:07:22 crc kubenswrapper[4764]: I0202 09:07:22.882822 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"408722c7f83d28919431532f8c797a1c795fd5d96736d14e7f06a05744899702"} Feb 02 09:07:22 crc kubenswrapper[4764]: I0202 09:07:22.882865 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"6d38c7cb5dcd22fb07524494ebe77aa1a7f4f808f566043563817e47e7ec12a9"} Feb 02 09:07:22 crc kubenswrapper[4764]: I0202 09:07:22.882876 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"0a4ac273a4f127d79e5fad7d4c30cd2044a346f5734dba5d33d68cd092c0c2d1"} Feb 02 09:07:22 crc kubenswrapper[4764]: I0202 09:07:22.882885 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"d92ac486f5dade46bc75de32099859a71fc2ed25a7abbe6ae8937df128e7df30"} Feb 02 09:07:22 crc kubenswrapper[4764]: I0202 09:07:22.882893 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"1f473b18cd8003f75bb1dafb82bcec45745a73ccd10434297356dac185a3d15c"} Feb 02 09:07:22 crc kubenswrapper[4764]: I0202 09:07:22.883034 4764 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 09:07:22 crc kubenswrapper[4764]: I0202 09:07:22.884152 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:22 crc kubenswrapper[4764]: I0202 09:07:22.884180 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:22 crc kubenswrapper[4764]: I0202 09:07:22.884192 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:22 crc kubenswrapper[4764]: I0202 09:07:22.884798 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Feb 02 09:07:22 crc kubenswrapper[4764]: I0202 09:07:22.886900 4764 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="b25bc9395831524c53f9e565b4efbe74f21792efafeb1dd507161bc91aa0ecdf" exitCode=255 Feb 02 09:07:22 crc kubenswrapper[4764]: I0202 09:07:22.887010 4764 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 09:07:22 crc kubenswrapper[4764]: I0202 09:07:22.887000 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"b25bc9395831524c53f9e565b4efbe74f21792efafeb1dd507161bc91aa0ecdf"} Feb 02 09:07:22 crc kubenswrapper[4764]: I0202 09:07:22.887995 4764 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 09:07:22 crc kubenswrapper[4764]: I0202 09:07:22.888170 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:22 crc kubenswrapper[4764]: I0202 09:07:22.888190 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:22 crc kubenswrapper[4764]: I0202 09:07:22.888201 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:22 crc kubenswrapper[4764]: I0202 09:07:22.888653 4764 scope.go:117] "RemoveContainer" containerID="b25bc9395831524c53f9e565b4efbe74f21792efafeb1dd507161bc91aa0ecdf" Feb 02 09:07:22 crc kubenswrapper[4764]: I0202 09:07:22.888686 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:22 crc kubenswrapper[4764]: I0202 09:07:22.888721 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:22 crc kubenswrapper[4764]: I0202 09:07:22.888732 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:22 crc kubenswrapper[4764]: I0202 09:07:22.900015 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 02 09:07:22 crc kubenswrapper[4764]: I0202 09:07:22.900234 4764 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 09:07:22 crc kubenswrapper[4764]: I0202 09:07:22.901667 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:22 crc kubenswrapper[4764]: I0202 09:07:22.901707 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:22 crc kubenswrapper[4764]: I0202 09:07:22.901717 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:23 crc kubenswrapper[4764]: I0202 09:07:23.714264 4764 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Feb 02 09:07:23 crc kubenswrapper[4764]: I0202 09:07:23.765421 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-30 23:41:35.368258335 +0000 UTC Feb 02 09:07:23 crc kubenswrapper[4764]: I0202 09:07:23.892452 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Feb 02 09:07:23 crc kubenswrapper[4764]: I0202 09:07:23.901239 4764 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 09:07:23 crc kubenswrapper[4764]: I0202 09:07:23.901696 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"269c5074371bd00f8363fb3888c11f1df383081a61497aea6578f08700881d50"} Feb 02 09:07:23 crc kubenswrapper[4764]: I0202 09:07:23.901874 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 02 09:07:23 crc kubenswrapper[4764]: I0202 09:07:23.901805 4764 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 09:07:23 crc kubenswrapper[4764]: I0202 09:07:23.902655 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:23 crc kubenswrapper[4764]: I0202 09:07:23.902715 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:23 crc kubenswrapper[4764]: I0202 09:07:23.902730 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:23 crc kubenswrapper[4764]: I0202 09:07:23.903511 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:23 crc kubenswrapper[4764]: I0202 09:07:23.903550 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:23 crc kubenswrapper[4764]: I0202 09:07:23.903565 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:23 crc kubenswrapper[4764]: I0202 09:07:23.936856 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 02 09:07:24 crc kubenswrapper[4764]: I0202 09:07:24.109138 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 02 09:07:24 crc kubenswrapper[4764]: I0202 09:07:24.133708 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 02 09:07:24 crc kubenswrapper[4764]: I0202 09:07:24.134015 4764 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 09:07:24 crc kubenswrapper[4764]: I0202 09:07:24.135444 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:24 crc kubenswrapper[4764]: I0202 09:07:24.135513 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:24 crc kubenswrapper[4764]: I0202 09:07:24.135532 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:24 crc kubenswrapper[4764]: I0202 09:07:24.141775 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 02 09:07:24 crc kubenswrapper[4764]: I0202 09:07:24.199655 4764 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 09:07:24 crc kubenswrapper[4764]: I0202 09:07:24.201047 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:24 crc kubenswrapper[4764]: I0202 09:07:24.201087 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:24 crc kubenswrapper[4764]: I0202 09:07:24.201098 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:24 crc kubenswrapper[4764]: I0202 09:07:24.201122 4764 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 02 09:07:24 crc kubenswrapper[4764]: I0202 09:07:24.766333 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-16 07:40:22.381753357 +0000 UTC Feb 02 09:07:24 crc kubenswrapper[4764]: I0202 09:07:24.904199 4764 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 09:07:24 crc kubenswrapper[4764]: I0202 09:07:24.904241 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 02 09:07:24 crc kubenswrapper[4764]: I0202 09:07:24.904331 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 02 09:07:24 crc kubenswrapper[4764]: I0202 09:07:24.904210 4764 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 09:07:24 crc kubenswrapper[4764]: I0202 09:07:24.906069 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:24 crc kubenswrapper[4764]: I0202 09:07:24.906133 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:24 crc kubenswrapper[4764]: I0202 09:07:24.906153 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:24 crc kubenswrapper[4764]: I0202 09:07:24.906550 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:24 crc kubenswrapper[4764]: I0202 09:07:24.906589 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:24 crc kubenswrapper[4764]: I0202 09:07:24.906602 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:25 crc kubenswrapper[4764]: I0202 09:07:25.767175 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-11 00:50:10.126988526 +0000 UTC Feb 02 09:07:25 crc kubenswrapper[4764]: I0202 09:07:25.789580 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 02 09:07:25 crc kubenswrapper[4764]: I0202 09:07:25.903580 4764 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 02 09:07:25 crc kubenswrapper[4764]: I0202 09:07:25.903716 4764 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Feb 02 09:07:25 crc kubenswrapper[4764]: I0202 09:07:25.907348 4764 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 09:07:25 crc kubenswrapper[4764]: I0202 09:07:25.907410 4764 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 09:07:25 crc kubenswrapper[4764]: I0202 09:07:25.909042 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:25 crc kubenswrapper[4764]: I0202 09:07:25.909098 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:25 crc kubenswrapper[4764]: I0202 09:07:25.909124 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:25 crc kubenswrapper[4764]: I0202 09:07:25.909324 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:25 crc kubenswrapper[4764]: I0202 09:07:25.909404 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:25 crc kubenswrapper[4764]: I0202 09:07:25.909424 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:26 crc kubenswrapper[4764]: I0202 09:07:26.768765 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-16 15:38:02.703096548 +0000 UTC Feb 02 09:07:26 crc kubenswrapper[4764]: I0202 09:07:26.910485 4764 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 09:07:26 crc kubenswrapper[4764]: I0202 09:07:26.911836 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:26 crc kubenswrapper[4764]: I0202 09:07:26.911906 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:26 crc kubenswrapper[4764]: I0202 09:07:26.911928 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:27 crc kubenswrapper[4764]: I0202 09:07:27.734395 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Feb 02 09:07:27 crc kubenswrapper[4764]: I0202 09:07:27.734638 4764 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 09:07:27 crc kubenswrapper[4764]: I0202 09:07:27.736053 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:27 crc kubenswrapper[4764]: I0202 09:07:27.736087 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:27 crc kubenswrapper[4764]: I0202 09:07:27.736097 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:27 crc kubenswrapper[4764]: I0202 09:07:27.769840 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-10 15:43:14.779900263 +0000 UTC Feb 02 09:07:27 crc kubenswrapper[4764]: E0202 09:07:27.897990 4764 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Feb 02 09:07:28 crc kubenswrapper[4764]: I0202 09:07:28.770206 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-12 23:30:56.724367791 +0000 UTC Feb 02 09:07:29 crc kubenswrapper[4764]: I0202 09:07:29.725887 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Feb 02 09:07:29 crc kubenswrapper[4764]: I0202 09:07:29.726433 4764 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 09:07:29 crc kubenswrapper[4764]: I0202 09:07:29.728028 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:29 crc kubenswrapper[4764]: I0202 09:07:29.728085 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:29 crc kubenswrapper[4764]: I0202 09:07:29.728096 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:29 crc kubenswrapper[4764]: I0202 09:07:29.771406 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-22 06:12:03.866123096 +0000 UTC Feb 02 09:07:30 crc kubenswrapper[4764]: I0202 09:07:30.334056 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 02 09:07:30 crc kubenswrapper[4764]: I0202 09:07:30.334356 4764 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 09:07:30 crc kubenswrapper[4764]: I0202 09:07:30.336331 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:30 crc kubenswrapper[4764]: I0202 09:07:30.336458 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:30 crc kubenswrapper[4764]: I0202 09:07:30.336542 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:30 crc kubenswrapper[4764]: I0202 09:07:30.772018 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-13 00:02:53.554416076 +0000 UTC Feb 02 09:07:31 crc kubenswrapper[4764]: I0202 09:07:31.773619 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-27 06:44:30.368238037 +0000 UTC Feb 02 09:07:32 crc kubenswrapper[4764]: I0202 09:07:32.761896 4764 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Feb 02 09:07:32 crc kubenswrapper[4764]: I0202 09:07:32.774298 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-04 00:30:55.465664939 +0000 UTC Feb 02 09:07:33 crc kubenswrapper[4764]: I0202 09:07:33.415710 4764 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Feb 02 09:07:33 crc kubenswrapper[4764]: I0202 09:07:33.415764 4764 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Feb 02 09:07:33 crc kubenswrapper[4764]: I0202 09:07:33.440397 4764 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Feb 02 09:07:33 crc kubenswrapper[4764]: I0202 09:07:33.440455 4764 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Feb 02 09:07:33 crc kubenswrapper[4764]: I0202 09:07:33.774431 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-02 23:24:35.996379947 +0000 UTC Feb 02 09:07:33 crc kubenswrapper[4764]: I0202 09:07:33.942243 4764 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Feb 02 09:07:33 crc kubenswrapper[4764]: [+]log ok Feb 02 09:07:33 crc kubenswrapper[4764]: [+]etcd ok Feb 02 09:07:33 crc kubenswrapper[4764]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Feb 02 09:07:33 crc kubenswrapper[4764]: [+]poststarthook/openshift.io-api-request-count-filter ok Feb 02 09:07:33 crc kubenswrapper[4764]: [+]poststarthook/openshift.io-startkubeinformers ok Feb 02 09:07:33 crc kubenswrapper[4764]: [+]poststarthook/openshift.io-openshift-apiserver-reachable ok Feb 02 09:07:33 crc kubenswrapper[4764]: [+]poststarthook/openshift.io-oauth-apiserver-reachable ok Feb 02 09:07:33 crc kubenswrapper[4764]: [+]poststarthook/start-apiserver-admission-initializer ok Feb 02 09:07:33 crc kubenswrapper[4764]: [+]poststarthook/generic-apiserver-start-informers ok Feb 02 09:07:33 crc kubenswrapper[4764]: [+]poststarthook/priority-and-fairness-config-consumer ok Feb 02 09:07:33 crc kubenswrapper[4764]: [+]poststarthook/priority-and-fairness-filter ok Feb 02 09:07:33 crc kubenswrapper[4764]: [+]poststarthook/storage-object-count-tracker-hook ok Feb 02 09:07:33 crc kubenswrapper[4764]: [+]poststarthook/start-apiextensions-informers ok Feb 02 09:07:33 crc kubenswrapper[4764]: [+]poststarthook/start-apiextensions-controllers ok Feb 02 09:07:33 crc kubenswrapper[4764]: [+]poststarthook/crd-informer-synced ok Feb 02 09:07:33 crc kubenswrapper[4764]: [+]poststarthook/start-system-namespaces-controller ok Feb 02 09:07:33 crc kubenswrapper[4764]: [+]poststarthook/start-cluster-authentication-info-controller ok Feb 02 09:07:33 crc kubenswrapper[4764]: [+]poststarthook/start-kube-apiserver-identity-lease-controller ok Feb 02 09:07:33 crc kubenswrapper[4764]: [+]poststarthook/start-kube-apiserver-identity-lease-garbage-collector ok Feb 02 09:07:33 crc kubenswrapper[4764]: [+]poststarthook/start-legacy-token-tracking-controller ok Feb 02 09:07:33 crc kubenswrapper[4764]: [+]poststarthook/start-service-ip-repair-controllers ok Feb 02 09:07:33 crc kubenswrapper[4764]: [-]poststarthook/rbac/bootstrap-roles failed: reason withheld Feb 02 09:07:33 crc kubenswrapper[4764]: [-]poststarthook/scheduling/bootstrap-system-priority-classes failed: reason withheld Feb 02 09:07:33 crc kubenswrapper[4764]: [+]poststarthook/priority-and-fairness-config-producer ok Feb 02 09:07:33 crc kubenswrapper[4764]: [+]poststarthook/bootstrap-controller ok Feb 02 09:07:33 crc kubenswrapper[4764]: [+]poststarthook/aggregator-reload-proxy-client-cert ok Feb 02 09:07:33 crc kubenswrapper[4764]: [+]poststarthook/start-kube-aggregator-informers ok Feb 02 09:07:33 crc kubenswrapper[4764]: [+]poststarthook/apiservice-status-local-available-controller ok Feb 02 09:07:33 crc kubenswrapper[4764]: [+]poststarthook/apiservice-status-remote-available-controller ok Feb 02 09:07:33 crc kubenswrapper[4764]: [+]poststarthook/apiservice-registration-controller ok Feb 02 09:07:33 crc kubenswrapper[4764]: [+]poststarthook/apiservice-wait-for-first-sync ok Feb 02 09:07:33 crc kubenswrapper[4764]: [+]poststarthook/apiservice-discovery-controller ok Feb 02 09:07:33 crc kubenswrapper[4764]: [+]poststarthook/kube-apiserver-autoregistration ok Feb 02 09:07:33 crc kubenswrapper[4764]: [+]autoregister-completion ok Feb 02 09:07:33 crc kubenswrapper[4764]: [+]poststarthook/apiservice-openapi-controller ok Feb 02 09:07:33 crc kubenswrapper[4764]: [+]poststarthook/apiservice-openapiv3-controller ok Feb 02 09:07:33 crc kubenswrapper[4764]: livez check failed Feb 02 09:07:33 crc kubenswrapper[4764]: I0202 09:07:33.942323 4764 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 02 09:07:34 crc kubenswrapper[4764]: I0202 09:07:34.775015 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-02 15:13:45.451937023 +0000 UTC Feb 02 09:07:35 crc kubenswrapper[4764]: I0202 09:07:35.775971 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-16 07:27:29.145532403 +0000 UTC Feb 02 09:07:35 crc kubenswrapper[4764]: I0202 09:07:35.901165 4764 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 02 09:07:35 crc kubenswrapper[4764]: I0202 09:07:35.901245 4764 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Feb 02 09:07:36 crc kubenswrapper[4764]: I0202 09:07:36.778059 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-18 09:03:36.147614205 +0000 UTC Feb 02 09:07:37 crc kubenswrapper[4764]: I0202 09:07:37.779007 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-04 12:41:53.708464646 +0000 UTC Feb 02 09:07:37 crc kubenswrapper[4764]: E0202 09:07:37.898279 4764 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.428120 4764 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Feb 02 09:07:38 crc kubenswrapper[4764]: E0202 09:07:38.432616 4764 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="6.4s" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.440481 4764 trace.go:236] Trace[1042452348]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (02-Feb-2026 09:07:28.237) (total time: 10202ms): Feb 02 09:07:38 crc kubenswrapper[4764]: Trace[1042452348]: ---"Objects listed" error: 10202ms (09:07:38.440) Feb 02 09:07:38 crc kubenswrapper[4764]: Trace[1042452348]: [10.20266858s] [10.20266858s] END Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.440750 4764 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.440517 4764 trace.go:236] Trace[564766397]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (02-Feb-2026 09:07:26.287) (total time: 12152ms): Feb 02 09:07:38 crc kubenswrapper[4764]: Trace[564766397]: ---"Objects listed" error: 12152ms (09:07:38.440) Feb 02 09:07:38 crc kubenswrapper[4764]: Trace[564766397]: [12.152838289s] [12.152838289s] END Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.440872 4764 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.442414 4764 trace.go:236] Trace[234948742]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (02-Feb-2026 09:07:26.021) (total time: 12421ms): Feb 02 09:07:38 crc kubenswrapper[4764]: Trace[234948742]: ---"Objects listed" error: 12420ms (09:07:38.442) Feb 02 09:07:38 crc kubenswrapper[4764]: Trace[234948742]: [12.421024089s] [12.421024089s] END Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.442435 4764 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.444563 4764 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Feb 02 09:07:38 crc kubenswrapper[4764]: E0202 09:07:38.445339 4764 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.446004 4764 trace.go:236] Trace[833755422]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (02-Feb-2026 09:07:24.906) (total time: 13539ms): Feb 02 09:07:38 crc kubenswrapper[4764]: Trace[833755422]: ---"Objects listed" error: 13539ms (09:07:38.445) Feb 02 09:07:38 crc kubenswrapper[4764]: Trace[833755422]: [13.539351472s] [13.539351472s] END Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.446142 4764 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.474277 4764 csr.go:261] certificate signing request csr-992cg is approved, waiting to be issued Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.488371 4764 csr.go:257] certificate signing request csr-992cg is issued Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.758276 4764 apiserver.go:52] "Watching apiserver" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.760879 4764 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.761100 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c"] Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.761366 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.762533 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.763836 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 09:07:38 crc kubenswrapper[4764]: E0202 09:07:38.763988 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.764278 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.765179 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 09:07:38 crc kubenswrapper[4764]: E0202 09:07:38.765258 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.761513 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 09:07:38 crc kubenswrapper[4764]: E0202 09:07:38.765523 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.766916 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.766993 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.767031 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.767054 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.767377 4764 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.767388 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.767477 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.767527 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.767569 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.768393 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.779993 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-10 23:24:14.746745483 +0000 UTC Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.800839 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.817790 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.830832 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.839241 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.845771 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.845806 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.845822 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.845841 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.845856 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.845870 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.845885 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.845902 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.845924 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.845957 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.845972 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.845988 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.846003 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.846019 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.846036 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.846052 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.846103 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.846122 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.846137 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.846153 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.846169 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.846162 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.846184 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.846248 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.846270 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.846286 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.846303 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.846318 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.846336 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.846352 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.846368 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.846383 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.846397 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.846403 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.846479 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.846498 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.846513 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.846528 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.846538 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.846546 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.846562 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.846581 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.846596 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.846611 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.846627 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.846642 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.846661 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.846668 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.846677 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.846694 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.846711 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.846728 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.846743 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.846757 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.846771 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.846787 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.846801 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.846815 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.846831 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.846865 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.846879 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.846895 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.846910 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.846924 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.846954 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.846968 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.846990 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.847005 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.847022 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.847038 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.847052 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.847067 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.847096 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.847111 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.847126 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.847139 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.847180 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.847196 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.847211 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.847227 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.847242 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.847258 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.847274 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.847289 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.847304 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.847319 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.847334 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.847348 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.847354 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.847363 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.847392 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.847407 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.847422 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.847436 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.847455 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.847485 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.847495 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.847509 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.847533 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.847553 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.847572 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.847587 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.847603 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.847619 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.847620 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.847635 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.847652 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.847667 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.847682 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.847697 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.847713 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.847739 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.847754 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.847768 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.847784 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.847799 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.847815 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.847832 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.847847 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.847866 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.847881 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.847898 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.847914 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.847949 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.847981 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.847988 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.847997 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.848017 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.848033 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.848052 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.848071 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.848089 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.848108 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.848127 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.848147 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.848152 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.848162 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.848179 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.848194 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.848209 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.848226 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.848242 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.848256 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.848271 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.848286 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.848301 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.848316 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.848331 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.848348 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.848373 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.848385 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.848389 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.848421 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.848441 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.848457 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.848473 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.848494 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.848512 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.848528 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.848544 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.848560 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.848578 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.848594 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.848613 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.848640 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.848656 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.848672 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.848689 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.848707 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.848725 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.848740 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.848756 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.848772 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.848790 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.848817 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.848835 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.848851 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.848867 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.848882 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.848897 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.848913 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.848928 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.848958 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.848975 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.848993 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.849009 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.849024 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.849041 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.849060 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.849078 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.849093 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.849109 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.849125 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.849142 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.849160 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.849179 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.849196 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.849214 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.849230 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.849247 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.849266 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.849282 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.849300 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.849316 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.849333 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.849371 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.849391 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.849409 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.849427 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.849444 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.849463 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.849481 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.849500 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.849516 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.849533 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.849552 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.849570 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.849586 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.849602 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.849663 4764 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.849676 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.849686 4764 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.849696 4764 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.849705 4764 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.849715 4764 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.849724 4764 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.849733 4764 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.849743 4764 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.849752 4764 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.849040 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.849160 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.849522 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.849866 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.850313 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.850557 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.850571 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.850727 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.851791 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.852668 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.852728 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.852761 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.852902 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.852745 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.853108 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.853136 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.853241 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.853359 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.853418 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.853438 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.853562 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.853665 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.853675 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.853713 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.854120 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.854146 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: E0202 09:07:38.854242 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 09:07:39.354218162 +0000 UTC m=+22.287942310 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.856777 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.856775 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.854421 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.854464 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.854507 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.854707 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.854733 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.854743 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.854851 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.854849 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.854920 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.854978 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.855127 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.855186 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.855189 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.855200 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.855315 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.855348 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.855500 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.855722 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.855819 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.856003 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.856203 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.857048 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.857196 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.857387 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.857491 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.857554 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.858009 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.858459 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.858547 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.858631 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.858789 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.858993 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.859358 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.859404 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.859403 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.859557 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.859775 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.859897 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.860153 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.860236 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.860163 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.860336 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.860367 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.860614 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.861418 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.861679 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.861863 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.861615 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.862056 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.862099 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.862185 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.862317 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.862438 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.862449 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.862567 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.862778 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.862800 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.862979 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.863006 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.863165 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.863337 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.863591 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.863711 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.863833 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.863427 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.864215 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.864064 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.864321 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.864575 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.864857 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.865099 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.865772 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.865846 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.865922 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.866067 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.866296 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.866431 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.866491 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.866649 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.866821 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.867067 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.867174 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.867226 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.867207 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.867557 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.867875 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.867981 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.868641 4764 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.869237 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.869528 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.870168 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.870353 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.870454 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.870529 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.870674 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.871053 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.871155 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.871766 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.874267 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.874323 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: E0202 09:07:38.874361 4764 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 02 09:07:38 crc kubenswrapper[4764]: E0202 09:07:38.874567 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-02 09:07:39.374547295 +0000 UTC m=+22.308271464 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 02 09:07:38 crc kubenswrapper[4764]: E0202 09:07:38.874650 4764 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 02 09:07:38 crc kubenswrapper[4764]: E0202 09:07:38.874687 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-02 09:07:39.374678669 +0000 UTC m=+22.308402857 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.874864 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.875375 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.875673 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.875714 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.875989 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.876627 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.876718 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.878950 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.879182 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.883477 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.886253 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.886678 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.886765 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.887059 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.887080 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.887325 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.889460 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.890194 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.890515 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.890844 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.891127 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.891839 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.892179 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.892258 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.892455 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.892506 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.854264 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.892578 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.892612 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.892808 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: E0202 09:07:38.892829 4764 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 02 09:07:38 crc kubenswrapper[4764]: E0202 09:07:38.892849 4764 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 02 09:07:38 crc kubenswrapper[4764]: E0202 09:07:38.892863 4764 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.892903 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: E0202 09:07:38.892962 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-02 09:07:39.39290357 +0000 UTC m=+22.326627748 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.893085 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.893184 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.897113 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.900394 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.903236 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.908191 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.908210 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.908538 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: E0202 09:07:38.908723 4764 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 02 09:07:38 crc kubenswrapper[4764]: E0202 09:07:38.908750 4764 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 02 09:07:38 crc kubenswrapper[4764]: E0202 09:07:38.908762 4764 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 02 09:07:38 crc kubenswrapper[4764]: E0202 09:07:38.908810 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-02 09:07:39.408793614 +0000 UTC m=+22.342517702 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.909145 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.909821 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.909997 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.910179 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.910251 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.912202 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.912449 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.912996 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.914199 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.915667 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.917404 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.920381 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.921270 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.921377 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.923180 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.924239 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.924380 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.928620 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.929005 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.931106 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.934763 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.936452 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.940772 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.941454 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.943629 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.945447 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.950109 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.951874 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.951953 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.952038 4764 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.952051 4764 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.952064 4764 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.952074 4764 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.952082 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.952091 4764 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.952101 4764 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.952110 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.952119 4764 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.952127 4764 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.952138 4764 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.952147 4764 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.952156 4764 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.952166 4764 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.952175 4764 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.952183 4764 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.952192 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.952202 4764 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.952200 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.952211 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.952247 4764 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.952257 4764 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.952271 4764 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.952282 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.952291 4764 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.952300 4764 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.952301 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.952312 4764 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.952322 4764 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.952331 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.952343 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.952352 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.952360 4764 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.952369 4764 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.952380 4764 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.952389 4764 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.952398 4764 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.952406 4764 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.952418 4764 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.952427 4764 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.952437 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.952448 4764 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.952457 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.952466 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.952476 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.952493 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.952509 4764 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.952519 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.952529 4764 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.952538 4764 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.952565 4764 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.952574 4764 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.952584 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.952594 4764 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.952605 4764 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.952615 4764 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.952623 4764 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.952633 4764 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.952642 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.952651 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.952660 4764 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.952671 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.952682 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.952690 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.952699 4764 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.952713 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.952725 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.952736 4764 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.952751 4764 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.952765 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.952775 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.952782 4764 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.952794 4764 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.952810 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.952831 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.952843 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.952921 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.953339 4764 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.953353 4764 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.953362 4764 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.953380 4764 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.953392 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.953403 4764 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.953413 4764 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.953428 4764 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.953443 4764 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.953457 4764 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.953468 4764 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.953481 4764 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.953490 4764 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.953498 4764 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.953509 4764 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.953518 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.953531 4764 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.953542 4764 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.953553 4764 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.953561 4764 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.953569 4764 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.953597 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.953609 4764 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.953617 4764 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.953625 4764 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.953636 4764 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.953645 4764 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.953654 4764 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.953662 4764 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.953674 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.953682 4764 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.953691 4764 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.953700 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.953711 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.953719 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.953727 4764 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.953737 4764 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.953748 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.953756 4764 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.953764 4764 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.953774 4764 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.953784 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.953792 4764 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.953799 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.953811 4764 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.953820 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.953830 4764 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.953838 4764 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.953851 4764 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.953859 4764 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.953868 4764 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.953877 4764 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.953887 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.953896 4764 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.953904 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.953916 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.953924 4764 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.953950 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.953959 4764 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.953969 4764 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.953977 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.953986 4764 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.953995 4764 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.954007 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.954017 4764 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.954024 4764 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.954035 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.954043 4764 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.954052 4764 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.954059 4764 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.954070 4764 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.954078 4764 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.954086 4764 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.954094 4764 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.954105 4764 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.954115 4764 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.954124 4764 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.954139 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.954152 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.954161 4764 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.954172 4764 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.954183 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.954190 4764 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.954199 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.954207 4764 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.954217 4764 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.954225 4764 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.954234 4764 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.954242 4764 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.954253 4764 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.954261 4764 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.954269 4764 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.954277 4764 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.954288 4764 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.954296 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.954305 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.953513 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.954317 4764 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.954351 4764 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.954361 4764 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.954370 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.954382 4764 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.954391 4764 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.954400 4764 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.954410 4764 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.954418 4764 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.954426 4764 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.954691 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.958230 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.965050 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.969704 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.975371 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 09:07:38 crc kubenswrapper[4764]: I0202 09:07:38.990765 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.001112 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.009970 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.018159 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.028875 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.037679 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.045743 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.055081 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.055183 4764 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.055217 4764 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.055227 4764 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.064099 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.073167 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.077310 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.082145 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.085704 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.092758 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 02 09:07:39 crc kubenswrapper[4764]: W0202 09:07:39.098781 4764 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-fafb2af40e0aed751bb0cb7cb646165b56399911a4c09ced1689eeed3a82b6cc WatchSource:0}: Error finding container fafb2af40e0aed751bb0cb7cb646165b56399911a4c09ced1689eeed3a82b6cc: Status 404 returned error can't find the container with id fafb2af40e0aed751bb0cb7cb646165b56399911a4c09ced1689eeed3a82b6cc Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.139716 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.155568 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.167576 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.201729 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.223123 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87d61aea-46f6-4ee6-a7a4-767a6a67fea4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e78b6d08e4d616cefbb73e3fd43b3c6886dcd8aa997c7b4948bf1ff63401293\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d468715c2da33996561c40043216233fd12e21539d0c58dd1f3b892fd9f91a9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b184f7c936dc4046f74c0018269cbc2c4b7ceddb650b4e871d3f8a707e0df88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://269c5074371bd00f8363fb3888c11f1df383081a61497aea6578f08700881d50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b25bc9395831524c53f9e565b4efbe74f21792efafeb1dd507161bc91aa0ecdf\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T09:07:22Z\\\",\\\"message\\\":\\\"W0202 09:07:21.597488 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0202 09:07:21.598117 1 crypto.go:601] Generating new CA for check-endpoints-signer@1770023241 cert, and key in /tmp/serving-cert-3856183724/serving-signer.crt, /tmp/serving-cert-3856183724/serving-signer.key\\\\nI0202 09:07:22.012184 1 observer_polling.go:159] Starting file observer\\\\nW0202 09:07:22.015493 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0202 09:07:22.015797 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 09:07:22.016596 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3856183724/tls.crt::/tmp/serving-cert-3856183724/tls.key\\\\\\\"\\\\nF0202 09:07:22.270605 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b44e17b8ca01f549bc10081dd5f109cb1b9fff364bf5bd4fd90d9f0e3eeb04db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfb06a5ec9262cf39b108a66014d15e325cdbd5441a2d9e7a811dd57265c99a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfb06a5ec9262cf39b108a66014d15e325cdbd5441a2d9e7a811dd57265c99a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.254820 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.278328 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.301984 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.356968 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 09:07:39 crc kubenswrapper[4764]: E0202 09:07:39.357105 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 09:07:40.357092481 +0000 UTC m=+23.290816569 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.458340 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.458398 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 09:07:39 crc kubenswrapper[4764]: E0202 09:07:39.458533 4764 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 02 09:07:39 crc kubenswrapper[4764]: E0202 09:07:39.458548 4764 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 02 09:07:39 crc kubenswrapper[4764]: E0202 09:07:39.458559 4764 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 02 09:07:39 crc kubenswrapper[4764]: E0202 09:07:39.458620 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-02 09:07:40.458607286 +0000 UTC m=+23.392331374 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 02 09:07:39 crc kubenswrapper[4764]: E0202 09:07:39.458636 4764 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 02 09:07:39 crc kubenswrapper[4764]: E0202 09:07:39.458727 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-02 09:07:40.458709108 +0000 UTC m=+23.392433246 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.459026 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.459072 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 09:07:39 crc kubenswrapper[4764]: E0202 09:07:39.459166 4764 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 02 09:07:39 crc kubenswrapper[4764]: E0202 09:07:39.459213 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-02 09:07:40.45918903 +0000 UTC m=+23.392913118 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 02 09:07:39 crc kubenswrapper[4764]: E0202 09:07:39.459249 4764 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 02 09:07:39 crc kubenswrapper[4764]: E0202 09:07:39.459258 4764 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 02 09:07:39 crc kubenswrapper[4764]: E0202 09:07:39.459265 4764 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 02 09:07:39 crc kubenswrapper[4764]: E0202 09:07:39.459308 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-02 09:07:40.459301383 +0000 UTC m=+23.393025471 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.489848 4764 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2027-02-02 09:02:38 +0000 UTC, rotation deadline is 2026-10-18 17:25:59.969597833 +0000 UTC Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.489918 4764 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 6200h18m20.479683852s for next certificate rotation Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.737972 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-6zxff"] Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.738262 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-6zxff" Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.740121 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.740302 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.741727 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.752281 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.761736 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.762267 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.772170 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.780694 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-06 17:05:09.426117453 +0000 UTC Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.782113 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6zxff" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aaba93da-12a5-43df-b7c7-4c8b988163c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqwb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6zxff\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.784486 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.794025 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87d61aea-46f6-4ee6-a7a4-767a6a67fea4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e78b6d08e4d616cefbb73e3fd43b3c6886dcd8aa997c7b4948bf1ff63401293\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d468715c2da33996561c40043216233fd12e21539d0c58dd1f3b892fd9f91a9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b184f7c936dc4046f74c0018269cbc2c4b7ceddb650b4e871d3f8a707e0df88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://269c5074371bd00f8363fb3888c11f1df383081a61497aea6578f08700881d50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b25bc9395831524c53f9e565b4efbe74f21792efafeb1dd507161bc91aa0ecdf\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T09:07:22Z\\\",\\\"message\\\":\\\"W0202 09:07:21.597488 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0202 09:07:21.598117 1 crypto.go:601] Generating new CA for check-endpoints-signer@1770023241 cert, and key in /tmp/serving-cert-3856183724/serving-signer.crt, /tmp/serving-cert-3856183724/serving-signer.key\\\\nI0202 09:07:22.012184 1 observer_polling.go:159] Starting file observer\\\\nW0202 09:07:22.015493 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0202 09:07:22.015797 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 09:07:22.016596 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3856183724/tls.crt::/tmp/serving-cert-3856183724/tls.key\\\\\\\"\\\\nF0202 09:07:22.270605 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b44e17b8ca01f549bc10081dd5f109cb1b9fff364bf5bd4fd90d9f0e3eeb04db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfb06a5ec9262cf39b108a66014d15e325cdbd5441a2d9e7a811dd57265c99a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfb06a5ec9262cf39b108a66014d15e325cdbd5441a2d9e7a811dd57265c99a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.806405 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.816602 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.828269 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.828658 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.829269 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.830134 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.830756 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.831327 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.831792 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.833904 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.834819 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.835601 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.836213 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.836752 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.837629 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.838825 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.839956 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.840634 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.841623 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.842232 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.842863 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.843658 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.844221 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.844774 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.845752 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.846330 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.847142 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.847834 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.848695 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.848989 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.849343 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.850347 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.850813 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.851381 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.852286 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.852742 4764 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.852846 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.854961 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.855536 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.856091 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.858290 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.859024 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.859714 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.859738 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.861123 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.862227 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/aaba93da-12a5-43df-b7c7-4c8b988163c6-hosts-file\") pod \"node-resolver-6zxff\" (UID: \"aaba93da-12a5-43df-b7c7-4c8b988163c6\") " pod="openshift-dns/node-resolver-6zxff" Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.862372 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xqwb5\" (UniqueName: \"kubernetes.io/projected/aaba93da-12a5-43df-b7c7-4c8b988163c6-kube-api-access-xqwb5\") pod \"node-resolver-6zxff\" (UID: \"aaba93da-12a5-43df-b7c7-4c8b988163c6\") " pod="openshift-dns/node-resolver-6zxff" Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.862403 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.863561 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.864384 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.865738 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.867006 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.867536 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.868772 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.869521 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.871250 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.871716 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.871995 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.873030 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.873604 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.874362 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.875573 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.876373 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.877021 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.881519 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.893435 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.901418 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6zxff" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aaba93da-12a5-43df-b7c7-4c8b988163c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqwb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6zxff\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.911598 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87d61aea-46f6-4ee6-a7a4-767a6a67fea4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e78b6d08e4d616cefbb73e3fd43b3c6886dcd8aa997c7b4948bf1ff63401293\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d468715c2da33996561c40043216233fd12e21539d0c58dd1f3b892fd9f91a9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b184f7c936dc4046f74c0018269cbc2c4b7ceddb650b4e871d3f8a707e0df88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://269c5074371bd00f8363fb3888c11f1df383081a61497aea6578f08700881d50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b25bc9395831524c53f9e565b4efbe74f21792efafeb1dd507161bc91aa0ecdf\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T09:07:22Z\\\",\\\"message\\\":\\\"W0202 09:07:21.597488 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0202 09:07:21.598117 1 crypto.go:601] Generating new CA for check-endpoints-signer@1770023241 cert, and key in /tmp/serving-cert-3856183724/serving-signer.crt, /tmp/serving-cert-3856183724/serving-signer.key\\\\nI0202 09:07:22.012184 1 observer_polling.go:159] Starting file observer\\\\nW0202 09:07:22.015493 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0202 09:07:22.015797 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 09:07:22.016596 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3856183724/tls.crt::/tmp/serving-cert-3856183724/tls.key\\\\\\\"\\\\nF0202 09:07:22.270605 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b44e17b8ca01f549bc10081dd5f109cb1b9fff364bf5bd4fd90d9f0e3eeb04db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfb06a5ec9262cf39b108a66014d15e325cdbd5441a2d9e7a811dd57265c99a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfb06a5ec9262cf39b108a66014d15e325cdbd5441a2d9e7a811dd57265c99a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.946284 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"aa9d037e66cb53b9dd698eff46284c9e718fd842cda6f7b525f0c79062d70bbf"} Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.946364 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"954af71078823b4fef7b230f232ef51296f38d685b8a7c9cafea637ca89676d2"} Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.947243 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"c55b340ba65dd8b95b1aedc0e82dbf44862ed7ebc2800c71456d0d963283ee60"} Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.949261 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"5528225d91ac2778bef4554abb0bd036ac5d3b6cdac7357bab69e70c3894b6f1"} Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.949295 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"071bbb80d0747cf3e109ba626d47f7d05f85f4cc0b3aba77276c795741f47ac6"} Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.949308 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"fafb2af40e0aed751bb0cb7cb646165b56399911a4c09ced1689eeed3a82b6cc"} Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.958327 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 09:07:39 crc kubenswrapper[4764]: E0202 09:07:39.958665 4764 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"etcd-crc\" already exists" pod="openshift-etcd/etcd-crc" Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.963099 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/aaba93da-12a5-43df-b7c7-4c8b988163c6-hosts-file\") pod \"node-resolver-6zxff\" (UID: \"aaba93da-12a5-43df-b7c7-4c8b988163c6\") " pod="openshift-dns/node-resolver-6zxff" Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.963230 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xqwb5\" (UniqueName: \"kubernetes.io/projected/aaba93da-12a5-43df-b7c7-4c8b988163c6-kube-api-access-xqwb5\") pod \"node-resolver-6zxff\" (UID: \"aaba93da-12a5-43df-b7c7-4c8b988163c6\") " pod="openshift-dns/node-resolver-6zxff" Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.963248 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/aaba93da-12a5-43df-b7c7-4c8b988163c6-hosts-file\") pod \"node-resolver-6zxff\" (UID: \"aaba93da-12a5-43df-b7c7-4c8b988163c6\") " pod="openshift-dns/node-resolver-6zxff" Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.976246 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d66ab3c0-9b5b-41d5-b6c0-91ccddccd7e9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d92ac486f5dade46bc75de32099859a71fc2ed25a7abbe6ae8937df128e7df30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a4ac273a4f127d79e5fad7d4c30cd2044a346f5734dba5d33d68cd092c0c2d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d38c7cb5dcd22fb07524494ebe77aa1a7f4f808f566043563817e47e7ec12a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://408722c7f83d28919431532f8c797a1c795fd5d96736d14e7f06a05744899702\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f473b18cd8003f75bb1dafb82bcec45745a73ccd10434297356dac185a3d15c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e229f484d19cbdabedf53d90a822a4434db9e00b1301538e885e26acc19d25f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e229f484d19cbdabedf53d90a822a4434db9e00b1301538e885e26acc19d25f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a9cc673be8ff245dba81278e8cd34cfe45ed36ff0b8bd5b30a0ddf333ec6371\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a9cc673be8ff245dba81278e8cd34cfe45ed36ff0b8bd5b30a0ddf333ec6371\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bc7e49c7a5ae66ea939060192fa6a346dc0b0b9aa9750b29879d3c426cc0d276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc7e49c7a5ae66ea939060192fa6a346dc0b0b9aa9750b29879d3c426cc0d276\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.978560 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xqwb5\" (UniqueName: \"kubernetes.io/projected/aaba93da-12a5-43df-b7c7-4c8b988163c6-kube-api-access-xqwb5\") pod \"node-resolver-6zxff\" (UID: \"aaba93da-12a5-43df-b7c7-4c8b988163c6\") " pod="openshift-dns/node-resolver-6zxff" Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.985713 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa9d037e66cb53b9dd698eff46284c9e718fd842cda6f7b525f0c79062d70bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 09:07:39 crc kubenswrapper[4764]: I0202 09:07:39.994276 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.005799 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.017612 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.025135 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6zxff" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aaba93da-12a5-43df-b7c7-4c8b988163c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqwb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6zxff\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.036171 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87d61aea-46f6-4ee6-a7a4-767a6a67fea4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e78b6d08e4d616cefbb73e3fd43b3c6886dcd8aa997c7b4948bf1ff63401293\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d468715c2da33996561c40043216233fd12e21539d0c58dd1f3b892fd9f91a9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b184f7c936dc4046f74c0018269cbc2c4b7ceddb650b4e871d3f8a707e0df88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://269c5074371bd00f8363fb3888c11f1df383081a61497aea6578f08700881d50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b25bc9395831524c53f9e565b4efbe74f21792efafeb1dd507161bc91aa0ecdf\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T09:07:22Z\\\",\\\"message\\\":\\\"W0202 09:07:21.597488 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0202 09:07:21.598117 1 crypto.go:601] Generating new CA for check-endpoints-signer@1770023241 cert, and key in /tmp/serving-cert-3856183724/serving-signer.crt, /tmp/serving-cert-3856183724/serving-signer.key\\\\nI0202 09:07:22.012184 1 observer_polling.go:159] Starting file observer\\\\nW0202 09:07:22.015493 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0202 09:07:22.015797 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 09:07:22.016596 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3856183724/tls.crt::/tmp/serving-cert-3856183724/tls.key\\\\\\\"\\\\nF0202 09:07:22.270605 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b44e17b8ca01f549bc10081dd5f109cb1b9fff364bf5bd4fd90d9f0e3eeb04db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfb06a5ec9262cf39b108a66014d15e325cdbd5441a2d9e7a811dd57265c99a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfb06a5ec9262cf39b108a66014d15e325cdbd5441a2d9e7a811dd57265c99a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.045990 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.049038 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-6zxff" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.059223 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6zxff" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aaba93da-12a5-43df-b7c7-4c8b988163c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqwb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6zxff\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.078012 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87d61aea-46f6-4ee6-a7a4-767a6a67fea4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e78b6d08e4d616cefbb73e3fd43b3c6886dcd8aa997c7b4948bf1ff63401293\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d468715c2da33996561c40043216233fd12e21539d0c58dd1f3b892fd9f91a9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b184f7c936dc4046f74c0018269cbc2c4b7ceddb650b4e871d3f8a707e0df88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://269c5074371bd00f8363fb3888c11f1df383081a61497aea6578f08700881d50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b25bc9395831524c53f9e565b4efbe74f21792efafeb1dd507161bc91aa0ecdf\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T09:07:22Z\\\",\\\"message\\\":\\\"W0202 09:07:21.597488 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0202 09:07:21.598117 1 crypto.go:601] Generating new CA for check-endpoints-signer@1770023241 cert, and key in /tmp/serving-cert-3856183724/serving-signer.crt, /tmp/serving-cert-3856183724/serving-signer.key\\\\nI0202 09:07:22.012184 1 observer_polling.go:159] Starting file observer\\\\nW0202 09:07:22.015493 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0202 09:07:22.015797 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 09:07:22.016596 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3856183724/tls.crt::/tmp/serving-cert-3856183724/tls.key\\\\\\\"\\\\nF0202 09:07:22.270605 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b44e17b8ca01f549bc10081dd5f109cb1b9fff364bf5bd4fd90d9f0e3eeb04db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfb06a5ec9262cf39b108a66014d15e325cdbd5441a2d9e7a811dd57265c99a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfb06a5ec9262cf39b108a66014d15e325cdbd5441a2d9e7a811dd57265c99a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.091420 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5528225d91ac2778bef4554abb0bd036ac5d3b6cdac7357bab69e70c3894b6f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://071bbb80d0747cf3e109ba626d47f7d05f85f4cc0b3aba77276c795741f47ac6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.106697 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:40Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.123254 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:40Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.123404 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-nc7qt"] Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.123614 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-4ndm4"] Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.123773 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-9p5dc"] Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.124074 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-nc7qt" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.124148 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.125358 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-9p5dc" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.126010 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.126794 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.127674 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.127719 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.127899 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.127915 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.128846 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.129285 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.129531 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.129649 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.129815 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.130006 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.162154 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d66ab3c0-9b5b-41d5-b6c0-91ccddccd7e9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d92ac486f5dade46bc75de32099859a71fc2ed25a7abbe6ae8937df128e7df30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a4ac273a4f127d79e5fad7d4c30cd2044a346f5734dba5d33d68cd092c0c2d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d38c7cb5dcd22fb07524494ebe77aa1a7f4f808f566043563817e47e7ec12a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://408722c7f83d28919431532f8c797a1c795fd5d96736d14e7f06a05744899702\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f473b18cd8003f75bb1dafb82bcec45745a73ccd10434297356dac185a3d15c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e229f484d19cbdabedf53d90a822a4434db9e00b1301538e885e26acc19d25f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e229f484d19cbdabedf53d90a822a4434db9e00b1301538e885e26acc19d25f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a9cc673be8ff245dba81278e8cd34cfe45ed36ff0b8bd5b30a0ddf333ec6371\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a9cc673be8ff245dba81278e8cd34cfe45ed36ff0b8bd5b30a0ddf333ec6371\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bc7e49c7a5ae66ea939060192fa6a346dc0b0b9aa9750b29879d3c426cc0d276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc7e49c7a5ae66ea939060192fa6a346dc0b0b9aa9750b29879d3c426cc0d276\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:40Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.175266 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa9d037e66cb53b9dd698eff46284c9e718fd842cda6f7b525f0c79062d70bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:40Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.190919 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:40Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.207003 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:40Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.219922 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:40Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.231307 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:40Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.243385 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nc7qt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2hz4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nc7qt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:40Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.258362 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9p5dc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"660ac006-2495-45c4-9fb3-e9c1dddcf7a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9p5dc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:40Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.266201 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e-host-run-netns\") pod \"multus-nc7qt\" (UID: \"6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e\") " pod="openshift-multus/multus-nc7qt" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.266245 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e-etc-kubernetes\") pod \"multus-nc7qt\" (UID: \"6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e\") " pod="openshift-multus/multus-nc7qt" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.266273 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e-multus-socket-dir-parent\") pod \"multus-nc7qt\" (UID: \"6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e\") " pod="openshift-multus/multus-nc7qt" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.266308 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e-cnibin\") pod \"multus-nc7qt\" (UID: \"6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e\") " pod="openshift-multus/multus-nc7qt" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.266337 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e-os-release\") pod \"multus-nc7qt\" (UID: \"6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e\") " pod="openshift-multus/multus-nc7qt" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.266368 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e-hostroot\") pod \"multus-nc7qt\" (UID: \"6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e\") " pod="openshift-multus/multus-nc7qt" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.266392 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/d192f670-9f9d-4539-9641-e4bed73acdd4-proxy-tls\") pod \"machine-config-daemon-4ndm4\" (UID: \"d192f670-9f9d-4539-9641-e4bed73acdd4\") " pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.266406 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e-system-cni-dir\") pod \"multus-nc7qt\" (UID: \"6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e\") " pod="openshift-multus/multus-nc7qt" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.266420 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e-host-var-lib-kubelet\") pod \"multus-nc7qt\" (UID: \"6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e\") " pod="openshift-multus/multus-nc7qt" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.266438 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fbksf\" (UniqueName: \"kubernetes.io/projected/d192f670-9f9d-4539-9641-e4bed73acdd4-kube-api-access-fbksf\") pod \"machine-config-daemon-4ndm4\" (UID: \"d192f670-9f9d-4539-9641-e4bed73acdd4\") " pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.266453 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e-cni-binary-copy\") pod \"multus-nc7qt\" (UID: \"6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e\") " pod="openshift-multus/multus-nc7qt" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.266467 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e-host-run-multus-certs\") pod \"multus-nc7qt\" (UID: \"6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e\") " pod="openshift-multus/multus-nc7qt" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.266482 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d192f670-9f9d-4539-9641-e4bed73acdd4-mcd-auth-proxy-config\") pod \"machine-config-daemon-4ndm4\" (UID: \"d192f670-9f9d-4539-9641-e4bed73acdd4\") " pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.266496 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/660ac006-2495-45c4-9fb3-e9c1dddcf7a0-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-9p5dc\" (UID: \"660ac006-2495-45c4-9fb3-e9c1dddcf7a0\") " pod="openshift-multus/multus-additional-cni-plugins-9p5dc" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.266510 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n2hz4\" (UniqueName: \"kubernetes.io/projected/6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e-kube-api-access-n2hz4\") pod \"multus-nc7qt\" (UID: \"6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e\") " pod="openshift-multus/multus-nc7qt" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.266526 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ch5jr\" (UniqueName: \"kubernetes.io/projected/660ac006-2495-45c4-9fb3-e9c1dddcf7a0-kube-api-access-ch5jr\") pod \"multus-additional-cni-plugins-9p5dc\" (UID: \"660ac006-2495-45c4-9fb3-e9c1dddcf7a0\") " pod="openshift-multus/multus-additional-cni-plugins-9p5dc" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.266561 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e-multus-cni-dir\") pod \"multus-nc7qt\" (UID: \"6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e\") " pod="openshift-multus/multus-nc7qt" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.266577 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e-host-var-lib-cni-bin\") pod \"multus-nc7qt\" (UID: \"6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e\") " pod="openshift-multus/multus-nc7qt" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.266591 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e-multus-daemon-config\") pod \"multus-nc7qt\" (UID: \"6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e\") " pod="openshift-multus/multus-nc7qt" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.266658 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/660ac006-2495-45c4-9fb3-e9c1dddcf7a0-cni-binary-copy\") pod \"multus-additional-cni-plugins-9p5dc\" (UID: \"660ac006-2495-45c4-9fb3-e9c1dddcf7a0\") " pod="openshift-multus/multus-additional-cni-plugins-9p5dc" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.266696 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e-host-var-lib-cni-multus\") pod \"multus-nc7qt\" (UID: \"6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e\") " pod="openshift-multus/multus-nc7qt" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.266729 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/660ac006-2495-45c4-9fb3-e9c1dddcf7a0-tuning-conf-dir\") pod \"multus-additional-cni-plugins-9p5dc\" (UID: \"660ac006-2495-45c4-9fb3-e9c1dddcf7a0\") " pod="openshift-multus/multus-additional-cni-plugins-9p5dc" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.266778 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e-multus-conf-dir\") pod \"multus-nc7qt\" (UID: \"6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e\") " pod="openshift-multus/multus-nc7qt" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.266797 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/660ac006-2495-45c4-9fb3-e9c1dddcf7a0-cnibin\") pod \"multus-additional-cni-plugins-9p5dc\" (UID: \"660ac006-2495-45c4-9fb3-e9c1dddcf7a0\") " pod="openshift-multus/multus-additional-cni-plugins-9p5dc" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.266829 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/660ac006-2495-45c4-9fb3-e9c1dddcf7a0-os-release\") pod \"multus-additional-cni-plugins-9p5dc\" (UID: \"660ac006-2495-45c4-9fb3-e9c1dddcf7a0\") " pod="openshift-multus/multus-additional-cni-plugins-9p5dc" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.266844 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/d192f670-9f9d-4539-9641-e4bed73acdd4-rootfs\") pod \"machine-config-daemon-4ndm4\" (UID: \"d192f670-9f9d-4539-9641-e4bed73acdd4\") " pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.266858 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/660ac006-2495-45c4-9fb3-e9c1dddcf7a0-system-cni-dir\") pod \"multus-additional-cni-plugins-9p5dc\" (UID: \"660ac006-2495-45c4-9fb3-e9c1dddcf7a0\") " pod="openshift-multus/multus-additional-cni-plugins-9p5dc" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.266872 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e-host-run-k8s-cni-cncf-io\") pod \"multus-nc7qt\" (UID: \"6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e\") " pod="openshift-multus/multus-nc7qt" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.271343 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87d61aea-46f6-4ee6-a7a4-767a6a67fea4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e78b6d08e4d616cefbb73e3fd43b3c6886dcd8aa997c7b4948bf1ff63401293\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d468715c2da33996561c40043216233fd12e21539d0c58dd1f3b892fd9f91a9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b184f7c936dc4046f74c0018269cbc2c4b7ceddb650b4e871d3f8a707e0df88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://269c5074371bd00f8363fb3888c11f1df383081a61497aea6578f08700881d50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b25bc9395831524c53f9e565b4efbe74f21792efafeb1dd507161bc91aa0ecdf\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T09:07:22Z\\\",\\\"message\\\":\\\"W0202 09:07:21.597488 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0202 09:07:21.598117 1 crypto.go:601] Generating new CA for check-endpoints-signer@1770023241 cert, and key in /tmp/serving-cert-3856183724/serving-signer.crt, /tmp/serving-cert-3856183724/serving-signer.key\\\\nI0202 09:07:22.012184 1 observer_polling.go:159] Starting file observer\\\\nW0202 09:07:22.015493 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0202 09:07:22.015797 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 09:07:22.016596 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3856183724/tls.crt::/tmp/serving-cert-3856183724/tls.key\\\\\\\"\\\\nF0202 09:07:22.270605 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b44e17b8ca01f549bc10081dd5f109cb1b9fff364bf5bd4fd90d9f0e3eeb04db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfb06a5ec9262cf39b108a66014d15e325cdbd5441a2d9e7a811dd57265c99a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfb06a5ec9262cf39b108a66014d15e325cdbd5441a2d9e7a811dd57265c99a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:40Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.284037 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5528225d91ac2778bef4554abb0bd036ac5d3b6cdac7357bab69e70c3894b6f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://071bbb80d0747cf3e109ba626d47f7d05f85f4cc0b3aba77276c795741f47ac6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:40Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.302417 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d66ab3c0-9b5b-41d5-b6c0-91ccddccd7e9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d92ac486f5dade46bc75de32099859a71fc2ed25a7abbe6ae8937df128e7df30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a4ac273a4f127d79e5fad7d4c30cd2044a346f5734dba5d33d68cd092c0c2d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d38c7cb5dcd22fb07524494ebe77aa1a7f4f808f566043563817e47e7ec12a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://408722c7f83d28919431532f8c797a1c795fd5d96736d14e7f06a05744899702\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f473b18cd8003f75bb1dafb82bcec45745a73ccd10434297356dac185a3d15c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e229f484d19cbdabedf53d90a822a4434db9e00b1301538e885e26acc19d25f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e229f484d19cbdabedf53d90a822a4434db9e00b1301538e885e26acc19d25f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a9cc673be8ff245dba81278e8cd34cfe45ed36ff0b8bd5b30a0ddf333ec6371\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a9cc673be8ff245dba81278e8cd34cfe45ed36ff0b8bd5b30a0ddf333ec6371\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bc7e49c7a5ae66ea939060192fa6a346dc0b0b9aa9750b29879d3c426cc0d276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc7e49c7a5ae66ea939060192fa6a346dc0b0b9aa9750b29879d3c426cc0d276\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:40Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.314896 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa9d037e66cb53b9dd698eff46284c9e718fd842cda6f7b525f0c79062d70bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:40Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.325899 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:40Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.336364 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d192f670-9f9d-4539-9641-e4bed73acdd4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbksf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbksf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4ndm4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:40Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.347388 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:40Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.367107 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.367214 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/660ac006-2495-45c4-9fb3-e9c1dddcf7a0-cni-binary-copy\") pod \"multus-additional-cni-plugins-9p5dc\" (UID: \"660ac006-2495-45c4-9fb3-e9c1dddcf7a0\") " pod="openshift-multus/multus-additional-cni-plugins-9p5dc" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.367243 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e-host-var-lib-cni-multus\") pod \"multus-nc7qt\" (UID: \"6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e\") " pod="openshift-multus/multus-nc7qt" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.367269 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/660ac006-2495-45c4-9fb3-e9c1dddcf7a0-tuning-conf-dir\") pod \"multus-additional-cni-plugins-9p5dc\" (UID: \"660ac006-2495-45c4-9fb3-e9c1dddcf7a0\") " pod="openshift-multus/multus-additional-cni-plugins-9p5dc" Feb 02 09:07:40 crc kubenswrapper[4764]: E0202 09:07:40.367292 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 09:07:42.367266268 +0000 UTC m=+25.300990356 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.367332 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e-multus-conf-dir\") pod \"multus-nc7qt\" (UID: \"6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e\") " pod="openshift-multus/multus-nc7qt" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.367367 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/660ac006-2495-45c4-9fb3-e9c1dddcf7a0-cnibin\") pod \"multus-additional-cni-plugins-9p5dc\" (UID: \"660ac006-2495-45c4-9fb3-e9c1dddcf7a0\") " pod="openshift-multus/multus-additional-cni-plugins-9p5dc" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.367382 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/660ac006-2495-45c4-9fb3-e9c1dddcf7a0-os-release\") pod \"multus-additional-cni-plugins-9p5dc\" (UID: \"660ac006-2495-45c4-9fb3-e9c1dddcf7a0\") " pod="openshift-multus/multus-additional-cni-plugins-9p5dc" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.367400 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/d192f670-9f9d-4539-9641-e4bed73acdd4-rootfs\") pod \"machine-config-daemon-4ndm4\" (UID: \"d192f670-9f9d-4539-9641-e4bed73acdd4\") " pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.367416 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/660ac006-2495-45c4-9fb3-e9c1dddcf7a0-system-cni-dir\") pod \"multus-additional-cni-plugins-9p5dc\" (UID: \"660ac006-2495-45c4-9fb3-e9c1dddcf7a0\") " pod="openshift-multus/multus-additional-cni-plugins-9p5dc" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.367431 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e-host-run-k8s-cni-cncf-io\") pod \"multus-nc7qt\" (UID: \"6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e\") " pod="openshift-multus/multus-nc7qt" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.367448 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e-host-run-netns\") pod \"multus-nc7qt\" (UID: \"6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e\") " pod="openshift-multus/multus-nc7qt" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.367462 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e-etc-kubernetes\") pod \"multus-nc7qt\" (UID: \"6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e\") " pod="openshift-multus/multus-nc7qt" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.367524 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e-multus-conf-dir\") pod \"multus-nc7qt\" (UID: \"6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e\") " pod="openshift-multus/multus-nc7qt" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.367549 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e-host-var-lib-cni-multus\") pod \"multus-nc7qt\" (UID: \"6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e\") " pod="openshift-multus/multus-nc7qt" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.367610 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e-multus-socket-dir-parent\") pod \"multus-nc7qt\" (UID: \"6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e\") " pod="openshift-multus/multus-nc7qt" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.367647 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/660ac006-2495-45c4-9fb3-e9c1dddcf7a0-os-release\") pod \"multus-additional-cni-plugins-9p5dc\" (UID: \"660ac006-2495-45c4-9fb3-e9c1dddcf7a0\") " pod="openshift-multus/multus-additional-cni-plugins-9p5dc" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.367680 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e-os-release\") pod \"multus-nc7qt\" (UID: \"6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e\") " pod="openshift-multus/multus-nc7qt" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.367691 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e-etc-kubernetes\") pod \"multus-nc7qt\" (UID: \"6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e\") " pod="openshift-multus/multus-nc7qt" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.367723 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e-hostroot\") pod \"multus-nc7qt\" (UID: \"6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e\") " pod="openshift-multus/multus-nc7qt" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.367742 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e-multus-socket-dir-parent\") pod \"multus-nc7qt\" (UID: \"6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e\") " pod="openshift-multus/multus-nc7qt" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.367745 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e-host-run-k8s-cni-cncf-io\") pod \"multus-nc7qt\" (UID: \"6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e\") " pod="openshift-multus/multus-nc7qt" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.367699 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e-hostroot\") pod \"multus-nc7qt\" (UID: \"6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e\") " pod="openshift-multus/multus-nc7qt" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.367783 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/660ac006-2495-45c4-9fb3-e9c1dddcf7a0-cnibin\") pod \"multus-additional-cni-plugins-9p5dc\" (UID: \"660ac006-2495-45c4-9fb3-e9c1dddcf7a0\") " pod="openshift-multus/multus-additional-cni-plugins-9p5dc" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.367789 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/d192f670-9f9d-4539-9641-e4bed73acdd4-proxy-tls\") pod \"machine-config-daemon-4ndm4\" (UID: \"d192f670-9f9d-4539-9641-e4bed73acdd4\") " pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.367807 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e-host-run-netns\") pod \"multus-nc7qt\" (UID: \"6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e\") " pod="openshift-multus/multus-nc7qt" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.367822 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e-os-release\") pod \"multus-nc7qt\" (UID: \"6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e\") " pod="openshift-multus/multus-nc7qt" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.367832 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/660ac006-2495-45c4-9fb3-e9c1dddcf7a0-system-cni-dir\") pod \"multus-additional-cni-plugins-9p5dc\" (UID: \"660ac006-2495-45c4-9fb3-e9c1dddcf7a0\") " pod="openshift-multus/multus-additional-cni-plugins-9p5dc" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.367836 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e-system-cni-dir\") pod \"multus-nc7qt\" (UID: \"6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e\") " pod="openshift-multus/multus-nc7qt" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.367864 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/d192f670-9f9d-4539-9641-e4bed73acdd4-rootfs\") pod \"machine-config-daemon-4ndm4\" (UID: \"d192f670-9f9d-4539-9641-e4bed73acdd4\") " pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.367875 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e-cnibin\") pod \"multus-nc7qt\" (UID: \"6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e\") " pod="openshift-multus/multus-nc7qt" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.367912 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e-host-var-lib-kubelet\") pod \"multus-nc7qt\" (UID: \"6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e\") " pod="openshift-multus/multus-nc7qt" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.367945 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e-cnibin\") pod \"multus-nc7qt\" (UID: \"6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e\") " pod="openshift-multus/multus-nc7qt" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.367956 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fbksf\" (UniqueName: \"kubernetes.io/projected/d192f670-9f9d-4539-9641-e4bed73acdd4-kube-api-access-fbksf\") pod \"machine-config-daemon-4ndm4\" (UID: \"d192f670-9f9d-4539-9641-e4bed73acdd4\") " pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.367992 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e-cni-binary-copy\") pod \"multus-nc7qt\" (UID: \"6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e\") " pod="openshift-multus/multus-nc7qt" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.368011 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e-host-run-multus-certs\") pod \"multus-nc7qt\" (UID: \"6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e\") " pod="openshift-multus/multus-nc7qt" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.367988 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e-host-var-lib-kubelet\") pod \"multus-nc7qt\" (UID: \"6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e\") " pod="openshift-multus/multus-nc7qt" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.368035 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d192f670-9f9d-4539-9641-e4bed73acdd4-mcd-auth-proxy-config\") pod \"machine-config-daemon-4ndm4\" (UID: \"d192f670-9f9d-4539-9641-e4bed73acdd4\") " pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.368068 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e-host-run-multus-certs\") pod \"multus-nc7qt\" (UID: \"6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e\") " pod="openshift-multus/multus-nc7qt" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.368070 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/660ac006-2495-45c4-9fb3-e9c1dddcf7a0-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-9p5dc\" (UID: \"660ac006-2495-45c4-9fb3-e9c1dddcf7a0\") " pod="openshift-multus/multus-additional-cni-plugins-9p5dc" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.368104 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ch5jr\" (UniqueName: \"kubernetes.io/projected/660ac006-2495-45c4-9fb3-e9c1dddcf7a0-kube-api-access-ch5jr\") pod \"multus-additional-cni-plugins-9p5dc\" (UID: \"660ac006-2495-45c4-9fb3-e9c1dddcf7a0\") " pod="openshift-multus/multus-additional-cni-plugins-9p5dc" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.368120 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e-system-cni-dir\") pod \"multus-nc7qt\" (UID: \"6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e\") " pod="openshift-multus/multus-nc7qt" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.368259 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e-multus-cni-dir\") pod \"multus-nc7qt\" (UID: \"6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e\") " pod="openshift-multus/multus-nc7qt" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.368813 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6zxff" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aaba93da-12a5-43df-b7c7-4c8b988163c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqwb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6zxff\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:40Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.368123 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e-multus-cni-dir\") pod \"multus-nc7qt\" (UID: \"6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e\") " pod="openshift-multus/multus-nc7qt" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.369001 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/660ac006-2495-45c4-9fb3-e9c1dddcf7a0-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-9p5dc\" (UID: \"660ac006-2495-45c4-9fb3-e9c1dddcf7a0\") " pod="openshift-multus/multus-additional-cni-plugins-9p5dc" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.369021 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e-host-var-lib-cni-bin\") pod \"multus-nc7qt\" (UID: \"6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e\") " pod="openshift-multus/multus-nc7qt" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.369049 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e-multus-daemon-config\") pod \"multus-nc7qt\" (UID: \"6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e\") " pod="openshift-multus/multus-nc7qt" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.369074 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n2hz4\" (UniqueName: \"kubernetes.io/projected/6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e-kube-api-access-n2hz4\") pod \"multus-nc7qt\" (UID: \"6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e\") " pod="openshift-multus/multus-nc7qt" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.369110 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e-host-var-lib-cni-bin\") pod \"multus-nc7qt\" (UID: \"6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e\") " pod="openshift-multus/multus-nc7qt" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.369269 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e-cni-binary-copy\") pod \"multus-nc7qt\" (UID: \"6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e\") " pod="openshift-multus/multus-nc7qt" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.369310 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d192f670-9f9d-4539-9641-e4bed73acdd4-mcd-auth-proxy-config\") pod \"machine-config-daemon-4ndm4\" (UID: \"d192f670-9f9d-4539-9641-e4bed73acdd4\") " pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.369657 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e-multus-daemon-config\") pod \"multus-nc7qt\" (UID: \"6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e\") " pod="openshift-multus/multus-nc7qt" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.369831 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/660ac006-2495-45c4-9fb3-e9c1dddcf7a0-tuning-conf-dir\") pod \"multus-additional-cni-plugins-9p5dc\" (UID: \"660ac006-2495-45c4-9fb3-e9c1dddcf7a0\") " pod="openshift-multus/multus-additional-cni-plugins-9p5dc" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.370050 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/660ac006-2495-45c4-9fb3-e9c1dddcf7a0-cni-binary-copy\") pod \"multus-additional-cni-plugins-9p5dc\" (UID: \"660ac006-2495-45c4-9fb3-e9c1dddcf7a0\") " pod="openshift-multus/multus-additional-cni-plugins-9p5dc" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.373833 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/d192f670-9f9d-4539-9641-e4bed73acdd4-proxy-tls\") pod \"machine-config-daemon-4ndm4\" (UID: \"d192f670-9f9d-4539-9641-e4bed73acdd4\") " pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.394241 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fbksf\" (UniqueName: \"kubernetes.io/projected/d192f670-9f9d-4539-9641-e4bed73acdd4-kube-api-access-fbksf\") pod \"machine-config-daemon-4ndm4\" (UID: \"d192f670-9f9d-4539-9641-e4bed73acdd4\") " pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.413215 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ch5jr\" (UniqueName: \"kubernetes.io/projected/660ac006-2495-45c4-9fb3-e9c1dddcf7a0-kube-api-access-ch5jr\") pod \"multus-additional-cni-plugins-9p5dc\" (UID: \"660ac006-2495-45c4-9fb3-e9c1dddcf7a0\") " pod="openshift-multus/multus-additional-cni-plugins-9p5dc" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.435289 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n2hz4\" (UniqueName: \"kubernetes.io/projected/6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e-kube-api-access-n2hz4\") pod \"multus-nc7qt\" (UID: \"6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e\") " pod="openshift-multus/multus-nc7qt" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.440926 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-nc7qt" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.448769 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" Feb 02 09:07:40 crc kubenswrapper[4764]: W0202 09:07:40.452199 4764 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6ae2eac6_1e25_4166_bb6b_7e0e0dd4be6e.slice/crio-9add6c44ccad318e1d730c7e24782c710b7c873914ba50783dada9aaa5e38a4a WatchSource:0}: Error finding container 9add6c44ccad318e1d730c7e24782c710b7c873914ba50783dada9aaa5e38a4a: Status 404 returned error can't find the container with id 9add6c44ccad318e1d730c7e24782c710b7c873914ba50783dada9aaa5e38a4a Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.457051 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-9p5dc" Feb 02 09:07:40 crc kubenswrapper[4764]: W0202 09:07:40.468659 4764 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd192f670_9f9d_4539_9641_e4bed73acdd4.slice/crio-39737477815bb7ac16890fc0da49704b233e0dd88343b75b7b34865f87189f60 WatchSource:0}: Error finding container 39737477815bb7ac16890fc0da49704b233e0dd88343b75b7b34865f87189f60: Status 404 returned error can't find the container with id 39737477815bb7ac16890fc0da49704b233e0dd88343b75b7b34865f87189f60 Feb 02 09:07:40 crc kubenswrapper[4764]: E0202 09:07:40.469644 4764 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 02 09:07:40 crc kubenswrapper[4764]: E0202 09:07:40.469709 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-02 09:07:42.469690425 +0000 UTC m=+25.403414513 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.469818 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.469855 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.469879 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.469900 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 09:07:40 crc kubenswrapper[4764]: E0202 09:07:40.470022 4764 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 02 09:07:40 crc kubenswrapper[4764]: E0202 09:07:40.470033 4764 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 02 09:07:40 crc kubenswrapper[4764]: E0202 09:07:40.470083 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-02 09:07:42.470067744 +0000 UTC m=+25.403791912 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 02 09:07:40 crc kubenswrapper[4764]: E0202 09:07:40.470086 4764 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 02 09:07:40 crc kubenswrapper[4764]: E0202 09:07:40.470135 4764 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 02 09:07:40 crc kubenswrapper[4764]: E0202 09:07:40.470149 4764 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 02 09:07:40 crc kubenswrapper[4764]: E0202 09:07:40.470181 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-02 09:07:42.470172067 +0000 UTC m=+25.403896155 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 02 09:07:40 crc kubenswrapper[4764]: E0202 09:07:40.470040 4764 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 02 09:07:40 crc kubenswrapper[4764]: E0202 09:07:40.470200 4764 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 02 09:07:40 crc kubenswrapper[4764]: E0202 09:07:40.470228 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-02 09:07:42.470220118 +0000 UTC m=+25.403944296 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 02 09:07:40 crc kubenswrapper[4764]: W0202 09:07:40.482971 4764 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod660ac006_2495_45c4_9fb3_e9c1dddcf7a0.slice/crio-c58f34d80eeafb1c789c4488b07addc1e32bfc1b7a0ab45c2873b2c4e5127e62 WatchSource:0}: Error finding container c58f34d80eeafb1c789c4488b07addc1e32bfc1b7a0ab45c2873b2c4e5127e62: Status 404 returned error can't find the container with id c58f34d80eeafb1c789c4488b07addc1e32bfc1b7a0ab45c2873b2c4e5127e62 Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.495784 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-zhn7j"] Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.496752 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.498761 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.499035 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.499175 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.499440 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.521418 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.540596 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.560164 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.607880 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5528225d91ac2778bef4554abb0bd036ac5d3b6cdac7357bab69e70c3894b6f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://071bbb80d0747cf3e109ba626d47f7d05f85f4cc0b3aba77276c795741f47ac6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:40Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.670212 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:40Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.672320 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/24632dda-6100-4ab6-a28e-214ddae4360c-host-kubelet\") pod \"ovnkube-node-zhn7j\" (UID: \"24632dda-6100-4ab6-a28e-214ddae4360c\") " pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.672366 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/24632dda-6100-4ab6-a28e-214ddae4360c-run-ovn\") pod \"ovnkube-node-zhn7j\" (UID: \"24632dda-6100-4ab6-a28e-214ddae4360c\") " pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.672388 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/24632dda-6100-4ab6-a28e-214ddae4360c-env-overrides\") pod \"ovnkube-node-zhn7j\" (UID: \"24632dda-6100-4ab6-a28e-214ddae4360c\") " pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.672416 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/24632dda-6100-4ab6-a28e-214ddae4360c-var-lib-openvswitch\") pod \"ovnkube-node-zhn7j\" (UID: \"24632dda-6100-4ab6-a28e-214ddae4360c\") " pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.672440 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/24632dda-6100-4ab6-a28e-214ddae4360c-log-socket\") pod \"ovnkube-node-zhn7j\" (UID: \"24632dda-6100-4ab6-a28e-214ddae4360c\") " pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.672475 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/24632dda-6100-4ab6-a28e-214ddae4360c-systemd-units\") pod \"ovnkube-node-zhn7j\" (UID: \"24632dda-6100-4ab6-a28e-214ddae4360c\") " pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.672494 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/24632dda-6100-4ab6-a28e-214ddae4360c-host-slash\") pod \"ovnkube-node-zhn7j\" (UID: \"24632dda-6100-4ab6-a28e-214ddae4360c\") " pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.672514 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/24632dda-6100-4ab6-a28e-214ddae4360c-host-cni-bin\") pod \"ovnkube-node-zhn7j\" (UID: \"24632dda-6100-4ab6-a28e-214ddae4360c\") " pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.672537 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/24632dda-6100-4ab6-a28e-214ddae4360c-ovnkube-config\") pod \"ovnkube-node-zhn7j\" (UID: \"24632dda-6100-4ab6-a28e-214ddae4360c\") " pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.672562 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/24632dda-6100-4ab6-a28e-214ddae4360c-ovn-node-metrics-cert\") pod \"ovnkube-node-zhn7j\" (UID: \"24632dda-6100-4ab6-a28e-214ddae4360c\") " pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.672628 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/24632dda-6100-4ab6-a28e-214ddae4360c-run-systemd\") pod \"ovnkube-node-zhn7j\" (UID: \"24632dda-6100-4ab6-a28e-214ddae4360c\") " pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.672669 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/24632dda-6100-4ab6-a28e-214ddae4360c-host-run-netns\") pod \"ovnkube-node-zhn7j\" (UID: \"24632dda-6100-4ab6-a28e-214ddae4360c\") " pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.672695 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/24632dda-6100-4ab6-a28e-214ddae4360c-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-zhn7j\" (UID: \"24632dda-6100-4ab6-a28e-214ddae4360c\") " pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.672716 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/24632dda-6100-4ab6-a28e-214ddae4360c-etc-openvswitch\") pod \"ovnkube-node-zhn7j\" (UID: \"24632dda-6100-4ab6-a28e-214ddae4360c\") " pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.672736 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/24632dda-6100-4ab6-a28e-214ddae4360c-node-log\") pod \"ovnkube-node-zhn7j\" (UID: \"24632dda-6100-4ab6-a28e-214ddae4360c\") " pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.672776 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7r4sk\" (UniqueName: \"kubernetes.io/projected/24632dda-6100-4ab6-a28e-214ddae4360c-kube-api-access-7r4sk\") pod \"ovnkube-node-zhn7j\" (UID: \"24632dda-6100-4ab6-a28e-214ddae4360c\") " pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.672799 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/24632dda-6100-4ab6-a28e-214ddae4360c-host-cni-netd\") pod \"ovnkube-node-zhn7j\" (UID: \"24632dda-6100-4ab6-a28e-214ddae4360c\") " pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.672860 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/24632dda-6100-4ab6-a28e-214ddae4360c-run-openvswitch\") pod \"ovnkube-node-zhn7j\" (UID: \"24632dda-6100-4ab6-a28e-214ddae4360c\") " pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.672904 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/24632dda-6100-4ab6-a28e-214ddae4360c-host-run-ovn-kubernetes\") pod \"ovnkube-node-zhn7j\" (UID: \"24632dda-6100-4ab6-a28e-214ddae4360c\") " pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.672918 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/24632dda-6100-4ab6-a28e-214ddae4360c-ovnkube-script-lib\") pod \"ovnkube-node-zhn7j\" (UID: \"24632dda-6100-4ab6-a28e-214ddae4360c\") " pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.704264 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nc7qt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2hz4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nc7qt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:40Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.741233 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9p5dc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"660ac006-2495-45c4-9fb3-e9c1dddcf7a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9p5dc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:40Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.769586 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87d61aea-46f6-4ee6-a7a4-767a6a67fea4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e78b6d08e4d616cefbb73e3fd43b3c6886dcd8aa997c7b4948bf1ff63401293\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d468715c2da33996561c40043216233fd12e21539d0c58dd1f3b892fd9f91a9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b184f7c936dc4046f74c0018269cbc2c4b7ceddb650b4e871d3f8a707e0df88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://269c5074371bd00f8363fb3888c11f1df383081a61497aea6578f08700881d50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b25bc9395831524c53f9e565b4efbe74f21792efafeb1dd507161bc91aa0ecdf\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T09:07:22Z\\\",\\\"message\\\":\\\"W0202 09:07:21.597488 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0202 09:07:21.598117 1 crypto.go:601] Generating new CA for check-endpoints-signer@1770023241 cert, and key in /tmp/serving-cert-3856183724/serving-signer.crt, /tmp/serving-cert-3856183724/serving-signer.key\\\\nI0202 09:07:22.012184 1 observer_polling.go:159] Starting file observer\\\\nW0202 09:07:22.015493 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0202 09:07:22.015797 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 09:07:22.016596 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3856183724/tls.crt::/tmp/serving-cert-3856183724/tls.key\\\\\\\"\\\\nF0202 09:07:22.270605 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b44e17b8ca01f549bc10081dd5f109cb1b9fff364bf5bd4fd90d9f0e3eeb04db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfb06a5ec9262cf39b108a66014d15e325cdbd5441a2d9e7a811dd57265c99a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfb06a5ec9262cf39b108a66014d15e325cdbd5441a2d9e7a811dd57265c99a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:40Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.773981 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/24632dda-6100-4ab6-a28e-214ddae4360c-run-systemd\") pod \"ovnkube-node-zhn7j\" (UID: \"24632dda-6100-4ab6-a28e-214ddae4360c\") " pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.774037 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/24632dda-6100-4ab6-a28e-214ddae4360c-host-run-netns\") pod \"ovnkube-node-zhn7j\" (UID: \"24632dda-6100-4ab6-a28e-214ddae4360c\") " pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.774068 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/24632dda-6100-4ab6-a28e-214ddae4360c-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-zhn7j\" (UID: \"24632dda-6100-4ab6-a28e-214ddae4360c\") " pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.774092 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/24632dda-6100-4ab6-a28e-214ddae4360c-etc-openvswitch\") pod \"ovnkube-node-zhn7j\" (UID: \"24632dda-6100-4ab6-a28e-214ddae4360c\") " pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.774118 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/24632dda-6100-4ab6-a28e-214ddae4360c-node-log\") pod \"ovnkube-node-zhn7j\" (UID: \"24632dda-6100-4ab6-a28e-214ddae4360c\") " pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.774123 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/24632dda-6100-4ab6-a28e-214ddae4360c-run-systemd\") pod \"ovnkube-node-zhn7j\" (UID: \"24632dda-6100-4ab6-a28e-214ddae4360c\") " pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.774141 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7r4sk\" (UniqueName: \"kubernetes.io/projected/24632dda-6100-4ab6-a28e-214ddae4360c-kube-api-access-7r4sk\") pod \"ovnkube-node-zhn7j\" (UID: \"24632dda-6100-4ab6-a28e-214ddae4360c\") " pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.774198 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/24632dda-6100-4ab6-a28e-214ddae4360c-host-run-netns\") pod \"ovnkube-node-zhn7j\" (UID: \"24632dda-6100-4ab6-a28e-214ddae4360c\") " pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.774229 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/24632dda-6100-4ab6-a28e-214ddae4360c-host-cni-netd\") pod \"ovnkube-node-zhn7j\" (UID: \"24632dda-6100-4ab6-a28e-214ddae4360c\") " pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.774198 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/24632dda-6100-4ab6-a28e-214ddae4360c-etc-openvswitch\") pod \"ovnkube-node-zhn7j\" (UID: \"24632dda-6100-4ab6-a28e-214ddae4360c\") " pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.774233 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/24632dda-6100-4ab6-a28e-214ddae4360c-node-log\") pod \"ovnkube-node-zhn7j\" (UID: \"24632dda-6100-4ab6-a28e-214ddae4360c\") " pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.774264 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/24632dda-6100-4ab6-a28e-214ddae4360c-host-cni-netd\") pod \"ovnkube-node-zhn7j\" (UID: \"24632dda-6100-4ab6-a28e-214ddae4360c\") " pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.774228 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/24632dda-6100-4ab6-a28e-214ddae4360c-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-zhn7j\" (UID: \"24632dda-6100-4ab6-a28e-214ddae4360c\") " pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.774300 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/24632dda-6100-4ab6-a28e-214ddae4360c-run-openvswitch\") pod \"ovnkube-node-zhn7j\" (UID: \"24632dda-6100-4ab6-a28e-214ddae4360c\") " pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.774326 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/24632dda-6100-4ab6-a28e-214ddae4360c-run-openvswitch\") pod \"ovnkube-node-zhn7j\" (UID: \"24632dda-6100-4ab6-a28e-214ddae4360c\") " pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.774365 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/24632dda-6100-4ab6-a28e-214ddae4360c-host-run-ovn-kubernetes\") pod \"ovnkube-node-zhn7j\" (UID: \"24632dda-6100-4ab6-a28e-214ddae4360c\") " pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.774391 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/24632dda-6100-4ab6-a28e-214ddae4360c-ovnkube-script-lib\") pod \"ovnkube-node-zhn7j\" (UID: \"24632dda-6100-4ab6-a28e-214ddae4360c\") " pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.774425 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/24632dda-6100-4ab6-a28e-214ddae4360c-host-kubelet\") pod \"ovnkube-node-zhn7j\" (UID: \"24632dda-6100-4ab6-a28e-214ddae4360c\") " pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.774445 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/24632dda-6100-4ab6-a28e-214ddae4360c-run-ovn\") pod \"ovnkube-node-zhn7j\" (UID: \"24632dda-6100-4ab6-a28e-214ddae4360c\") " pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.774465 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/24632dda-6100-4ab6-a28e-214ddae4360c-env-overrides\") pod \"ovnkube-node-zhn7j\" (UID: \"24632dda-6100-4ab6-a28e-214ddae4360c\") " pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.774485 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/24632dda-6100-4ab6-a28e-214ddae4360c-var-lib-openvswitch\") pod \"ovnkube-node-zhn7j\" (UID: \"24632dda-6100-4ab6-a28e-214ddae4360c\") " pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.774505 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/24632dda-6100-4ab6-a28e-214ddae4360c-log-socket\") pod \"ovnkube-node-zhn7j\" (UID: \"24632dda-6100-4ab6-a28e-214ddae4360c\") " pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.774544 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/24632dda-6100-4ab6-a28e-214ddae4360c-ovnkube-config\") pod \"ovnkube-node-zhn7j\" (UID: \"24632dda-6100-4ab6-a28e-214ddae4360c\") " pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.774586 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/24632dda-6100-4ab6-a28e-214ddae4360c-ovn-node-metrics-cert\") pod \"ovnkube-node-zhn7j\" (UID: \"24632dda-6100-4ab6-a28e-214ddae4360c\") " pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.774606 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/24632dda-6100-4ab6-a28e-214ddae4360c-systemd-units\") pod \"ovnkube-node-zhn7j\" (UID: \"24632dda-6100-4ab6-a28e-214ddae4360c\") " pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.774639 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/24632dda-6100-4ab6-a28e-214ddae4360c-host-slash\") pod \"ovnkube-node-zhn7j\" (UID: \"24632dda-6100-4ab6-a28e-214ddae4360c\") " pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.774657 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/24632dda-6100-4ab6-a28e-214ddae4360c-host-cni-bin\") pod \"ovnkube-node-zhn7j\" (UID: \"24632dda-6100-4ab6-a28e-214ddae4360c\") " pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.774693 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/24632dda-6100-4ab6-a28e-214ddae4360c-var-lib-openvswitch\") pod \"ovnkube-node-zhn7j\" (UID: \"24632dda-6100-4ab6-a28e-214ddae4360c\") " pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.774713 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/24632dda-6100-4ab6-a28e-214ddae4360c-host-cni-bin\") pod \"ovnkube-node-zhn7j\" (UID: \"24632dda-6100-4ab6-a28e-214ddae4360c\") " pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.774724 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/24632dda-6100-4ab6-a28e-214ddae4360c-host-run-ovn-kubernetes\") pod \"ovnkube-node-zhn7j\" (UID: \"24632dda-6100-4ab6-a28e-214ddae4360c\") " pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.774779 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/24632dda-6100-4ab6-a28e-214ddae4360c-log-socket\") pod \"ovnkube-node-zhn7j\" (UID: \"24632dda-6100-4ab6-a28e-214ddae4360c\") " pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.775080 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/24632dda-6100-4ab6-a28e-214ddae4360c-run-ovn\") pod \"ovnkube-node-zhn7j\" (UID: \"24632dda-6100-4ab6-a28e-214ddae4360c\") " pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.775112 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/24632dda-6100-4ab6-a28e-214ddae4360c-host-kubelet\") pod \"ovnkube-node-zhn7j\" (UID: \"24632dda-6100-4ab6-a28e-214ddae4360c\") " pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.775125 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/24632dda-6100-4ab6-a28e-214ddae4360c-systemd-units\") pod \"ovnkube-node-zhn7j\" (UID: \"24632dda-6100-4ab6-a28e-214ddae4360c\") " pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.775151 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/24632dda-6100-4ab6-a28e-214ddae4360c-host-slash\") pod \"ovnkube-node-zhn7j\" (UID: \"24632dda-6100-4ab6-a28e-214ddae4360c\") " pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.775382 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/24632dda-6100-4ab6-a28e-214ddae4360c-ovnkube-script-lib\") pod \"ovnkube-node-zhn7j\" (UID: \"24632dda-6100-4ab6-a28e-214ddae4360c\") " pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.775515 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/24632dda-6100-4ab6-a28e-214ddae4360c-ovnkube-config\") pod \"ovnkube-node-zhn7j\" (UID: \"24632dda-6100-4ab6-a28e-214ddae4360c\") " pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.775666 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/24632dda-6100-4ab6-a28e-214ddae4360c-env-overrides\") pod \"ovnkube-node-zhn7j\" (UID: \"24632dda-6100-4ab6-a28e-214ddae4360c\") " pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.778499 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/24632dda-6100-4ab6-a28e-214ddae4360c-ovn-node-metrics-cert\") pod \"ovnkube-node-zhn7j\" (UID: \"24632dda-6100-4ab6-a28e-214ddae4360c\") " pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.781284 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-30 08:09:29.42342845 +0000 UTC Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.818492 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7r4sk\" (UniqueName: \"kubernetes.io/projected/24632dda-6100-4ab6-a28e-214ddae4360c-kube-api-access-7r4sk\") pod \"ovnkube-node-zhn7j\" (UID: \"24632dda-6100-4ab6-a28e-214ddae4360c\") " pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.820289 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.824600 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.824622 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.824611 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 09:07:40 crc kubenswrapper[4764]: E0202 09:07:40.824741 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 09:07:40 crc kubenswrapper[4764]: E0202 09:07:40.824844 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 09:07:40 crc kubenswrapper[4764]: E0202 09:07:40.824964 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.830487 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:40Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.878346 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24632dda-6100-4ab6-a28e-214ddae4360c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zhn7j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:40Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.915177 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d66ab3c0-9b5b-41d5-b6c0-91ccddccd7e9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d92ac486f5dade46bc75de32099859a71fc2ed25a7abbe6ae8937df128e7df30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a4ac273a4f127d79e5fad7d4c30cd2044a346f5734dba5d33d68cd092c0c2d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d38c7cb5dcd22fb07524494ebe77aa1a7f4f808f566043563817e47e7ec12a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://408722c7f83d28919431532f8c797a1c795fd5d96736d14e7f06a05744899702\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f473b18cd8003f75bb1dafb82bcec45745a73ccd10434297356dac185a3d15c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e229f484d19cbdabedf53d90a822a4434db9e00b1301538e885e26acc19d25f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e229f484d19cbdabedf53d90a822a4434db9e00b1301538e885e26acc19d25f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a9cc673be8ff245dba81278e8cd34cfe45ed36ff0b8bd5b30a0ddf333ec6371\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a9cc673be8ff245dba81278e8cd34cfe45ed36ff0b8bd5b30a0ddf333ec6371\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bc7e49c7a5ae66ea939060192fa6a346dc0b0b9aa9750b29879d3c426cc0d276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc7e49c7a5ae66ea939060192fa6a346dc0b0b9aa9750b29879d3c426cc0d276\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:40Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.949756 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa9d037e66cb53b9dd698eff46284c9e718fd842cda6f7b525f0c79062d70bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:40Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.953416 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-nc7qt" event={"ID":"6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e","Type":"ContainerStarted","Data":"b1a2c7b22068b7165927fc158885d48bf9952b9eba521d8ae9a599eee53dc305"} Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.953467 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-nc7qt" event={"ID":"6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e","Type":"ContainerStarted","Data":"9add6c44ccad318e1d730c7e24782c710b7c873914ba50783dada9aaa5e38a4a"} Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.954927 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-6zxff" event={"ID":"aaba93da-12a5-43df-b7c7-4c8b988163c6","Type":"ContainerStarted","Data":"3db293678a907cd55b31b1d0c0bd7c30e1ba75ecd5e13cbc68ab98fbf0fbd9d6"} Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.954980 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-6zxff" event={"ID":"aaba93da-12a5-43df-b7c7-4c8b988163c6","Type":"ContainerStarted","Data":"e749516f13004ad38787177acf4c421438a4eadb8c439510f86da7c2c881cff8"} Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.956166 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" event={"ID":"24632dda-6100-4ab6-a28e-214ddae4360c","Type":"ContainerStarted","Data":"579903eadb183466a2564737e6b678e04308239c0cd75b6808892053b97afbc0"} Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.956200 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" event={"ID":"24632dda-6100-4ab6-a28e-214ddae4360c","Type":"ContainerStarted","Data":"56f8167826c412be9cb1f1f95757c0e5a8d5cc8a132658744a99a4bbe5ca1c84"} Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.957268 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-9p5dc" event={"ID":"660ac006-2495-45c4-9fb3-e9c1dddcf7a0","Type":"ContainerStarted","Data":"23e57e65a6f5c3059506b9db21e9ac2492c30bd9191a98db0281915e9ce39797"} Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.957395 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-9p5dc" event={"ID":"660ac006-2495-45c4-9fb3-e9c1dddcf7a0","Type":"ContainerStarted","Data":"c58f34d80eeafb1c789c4488b07addc1e32bfc1b7a0ab45c2873b2c4e5127e62"} Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.959519 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" event={"ID":"d192f670-9f9d-4539-9641-e4bed73acdd4","Type":"ContainerStarted","Data":"040c5e8a5659f85f0d52d1a1e064154d998ae104f711101d2204c0b1caebbb30"} Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.959555 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" event={"ID":"d192f670-9f9d-4539-9641-e4bed73acdd4","Type":"ContainerStarted","Data":"5fdf06410349b7c99aa166e87fbe80bf68e2ae923635a6a978b7e2cd50383f7b"} Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.959567 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" event={"ID":"d192f670-9f9d-4539-9641-e4bed73acdd4","Type":"ContainerStarted","Data":"39737477815bb7ac16890fc0da49704b233e0dd88343b75b7b34865f87189f60"} Feb 02 09:07:40 crc kubenswrapper[4764]: I0202 09:07:40.988353 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6zxff" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aaba93da-12a5-43df-b7c7-4c8b988163c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqwb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6zxff\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:40Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:41 crc kubenswrapper[4764]: I0202 09:07:41.030347 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d192f670-9f9d-4539-9641-e4bed73acdd4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbksf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbksf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4ndm4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:41Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:41 crc kubenswrapper[4764]: I0202 09:07:41.068661 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:41Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:41 crc kubenswrapper[4764]: I0202 09:07:41.109805 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:41Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:41 crc kubenswrapper[4764]: I0202 09:07:41.162495 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d66ab3c0-9b5b-41d5-b6c0-91ccddccd7e9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d92ac486f5dade46bc75de32099859a71fc2ed25a7abbe6ae8937df128e7df30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a4ac273a4f127d79e5fad7d4c30cd2044a346f5734dba5d33d68cd092c0c2d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d38c7cb5dcd22fb07524494ebe77aa1a7f4f808f566043563817e47e7ec12a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://408722c7f83d28919431532f8c797a1c795fd5d96736d14e7f06a05744899702\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f473b18cd8003f75bb1dafb82bcec45745a73ccd10434297356dac185a3d15c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e229f484d19cbdabedf53d90a822a4434db9e00b1301538e885e26acc19d25f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e229f484d19cbdabedf53d90a822a4434db9e00b1301538e885e26acc19d25f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a9cc673be8ff245dba81278e8cd34cfe45ed36ff0b8bd5b30a0ddf333ec6371\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a9cc673be8ff245dba81278e8cd34cfe45ed36ff0b8bd5b30a0ddf333ec6371\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bc7e49c7a5ae66ea939060192fa6a346dc0b0b9aa9750b29879d3c426cc0d276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc7e49c7a5ae66ea939060192fa6a346dc0b0b9aa9750b29879d3c426cc0d276\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:41Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:41 crc kubenswrapper[4764]: I0202 09:07:41.190128 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa9d037e66cb53b9dd698eff46284c9e718fd842cda6f7b525f0c79062d70bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:41Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:41 crc kubenswrapper[4764]: I0202 09:07:41.230569 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:41Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:41 crc kubenswrapper[4764]: I0202 09:07:41.277203 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24632dda-6100-4ab6-a28e-214ddae4360c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zhn7j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:41Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:41 crc kubenswrapper[4764]: I0202 09:07:41.308965 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:41Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:41 crc kubenswrapper[4764]: I0202 09:07:41.347289 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6zxff" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aaba93da-12a5-43df-b7c7-4c8b988163c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3db293678a907cd55b31b1d0c0bd7c30e1ba75ecd5e13cbc68ab98fbf0fbd9d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqwb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6zxff\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:41Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:41 crc kubenswrapper[4764]: I0202 09:07:41.387720 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d192f670-9f9d-4539-9641-e4bed73acdd4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://040c5e8a5659f85f0d52d1a1e064154d998ae104f711101d2204c0b1caebbb30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbksf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fdf06410349b7c99aa166e87fbe80bf68e2ae923635a6a978b7e2cd50383f7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbksf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4ndm4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:41Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:41 crc kubenswrapper[4764]: I0202 09:07:41.428234 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:41Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:41 crc kubenswrapper[4764]: I0202 09:07:41.469790 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9p5dc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"660ac006-2495-45c4-9fb3-e9c1dddcf7a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23e57e65a6f5c3059506b9db21e9ac2492c30bd9191a98db0281915e9ce39797\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9p5dc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:41Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:41 crc kubenswrapper[4764]: I0202 09:07:41.511301 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87d61aea-46f6-4ee6-a7a4-767a6a67fea4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e78b6d08e4d616cefbb73e3fd43b3c6886dcd8aa997c7b4948bf1ff63401293\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d468715c2da33996561c40043216233fd12e21539d0c58dd1f3b892fd9f91a9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b184f7c936dc4046f74c0018269cbc2c4b7ceddb650b4e871d3f8a707e0df88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://269c5074371bd00f8363fb3888c11f1df383081a61497aea6578f08700881d50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b25bc9395831524c53f9e565b4efbe74f21792efafeb1dd507161bc91aa0ecdf\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T09:07:22Z\\\",\\\"message\\\":\\\"W0202 09:07:21.597488 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0202 09:07:21.598117 1 crypto.go:601] Generating new CA for check-endpoints-signer@1770023241 cert, and key in /tmp/serving-cert-3856183724/serving-signer.crt, /tmp/serving-cert-3856183724/serving-signer.key\\\\nI0202 09:07:22.012184 1 observer_polling.go:159] Starting file observer\\\\nW0202 09:07:22.015493 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0202 09:07:22.015797 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 09:07:22.016596 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3856183724/tls.crt::/tmp/serving-cert-3856183724/tls.key\\\\\\\"\\\\nF0202 09:07:22.270605 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b44e17b8ca01f549bc10081dd5f109cb1b9fff364bf5bd4fd90d9f0e3eeb04db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfb06a5ec9262cf39b108a66014d15e325cdbd5441a2d9e7a811dd57265c99a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfb06a5ec9262cf39b108a66014d15e325cdbd5441a2d9e7a811dd57265c99a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:41Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:41 crc kubenswrapper[4764]: I0202 09:07:41.551980 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5528225d91ac2778bef4554abb0bd036ac5d3b6cdac7357bab69e70c3894b6f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://071bbb80d0747cf3e109ba626d47f7d05f85f4cc0b3aba77276c795741f47ac6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:41Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:41 crc kubenswrapper[4764]: I0202 09:07:41.600166 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:41Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:41 crc kubenswrapper[4764]: I0202 09:07:41.629500 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nc7qt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1a2c7b22068b7165927fc158885d48bf9952b9eba521d8ae9a599eee53dc305\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2hz4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nc7qt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:41Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:41 crc kubenswrapper[4764]: I0202 09:07:41.781772 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-14 06:10:44.902817901 +0000 UTC Feb 02 09:07:41 crc kubenswrapper[4764]: I0202 09:07:41.963843 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"8f6904228e052fff69ad1fd66df97758b6b60ba4351e97a7fbd03eee4df1b653"} Feb 02 09:07:41 crc kubenswrapper[4764]: I0202 09:07:41.965385 4764 generic.go:334] "Generic (PLEG): container finished" podID="24632dda-6100-4ab6-a28e-214ddae4360c" containerID="579903eadb183466a2564737e6b678e04308239c0cd75b6808892053b97afbc0" exitCode=0 Feb 02 09:07:41 crc kubenswrapper[4764]: I0202 09:07:41.965442 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" event={"ID":"24632dda-6100-4ab6-a28e-214ddae4360c","Type":"ContainerDied","Data":"579903eadb183466a2564737e6b678e04308239c0cd75b6808892053b97afbc0"} Feb 02 09:07:41 crc kubenswrapper[4764]: I0202 09:07:41.967214 4764 generic.go:334] "Generic (PLEG): container finished" podID="660ac006-2495-45c4-9fb3-e9c1dddcf7a0" containerID="23e57e65a6f5c3059506b9db21e9ac2492c30bd9191a98db0281915e9ce39797" exitCode=0 Feb 02 09:07:41 crc kubenswrapper[4764]: I0202 09:07:41.967845 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-9p5dc" event={"ID":"660ac006-2495-45c4-9fb3-e9c1dddcf7a0","Type":"ContainerDied","Data":"23e57e65a6f5c3059506b9db21e9ac2492c30bd9191a98db0281915e9ce39797"} Feb 02 09:07:42 crc kubenswrapper[4764]: I0202 09:07:42.000955 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:41Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:42 crc kubenswrapper[4764]: I0202 09:07:42.035890 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5528225d91ac2778bef4554abb0bd036ac5d3b6cdac7357bab69e70c3894b6f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://071bbb80d0747cf3e109ba626d47f7d05f85f4cc0b3aba77276c795741f47ac6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:42Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:42 crc kubenswrapper[4764]: I0202 09:07:42.057416 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:42Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:42 crc kubenswrapper[4764]: I0202 09:07:42.081400 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nc7qt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1a2c7b22068b7165927fc158885d48bf9952b9eba521d8ae9a599eee53dc305\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2hz4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nc7qt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:42Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:42 crc kubenswrapper[4764]: I0202 09:07:42.119203 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9p5dc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"660ac006-2495-45c4-9fb3-e9c1dddcf7a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23e57e65a6f5c3059506b9db21e9ac2492c30bd9191a98db0281915e9ce39797\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9p5dc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:42Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:42 crc kubenswrapper[4764]: I0202 09:07:42.165658 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87d61aea-46f6-4ee6-a7a4-767a6a67fea4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e78b6d08e4d616cefbb73e3fd43b3c6886dcd8aa997c7b4948bf1ff63401293\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d468715c2da33996561c40043216233fd12e21539d0c58dd1f3b892fd9f91a9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b184f7c936dc4046f74c0018269cbc2c4b7ceddb650b4e871d3f8a707e0df88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://269c5074371bd00f8363fb3888c11f1df383081a61497aea6578f08700881d50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b25bc9395831524c53f9e565b4efbe74f21792efafeb1dd507161bc91aa0ecdf\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T09:07:22Z\\\",\\\"message\\\":\\\"W0202 09:07:21.597488 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0202 09:07:21.598117 1 crypto.go:601] Generating new CA for check-endpoints-signer@1770023241 cert, and key in /tmp/serving-cert-3856183724/serving-signer.crt, /tmp/serving-cert-3856183724/serving-signer.key\\\\nI0202 09:07:22.012184 1 observer_polling.go:159] Starting file observer\\\\nW0202 09:07:22.015493 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0202 09:07:22.015797 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 09:07:22.016596 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3856183724/tls.crt::/tmp/serving-cert-3856183724/tls.key\\\\\\\"\\\\nF0202 09:07:22.270605 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b44e17b8ca01f549bc10081dd5f109cb1b9fff364bf5bd4fd90d9f0e3eeb04db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfb06a5ec9262cf39b108a66014d15e325cdbd5441a2d9e7a811dd57265c99a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfb06a5ec9262cf39b108a66014d15e325cdbd5441a2d9e7a811dd57265c99a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:42Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:42 crc kubenswrapper[4764]: I0202 09:07:42.192138 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa9d037e66cb53b9dd698eff46284c9e718fd842cda6f7b525f0c79062d70bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:42Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:42 crc kubenswrapper[4764]: I0202 09:07:42.215189 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:42Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:42 crc kubenswrapper[4764]: I0202 09:07:42.238789 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24632dda-6100-4ab6-a28e-214ddae4360c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zhn7j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:42Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:42 crc kubenswrapper[4764]: I0202 09:07:42.260658 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d66ab3c0-9b5b-41d5-b6c0-91ccddccd7e9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d92ac486f5dade46bc75de32099859a71fc2ed25a7abbe6ae8937df128e7df30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a4ac273a4f127d79e5fad7d4c30cd2044a346f5734dba5d33d68cd092c0c2d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d38c7cb5dcd22fb07524494ebe77aa1a7f4f808f566043563817e47e7ec12a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://408722c7f83d28919431532f8c797a1c795fd5d96736d14e7f06a05744899702\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f473b18cd8003f75bb1dafb82bcec45745a73ccd10434297356dac185a3d15c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e229f484d19cbdabedf53d90a822a4434db9e00b1301538e885e26acc19d25f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e229f484d19cbdabedf53d90a822a4434db9e00b1301538e885e26acc19d25f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a9cc673be8ff245dba81278e8cd34cfe45ed36ff0b8bd5b30a0ddf333ec6371\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a9cc673be8ff245dba81278e8cd34cfe45ed36ff0b8bd5b30a0ddf333ec6371\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bc7e49c7a5ae66ea939060192fa6a346dc0b0b9aa9750b29879d3c426cc0d276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc7e49c7a5ae66ea939060192fa6a346dc0b0b9aa9750b29879d3c426cc0d276\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:42Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:42 crc kubenswrapper[4764]: I0202 09:07:42.279183 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f6904228e052fff69ad1fd66df97758b6b60ba4351e97a7fbd03eee4df1b653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:42Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:42 crc kubenswrapper[4764]: I0202 09:07:42.296282 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6zxff" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aaba93da-12a5-43df-b7c7-4c8b988163c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3db293678a907cd55b31b1d0c0bd7c30e1ba75ecd5e13cbc68ab98fbf0fbd9d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqwb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6zxff\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:42Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:42 crc kubenswrapper[4764]: I0202 09:07:42.309323 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d192f670-9f9d-4539-9641-e4bed73acdd4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://040c5e8a5659f85f0d52d1a1e064154d998ae104f711101d2204c0b1caebbb30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbksf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fdf06410349b7c99aa166e87fbe80bf68e2ae923635a6a978b7e2cd50383f7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbksf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4ndm4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:42Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:42 crc kubenswrapper[4764]: I0202 09:07:42.330023 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:42Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:42 crc kubenswrapper[4764]: I0202 09:07:42.344028 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nc7qt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1a2c7b22068b7165927fc158885d48bf9952b9eba521d8ae9a599eee53dc305\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2hz4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nc7qt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:42Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:42 crc kubenswrapper[4764]: I0202 09:07:42.364126 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9p5dc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"660ac006-2495-45c4-9fb3-e9c1dddcf7a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23e57e65a6f5c3059506b9db21e9ac2492c30bd9191a98db0281915e9ce39797\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23e57e65a6f5c3059506b9db21e9ac2492c30bd9191a98db0281915e9ce39797\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9p5dc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:42Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:42 crc kubenswrapper[4764]: I0202 09:07:42.385091 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87d61aea-46f6-4ee6-a7a4-767a6a67fea4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e78b6d08e4d616cefbb73e3fd43b3c6886dcd8aa997c7b4948bf1ff63401293\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d468715c2da33996561c40043216233fd12e21539d0c58dd1f3b892fd9f91a9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b184f7c936dc4046f74c0018269cbc2c4b7ceddb650b4e871d3f8a707e0df88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://269c5074371bd00f8363fb3888c11f1df383081a61497aea6578f08700881d50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b25bc9395831524c53f9e565b4efbe74f21792efafeb1dd507161bc91aa0ecdf\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T09:07:22Z\\\",\\\"message\\\":\\\"W0202 09:07:21.597488 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0202 09:07:21.598117 1 crypto.go:601] Generating new CA for check-endpoints-signer@1770023241 cert, and key in /tmp/serving-cert-3856183724/serving-signer.crt, /tmp/serving-cert-3856183724/serving-signer.key\\\\nI0202 09:07:22.012184 1 observer_polling.go:159] Starting file observer\\\\nW0202 09:07:22.015493 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0202 09:07:22.015797 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 09:07:22.016596 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3856183724/tls.crt::/tmp/serving-cert-3856183724/tls.key\\\\\\\"\\\\nF0202 09:07:22.270605 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b44e17b8ca01f549bc10081dd5f109cb1b9fff364bf5bd4fd90d9f0e3eeb04db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfb06a5ec9262cf39b108a66014d15e325cdbd5441a2d9e7a811dd57265c99a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfb06a5ec9262cf39b108a66014d15e325cdbd5441a2d9e7a811dd57265c99a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:42Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:42 crc kubenswrapper[4764]: I0202 09:07:42.389558 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 09:07:42 crc kubenswrapper[4764]: E0202 09:07:42.389766 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 09:07:46.389733853 +0000 UTC m=+29.323457931 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:07:42 crc kubenswrapper[4764]: I0202 09:07:42.402353 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5528225d91ac2778bef4554abb0bd036ac5d3b6cdac7357bab69e70c3894b6f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://071bbb80d0747cf3e109ba626d47f7d05f85f4cc0b3aba77276c795741f47ac6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:42Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:42 crc kubenswrapper[4764]: I0202 09:07:42.428341 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24632dda-6100-4ab6-a28e-214ddae4360c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://579903eadb183466a2564737e6b678e04308239c0cd75b6808892053b97afbc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://579903eadb183466a2564737e6b678e04308239c0cd75b6808892053b97afbc0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zhn7j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:42Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:42 crc kubenswrapper[4764]: I0202 09:07:42.452162 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d66ab3c0-9b5b-41d5-b6c0-91ccddccd7e9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d92ac486f5dade46bc75de32099859a71fc2ed25a7abbe6ae8937df128e7df30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a4ac273a4f127d79e5fad7d4c30cd2044a346f5734dba5d33d68cd092c0c2d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d38c7cb5dcd22fb07524494ebe77aa1a7f4f808f566043563817e47e7ec12a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://408722c7f83d28919431532f8c797a1c795fd5d96736d14e7f06a05744899702\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f473b18cd8003f75bb1dafb82bcec45745a73ccd10434297356dac185a3d15c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e229f484d19cbdabedf53d90a822a4434db9e00b1301538e885e26acc19d25f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e229f484d19cbdabedf53d90a822a4434db9e00b1301538e885e26acc19d25f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a9cc673be8ff245dba81278e8cd34cfe45ed36ff0b8bd5b30a0ddf333ec6371\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a9cc673be8ff245dba81278e8cd34cfe45ed36ff0b8bd5b30a0ddf333ec6371\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bc7e49c7a5ae66ea939060192fa6a346dc0b0b9aa9750b29879d3c426cc0d276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc7e49c7a5ae66ea939060192fa6a346dc0b0b9aa9750b29879d3c426cc0d276\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:42Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:42 crc kubenswrapper[4764]: I0202 09:07:42.481834 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa9d037e66cb53b9dd698eff46284c9e718fd842cda6f7b525f0c79062d70bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:42Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:42 crc kubenswrapper[4764]: I0202 09:07:42.491006 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 09:07:42 crc kubenswrapper[4764]: I0202 09:07:42.491073 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 09:07:42 crc kubenswrapper[4764]: I0202 09:07:42.491106 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 09:07:42 crc kubenswrapper[4764]: I0202 09:07:42.491143 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 09:07:42 crc kubenswrapper[4764]: E0202 09:07:42.491311 4764 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 02 09:07:42 crc kubenswrapper[4764]: E0202 09:07:42.491340 4764 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 02 09:07:42 crc kubenswrapper[4764]: E0202 09:07:42.491354 4764 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 02 09:07:42 crc kubenswrapper[4764]: E0202 09:07:42.491415 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-02 09:07:46.491396862 +0000 UTC m=+29.425120950 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 02 09:07:42 crc kubenswrapper[4764]: E0202 09:07:42.491816 4764 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 02 09:07:42 crc kubenswrapper[4764]: E0202 09:07:42.491851 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-02 09:07:46.491843873 +0000 UTC m=+29.425567961 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 02 09:07:42 crc kubenswrapper[4764]: E0202 09:07:42.491922 4764 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 02 09:07:42 crc kubenswrapper[4764]: E0202 09:07:42.492101 4764 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 02 09:07:42 crc kubenswrapper[4764]: E0202 09:07:42.492126 4764 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 02 09:07:42 crc kubenswrapper[4764]: E0202 09:07:42.492140 4764 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 02 09:07:42 crc kubenswrapper[4764]: E0202 09:07:42.492184 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-02 09:07:46.491968006 +0000 UTC m=+29.425692094 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 02 09:07:42 crc kubenswrapper[4764]: E0202 09:07:42.492205 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-02 09:07:46.492194601 +0000 UTC m=+29.425918689 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 02 09:07:42 crc kubenswrapper[4764]: I0202 09:07:42.512465 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:42Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:42 crc kubenswrapper[4764]: I0202 09:07:42.554984 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d192f670-9f9d-4539-9641-e4bed73acdd4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://040c5e8a5659f85f0d52d1a1e064154d998ae104f711101d2204c0b1caebbb30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbksf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fdf06410349b7c99aa166e87fbe80bf68e2ae923635a6a978b7e2cd50383f7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbksf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4ndm4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:42Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:42 crc kubenswrapper[4764]: I0202 09:07:42.587441 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f6904228e052fff69ad1fd66df97758b6b60ba4351e97a7fbd03eee4df1b653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:42Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:42 crc kubenswrapper[4764]: I0202 09:07:42.628440 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6zxff" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aaba93da-12a5-43df-b7c7-4c8b988163c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3db293678a907cd55b31b1d0c0bd7c30e1ba75ecd5e13cbc68ab98fbf0fbd9d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqwb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6zxff\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:42Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:42 crc kubenswrapper[4764]: I0202 09:07:42.672423 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:42Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:42 crc kubenswrapper[4764]: I0202 09:07:42.782333 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-06 00:58:36.789721908 +0000 UTC Feb 02 09:07:42 crc kubenswrapper[4764]: I0202 09:07:42.825415 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 09:07:42 crc kubenswrapper[4764]: E0202 09:07:42.825631 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 09:07:42 crc kubenswrapper[4764]: I0202 09:07:42.826251 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 09:07:42 crc kubenswrapper[4764]: E0202 09:07:42.826402 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 09:07:42 crc kubenswrapper[4764]: I0202 09:07:42.826487 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 09:07:42 crc kubenswrapper[4764]: E0202 09:07:42.826616 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 09:07:42 crc kubenswrapper[4764]: I0202 09:07:42.904908 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 02 09:07:42 crc kubenswrapper[4764]: I0202 09:07:42.911768 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 02 09:07:42 crc kubenswrapper[4764]: I0202 09:07:42.913202 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Feb 02 09:07:42 crc kubenswrapper[4764]: I0202 09:07:42.917797 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f6904228e052fff69ad1fd66df97758b6b60ba4351e97a7fbd03eee4df1b653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:42Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:42 crc kubenswrapper[4764]: I0202 09:07:42.930157 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6zxff" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aaba93da-12a5-43df-b7c7-4c8b988163c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3db293678a907cd55b31b1d0c0bd7c30e1ba75ecd5e13cbc68ab98fbf0fbd9d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqwb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6zxff\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:42Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:42 crc kubenswrapper[4764]: I0202 09:07:42.942249 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d192f670-9f9d-4539-9641-e4bed73acdd4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://040c5e8a5659f85f0d52d1a1e064154d998ae104f711101d2204c0b1caebbb30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbksf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fdf06410349b7c99aa166e87fbe80bf68e2ae923635a6a978b7e2cd50383f7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbksf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4ndm4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:42Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:42 crc kubenswrapper[4764]: I0202 09:07:42.953886 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:42Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:42 crc kubenswrapper[4764]: I0202 09:07:42.966251 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nc7qt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1a2c7b22068b7165927fc158885d48bf9952b9eba521d8ae9a599eee53dc305\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2hz4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nc7qt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:42Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:42 crc kubenswrapper[4764]: I0202 09:07:42.974805 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" event={"ID":"24632dda-6100-4ab6-a28e-214ddae4360c","Type":"ContainerStarted","Data":"bffda48d7c881c1470251b501661324859373246cc1a21811bd36ca24887c7aa"} Feb 02 09:07:42 crc kubenswrapper[4764]: I0202 09:07:42.974847 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" event={"ID":"24632dda-6100-4ab6-a28e-214ddae4360c","Type":"ContainerStarted","Data":"0a6c67e9b7708cc0d85b0b6ffbbade0d34867e159a9c5487ca2606475c667a50"} Feb 02 09:07:42 crc kubenswrapper[4764]: I0202 09:07:42.974857 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" event={"ID":"24632dda-6100-4ab6-a28e-214ddae4360c","Type":"ContainerStarted","Data":"c380383b6a014d1b694a13ddc22f54b20b9e4f9d5c9b9000d76187621dce0a21"} Feb 02 09:07:42 crc kubenswrapper[4764]: I0202 09:07:42.974867 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" event={"ID":"24632dda-6100-4ab6-a28e-214ddae4360c","Type":"ContainerStarted","Data":"5529f2bff87753428e9173826fc1880acf96a4c134fa9ecad195b15697cad73a"} Feb 02 09:07:42 crc kubenswrapper[4764]: I0202 09:07:42.977656 4764 generic.go:334] "Generic (PLEG): container finished" podID="660ac006-2495-45c4-9fb3-e9c1dddcf7a0" containerID="2417b66816da1a6cd1791a4b91eca4a3c4f3694462e066e7f0722f1b9c2105c8" exitCode=0 Feb 02 09:07:42 crc kubenswrapper[4764]: I0202 09:07:42.978212 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-9p5dc" event={"ID":"660ac006-2495-45c4-9fb3-e9c1dddcf7a0","Type":"ContainerDied","Data":"2417b66816da1a6cd1791a4b91eca4a3c4f3694462e066e7f0722f1b9c2105c8"} Feb 02 09:07:42 crc kubenswrapper[4764]: I0202 09:07:42.986011 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9p5dc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"660ac006-2495-45c4-9fb3-e9c1dddcf7a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23e57e65a6f5c3059506b9db21e9ac2492c30bd9191a98db0281915e9ce39797\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23e57e65a6f5c3059506b9db21e9ac2492c30bd9191a98db0281915e9ce39797\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9p5dc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:42Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:43 crc kubenswrapper[4764]: I0202 09:07:43.001407 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87d61aea-46f6-4ee6-a7a4-767a6a67fea4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e78b6d08e4d616cefbb73e3fd43b3c6886dcd8aa997c7b4948bf1ff63401293\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d468715c2da33996561c40043216233fd12e21539d0c58dd1f3b892fd9f91a9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b184f7c936dc4046f74c0018269cbc2c4b7ceddb650b4e871d3f8a707e0df88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://269c5074371bd00f8363fb3888c11f1df383081a61497aea6578f08700881d50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b25bc9395831524c53f9e565b4efbe74f21792efafeb1dd507161bc91aa0ecdf\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T09:07:22Z\\\",\\\"message\\\":\\\"W0202 09:07:21.597488 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0202 09:07:21.598117 1 crypto.go:601] Generating new CA for check-endpoints-signer@1770023241 cert, and key in /tmp/serving-cert-3856183724/serving-signer.crt, /tmp/serving-cert-3856183724/serving-signer.key\\\\nI0202 09:07:22.012184 1 observer_polling.go:159] Starting file observer\\\\nW0202 09:07:22.015493 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0202 09:07:22.015797 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 09:07:22.016596 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3856183724/tls.crt::/tmp/serving-cert-3856183724/tls.key\\\\\\\"\\\\nF0202 09:07:22.270605 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b44e17b8ca01f549bc10081dd5f109cb1b9fff364bf5bd4fd90d9f0e3eeb04db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfb06a5ec9262cf39b108a66014d15e325cdbd5441a2d9e7a811dd57265c99a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfb06a5ec9262cf39b108a66014d15e325cdbd5441a2d9e7a811dd57265c99a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:42Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:43 crc kubenswrapper[4764]: I0202 09:07:43.014377 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5528225d91ac2778bef4554abb0bd036ac5d3b6cdac7357bab69e70c3894b6f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://071bbb80d0747cf3e109ba626d47f7d05f85f4cc0b3aba77276c795741f47ac6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:43Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:43 crc kubenswrapper[4764]: I0202 09:07:43.054502 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:43Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:43 crc kubenswrapper[4764]: I0202 09:07:43.102493 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d66ab3c0-9b5b-41d5-b6c0-91ccddccd7e9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d92ac486f5dade46bc75de32099859a71fc2ed25a7abbe6ae8937df128e7df30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a4ac273a4f127d79e5fad7d4c30cd2044a346f5734dba5d33d68cd092c0c2d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d38c7cb5dcd22fb07524494ebe77aa1a7f4f808f566043563817e47e7ec12a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://408722c7f83d28919431532f8c797a1c795fd5d96736d14e7f06a05744899702\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f473b18cd8003f75bb1dafb82bcec45745a73ccd10434297356dac185a3d15c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e229f484d19cbdabedf53d90a822a4434db9e00b1301538e885e26acc19d25f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e229f484d19cbdabedf53d90a822a4434db9e00b1301538e885e26acc19d25f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a9cc673be8ff245dba81278e8cd34cfe45ed36ff0b8bd5b30a0ddf333ec6371\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a9cc673be8ff245dba81278e8cd34cfe45ed36ff0b8bd5b30a0ddf333ec6371\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bc7e49c7a5ae66ea939060192fa6a346dc0b0b9aa9750b29879d3c426cc0d276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc7e49c7a5ae66ea939060192fa6a346dc0b0b9aa9750b29879d3c426cc0d276\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:43Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:43 crc kubenswrapper[4764]: I0202 09:07:43.128325 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa9d037e66cb53b9dd698eff46284c9e718fd842cda6f7b525f0c79062d70bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:43Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:43 crc kubenswrapper[4764]: I0202 09:07:43.172824 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:43Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:43 crc kubenswrapper[4764]: I0202 09:07:43.197189 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-pcjmn"] Feb 02 09:07:43 crc kubenswrapper[4764]: I0202 09:07:43.197753 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-pcjmn" Feb 02 09:07:43 crc kubenswrapper[4764]: I0202 09:07:43.219344 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24632dda-6100-4ab6-a28e-214ddae4360c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://579903eadb183466a2564737e6b678e04308239c0cd75b6808892053b97afbc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://579903eadb183466a2564737e6b678e04308239c0cd75b6808892053b97afbc0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zhn7j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:43Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:43 crc kubenswrapper[4764]: I0202 09:07:43.221114 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Feb 02 09:07:43 crc kubenswrapper[4764]: I0202 09:07:43.240947 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Feb 02 09:07:43 crc kubenswrapper[4764]: I0202 09:07:43.260130 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Feb 02 09:07:43 crc kubenswrapper[4764]: I0202 09:07:43.280716 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Feb 02 09:07:43 crc kubenswrapper[4764]: I0202 09:07:43.304312 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/6e5ccd3b-b657-47a2-9c4d-bc9adb126533-serviceca\") pod \"node-ca-pcjmn\" (UID: \"6e5ccd3b-b657-47a2-9c4d-bc9adb126533\") " pod="openshift-image-registry/node-ca-pcjmn" Feb 02 09:07:43 crc kubenswrapper[4764]: I0202 09:07:43.304549 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2f7zl\" (UniqueName: \"kubernetes.io/projected/6e5ccd3b-b657-47a2-9c4d-bc9adb126533-kube-api-access-2f7zl\") pod \"node-ca-pcjmn\" (UID: \"6e5ccd3b-b657-47a2-9c4d-bc9adb126533\") " pod="openshift-image-registry/node-ca-pcjmn" Feb 02 09:07:43 crc kubenswrapper[4764]: I0202 09:07:43.304653 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/6e5ccd3b-b657-47a2-9c4d-bc9adb126533-host\") pod \"node-ca-pcjmn\" (UID: \"6e5ccd3b-b657-47a2-9c4d-bc9adb126533\") " pod="openshift-image-registry/node-ca-pcjmn" Feb 02 09:07:43 crc kubenswrapper[4764]: I0202 09:07:43.329214 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nc7qt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1a2c7b22068b7165927fc158885d48bf9952b9eba521d8ae9a599eee53dc305\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2hz4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nc7qt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:43Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:43 crc kubenswrapper[4764]: I0202 09:07:43.370242 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9p5dc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"660ac006-2495-45c4-9fb3-e9c1dddcf7a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23e57e65a6f5c3059506b9db21e9ac2492c30bd9191a98db0281915e9ce39797\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23e57e65a6f5c3059506b9db21e9ac2492c30bd9191a98db0281915e9ce39797\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2417b66816da1a6cd1791a4b91eca4a3c4f3694462e066e7f0722f1b9c2105c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2417b66816da1a6cd1791a4b91eca4a3c4f3694462e066e7f0722f1b9c2105c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9p5dc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:43Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:43 crc kubenswrapper[4764]: I0202 09:07:43.406169 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/6e5ccd3b-b657-47a2-9c4d-bc9adb126533-serviceca\") pod \"node-ca-pcjmn\" (UID: \"6e5ccd3b-b657-47a2-9c4d-bc9adb126533\") " pod="openshift-image-registry/node-ca-pcjmn" Feb 02 09:07:43 crc kubenswrapper[4764]: I0202 09:07:43.406221 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2f7zl\" (UniqueName: \"kubernetes.io/projected/6e5ccd3b-b657-47a2-9c4d-bc9adb126533-kube-api-access-2f7zl\") pod \"node-ca-pcjmn\" (UID: \"6e5ccd3b-b657-47a2-9c4d-bc9adb126533\") " pod="openshift-image-registry/node-ca-pcjmn" Feb 02 09:07:43 crc kubenswrapper[4764]: I0202 09:07:43.406252 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/6e5ccd3b-b657-47a2-9c4d-bc9adb126533-host\") pod \"node-ca-pcjmn\" (UID: \"6e5ccd3b-b657-47a2-9c4d-bc9adb126533\") " pod="openshift-image-registry/node-ca-pcjmn" Feb 02 09:07:43 crc kubenswrapper[4764]: I0202 09:07:43.406311 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/6e5ccd3b-b657-47a2-9c4d-bc9adb126533-host\") pod \"node-ca-pcjmn\" (UID: \"6e5ccd3b-b657-47a2-9c4d-bc9adb126533\") " pod="openshift-image-registry/node-ca-pcjmn" Feb 02 09:07:43 crc kubenswrapper[4764]: I0202 09:07:43.407302 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/6e5ccd3b-b657-47a2-9c4d-bc9adb126533-serviceca\") pod \"node-ca-pcjmn\" (UID: \"6e5ccd3b-b657-47a2-9c4d-bc9adb126533\") " pod="openshift-image-registry/node-ca-pcjmn" Feb 02 09:07:43 crc kubenswrapper[4764]: I0202 09:07:43.411325 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87d61aea-46f6-4ee6-a7a4-767a6a67fea4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e78b6d08e4d616cefbb73e3fd43b3c6886dcd8aa997c7b4948bf1ff63401293\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d468715c2da33996561c40043216233fd12e21539d0c58dd1f3b892fd9f91a9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b184f7c936dc4046f74c0018269cbc2c4b7ceddb650b4e871d3f8a707e0df88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://269c5074371bd00f8363fb3888c11f1df383081a61497aea6578f08700881d50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b25bc9395831524c53f9e565b4efbe74f21792efafeb1dd507161bc91aa0ecdf\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T09:07:22Z\\\",\\\"message\\\":\\\"W0202 09:07:21.597488 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0202 09:07:21.598117 1 crypto.go:601] Generating new CA for check-endpoints-signer@1770023241 cert, and key in /tmp/serving-cert-3856183724/serving-signer.crt, /tmp/serving-cert-3856183724/serving-signer.key\\\\nI0202 09:07:22.012184 1 observer_polling.go:159] Starting file observer\\\\nW0202 09:07:22.015493 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0202 09:07:22.015797 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 09:07:22.016596 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3856183724/tls.crt::/tmp/serving-cert-3856183724/tls.key\\\\\\\"\\\\nF0202 09:07:22.270605 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b44e17b8ca01f549bc10081dd5f109cb1b9fff364bf5bd4fd90d9f0e3eeb04db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfb06a5ec9262cf39b108a66014d15e325cdbd5441a2d9e7a811dd57265c99a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfb06a5ec9262cf39b108a66014d15e325cdbd5441a2d9e7a811dd57265c99a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:43Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:43 crc kubenswrapper[4764]: I0202 09:07:43.437638 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2f7zl\" (UniqueName: \"kubernetes.io/projected/6e5ccd3b-b657-47a2-9c4d-bc9adb126533-kube-api-access-2f7zl\") pod \"node-ca-pcjmn\" (UID: \"6e5ccd3b-b657-47a2-9c4d-bc9adb126533\") " pod="openshift-image-registry/node-ca-pcjmn" Feb 02 09:07:43 crc kubenswrapper[4764]: I0202 09:07:43.469778 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5528225d91ac2778bef4554abb0bd036ac5d3b6cdac7357bab69e70c3894b6f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://071bbb80d0747cf3e109ba626d47f7d05f85f4cc0b3aba77276c795741f47ac6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:43Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:43 crc kubenswrapper[4764]: I0202 09:07:43.508790 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:43Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:43 crc kubenswrapper[4764]: I0202 09:07:43.509816 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-pcjmn" Feb 02 09:07:43 crc kubenswrapper[4764]: W0202 09:07:43.523551 4764 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6e5ccd3b_b657_47a2_9c4d_bc9adb126533.slice/crio-cd418fd0269aa7978c3929aff74f6e6763c9b8eeb47fb50781485c09eb1185b1 WatchSource:0}: Error finding container cd418fd0269aa7978c3929aff74f6e6763c9b8eeb47fb50781485c09eb1185b1: Status 404 returned error can't find the container with id cd418fd0269aa7978c3929aff74f6e6763c9b8eeb47fb50781485c09eb1185b1 Feb 02 09:07:43 crc kubenswrapper[4764]: I0202 09:07:43.557273 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d66ab3c0-9b5b-41d5-b6c0-91ccddccd7e9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d92ac486f5dade46bc75de32099859a71fc2ed25a7abbe6ae8937df128e7df30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a4ac273a4f127d79e5fad7d4c30cd2044a346f5734dba5d33d68cd092c0c2d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d38c7cb5dcd22fb07524494ebe77aa1a7f4f808f566043563817e47e7ec12a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://408722c7f83d28919431532f8c797a1c795fd5d96736d14e7f06a05744899702\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f473b18cd8003f75bb1dafb82bcec45745a73ccd10434297356dac185a3d15c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e229f484d19cbdabedf53d90a822a4434db9e00b1301538e885e26acc19d25f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e229f484d19cbdabedf53d90a822a4434db9e00b1301538e885e26acc19d25f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a9cc673be8ff245dba81278e8cd34cfe45ed36ff0b8bd5b30a0ddf333ec6371\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a9cc673be8ff245dba81278e8cd34cfe45ed36ff0b8bd5b30a0ddf333ec6371\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bc7e49c7a5ae66ea939060192fa6a346dc0b0b9aa9750b29879d3c426cc0d276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc7e49c7a5ae66ea939060192fa6a346dc0b0b9aa9750b29879d3c426cc0d276\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:43Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:43 crc kubenswrapper[4764]: I0202 09:07:43.589242 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa9d037e66cb53b9dd698eff46284c9e718fd842cda6f7b525f0c79062d70bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:43Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:43 crc kubenswrapper[4764]: I0202 09:07:43.630462 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:43Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:43 crc kubenswrapper[4764]: I0202 09:07:43.674413 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24632dda-6100-4ab6-a28e-214ddae4360c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://579903eadb183466a2564737e6b678e04308239c0cd75b6808892053b97afbc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://579903eadb183466a2564737e6b678e04308239c0cd75b6808892053b97afbc0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zhn7j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:43Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:43 crc kubenswrapper[4764]: I0202 09:07:43.707090 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pcjmn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6e5ccd3b-b657-47a2-9c4d-bc9adb126533\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2f7zl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pcjmn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:43Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:43 crc kubenswrapper[4764]: I0202 09:07:43.750541 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed747fd8-e3ce-4319-a6da-bda0fd51b85a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6f064bdc66470aace3180d4988244026b54009a63cff199e41c5d02a1d18479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://60258f84a1d8dcfe48b7e9688ed50f077e201971f20f3f6663fbf1ba58a4611a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f84be2e08217a2ac3edde858d78b4410376ed9bae3bcabe2a7f9b2fbe2f69c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a69d25d352d7ceac5c735bc1fcf446a1fea6dc22c65a5296cfab828cea17377\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:43Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:43 crc kubenswrapper[4764]: I0202 09:07:43.782665 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-12 18:00:21.258147548 +0000 UTC Feb 02 09:07:43 crc kubenswrapper[4764]: I0202 09:07:43.787976 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f6904228e052fff69ad1fd66df97758b6b60ba4351e97a7fbd03eee4df1b653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:43Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:43 crc kubenswrapper[4764]: I0202 09:07:43.825407 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6zxff" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aaba93da-12a5-43df-b7c7-4c8b988163c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3db293678a907cd55b31b1d0c0bd7c30e1ba75ecd5e13cbc68ab98fbf0fbd9d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqwb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6zxff\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:43Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:43 crc kubenswrapper[4764]: I0202 09:07:43.868783 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d192f670-9f9d-4539-9641-e4bed73acdd4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://040c5e8a5659f85f0d52d1a1e064154d998ae104f711101d2204c0b1caebbb30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbksf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fdf06410349b7c99aa166e87fbe80bf68e2ae923635a6a978b7e2cd50383f7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbksf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4ndm4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:43Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:43 crc kubenswrapper[4764]: I0202 09:07:43.909109 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:43Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:43 crc kubenswrapper[4764]: I0202 09:07:43.983069 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-pcjmn" event={"ID":"6e5ccd3b-b657-47a2-9c4d-bc9adb126533","Type":"ContainerStarted","Data":"72863eef568d92a1abeb3cde2e89630a4f5fb7626c6949448ff5195774c98eab"} Feb 02 09:07:43 crc kubenswrapper[4764]: I0202 09:07:43.983116 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-pcjmn" event={"ID":"6e5ccd3b-b657-47a2-9c4d-bc9adb126533","Type":"ContainerStarted","Data":"cd418fd0269aa7978c3929aff74f6e6763c9b8eeb47fb50781485c09eb1185b1"} Feb 02 09:07:43 crc kubenswrapper[4764]: I0202 09:07:43.987278 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" event={"ID":"24632dda-6100-4ab6-a28e-214ddae4360c","Type":"ContainerStarted","Data":"8a989645cf485cd8ff80ffff1a86cd7808e1564eba5667535e161cc68db13b26"} Feb 02 09:07:43 crc kubenswrapper[4764]: I0202 09:07:43.987341 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" event={"ID":"24632dda-6100-4ab6-a28e-214ddae4360c","Type":"ContainerStarted","Data":"20b71d3ee7a381baa51b0b0d0470283022cfede30f491135737b5e82d631d763"} Feb 02 09:07:43 crc kubenswrapper[4764]: I0202 09:07:43.989251 4764 generic.go:334] "Generic (PLEG): container finished" podID="660ac006-2495-45c4-9fb3-e9c1dddcf7a0" containerID="883641f29e9120c2cea6b272bacf1dc63c752f55ab0408cdaf73db85c85ac369" exitCode=0 Feb 02 09:07:43 crc kubenswrapper[4764]: I0202 09:07:43.989280 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-9p5dc" event={"ID":"660ac006-2495-45c4-9fb3-e9c1dddcf7a0","Type":"ContainerDied","Data":"883641f29e9120c2cea6b272bacf1dc63c752f55ab0408cdaf73db85c85ac369"} Feb 02 09:07:44 crc kubenswrapper[4764]: I0202 09:07:44.009284 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d66ab3c0-9b5b-41d5-b6c0-91ccddccd7e9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d92ac486f5dade46bc75de32099859a71fc2ed25a7abbe6ae8937df128e7df30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a4ac273a4f127d79e5fad7d4c30cd2044a346f5734dba5d33d68cd092c0c2d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d38c7cb5dcd22fb07524494ebe77aa1a7f4f808f566043563817e47e7ec12a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://408722c7f83d28919431532f8c797a1c795fd5d96736d14e7f06a05744899702\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f473b18cd8003f75bb1dafb82bcec45745a73ccd10434297356dac185a3d15c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e229f484d19cbdabedf53d90a822a4434db9e00b1301538e885e26acc19d25f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e229f484d19cbdabedf53d90a822a4434db9e00b1301538e885e26acc19d25f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a9cc673be8ff245dba81278e8cd34cfe45ed36ff0b8bd5b30a0ddf333ec6371\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a9cc673be8ff245dba81278e8cd34cfe45ed36ff0b8bd5b30a0ddf333ec6371\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bc7e49c7a5ae66ea939060192fa6a346dc0b0b9aa9750b29879d3c426cc0d276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc7e49c7a5ae66ea939060192fa6a346dc0b0b9aa9750b29879d3c426cc0d276\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:44Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:44 crc kubenswrapper[4764]: I0202 09:07:44.026198 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa9d037e66cb53b9dd698eff46284c9e718fd842cda6f7b525f0c79062d70bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:44Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:44 crc kubenswrapper[4764]: I0202 09:07:44.038162 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:44Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:44 crc kubenswrapper[4764]: I0202 09:07:44.073060 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24632dda-6100-4ab6-a28e-214ddae4360c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://579903eadb183466a2564737e6b678e04308239c0cd75b6808892053b97afbc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://579903eadb183466a2564737e6b678e04308239c0cd75b6808892053b97afbc0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zhn7j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:44Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:44 crc kubenswrapper[4764]: I0202 09:07:44.111227 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed747fd8-e3ce-4319-a6da-bda0fd51b85a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6f064bdc66470aace3180d4988244026b54009a63cff199e41c5d02a1d18479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://60258f84a1d8dcfe48b7e9688ed50f077e201971f20f3f6663fbf1ba58a4611a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f84be2e08217a2ac3edde858d78b4410376ed9bae3bcabe2a7f9b2fbe2f69c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a69d25d352d7ceac5c735bc1fcf446a1fea6dc22c65a5296cfab828cea17377\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:44Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:44 crc kubenswrapper[4764]: I0202 09:07:44.155231 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f6904228e052fff69ad1fd66df97758b6b60ba4351e97a7fbd03eee4df1b653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:44Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:44 crc kubenswrapper[4764]: I0202 09:07:44.193289 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6zxff" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aaba93da-12a5-43df-b7c7-4c8b988163c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3db293678a907cd55b31b1d0c0bd7c30e1ba75ecd5e13cbc68ab98fbf0fbd9d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqwb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6zxff\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:44Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:44 crc kubenswrapper[4764]: I0202 09:07:44.228125 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d192f670-9f9d-4539-9641-e4bed73acdd4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://040c5e8a5659f85f0d52d1a1e064154d998ae104f711101d2204c0b1caebbb30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbksf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fdf06410349b7c99aa166e87fbe80bf68e2ae923635a6a978b7e2cd50383f7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbksf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4ndm4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:44Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:44 crc kubenswrapper[4764]: I0202 09:07:44.268093 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pcjmn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6e5ccd3b-b657-47a2-9c4d-bc9adb126533\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72863eef568d92a1abeb3cde2e89630a4f5fb7626c6949448ff5195774c98eab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2f7zl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pcjmn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:44Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:44 crc kubenswrapper[4764]: I0202 09:07:44.308199 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:44Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:44 crc kubenswrapper[4764]: I0202 09:07:44.350409 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87d61aea-46f6-4ee6-a7a4-767a6a67fea4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e78b6d08e4d616cefbb73e3fd43b3c6886dcd8aa997c7b4948bf1ff63401293\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d468715c2da33996561c40043216233fd12e21539d0c58dd1f3b892fd9f91a9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b184f7c936dc4046f74c0018269cbc2c4b7ceddb650b4e871d3f8a707e0df88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://269c5074371bd00f8363fb3888c11f1df383081a61497aea6578f08700881d50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b25bc9395831524c53f9e565b4efbe74f21792efafeb1dd507161bc91aa0ecdf\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T09:07:22Z\\\",\\\"message\\\":\\\"W0202 09:07:21.597488 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0202 09:07:21.598117 1 crypto.go:601] Generating new CA for check-endpoints-signer@1770023241 cert, and key in /tmp/serving-cert-3856183724/serving-signer.crt, /tmp/serving-cert-3856183724/serving-signer.key\\\\nI0202 09:07:22.012184 1 observer_polling.go:159] Starting file observer\\\\nW0202 09:07:22.015493 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0202 09:07:22.015797 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 09:07:22.016596 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3856183724/tls.crt::/tmp/serving-cert-3856183724/tls.key\\\\\\\"\\\\nF0202 09:07:22.270605 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b44e17b8ca01f549bc10081dd5f109cb1b9fff364bf5bd4fd90d9f0e3eeb04db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfb06a5ec9262cf39b108a66014d15e325cdbd5441a2d9e7a811dd57265c99a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfb06a5ec9262cf39b108a66014d15e325cdbd5441a2d9e7a811dd57265c99a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:44Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:44 crc kubenswrapper[4764]: I0202 09:07:44.387364 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5528225d91ac2778bef4554abb0bd036ac5d3b6cdac7357bab69e70c3894b6f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://071bbb80d0747cf3e109ba626d47f7d05f85f4cc0b3aba77276c795741f47ac6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:44Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:44 crc kubenswrapper[4764]: I0202 09:07:44.432764 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:44Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:44 crc kubenswrapper[4764]: I0202 09:07:44.469888 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nc7qt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1a2c7b22068b7165927fc158885d48bf9952b9eba521d8ae9a599eee53dc305\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2hz4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nc7qt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:44Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:44 crc kubenswrapper[4764]: I0202 09:07:44.510540 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9p5dc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"660ac006-2495-45c4-9fb3-e9c1dddcf7a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23e57e65a6f5c3059506b9db21e9ac2492c30bd9191a98db0281915e9ce39797\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23e57e65a6f5c3059506b9db21e9ac2492c30bd9191a98db0281915e9ce39797\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2417b66816da1a6cd1791a4b91eca4a3c4f3694462e066e7f0722f1b9c2105c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2417b66816da1a6cd1791a4b91eca4a3c4f3694462e066e7f0722f1b9c2105c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9p5dc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:44Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:44 crc kubenswrapper[4764]: I0202 09:07:44.550076 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87d61aea-46f6-4ee6-a7a4-767a6a67fea4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e78b6d08e4d616cefbb73e3fd43b3c6886dcd8aa997c7b4948bf1ff63401293\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d468715c2da33996561c40043216233fd12e21539d0c58dd1f3b892fd9f91a9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b184f7c936dc4046f74c0018269cbc2c4b7ceddb650b4e871d3f8a707e0df88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://269c5074371bd00f8363fb3888c11f1df383081a61497aea6578f08700881d50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b25bc9395831524c53f9e565b4efbe74f21792efafeb1dd507161bc91aa0ecdf\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T09:07:22Z\\\",\\\"message\\\":\\\"W0202 09:07:21.597488 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0202 09:07:21.598117 1 crypto.go:601] Generating new CA for check-endpoints-signer@1770023241 cert, and key in /tmp/serving-cert-3856183724/serving-signer.crt, /tmp/serving-cert-3856183724/serving-signer.key\\\\nI0202 09:07:22.012184 1 observer_polling.go:159] Starting file observer\\\\nW0202 09:07:22.015493 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0202 09:07:22.015797 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 09:07:22.016596 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3856183724/tls.crt::/tmp/serving-cert-3856183724/tls.key\\\\\\\"\\\\nF0202 09:07:22.270605 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b44e17b8ca01f549bc10081dd5f109cb1b9fff364bf5bd4fd90d9f0e3eeb04db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfb06a5ec9262cf39b108a66014d15e325cdbd5441a2d9e7a811dd57265c99a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfb06a5ec9262cf39b108a66014d15e325cdbd5441a2d9e7a811dd57265c99a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:44Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:44 crc kubenswrapper[4764]: I0202 09:07:44.589231 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5528225d91ac2778bef4554abb0bd036ac5d3b6cdac7357bab69e70c3894b6f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://071bbb80d0747cf3e109ba626d47f7d05f85f4cc0b3aba77276c795741f47ac6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:44Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:44 crc kubenswrapper[4764]: I0202 09:07:44.631460 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:44Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:44 crc kubenswrapper[4764]: I0202 09:07:44.672438 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nc7qt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1a2c7b22068b7165927fc158885d48bf9952b9eba521d8ae9a599eee53dc305\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2hz4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nc7qt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:44Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:44 crc kubenswrapper[4764]: I0202 09:07:44.711145 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9p5dc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"660ac006-2495-45c4-9fb3-e9c1dddcf7a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23e57e65a6f5c3059506b9db21e9ac2492c30bd9191a98db0281915e9ce39797\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23e57e65a6f5c3059506b9db21e9ac2492c30bd9191a98db0281915e9ce39797\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2417b66816da1a6cd1791a4b91eca4a3c4f3694462e066e7f0722f1b9c2105c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2417b66816da1a6cd1791a4b91eca4a3c4f3694462e066e7f0722f1b9c2105c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://883641f29e9120c2cea6b272bacf1dc63c752f55ab0408cdaf73db85c85ac369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://883641f29e9120c2cea6b272bacf1dc63c752f55ab0408cdaf73db85c85ac369\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9p5dc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:44Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:44 crc kubenswrapper[4764]: I0202 09:07:44.754218 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d66ab3c0-9b5b-41d5-b6c0-91ccddccd7e9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d92ac486f5dade46bc75de32099859a71fc2ed25a7abbe6ae8937df128e7df30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a4ac273a4f127d79e5fad7d4c30cd2044a346f5734dba5d33d68cd092c0c2d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d38c7cb5dcd22fb07524494ebe77aa1a7f4f808f566043563817e47e7ec12a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://408722c7f83d28919431532f8c797a1c795fd5d96736d14e7f06a05744899702\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f473b18cd8003f75bb1dafb82bcec45745a73ccd10434297356dac185a3d15c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e229f484d19cbdabedf53d90a822a4434db9e00b1301538e885e26acc19d25f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e229f484d19cbdabedf53d90a822a4434db9e00b1301538e885e26acc19d25f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a9cc673be8ff245dba81278e8cd34cfe45ed36ff0b8bd5b30a0ddf333ec6371\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a9cc673be8ff245dba81278e8cd34cfe45ed36ff0b8bd5b30a0ddf333ec6371\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bc7e49c7a5ae66ea939060192fa6a346dc0b0b9aa9750b29879d3c426cc0d276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc7e49c7a5ae66ea939060192fa6a346dc0b0b9aa9750b29879d3c426cc0d276\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:44Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:44 crc kubenswrapper[4764]: I0202 09:07:44.783301 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-18 19:31:37.585624383 +0000 UTC Feb 02 09:07:44 crc kubenswrapper[4764]: I0202 09:07:44.790505 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa9d037e66cb53b9dd698eff46284c9e718fd842cda6f7b525f0c79062d70bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:44Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:44 crc kubenswrapper[4764]: I0202 09:07:44.824847 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 09:07:44 crc kubenswrapper[4764]: I0202 09:07:44.824888 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 09:07:44 crc kubenswrapper[4764]: I0202 09:07:44.824847 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 09:07:44 crc kubenswrapper[4764]: E0202 09:07:44.824985 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 09:07:44 crc kubenswrapper[4764]: E0202 09:07:44.825075 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 09:07:44 crc kubenswrapper[4764]: E0202 09:07:44.825132 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 09:07:44 crc kubenswrapper[4764]: I0202 09:07:44.828301 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:44Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:44 crc kubenswrapper[4764]: I0202 09:07:44.845894 4764 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 09:07:44 crc kubenswrapper[4764]: I0202 09:07:44.848272 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:44 crc kubenswrapper[4764]: I0202 09:07:44.848309 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:44 crc kubenswrapper[4764]: I0202 09:07:44.848319 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:44 crc kubenswrapper[4764]: I0202 09:07:44.848415 4764 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 02 09:07:44 crc kubenswrapper[4764]: I0202 09:07:44.872176 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24632dda-6100-4ab6-a28e-214ddae4360c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://579903eadb183466a2564737e6b678e04308239c0cd75b6808892053b97afbc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://579903eadb183466a2564737e6b678e04308239c0cd75b6808892053b97afbc0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zhn7j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:44Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:44 crc kubenswrapper[4764]: I0202 09:07:44.921182 4764 kubelet_node_status.go:115] "Node was previously registered" node="crc" Feb 02 09:07:44 crc kubenswrapper[4764]: I0202 09:07:44.921442 4764 kubelet_node_status.go:79] "Successfully registered node" node="crc" Feb 02 09:07:44 crc kubenswrapper[4764]: I0202 09:07:44.922352 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:44 crc kubenswrapper[4764]: I0202 09:07:44.922374 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:44 crc kubenswrapper[4764]: I0202 09:07:44.922382 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:44 crc kubenswrapper[4764]: I0202 09:07:44.922396 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:44 crc kubenswrapper[4764]: I0202 09:07:44.922405 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:44Z","lastTransitionTime":"2026-02-02T09:07:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:44 crc kubenswrapper[4764]: E0202 09:07:44.934289 4764 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:07:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:07:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:07:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:07:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"31973a70-ce17-4811-9a14-f5476c6979db\\\",\\\"systemUUID\\\":\\\"ca92679c-8580-43c3-be51-62441e93d22b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:44Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:44 crc kubenswrapper[4764]: I0202 09:07:44.937329 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:44 crc kubenswrapper[4764]: I0202 09:07:44.937384 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:44 crc kubenswrapper[4764]: I0202 09:07:44.937392 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:44 crc kubenswrapper[4764]: I0202 09:07:44.937405 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:44 crc kubenswrapper[4764]: I0202 09:07:44.937413 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:44Z","lastTransitionTime":"2026-02-02T09:07:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:44 crc kubenswrapper[4764]: I0202 09:07:44.947768 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed747fd8-e3ce-4319-a6da-bda0fd51b85a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6f064bdc66470aace3180d4988244026b54009a63cff199e41c5d02a1d18479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://60258f84a1d8dcfe48b7e9688ed50f077e201971f20f3f6663fbf1ba58a4611a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f84be2e08217a2ac3edde858d78b4410376ed9bae3bcabe2a7f9b2fbe2f69c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a69d25d352d7ceac5c735bc1fcf446a1fea6dc22c65a5296cfab828cea17377\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:44Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:44 crc kubenswrapper[4764]: E0202 09:07:44.951389 4764 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:07:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:07:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:07:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:07:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"31973a70-ce17-4811-9a14-f5476c6979db\\\",\\\"systemUUID\\\":\\\"ca92679c-8580-43c3-be51-62441e93d22b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:44Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:44 crc kubenswrapper[4764]: I0202 09:07:44.955466 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:44 crc kubenswrapper[4764]: I0202 09:07:44.955497 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:44 crc kubenswrapper[4764]: I0202 09:07:44.955508 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:44 crc kubenswrapper[4764]: I0202 09:07:44.955523 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:44 crc kubenswrapper[4764]: I0202 09:07:44.955540 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:44Z","lastTransitionTime":"2026-02-02T09:07:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:44 crc kubenswrapper[4764]: E0202 09:07:44.969195 4764 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:07:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:07:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:07:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:07:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"31973a70-ce17-4811-9a14-f5476c6979db\\\",\\\"systemUUID\\\":\\\"ca92679c-8580-43c3-be51-62441e93d22b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:44Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:44 crc kubenswrapper[4764]: I0202 09:07:44.973132 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:44 crc kubenswrapper[4764]: I0202 09:07:44.973168 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:44 crc kubenswrapper[4764]: I0202 09:07:44.973179 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:44 crc kubenswrapper[4764]: I0202 09:07:44.973196 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:44 crc kubenswrapper[4764]: I0202 09:07:44.973209 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:44Z","lastTransitionTime":"2026-02-02T09:07:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:44 crc kubenswrapper[4764]: E0202 09:07:44.986399 4764 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:07:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:07:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:07:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:07:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"31973a70-ce17-4811-9a14-f5476c6979db\\\",\\\"systemUUID\\\":\\\"ca92679c-8580-43c3-be51-62441e93d22b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:44Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:44 crc kubenswrapper[4764]: I0202 09:07:44.992911 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f6904228e052fff69ad1fd66df97758b6b60ba4351e97a7fbd03eee4df1b653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:44Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:44 crc kubenswrapper[4764]: I0202 09:07:44.993395 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:44 crc kubenswrapper[4764]: I0202 09:07:44.993435 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:44 crc kubenswrapper[4764]: I0202 09:07:44.993446 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:44 crc kubenswrapper[4764]: I0202 09:07:44.993463 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:44 crc kubenswrapper[4764]: I0202 09:07:44.993474 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:44Z","lastTransitionTime":"2026-02-02T09:07:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:44 crc kubenswrapper[4764]: I0202 09:07:44.998808 4764 generic.go:334] "Generic (PLEG): container finished" podID="660ac006-2495-45c4-9fb3-e9c1dddcf7a0" containerID="2745baba1188ba5e8f820616e4117a12c50ecee5b68e71ad9cabbb621b6e131f" exitCode=0 Feb 02 09:07:44 crc kubenswrapper[4764]: I0202 09:07:44.998846 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-9p5dc" event={"ID":"660ac006-2495-45c4-9fb3-e9c1dddcf7a0","Type":"ContainerDied","Data":"2745baba1188ba5e8f820616e4117a12c50ecee5b68e71ad9cabbb621b6e131f"} Feb 02 09:07:45 crc kubenswrapper[4764]: E0202 09:07:45.005357 4764 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:07:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:07:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:07:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:07:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"31973a70-ce17-4811-9a14-f5476c6979db\\\",\\\"systemUUID\\\":\\\"ca92679c-8580-43c3-be51-62441e93d22b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:45Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:45 crc kubenswrapper[4764]: E0202 09:07:45.005489 4764 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 02 09:07:45 crc kubenswrapper[4764]: I0202 09:07:45.006680 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:45 crc kubenswrapper[4764]: I0202 09:07:45.006783 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:45 crc kubenswrapper[4764]: I0202 09:07:45.006851 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:45 crc kubenswrapper[4764]: I0202 09:07:45.006943 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:45 crc kubenswrapper[4764]: I0202 09:07:45.007007 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:45Z","lastTransitionTime":"2026-02-02T09:07:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:45 crc kubenswrapper[4764]: I0202 09:07:45.028481 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6zxff" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aaba93da-12a5-43df-b7c7-4c8b988163c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3db293678a907cd55b31b1d0c0bd7c30e1ba75ecd5e13cbc68ab98fbf0fbd9d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqwb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6zxff\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:45Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:45 crc kubenswrapper[4764]: I0202 09:07:45.068367 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d192f670-9f9d-4539-9641-e4bed73acdd4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://040c5e8a5659f85f0d52d1a1e064154d998ae104f711101d2204c0b1caebbb30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbksf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fdf06410349b7c99aa166e87fbe80bf68e2ae923635a6a978b7e2cd50383f7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbksf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4ndm4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:45Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:45 crc kubenswrapper[4764]: I0202 09:07:45.107520 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pcjmn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6e5ccd3b-b657-47a2-9c4d-bc9adb126533\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72863eef568d92a1abeb3cde2e89630a4f5fb7626c6949448ff5195774c98eab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2f7zl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pcjmn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:45Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:45 crc kubenswrapper[4764]: I0202 09:07:45.108653 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:45 crc kubenswrapper[4764]: I0202 09:07:45.108698 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:45 crc kubenswrapper[4764]: I0202 09:07:45.108710 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:45 crc kubenswrapper[4764]: I0202 09:07:45.108726 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:45 crc kubenswrapper[4764]: I0202 09:07:45.108736 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:45Z","lastTransitionTime":"2026-02-02T09:07:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:45 crc kubenswrapper[4764]: I0202 09:07:45.148671 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:45Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:45 crc kubenswrapper[4764]: I0202 09:07:45.197011 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d66ab3c0-9b5b-41d5-b6c0-91ccddccd7e9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d92ac486f5dade46bc75de32099859a71fc2ed25a7abbe6ae8937df128e7df30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a4ac273a4f127d79e5fad7d4c30cd2044a346f5734dba5d33d68cd092c0c2d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d38c7cb5dcd22fb07524494ebe77aa1a7f4f808f566043563817e47e7ec12a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://408722c7f83d28919431532f8c797a1c795fd5d96736d14e7f06a05744899702\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f473b18cd8003f75bb1dafb82bcec45745a73ccd10434297356dac185a3d15c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e229f484d19cbdabedf53d90a822a4434db9e00b1301538e885e26acc19d25f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e229f484d19cbdabedf53d90a822a4434db9e00b1301538e885e26acc19d25f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a9cc673be8ff245dba81278e8cd34cfe45ed36ff0b8bd5b30a0ddf333ec6371\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a9cc673be8ff245dba81278e8cd34cfe45ed36ff0b8bd5b30a0ddf333ec6371\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bc7e49c7a5ae66ea939060192fa6a346dc0b0b9aa9750b29879d3c426cc0d276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc7e49c7a5ae66ea939060192fa6a346dc0b0b9aa9750b29879d3c426cc0d276\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:45Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:45 crc kubenswrapper[4764]: I0202 09:07:45.210438 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:45 crc kubenswrapper[4764]: I0202 09:07:45.210474 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:45 crc kubenswrapper[4764]: I0202 09:07:45.210486 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:45 crc kubenswrapper[4764]: I0202 09:07:45.210500 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:45 crc kubenswrapper[4764]: I0202 09:07:45.210511 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:45Z","lastTransitionTime":"2026-02-02T09:07:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:45 crc kubenswrapper[4764]: I0202 09:07:45.230819 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa9d037e66cb53b9dd698eff46284c9e718fd842cda6f7b525f0c79062d70bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:45Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:45 crc kubenswrapper[4764]: I0202 09:07:45.270563 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:45Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:45 crc kubenswrapper[4764]: I0202 09:07:45.312276 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:45 crc kubenswrapper[4764]: I0202 09:07:45.312316 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:45 crc kubenswrapper[4764]: I0202 09:07:45.312325 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:45 crc kubenswrapper[4764]: I0202 09:07:45.312340 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:45 crc kubenswrapper[4764]: I0202 09:07:45.312351 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:45Z","lastTransitionTime":"2026-02-02T09:07:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:45 crc kubenswrapper[4764]: I0202 09:07:45.315956 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24632dda-6100-4ab6-a28e-214ddae4360c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://579903eadb183466a2564737e6b678e04308239c0cd75b6808892053b97afbc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://579903eadb183466a2564737e6b678e04308239c0cd75b6808892053b97afbc0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zhn7j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:45Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:45 crc kubenswrapper[4764]: I0202 09:07:45.349644 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed747fd8-e3ce-4319-a6da-bda0fd51b85a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6f064bdc66470aace3180d4988244026b54009a63cff199e41c5d02a1d18479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://60258f84a1d8dcfe48b7e9688ed50f077e201971f20f3f6663fbf1ba58a4611a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f84be2e08217a2ac3edde858d78b4410376ed9bae3bcabe2a7f9b2fbe2f69c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a69d25d352d7ceac5c735bc1fcf446a1fea6dc22c65a5296cfab828cea17377\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:45Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:45 crc kubenswrapper[4764]: I0202 09:07:45.388640 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f6904228e052fff69ad1fd66df97758b6b60ba4351e97a7fbd03eee4df1b653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:45Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:45 crc kubenswrapper[4764]: I0202 09:07:45.414810 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:45 crc kubenswrapper[4764]: I0202 09:07:45.414868 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:45 crc kubenswrapper[4764]: I0202 09:07:45.414884 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:45 crc kubenswrapper[4764]: I0202 09:07:45.414906 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:45 crc kubenswrapper[4764]: I0202 09:07:45.414922 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:45Z","lastTransitionTime":"2026-02-02T09:07:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:45 crc kubenswrapper[4764]: I0202 09:07:45.427905 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6zxff" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aaba93da-12a5-43df-b7c7-4c8b988163c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3db293678a907cd55b31b1d0c0bd7c30e1ba75ecd5e13cbc68ab98fbf0fbd9d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqwb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6zxff\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:45Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:45 crc kubenswrapper[4764]: I0202 09:07:45.473096 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d192f670-9f9d-4539-9641-e4bed73acdd4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://040c5e8a5659f85f0d52d1a1e064154d998ae104f711101d2204c0b1caebbb30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbksf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fdf06410349b7c99aa166e87fbe80bf68e2ae923635a6a978b7e2cd50383f7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbksf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4ndm4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:45Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:45 crc kubenswrapper[4764]: I0202 09:07:45.506730 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pcjmn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6e5ccd3b-b657-47a2-9c4d-bc9adb126533\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72863eef568d92a1abeb3cde2e89630a4f5fb7626c6949448ff5195774c98eab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2f7zl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pcjmn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:45Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:45 crc kubenswrapper[4764]: I0202 09:07:45.517333 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:45 crc kubenswrapper[4764]: I0202 09:07:45.517359 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:45 crc kubenswrapper[4764]: I0202 09:07:45.517367 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:45 crc kubenswrapper[4764]: I0202 09:07:45.517383 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:45 crc kubenswrapper[4764]: I0202 09:07:45.517394 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:45Z","lastTransitionTime":"2026-02-02T09:07:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:45 crc kubenswrapper[4764]: I0202 09:07:45.549663 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:45Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:45 crc kubenswrapper[4764]: I0202 09:07:45.591207 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87d61aea-46f6-4ee6-a7a4-767a6a67fea4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e78b6d08e4d616cefbb73e3fd43b3c6886dcd8aa997c7b4948bf1ff63401293\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d468715c2da33996561c40043216233fd12e21539d0c58dd1f3b892fd9f91a9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b184f7c936dc4046f74c0018269cbc2c4b7ceddb650b4e871d3f8a707e0df88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://269c5074371bd00f8363fb3888c11f1df383081a61497aea6578f08700881d50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b25bc9395831524c53f9e565b4efbe74f21792efafeb1dd507161bc91aa0ecdf\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T09:07:22Z\\\",\\\"message\\\":\\\"W0202 09:07:21.597488 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0202 09:07:21.598117 1 crypto.go:601] Generating new CA for check-endpoints-signer@1770023241 cert, and key in /tmp/serving-cert-3856183724/serving-signer.crt, /tmp/serving-cert-3856183724/serving-signer.key\\\\nI0202 09:07:22.012184 1 observer_polling.go:159] Starting file observer\\\\nW0202 09:07:22.015493 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0202 09:07:22.015797 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 09:07:22.016596 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3856183724/tls.crt::/tmp/serving-cert-3856183724/tls.key\\\\\\\"\\\\nF0202 09:07:22.270605 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b44e17b8ca01f549bc10081dd5f109cb1b9fff364bf5bd4fd90d9f0e3eeb04db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfb06a5ec9262cf39b108a66014d15e325cdbd5441a2d9e7a811dd57265c99a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfb06a5ec9262cf39b108a66014d15e325cdbd5441a2d9e7a811dd57265c99a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:45Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:45 crc kubenswrapper[4764]: I0202 09:07:45.619883 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:45 crc kubenswrapper[4764]: I0202 09:07:45.619967 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:45 crc kubenswrapper[4764]: I0202 09:07:45.619987 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:45 crc kubenswrapper[4764]: I0202 09:07:45.620010 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:45 crc kubenswrapper[4764]: I0202 09:07:45.620027 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:45Z","lastTransitionTime":"2026-02-02T09:07:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:45 crc kubenswrapper[4764]: I0202 09:07:45.633686 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5528225d91ac2778bef4554abb0bd036ac5d3b6cdac7357bab69e70c3894b6f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://071bbb80d0747cf3e109ba626d47f7d05f85f4cc0b3aba77276c795741f47ac6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:45Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:45 crc kubenswrapper[4764]: I0202 09:07:45.673499 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:45Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:45 crc kubenswrapper[4764]: I0202 09:07:45.711063 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nc7qt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1a2c7b22068b7165927fc158885d48bf9952b9eba521d8ae9a599eee53dc305\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2hz4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nc7qt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:45Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:45 crc kubenswrapper[4764]: I0202 09:07:45.722246 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:45 crc kubenswrapper[4764]: I0202 09:07:45.722291 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:45 crc kubenswrapper[4764]: I0202 09:07:45.722303 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:45 crc kubenswrapper[4764]: I0202 09:07:45.722318 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:45 crc kubenswrapper[4764]: I0202 09:07:45.722331 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:45Z","lastTransitionTime":"2026-02-02T09:07:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:45 crc kubenswrapper[4764]: I0202 09:07:45.751303 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9p5dc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"660ac006-2495-45c4-9fb3-e9c1dddcf7a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23e57e65a6f5c3059506b9db21e9ac2492c30bd9191a98db0281915e9ce39797\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23e57e65a6f5c3059506b9db21e9ac2492c30bd9191a98db0281915e9ce39797\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2417b66816da1a6cd1791a4b91eca4a3c4f3694462e066e7f0722f1b9c2105c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2417b66816da1a6cd1791a4b91eca4a3c4f3694462e066e7f0722f1b9c2105c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://883641f29e9120c2cea6b272bacf1dc63c752f55ab0408cdaf73db85c85ac369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://883641f29e9120c2cea6b272bacf1dc63c752f55ab0408cdaf73db85c85ac369\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2745baba1188ba5e8f820616e4117a12c50ecee5b68e71ad9cabbb621b6e131f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2745baba1188ba5e8f820616e4117a12c50ecee5b68e71ad9cabbb621b6e131f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9p5dc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:45Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:45 crc kubenswrapper[4764]: I0202 09:07:45.784329 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-06 12:25:21.483041884 +0000 UTC Feb 02 09:07:45 crc kubenswrapper[4764]: I0202 09:07:45.825172 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:45 crc kubenswrapper[4764]: I0202 09:07:45.825239 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:45 crc kubenswrapper[4764]: I0202 09:07:45.825262 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:45 crc kubenswrapper[4764]: I0202 09:07:45.825288 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:45 crc kubenswrapper[4764]: I0202 09:07:45.825310 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:45Z","lastTransitionTime":"2026-02-02T09:07:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:45 crc kubenswrapper[4764]: I0202 09:07:45.928084 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:45 crc kubenswrapper[4764]: I0202 09:07:45.928145 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:45 crc kubenswrapper[4764]: I0202 09:07:45.928162 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:45 crc kubenswrapper[4764]: I0202 09:07:45.928185 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:45 crc kubenswrapper[4764]: I0202 09:07:45.928202 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:45Z","lastTransitionTime":"2026-02-02T09:07:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:46 crc kubenswrapper[4764]: I0202 09:07:46.009593 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" event={"ID":"24632dda-6100-4ab6-a28e-214ddae4360c","Type":"ContainerStarted","Data":"4b4cfbd4e682bcdcaff785013bdb72580ce36fced733e235538536827036b93d"} Feb 02 09:07:46 crc kubenswrapper[4764]: I0202 09:07:46.014579 4764 generic.go:334] "Generic (PLEG): container finished" podID="660ac006-2495-45c4-9fb3-e9c1dddcf7a0" containerID="5a8925e435be4d48d86824fc35f8b8819fb35d63e385164c4a20edb3bf3035c3" exitCode=0 Feb 02 09:07:46 crc kubenswrapper[4764]: I0202 09:07:46.014621 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-9p5dc" event={"ID":"660ac006-2495-45c4-9fb3-e9c1dddcf7a0","Type":"ContainerDied","Data":"5a8925e435be4d48d86824fc35f8b8819fb35d63e385164c4a20edb3bf3035c3"} Feb 02 09:07:46 crc kubenswrapper[4764]: I0202 09:07:46.030635 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:46Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:46 crc kubenswrapper[4764]: I0202 09:07:46.031629 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:46 crc kubenswrapper[4764]: I0202 09:07:46.031683 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:46 crc kubenswrapper[4764]: I0202 09:07:46.031699 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:46 crc kubenswrapper[4764]: I0202 09:07:46.031724 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:46 crc kubenswrapper[4764]: I0202 09:07:46.031739 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:46Z","lastTransitionTime":"2026-02-02T09:07:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:46 crc kubenswrapper[4764]: I0202 09:07:46.053375 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nc7qt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1a2c7b22068b7165927fc158885d48bf9952b9eba521d8ae9a599eee53dc305\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2hz4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nc7qt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:46Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:46 crc kubenswrapper[4764]: I0202 09:07:46.082149 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9p5dc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"660ac006-2495-45c4-9fb3-e9c1dddcf7a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23e57e65a6f5c3059506b9db21e9ac2492c30bd9191a98db0281915e9ce39797\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23e57e65a6f5c3059506b9db21e9ac2492c30bd9191a98db0281915e9ce39797\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2417b66816da1a6cd1791a4b91eca4a3c4f3694462e066e7f0722f1b9c2105c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2417b66816da1a6cd1791a4b91eca4a3c4f3694462e066e7f0722f1b9c2105c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://883641f29e9120c2cea6b272bacf1dc63c752f55ab0408cdaf73db85c85ac369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://883641f29e9120c2cea6b272bacf1dc63c752f55ab0408cdaf73db85c85ac369\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2745baba1188ba5e8f820616e4117a12c50ecee5b68e71ad9cabbb621b6e131f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2745baba1188ba5e8f820616e4117a12c50ecee5b68e71ad9cabbb621b6e131f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8925e435be4d48d86824fc35f8b8819fb35d63e385164c4a20edb3bf3035c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a8925e435be4d48d86824fc35f8b8819fb35d63e385164c4a20edb3bf3035c3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9p5dc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:46Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:46 crc kubenswrapper[4764]: I0202 09:07:46.106895 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87d61aea-46f6-4ee6-a7a4-767a6a67fea4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e78b6d08e4d616cefbb73e3fd43b3c6886dcd8aa997c7b4948bf1ff63401293\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d468715c2da33996561c40043216233fd12e21539d0c58dd1f3b892fd9f91a9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b184f7c936dc4046f74c0018269cbc2c4b7ceddb650b4e871d3f8a707e0df88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://269c5074371bd00f8363fb3888c11f1df383081a61497aea6578f08700881d50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b25bc9395831524c53f9e565b4efbe74f21792efafeb1dd507161bc91aa0ecdf\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T09:07:22Z\\\",\\\"message\\\":\\\"W0202 09:07:21.597488 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0202 09:07:21.598117 1 crypto.go:601] Generating new CA for check-endpoints-signer@1770023241 cert, and key in /tmp/serving-cert-3856183724/serving-signer.crt, /tmp/serving-cert-3856183724/serving-signer.key\\\\nI0202 09:07:22.012184 1 observer_polling.go:159] Starting file observer\\\\nW0202 09:07:22.015493 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0202 09:07:22.015797 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 09:07:22.016596 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3856183724/tls.crt::/tmp/serving-cert-3856183724/tls.key\\\\\\\"\\\\nF0202 09:07:22.270605 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b44e17b8ca01f549bc10081dd5f109cb1b9fff364bf5bd4fd90d9f0e3eeb04db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfb06a5ec9262cf39b108a66014d15e325cdbd5441a2d9e7a811dd57265c99a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfb06a5ec9262cf39b108a66014d15e325cdbd5441a2d9e7a811dd57265c99a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:46Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:46 crc kubenswrapper[4764]: I0202 09:07:46.125832 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5528225d91ac2778bef4554abb0bd036ac5d3b6cdac7357bab69e70c3894b6f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://071bbb80d0747cf3e109ba626d47f7d05f85f4cc0b3aba77276c795741f47ac6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:46Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:46 crc kubenswrapper[4764]: I0202 09:07:46.134393 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:46 crc kubenswrapper[4764]: I0202 09:07:46.134438 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:46 crc kubenswrapper[4764]: I0202 09:07:46.134451 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:46 crc kubenswrapper[4764]: I0202 09:07:46.134471 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:46 crc kubenswrapper[4764]: I0202 09:07:46.134483 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:46Z","lastTransitionTime":"2026-02-02T09:07:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:46 crc kubenswrapper[4764]: I0202 09:07:46.157050 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24632dda-6100-4ab6-a28e-214ddae4360c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://579903eadb183466a2564737e6b678e04308239c0cd75b6808892053b97afbc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://579903eadb183466a2564737e6b678e04308239c0cd75b6808892053b97afbc0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zhn7j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:46Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:46 crc kubenswrapper[4764]: I0202 09:07:46.183329 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d66ab3c0-9b5b-41d5-b6c0-91ccddccd7e9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d92ac486f5dade46bc75de32099859a71fc2ed25a7abbe6ae8937df128e7df30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a4ac273a4f127d79e5fad7d4c30cd2044a346f5734dba5d33d68cd092c0c2d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d38c7cb5dcd22fb07524494ebe77aa1a7f4f808f566043563817e47e7ec12a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://408722c7f83d28919431532f8c797a1c795fd5d96736d14e7f06a05744899702\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f473b18cd8003f75bb1dafb82bcec45745a73ccd10434297356dac185a3d15c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e229f484d19cbdabedf53d90a822a4434db9e00b1301538e885e26acc19d25f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e229f484d19cbdabedf53d90a822a4434db9e00b1301538e885e26acc19d25f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a9cc673be8ff245dba81278e8cd34cfe45ed36ff0b8bd5b30a0ddf333ec6371\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a9cc673be8ff245dba81278e8cd34cfe45ed36ff0b8bd5b30a0ddf333ec6371\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bc7e49c7a5ae66ea939060192fa6a346dc0b0b9aa9750b29879d3c426cc0d276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc7e49c7a5ae66ea939060192fa6a346dc0b0b9aa9750b29879d3c426cc0d276\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:46Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:46 crc kubenswrapper[4764]: I0202 09:07:46.201085 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa9d037e66cb53b9dd698eff46284c9e718fd842cda6f7b525f0c79062d70bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:46Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:46 crc kubenswrapper[4764]: I0202 09:07:46.215112 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:46Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:46 crc kubenswrapper[4764]: I0202 09:07:46.228985 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d192f670-9f9d-4539-9641-e4bed73acdd4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://040c5e8a5659f85f0d52d1a1e064154d998ae104f711101d2204c0b1caebbb30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbksf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fdf06410349b7c99aa166e87fbe80bf68e2ae923635a6a978b7e2cd50383f7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbksf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4ndm4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:46Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:46 crc kubenswrapper[4764]: I0202 09:07:46.236951 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:46 crc kubenswrapper[4764]: I0202 09:07:46.236992 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:46 crc kubenswrapper[4764]: I0202 09:07:46.237005 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:46 crc kubenswrapper[4764]: I0202 09:07:46.237026 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:46 crc kubenswrapper[4764]: I0202 09:07:46.237038 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:46Z","lastTransitionTime":"2026-02-02T09:07:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:46 crc kubenswrapper[4764]: I0202 09:07:46.240983 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pcjmn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6e5ccd3b-b657-47a2-9c4d-bc9adb126533\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72863eef568d92a1abeb3cde2e89630a4f5fb7626c6949448ff5195774c98eab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2f7zl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pcjmn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:46Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:46 crc kubenswrapper[4764]: I0202 09:07:46.261291 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed747fd8-e3ce-4319-a6da-bda0fd51b85a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6f064bdc66470aace3180d4988244026b54009a63cff199e41c5d02a1d18479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://60258f84a1d8dcfe48b7e9688ed50f077e201971f20f3f6663fbf1ba58a4611a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f84be2e08217a2ac3edde858d78b4410376ed9bae3bcabe2a7f9b2fbe2f69c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a69d25d352d7ceac5c735bc1fcf446a1fea6dc22c65a5296cfab828cea17377\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:46Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:46 crc kubenswrapper[4764]: I0202 09:07:46.273320 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f6904228e052fff69ad1fd66df97758b6b60ba4351e97a7fbd03eee4df1b653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:46Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:46 crc kubenswrapper[4764]: I0202 09:07:46.308361 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6zxff" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aaba93da-12a5-43df-b7c7-4c8b988163c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3db293678a907cd55b31b1d0c0bd7c30e1ba75ecd5e13cbc68ab98fbf0fbd9d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqwb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6zxff\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:46Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:46 crc kubenswrapper[4764]: I0202 09:07:46.340286 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:46 crc kubenswrapper[4764]: I0202 09:07:46.340350 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:46 crc kubenswrapper[4764]: I0202 09:07:46.340361 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:46 crc kubenswrapper[4764]: I0202 09:07:46.340376 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:46 crc kubenswrapper[4764]: I0202 09:07:46.340386 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:46Z","lastTransitionTime":"2026-02-02T09:07:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:46 crc kubenswrapper[4764]: I0202 09:07:46.351891 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:46Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:46 crc kubenswrapper[4764]: I0202 09:07:46.434751 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 09:07:46 crc kubenswrapper[4764]: E0202 09:07:46.435173 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 09:07:54.435113496 +0000 UTC m=+37.368837624 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:07:46 crc kubenswrapper[4764]: I0202 09:07:46.442583 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:46 crc kubenswrapper[4764]: I0202 09:07:46.442633 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:46 crc kubenswrapper[4764]: I0202 09:07:46.442645 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:46 crc kubenswrapper[4764]: I0202 09:07:46.442664 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:46 crc kubenswrapper[4764]: I0202 09:07:46.442679 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:46Z","lastTransitionTime":"2026-02-02T09:07:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:46 crc kubenswrapper[4764]: I0202 09:07:46.536081 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 09:07:46 crc kubenswrapper[4764]: I0202 09:07:46.536221 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 09:07:46 crc kubenswrapper[4764]: I0202 09:07:46.536274 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 09:07:46 crc kubenswrapper[4764]: I0202 09:07:46.536314 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 09:07:46 crc kubenswrapper[4764]: E0202 09:07:46.536393 4764 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 02 09:07:46 crc kubenswrapper[4764]: E0202 09:07:46.536478 4764 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 02 09:07:46 crc kubenswrapper[4764]: E0202 09:07:46.536498 4764 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 02 09:07:46 crc kubenswrapper[4764]: E0202 09:07:46.536535 4764 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 02 09:07:46 crc kubenswrapper[4764]: E0202 09:07:46.536587 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-02 09:07:54.536556879 +0000 UTC m=+37.470281167 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 02 09:07:46 crc kubenswrapper[4764]: E0202 09:07:46.536603 4764 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 02 09:07:46 crc kubenswrapper[4764]: E0202 09:07:46.536643 4764 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 02 09:07:46 crc kubenswrapper[4764]: E0202 09:07:46.536650 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-02 09:07:54.536620381 +0000 UTC m=+37.470344509 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 02 09:07:46 crc kubenswrapper[4764]: E0202 09:07:46.536542 4764 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 02 09:07:46 crc kubenswrapper[4764]: E0202 09:07:46.536667 4764 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 02 09:07:46 crc kubenswrapper[4764]: E0202 09:07:46.536726 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-02 09:07:54.536708523 +0000 UTC m=+37.470432651 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 02 09:07:46 crc kubenswrapper[4764]: E0202 09:07:46.536768 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-02 09:07:54.536741904 +0000 UTC m=+37.470466032 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 02 09:07:46 crc kubenswrapper[4764]: I0202 09:07:46.545418 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:46 crc kubenswrapper[4764]: I0202 09:07:46.545490 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:46 crc kubenswrapper[4764]: I0202 09:07:46.545509 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:46 crc kubenswrapper[4764]: I0202 09:07:46.545535 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:46 crc kubenswrapper[4764]: I0202 09:07:46.545551 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:46Z","lastTransitionTime":"2026-02-02T09:07:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:46 crc kubenswrapper[4764]: I0202 09:07:46.648549 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:46 crc kubenswrapper[4764]: I0202 09:07:46.648592 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:46 crc kubenswrapper[4764]: I0202 09:07:46.648603 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:46 crc kubenswrapper[4764]: I0202 09:07:46.648624 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:46 crc kubenswrapper[4764]: I0202 09:07:46.648637 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:46Z","lastTransitionTime":"2026-02-02T09:07:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:46 crc kubenswrapper[4764]: I0202 09:07:46.750739 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:46 crc kubenswrapper[4764]: I0202 09:07:46.750782 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:46 crc kubenswrapper[4764]: I0202 09:07:46.750790 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:46 crc kubenswrapper[4764]: I0202 09:07:46.750805 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:46 crc kubenswrapper[4764]: I0202 09:07:46.750814 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:46Z","lastTransitionTime":"2026-02-02T09:07:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:46 crc kubenswrapper[4764]: I0202 09:07:46.785127 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-04 03:41:12.560615278 +0000 UTC Feb 02 09:07:46 crc kubenswrapper[4764]: I0202 09:07:46.824617 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 09:07:46 crc kubenswrapper[4764]: I0202 09:07:46.824663 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 09:07:46 crc kubenswrapper[4764]: I0202 09:07:46.824678 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 09:07:46 crc kubenswrapper[4764]: E0202 09:07:46.824779 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 09:07:46 crc kubenswrapper[4764]: E0202 09:07:46.824911 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 09:07:46 crc kubenswrapper[4764]: E0202 09:07:46.825392 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 09:07:46 crc kubenswrapper[4764]: I0202 09:07:46.854290 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:46 crc kubenswrapper[4764]: I0202 09:07:46.854601 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:46 crc kubenswrapper[4764]: I0202 09:07:46.854615 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:46 crc kubenswrapper[4764]: I0202 09:07:46.854644 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:46 crc kubenswrapper[4764]: I0202 09:07:46.854656 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:46Z","lastTransitionTime":"2026-02-02T09:07:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:46 crc kubenswrapper[4764]: I0202 09:07:46.957315 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:46 crc kubenswrapper[4764]: I0202 09:07:46.957388 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:46 crc kubenswrapper[4764]: I0202 09:07:46.957408 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:46 crc kubenswrapper[4764]: I0202 09:07:46.957432 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:46 crc kubenswrapper[4764]: I0202 09:07:46.957448 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:46Z","lastTransitionTime":"2026-02-02T09:07:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:47 crc kubenswrapper[4764]: I0202 09:07:47.020597 4764 generic.go:334] "Generic (PLEG): container finished" podID="660ac006-2495-45c4-9fb3-e9c1dddcf7a0" containerID="6d3fbf668ae7d04c8d9e574f43c2fe90c5604204ba38a3c60847afdfbceb7a57" exitCode=0 Feb 02 09:07:47 crc kubenswrapper[4764]: I0202 09:07:47.020637 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-9p5dc" event={"ID":"660ac006-2495-45c4-9fb3-e9c1dddcf7a0","Type":"ContainerDied","Data":"6d3fbf668ae7d04c8d9e574f43c2fe90c5604204ba38a3c60847afdfbceb7a57"} Feb 02 09:07:47 crc kubenswrapper[4764]: I0202 09:07:47.034331 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pcjmn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6e5ccd3b-b657-47a2-9c4d-bc9adb126533\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72863eef568d92a1abeb3cde2e89630a4f5fb7626c6949448ff5195774c98eab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2f7zl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pcjmn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:47Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:47 crc kubenswrapper[4764]: I0202 09:07:47.052580 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed747fd8-e3ce-4319-a6da-bda0fd51b85a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6f064bdc66470aace3180d4988244026b54009a63cff199e41c5d02a1d18479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://60258f84a1d8dcfe48b7e9688ed50f077e201971f20f3f6663fbf1ba58a4611a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f84be2e08217a2ac3edde858d78b4410376ed9bae3bcabe2a7f9b2fbe2f69c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a69d25d352d7ceac5c735bc1fcf446a1fea6dc22c65a5296cfab828cea17377\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:47Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:47 crc kubenswrapper[4764]: I0202 09:07:47.060016 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:47 crc kubenswrapper[4764]: I0202 09:07:47.060050 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:47 crc kubenswrapper[4764]: I0202 09:07:47.060059 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:47 crc kubenswrapper[4764]: I0202 09:07:47.060075 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:47 crc kubenswrapper[4764]: I0202 09:07:47.060087 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:47Z","lastTransitionTime":"2026-02-02T09:07:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:47 crc kubenswrapper[4764]: I0202 09:07:47.066782 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f6904228e052fff69ad1fd66df97758b6b60ba4351e97a7fbd03eee4df1b653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:47Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:47 crc kubenswrapper[4764]: I0202 09:07:47.076236 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6zxff" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aaba93da-12a5-43df-b7c7-4c8b988163c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3db293678a907cd55b31b1d0c0bd7c30e1ba75ecd5e13cbc68ab98fbf0fbd9d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqwb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6zxff\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:47Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:47 crc kubenswrapper[4764]: I0202 09:07:47.091261 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d192f670-9f9d-4539-9641-e4bed73acdd4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://040c5e8a5659f85f0d52d1a1e064154d998ae104f711101d2204c0b1caebbb30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbksf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fdf06410349b7c99aa166e87fbe80bf68e2ae923635a6a978b7e2cd50383f7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbksf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4ndm4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:47Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:47 crc kubenswrapper[4764]: I0202 09:07:47.103502 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:47Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:47 crc kubenswrapper[4764]: I0202 09:07:47.120949 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nc7qt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1a2c7b22068b7165927fc158885d48bf9952b9eba521d8ae9a599eee53dc305\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2hz4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nc7qt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:47Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:47 crc kubenswrapper[4764]: I0202 09:07:47.135231 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9p5dc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"660ac006-2495-45c4-9fb3-e9c1dddcf7a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23e57e65a6f5c3059506b9db21e9ac2492c30bd9191a98db0281915e9ce39797\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23e57e65a6f5c3059506b9db21e9ac2492c30bd9191a98db0281915e9ce39797\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2417b66816da1a6cd1791a4b91eca4a3c4f3694462e066e7f0722f1b9c2105c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2417b66816da1a6cd1791a4b91eca4a3c4f3694462e066e7f0722f1b9c2105c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://883641f29e9120c2cea6b272bacf1dc63c752f55ab0408cdaf73db85c85ac369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://883641f29e9120c2cea6b272bacf1dc63c752f55ab0408cdaf73db85c85ac369\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2745baba1188ba5e8f820616e4117a12c50ecee5b68e71ad9cabbb621b6e131f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2745baba1188ba5e8f820616e4117a12c50ecee5b68e71ad9cabbb621b6e131f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8925e435be4d48d86824fc35f8b8819fb35d63e385164c4a20edb3bf3035c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a8925e435be4d48d86824fc35f8b8819fb35d63e385164c4a20edb3bf3035c3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d3fbf668ae7d04c8d9e574f43c2fe90c5604204ba38a3c60847afdfbceb7a57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d3fbf668ae7d04c8d9e574f43c2fe90c5604204ba38a3c60847afdfbceb7a57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9p5dc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:47Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:47 crc kubenswrapper[4764]: I0202 09:07:47.146795 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87d61aea-46f6-4ee6-a7a4-767a6a67fea4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e78b6d08e4d616cefbb73e3fd43b3c6886dcd8aa997c7b4948bf1ff63401293\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d468715c2da33996561c40043216233fd12e21539d0c58dd1f3b892fd9f91a9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b184f7c936dc4046f74c0018269cbc2c4b7ceddb650b4e871d3f8a707e0df88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://269c5074371bd00f8363fb3888c11f1df383081a61497aea6578f08700881d50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b25bc9395831524c53f9e565b4efbe74f21792efafeb1dd507161bc91aa0ecdf\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T09:07:22Z\\\",\\\"message\\\":\\\"W0202 09:07:21.597488 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0202 09:07:21.598117 1 crypto.go:601] Generating new CA for check-endpoints-signer@1770023241 cert, and key in /tmp/serving-cert-3856183724/serving-signer.crt, /tmp/serving-cert-3856183724/serving-signer.key\\\\nI0202 09:07:22.012184 1 observer_polling.go:159] Starting file observer\\\\nW0202 09:07:22.015493 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0202 09:07:22.015797 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 09:07:22.016596 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3856183724/tls.crt::/tmp/serving-cert-3856183724/tls.key\\\\\\\"\\\\nF0202 09:07:22.270605 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b44e17b8ca01f549bc10081dd5f109cb1b9fff364bf5bd4fd90d9f0e3eeb04db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfb06a5ec9262cf39b108a66014d15e325cdbd5441a2d9e7a811dd57265c99a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfb06a5ec9262cf39b108a66014d15e325cdbd5441a2d9e7a811dd57265c99a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:47Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:47 crc kubenswrapper[4764]: I0202 09:07:47.160493 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5528225d91ac2778bef4554abb0bd036ac5d3b6cdac7357bab69e70c3894b6f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://071bbb80d0747cf3e109ba626d47f7d05f85f4cc0b3aba77276c795741f47ac6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:47Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:47 crc kubenswrapper[4764]: I0202 09:07:47.163003 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:47 crc kubenswrapper[4764]: I0202 09:07:47.163032 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:47 crc kubenswrapper[4764]: I0202 09:07:47.163040 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:47 crc kubenswrapper[4764]: I0202 09:07:47.163056 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:47 crc kubenswrapper[4764]: I0202 09:07:47.163065 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:47Z","lastTransitionTime":"2026-02-02T09:07:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:47 crc kubenswrapper[4764]: I0202 09:07:47.174787 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:47Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:47 crc kubenswrapper[4764]: I0202 09:07:47.200569 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d66ab3c0-9b5b-41d5-b6c0-91ccddccd7e9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d92ac486f5dade46bc75de32099859a71fc2ed25a7abbe6ae8937df128e7df30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a4ac273a4f127d79e5fad7d4c30cd2044a346f5734dba5d33d68cd092c0c2d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d38c7cb5dcd22fb07524494ebe77aa1a7f4f808f566043563817e47e7ec12a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://408722c7f83d28919431532f8c797a1c795fd5d96736d14e7f06a05744899702\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f473b18cd8003f75bb1dafb82bcec45745a73ccd10434297356dac185a3d15c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e229f484d19cbdabedf53d90a822a4434db9e00b1301538e885e26acc19d25f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e229f484d19cbdabedf53d90a822a4434db9e00b1301538e885e26acc19d25f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a9cc673be8ff245dba81278e8cd34cfe45ed36ff0b8bd5b30a0ddf333ec6371\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a9cc673be8ff245dba81278e8cd34cfe45ed36ff0b8bd5b30a0ddf333ec6371\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bc7e49c7a5ae66ea939060192fa6a346dc0b0b9aa9750b29879d3c426cc0d276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc7e49c7a5ae66ea939060192fa6a346dc0b0b9aa9750b29879d3c426cc0d276\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:47Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:47 crc kubenswrapper[4764]: I0202 09:07:47.212204 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa9d037e66cb53b9dd698eff46284c9e718fd842cda6f7b525f0c79062d70bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:47Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:47 crc kubenswrapper[4764]: I0202 09:07:47.227019 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:47Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:47 crc kubenswrapper[4764]: I0202 09:07:47.244312 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24632dda-6100-4ab6-a28e-214ddae4360c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://579903eadb183466a2564737e6b678e04308239c0cd75b6808892053b97afbc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://579903eadb183466a2564737e6b678e04308239c0cd75b6808892053b97afbc0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zhn7j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:47Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:47 crc kubenswrapper[4764]: I0202 09:07:47.264728 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:47 crc kubenswrapper[4764]: I0202 09:07:47.264763 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:47 crc kubenswrapper[4764]: I0202 09:07:47.264774 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:47 crc kubenswrapper[4764]: I0202 09:07:47.264788 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:47 crc kubenswrapper[4764]: I0202 09:07:47.264800 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:47Z","lastTransitionTime":"2026-02-02T09:07:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:47 crc kubenswrapper[4764]: I0202 09:07:47.367462 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:47 crc kubenswrapper[4764]: I0202 09:07:47.367644 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:47 crc kubenswrapper[4764]: I0202 09:07:47.367830 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:47 crc kubenswrapper[4764]: I0202 09:07:47.368041 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:47 crc kubenswrapper[4764]: I0202 09:07:47.368109 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:47Z","lastTransitionTime":"2026-02-02T09:07:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:47 crc kubenswrapper[4764]: I0202 09:07:47.470795 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:47 crc kubenswrapper[4764]: I0202 09:07:47.470823 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:47 crc kubenswrapper[4764]: I0202 09:07:47.470831 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:47 crc kubenswrapper[4764]: I0202 09:07:47.470843 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:47 crc kubenswrapper[4764]: I0202 09:07:47.470852 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:47Z","lastTransitionTime":"2026-02-02T09:07:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:47 crc kubenswrapper[4764]: I0202 09:07:47.575104 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:47 crc kubenswrapper[4764]: I0202 09:07:47.575155 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:47 crc kubenswrapper[4764]: I0202 09:07:47.575174 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:47 crc kubenswrapper[4764]: I0202 09:07:47.575191 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:47 crc kubenswrapper[4764]: I0202 09:07:47.575202 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:47Z","lastTransitionTime":"2026-02-02T09:07:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:47 crc kubenswrapper[4764]: I0202 09:07:47.633747 4764 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Feb 02 09:07:47 crc kubenswrapper[4764]: I0202 09:07:47.677383 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:47 crc kubenswrapper[4764]: I0202 09:07:47.677454 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:47 crc kubenswrapper[4764]: I0202 09:07:47.677469 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:47 crc kubenswrapper[4764]: I0202 09:07:47.677486 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:47 crc kubenswrapper[4764]: I0202 09:07:47.677523 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:47Z","lastTransitionTime":"2026-02-02T09:07:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:47 crc kubenswrapper[4764]: I0202 09:07:47.779916 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:47 crc kubenswrapper[4764]: I0202 09:07:47.779966 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:47 crc kubenswrapper[4764]: I0202 09:07:47.779977 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:47 crc kubenswrapper[4764]: I0202 09:07:47.779998 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:47 crc kubenswrapper[4764]: I0202 09:07:47.780007 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:47Z","lastTransitionTime":"2026-02-02T09:07:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:47 crc kubenswrapper[4764]: I0202 09:07:47.786123 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-17 11:32:19.276074154 +0000 UTC Feb 02 09:07:47 crc kubenswrapper[4764]: I0202 09:07:47.846742 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d66ab3c0-9b5b-41d5-b6c0-91ccddccd7e9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d92ac486f5dade46bc75de32099859a71fc2ed25a7abbe6ae8937df128e7df30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a4ac273a4f127d79e5fad7d4c30cd2044a346f5734dba5d33d68cd092c0c2d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d38c7cb5dcd22fb07524494ebe77aa1a7f4f808f566043563817e47e7ec12a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://408722c7f83d28919431532f8c797a1c795fd5d96736d14e7f06a05744899702\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f473b18cd8003f75bb1dafb82bcec45745a73ccd10434297356dac185a3d15c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e229f484d19cbdabedf53d90a822a4434db9e00b1301538e885e26acc19d25f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e229f484d19cbdabedf53d90a822a4434db9e00b1301538e885e26acc19d25f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a9cc673be8ff245dba81278e8cd34cfe45ed36ff0b8bd5b30a0ddf333ec6371\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a9cc673be8ff245dba81278e8cd34cfe45ed36ff0b8bd5b30a0ddf333ec6371\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bc7e49c7a5ae66ea939060192fa6a346dc0b0b9aa9750b29879d3c426cc0d276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc7e49c7a5ae66ea939060192fa6a346dc0b0b9aa9750b29879d3c426cc0d276\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:47Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:47 crc kubenswrapper[4764]: I0202 09:07:47.859425 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa9d037e66cb53b9dd698eff46284c9e718fd842cda6f7b525f0c79062d70bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:47Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:47 crc kubenswrapper[4764]: I0202 09:07:47.868777 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:47Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:47 crc kubenswrapper[4764]: I0202 09:07:47.882045 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:47 crc kubenswrapper[4764]: I0202 09:07:47.882106 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:47 crc kubenswrapper[4764]: I0202 09:07:47.882115 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:47 crc kubenswrapper[4764]: I0202 09:07:47.882128 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:47 crc kubenswrapper[4764]: I0202 09:07:47.882137 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:47Z","lastTransitionTime":"2026-02-02T09:07:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:47 crc kubenswrapper[4764]: I0202 09:07:47.885456 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24632dda-6100-4ab6-a28e-214ddae4360c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://579903eadb183466a2564737e6b678e04308239c0cd75b6808892053b97afbc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://579903eadb183466a2564737e6b678e04308239c0cd75b6808892053b97afbc0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zhn7j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:47Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:47 crc kubenswrapper[4764]: I0202 09:07:47.907081 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed747fd8-e3ce-4319-a6da-bda0fd51b85a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6f064bdc66470aace3180d4988244026b54009a63cff199e41c5d02a1d18479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://60258f84a1d8dcfe48b7e9688ed50f077e201971f20f3f6663fbf1ba58a4611a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f84be2e08217a2ac3edde858d78b4410376ed9bae3bcabe2a7f9b2fbe2f69c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a69d25d352d7ceac5c735bc1fcf446a1fea6dc22c65a5296cfab828cea17377\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:47Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:47 crc kubenswrapper[4764]: I0202 09:07:47.920027 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f6904228e052fff69ad1fd66df97758b6b60ba4351e97a7fbd03eee4df1b653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:47Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:47 crc kubenswrapper[4764]: I0202 09:07:47.931504 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6zxff" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aaba93da-12a5-43df-b7c7-4c8b988163c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3db293678a907cd55b31b1d0c0bd7c30e1ba75ecd5e13cbc68ab98fbf0fbd9d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqwb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6zxff\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:47Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:47 crc kubenswrapper[4764]: I0202 09:07:47.946364 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d192f670-9f9d-4539-9641-e4bed73acdd4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://040c5e8a5659f85f0d52d1a1e064154d998ae104f711101d2204c0b1caebbb30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbksf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fdf06410349b7c99aa166e87fbe80bf68e2ae923635a6a978b7e2cd50383f7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbksf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4ndm4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:47Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:47 crc kubenswrapper[4764]: I0202 09:07:47.959029 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pcjmn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6e5ccd3b-b657-47a2-9c4d-bc9adb126533\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72863eef568d92a1abeb3cde2e89630a4f5fb7626c6949448ff5195774c98eab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2f7zl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pcjmn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:47Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:47 crc kubenswrapper[4764]: I0202 09:07:47.971578 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:47Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:47 crc kubenswrapper[4764]: I0202 09:07:47.984239 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87d61aea-46f6-4ee6-a7a4-767a6a67fea4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e78b6d08e4d616cefbb73e3fd43b3c6886dcd8aa997c7b4948bf1ff63401293\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d468715c2da33996561c40043216233fd12e21539d0c58dd1f3b892fd9f91a9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b184f7c936dc4046f74c0018269cbc2c4b7ceddb650b4e871d3f8a707e0df88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://269c5074371bd00f8363fb3888c11f1df383081a61497aea6578f08700881d50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b25bc9395831524c53f9e565b4efbe74f21792efafeb1dd507161bc91aa0ecdf\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T09:07:22Z\\\",\\\"message\\\":\\\"W0202 09:07:21.597488 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0202 09:07:21.598117 1 crypto.go:601] Generating new CA for check-endpoints-signer@1770023241 cert, and key in /tmp/serving-cert-3856183724/serving-signer.crt, /tmp/serving-cert-3856183724/serving-signer.key\\\\nI0202 09:07:22.012184 1 observer_polling.go:159] Starting file observer\\\\nW0202 09:07:22.015493 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0202 09:07:22.015797 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 09:07:22.016596 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3856183724/tls.crt::/tmp/serving-cert-3856183724/tls.key\\\\\\\"\\\\nF0202 09:07:22.270605 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b44e17b8ca01f549bc10081dd5f109cb1b9fff364bf5bd4fd90d9f0e3eeb04db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfb06a5ec9262cf39b108a66014d15e325cdbd5441a2d9e7a811dd57265c99a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfb06a5ec9262cf39b108a66014d15e325cdbd5441a2d9e7a811dd57265c99a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:47Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:47 crc kubenswrapper[4764]: I0202 09:07:47.984974 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:47 crc kubenswrapper[4764]: I0202 09:07:47.985027 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:47 crc kubenswrapper[4764]: I0202 09:07:47.985042 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:47 crc kubenswrapper[4764]: I0202 09:07:47.985058 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:47 crc kubenswrapper[4764]: I0202 09:07:47.985069 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:47Z","lastTransitionTime":"2026-02-02T09:07:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:47 crc kubenswrapper[4764]: I0202 09:07:47.997636 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5528225d91ac2778bef4554abb0bd036ac5d3b6cdac7357bab69e70c3894b6f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://071bbb80d0747cf3e109ba626d47f7d05f85f4cc0b3aba77276c795741f47ac6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:47Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:48 crc kubenswrapper[4764]: I0202 09:07:48.008957 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:48Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:48 crc kubenswrapper[4764]: I0202 09:07:48.024014 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nc7qt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1a2c7b22068b7165927fc158885d48bf9952b9eba521d8ae9a599eee53dc305\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2hz4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nc7qt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:48Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:48 crc kubenswrapper[4764]: I0202 09:07:48.038859 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" event={"ID":"24632dda-6100-4ab6-a28e-214ddae4360c","Type":"ContainerStarted","Data":"469f42001223aec6a9629296820e4925288790b1a17d6c7ac106dab6989bfc77"} Feb 02 09:07:48 crc kubenswrapper[4764]: I0202 09:07:48.044046 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9p5dc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"660ac006-2495-45c4-9fb3-e9c1dddcf7a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23e57e65a6f5c3059506b9db21e9ac2492c30bd9191a98db0281915e9ce39797\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23e57e65a6f5c3059506b9db21e9ac2492c30bd9191a98db0281915e9ce39797\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2417b66816da1a6cd1791a4b91eca4a3c4f3694462e066e7f0722f1b9c2105c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2417b66816da1a6cd1791a4b91eca4a3c4f3694462e066e7f0722f1b9c2105c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://883641f29e9120c2cea6b272bacf1dc63c752f55ab0408cdaf73db85c85ac369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://883641f29e9120c2cea6b272bacf1dc63c752f55ab0408cdaf73db85c85ac369\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2745baba1188ba5e8f820616e4117a12c50ecee5b68e71ad9cabbb621b6e131f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2745baba1188ba5e8f820616e4117a12c50ecee5b68e71ad9cabbb621b6e131f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8925e435be4d48d86824fc35f8b8819fb35d63e385164c4a20edb3bf3035c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a8925e435be4d48d86824fc35f8b8819fb35d63e385164c4a20edb3bf3035c3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d3fbf668ae7d04c8d9e574f43c2fe90c5604204ba38a3c60847afdfbceb7a57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d3fbf668ae7d04c8d9e574f43c2fe90c5604204ba38a3c60847afdfbceb7a57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9p5dc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:48Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:48 crc kubenswrapper[4764]: I0202 09:07:48.044147 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-9p5dc" event={"ID":"660ac006-2495-45c4-9fb3-e9c1dddcf7a0","Type":"ContainerStarted","Data":"da812b124c97aa8ad1563f4bcfac8f3bfa025096fa4a2561e5847b1bf9cb30b8"} Feb 02 09:07:48 crc kubenswrapper[4764]: I0202 09:07:48.058793 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nc7qt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1a2c7b22068b7165927fc158885d48bf9952b9eba521d8ae9a599eee53dc305\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2hz4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nc7qt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:48Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:48 crc kubenswrapper[4764]: I0202 09:07:48.072825 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9p5dc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"660ac006-2495-45c4-9fb3-e9c1dddcf7a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23e57e65a6f5c3059506b9db21e9ac2492c30bd9191a98db0281915e9ce39797\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23e57e65a6f5c3059506b9db21e9ac2492c30bd9191a98db0281915e9ce39797\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2417b66816da1a6cd1791a4b91eca4a3c4f3694462e066e7f0722f1b9c2105c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2417b66816da1a6cd1791a4b91eca4a3c4f3694462e066e7f0722f1b9c2105c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://883641f29e9120c2cea6b272bacf1dc63c752f55ab0408cdaf73db85c85ac369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://883641f29e9120c2cea6b272bacf1dc63c752f55ab0408cdaf73db85c85ac369\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2745baba1188ba5e8f820616e4117a12c50ecee5b68e71ad9cabbb621b6e131f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2745baba1188ba5e8f820616e4117a12c50ecee5b68e71ad9cabbb621b6e131f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8925e435be4d48d86824fc35f8b8819fb35d63e385164c4a20edb3bf3035c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a8925e435be4d48d86824fc35f8b8819fb35d63e385164c4a20edb3bf3035c3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d3fbf668ae7d04c8d9e574f43c2fe90c5604204ba38a3c60847afdfbceb7a57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d3fbf668ae7d04c8d9e574f43c2fe90c5604204ba38a3c60847afdfbceb7a57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9p5dc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:48Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:48 crc kubenswrapper[4764]: I0202 09:07:48.087442 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87d61aea-46f6-4ee6-a7a4-767a6a67fea4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e78b6d08e4d616cefbb73e3fd43b3c6886dcd8aa997c7b4948bf1ff63401293\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d468715c2da33996561c40043216233fd12e21539d0c58dd1f3b892fd9f91a9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b184f7c936dc4046f74c0018269cbc2c4b7ceddb650b4e871d3f8a707e0df88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://269c5074371bd00f8363fb3888c11f1df383081a61497aea6578f08700881d50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b25bc9395831524c53f9e565b4efbe74f21792efafeb1dd507161bc91aa0ecdf\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T09:07:22Z\\\",\\\"message\\\":\\\"W0202 09:07:21.597488 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0202 09:07:21.598117 1 crypto.go:601] Generating new CA for check-endpoints-signer@1770023241 cert, and key in /tmp/serving-cert-3856183724/serving-signer.crt, /tmp/serving-cert-3856183724/serving-signer.key\\\\nI0202 09:07:22.012184 1 observer_polling.go:159] Starting file observer\\\\nW0202 09:07:22.015493 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0202 09:07:22.015797 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 09:07:22.016596 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3856183724/tls.crt::/tmp/serving-cert-3856183724/tls.key\\\\\\\"\\\\nF0202 09:07:22.270605 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b44e17b8ca01f549bc10081dd5f109cb1b9fff364bf5bd4fd90d9f0e3eeb04db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfb06a5ec9262cf39b108a66014d15e325cdbd5441a2d9e7a811dd57265c99a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfb06a5ec9262cf39b108a66014d15e325cdbd5441a2d9e7a811dd57265c99a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:48Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:48 crc kubenswrapper[4764]: I0202 09:07:48.088109 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:48 crc kubenswrapper[4764]: I0202 09:07:48.088160 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:48 crc kubenswrapper[4764]: I0202 09:07:48.088172 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:48 crc kubenswrapper[4764]: I0202 09:07:48.088193 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:48 crc kubenswrapper[4764]: I0202 09:07:48.088205 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:48Z","lastTransitionTime":"2026-02-02T09:07:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:48 crc kubenswrapper[4764]: I0202 09:07:48.101355 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5528225d91ac2778bef4554abb0bd036ac5d3b6cdac7357bab69e70c3894b6f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://071bbb80d0747cf3e109ba626d47f7d05f85f4cc0b3aba77276c795741f47ac6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:48Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:48 crc kubenswrapper[4764]: I0202 09:07:48.115727 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:48Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:48 crc kubenswrapper[4764]: I0202 09:07:48.136028 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d66ab3c0-9b5b-41d5-b6c0-91ccddccd7e9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d92ac486f5dade46bc75de32099859a71fc2ed25a7abbe6ae8937df128e7df30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a4ac273a4f127d79e5fad7d4c30cd2044a346f5734dba5d33d68cd092c0c2d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d38c7cb5dcd22fb07524494ebe77aa1a7f4f808f566043563817e47e7ec12a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://408722c7f83d28919431532f8c797a1c795fd5d96736d14e7f06a05744899702\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f473b18cd8003f75bb1dafb82bcec45745a73ccd10434297356dac185a3d15c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e229f484d19cbdabedf53d90a822a4434db9e00b1301538e885e26acc19d25f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e229f484d19cbdabedf53d90a822a4434db9e00b1301538e885e26acc19d25f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a9cc673be8ff245dba81278e8cd34cfe45ed36ff0b8bd5b30a0ddf333ec6371\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a9cc673be8ff245dba81278e8cd34cfe45ed36ff0b8bd5b30a0ddf333ec6371\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bc7e49c7a5ae66ea939060192fa6a346dc0b0b9aa9750b29879d3c426cc0d276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc7e49c7a5ae66ea939060192fa6a346dc0b0b9aa9750b29879d3c426cc0d276\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:48Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:48 crc kubenswrapper[4764]: I0202 09:07:48.148159 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa9d037e66cb53b9dd698eff46284c9e718fd842cda6f7b525f0c79062d70bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:48Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:48 crc kubenswrapper[4764]: I0202 09:07:48.162191 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:48Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:48 crc kubenswrapper[4764]: I0202 09:07:48.178417 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24632dda-6100-4ab6-a28e-214ddae4360c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a6c67e9b7708cc0d85b0b6ffbbade0d34867e159a9c5487ca2606475c667a50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bffda48d7c881c1470251b501661324859373246cc1a21811bd36ca24887c7aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a989645cf485cd8ff80ffff1a86cd7808e1564eba5667535e161cc68db13b26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20b71d3ee7a381baa51b0b0d0470283022cfede30f491135737b5e82d631d763\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c380383b6a014d1b694a13ddc22f54b20b9e4f9d5c9b9000d76187621dce0a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5529f2bff87753428e9173826fc1880acf96a4c134fa9ecad195b15697cad73a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://469f42001223aec6a9629296820e4925288790b1a17d6c7ac106dab6989bfc77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b4cfbd4e682bcdcaff785013bdb72580ce36fced733e235538536827036b93d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://579903eadb183466a2564737e6b678e04308239c0cd75b6808892053b97afbc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://579903eadb183466a2564737e6b678e04308239c0cd75b6808892053b97afbc0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zhn7j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:48Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:48 crc kubenswrapper[4764]: I0202 09:07:48.187060 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pcjmn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6e5ccd3b-b657-47a2-9c4d-bc9adb126533\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72863eef568d92a1abeb3cde2e89630a4f5fb7626c6949448ff5195774c98eab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2f7zl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pcjmn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:48Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:48 crc kubenswrapper[4764]: I0202 09:07:48.189946 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:48 crc kubenswrapper[4764]: I0202 09:07:48.189977 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:48 crc kubenswrapper[4764]: I0202 09:07:48.189986 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:48 crc kubenswrapper[4764]: I0202 09:07:48.189999 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:48 crc kubenswrapper[4764]: I0202 09:07:48.190009 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:48Z","lastTransitionTime":"2026-02-02T09:07:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:48 crc kubenswrapper[4764]: I0202 09:07:48.198179 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed747fd8-e3ce-4319-a6da-bda0fd51b85a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6f064bdc66470aace3180d4988244026b54009a63cff199e41c5d02a1d18479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://60258f84a1d8dcfe48b7e9688ed50f077e201971f20f3f6663fbf1ba58a4611a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f84be2e08217a2ac3edde858d78b4410376ed9bae3bcabe2a7f9b2fbe2f69c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a69d25d352d7ceac5c735bc1fcf446a1fea6dc22c65a5296cfab828cea17377\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:48Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:48 crc kubenswrapper[4764]: I0202 09:07:48.209386 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f6904228e052fff69ad1fd66df97758b6b60ba4351e97a7fbd03eee4df1b653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:48Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:48 crc kubenswrapper[4764]: I0202 09:07:48.219921 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6zxff" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aaba93da-12a5-43df-b7c7-4c8b988163c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3db293678a907cd55b31b1d0c0bd7c30e1ba75ecd5e13cbc68ab98fbf0fbd9d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqwb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6zxff\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:48Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:48 crc kubenswrapper[4764]: I0202 09:07:48.229846 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d192f670-9f9d-4539-9641-e4bed73acdd4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://040c5e8a5659f85f0d52d1a1e064154d998ae104f711101d2204c0b1caebbb30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbksf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fdf06410349b7c99aa166e87fbe80bf68e2ae923635a6a978b7e2cd50383f7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbksf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4ndm4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:48Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:48 crc kubenswrapper[4764]: I0202 09:07:48.240151 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:48Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:48 crc kubenswrapper[4764]: I0202 09:07:48.258301 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d66ab3c0-9b5b-41d5-b6c0-91ccddccd7e9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d92ac486f5dade46bc75de32099859a71fc2ed25a7abbe6ae8937df128e7df30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a4ac273a4f127d79e5fad7d4c30cd2044a346f5734dba5d33d68cd092c0c2d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d38c7cb5dcd22fb07524494ebe77aa1a7f4f808f566043563817e47e7ec12a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://408722c7f83d28919431532f8c797a1c795fd5d96736d14e7f06a05744899702\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f473b18cd8003f75bb1dafb82bcec45745a73ccd10434297356dac185a3d15c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e229f484d19cbdabedf53d90a822a4434db9e00b1301538e885e26acc19d25f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e229f484d19cbdabedf53d90a822a4434db9e00b1301538e885e26acc19d25f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a9cc673be8ff245dba81278e8cd34cfe45ed36ff0b8bd5b30a0ddf333ec6371\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a9cc673be8ff245dba81278e8cd34cfe45ed36ff0b8bd5b30a0ddf333ec6371\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bc7e49c7a5ae66ea939060192fa6a346dc0b0b9aa9750b29879d3c426cc0d276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc7e49c7a5ae66ea939060192fa6a346dc0b0b9aa9750b29879d3c426cc0d276\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:48Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:48 crc kubenswrapper[4764]: I0202 09:07:48.274917 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa9d037e66cb53b9dd698eff46284c9e718fd842cda6f7b525f0c79062d70bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:48Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:48 crc kubenswrapper[4764]: I0202 09:07:48.287203 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:48Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:48 crc kubenswrapper[4764]: I0202 09:07:48.291819 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:48 crc kubenswrapper[4764]: I0202 09:07:48.291865 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:48 crc kubenswrapper[4764]: I0202 09:07:48.291877 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:48 crc kubenswrapper[4764]: I0202 09:07:48.291896 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:48 crc kubenswrapper[4764]: I0202 09:07:48.291909 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:48Z","lastTransitionTime":"2026-02-02T09:07:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:48 crc kubenswrapper[4764]: I0202 09:07:48.316995 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24632dda-6100-4ab6-a28e-214ddae4360c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a6c67e9b7708cc0d85b0b6ffbbade0d34867e159a9c5487ca2606475c667a50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bffda48d7c881c1470251b501661324859373246cc1a21811bd36ca24887c7aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a989645cf485cd8ff80ffff1a86cd7808e1564eba5667535e161cc68db13b26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20b71d3ee7a381baa51b0b0d0470283022cfede30f491135737b5e82d631d763\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c380383b6a014d1b694a13ddc22f54b20b9e4f9d5c9b9000d76187621dce0a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5529f2bff87753428e9173826fc1880acf96a4c134fa9ecad195b15697cad73a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://469f42001223aec6a9629296820e4925288790b1a17d6c7ac106dab6989bfc77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b4cfbd4e682bcdcaff785013bdb72580ce36fced733e235538536827036b93d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://579903eadb183466a2564737e6b678e04308239c0cd75b6808892053b97afbc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://579903eadb183466a2564737e6b678e04308239c0cd75b6808892053b97afbc0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zhn7j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:48Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:48 crc kubenswrapper[4764]: I0202 09:07:48.350756 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed747fd8-e3ce-4319-a6da-bda0fd51b85a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6f064bdc66470aace3180d4988244026b54009a63cff199e41c5d02a1d18479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://60258f84a1d8dcfe48b7e9688ed50f077e201971f20f3f6663fbf1ba58a4611a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f84be2e08217a2ac3edde858d78b4410376ed9bae3bcabe2a7f9b2fbe2f69c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a69d25d352d7ceac5c735bc1fcf446a1fea6dc22c65a5296cfab828cea17377\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:48Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:48 crc kubenswrapper[4764]: I0202 09:07:48.394477 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:48 crc kubenswrapper[4764]: I0202 09:07:48.394518 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:48 crc kubenswrapper[4764]: I0202 09:07:48.394532 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:48 crc kubenswrapper[4764]: I0202 09:07:48.394551 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:48 crc kubenswrapper[4764]: I0202 09:07:48.394566 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:48Z","lastTransitionTime":"2026-02-02T09:07:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:48 crc kubenswrapper[4764]: I0202 09:07:48.404014 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f6904228e052fff69ad1fd66df97758b6b60ba4351e97a7fbd03eee4df1b653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:48Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:48 crc kubenswrapper[4764]: I0202 09:07:48.442525 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6zxff" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aaba93da-12a5-43df-b7c7-4c8b988163c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3db293678a907cd55b31b1d0c0bd7c30e1ba75ecd5e13cbc68ab98fbf0fbd9d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqwb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6zxff\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:48Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:48 crc kubenswrapper[4764]: I0202 09:07:48.469232 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d192f670-9f9d-4539-9641-e4bed73acdd4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://040c5e8a5659f85f0d52d1a1e064154d998ae104f711101d2204c0b1caebbb30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbksf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fdf06410349b7c99aa166e87fbe80bf68e2ae923635a6a978b7e2cd50383f7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbksf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4ndm4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:48Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:48 crc kubenswrapper[4764]: I0202 09:07:48.497186 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:48 crc kubenswrapper[4764]: I0202 09:07:48.497226 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:48 crc kubenswrapper[4764]: I0202 09:07:48.497237 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:48 crc kubenswrapper[4764]: I0202 09:07:48.497252 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:48 crc kubenswrapper[4764]: I0202 09:07:48.497263 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:48Z","lastTransitionTime":"2026-02-02T09:07:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:48 crc kubenswrapper[4764]: I0202 09:07:48.506813 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pcjmn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6e5ccd3b-b657-47a2-9c4d-bc9adb126533\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72863eef568d92a1abeb3cde2e89630a4f5fb7626c6949448ff5195774c98eab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2f7zl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pcjmn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:48Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:48 crc kubenswrapper[4764]: I0202 09:07:48.548227 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:48Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:48 crc kubenswrapper[4764]: I0202 09:07:48.589670 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9p5dc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"660ac006-2495-45c4-9fb3-e9c1dddcf7a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da812b124c97aa8ad1563f4bcfac8f3bfa025096fa4a2561e5847b1bf9cb30b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23e57e65a6f5c3059506b9db21e9ac2492c30bd9191a98db0281915e9ce39797\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23e57e65a6f5c3059506b9db21e9ac2492c30bd9191a98db0281915e9ce39797\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2417b66816da1a6cd1791a4b91eca4a3c4f3694462e066e7f0722f1b9c2105c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2417b66816da1a6cd1791a4b91eca4a3c4f3694462e066e7f0722f1b9c2105c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://883641f29e9120c2cea6b272bacf1dc63c752f55ab0408cdaf73db85c85ac369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://883641f29e9120c2cea6b272bacf1dc63c752f55ab0408cdaf73db85c85ac369\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2745baba1188ba5e8f820616e4117a12c50ecee5b68e71ad9cabbb621b6e131f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2745baba1188ba5e8f820616e4117a12c50ecee5b68e71ad9cabbb621b6e131f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8925e435be4d48d86824fc35f8b8819fb35d63e385164c4a20edb3bf3035c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a8925e435be4d48d86824fc35f8b8819fb35d63e385164c4a20edb3bf3035c3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d3fbf668ae7d04c8d9e574f43c2fe90c5604204ba38a3c60847afdfbceb7a57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d3fbf668ae7d04c8d9e574f43c2fe90c5604204ba38a3c60847afdfbceb7a57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9p5dc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:48Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:48 crc kubenswrapper[4764]: I0202 09:07:48.599319 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:48 crc kubenswrapper[4764]: I0202 09:07:48.599348 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:48 crc kubenswrapper[4764]: I0202 09:07:48.599356 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:48 crc kubenswrapper[4764]: I0202 09:07:48.599369 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:48 crc kubenswrapper[4764]: I0202 09:07:48.599378 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:48Z","lastTransitionTime":"2026-02-02T09:07:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:48 crc kubenswrapper[4764]: I0202 09:07:48.628888 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87d61aea-46f6-4ee6-a7a4-767a6a67fea4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e78b6d08e4d616cefbb73e3fd43b3c6886dcd8aa997c7b4948bf1ff63401293\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d468715c2da33996561c40043216233fd12e21539d0c58dd1f3b892fd9f91a9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b184f7c936dc4046f74c0018269cbc2c4b7ceddb650b4e871d3f8a707e0df88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://269c5074371bd00f8363fb3888c11f1df383081a61497aea6578f08700881d50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b25bc9395831524c53f9e565b4efbe74f21792efafeb1dd507161bc91aa0ecdf\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T09:07:22Z\\\",\\\"message\\\":\\\"W0202 09:07:21.597488 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0202 09:07:21.598117 1 crypto.go:601] Generating new CA for check-endpoints-signer@1770023241 cert, and key in /tmp/serving-cert-3856183724/serving-signer.crt, /tmp/serving-cert-3856183724/serving-signer.key\\\\nI0202 09:07:22.012184 1 observer_polling.go:159] Starting file observer\\\\nW0202 09:07:22.015493 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0202 09:07:22.015797 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 09:07:22.016596 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3856183724/tls.crt::/tmp/serving-cert-3856183724/tls.key\\\\\\\"\\\\nF0202 09:07:22.270605 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b44e17b8ca01f549bc10081dd5f109cb1b9fff364bf5bd4fd90d9f0e3eeb04db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfb06a5ec9262cf39b108a66014d15e325cdbd5441a2d9e7a811dd57265c99a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfb06a5ec9262cf39b108a66014d15e325cdbd5441a2d9e7a811dd57265c99a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:48Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:48 crc kubenswrapper[4764]: I0202 09:07:48.669248 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5528225d91ac2778bef4554abb0bd036ac5d3b6cdac7357bab69e70c3894b6f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://071bbb80d0747cf3e109ba626d47f7d05f85f4cc0b3aba77276c795741f47ac6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:48Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:48 crc kubenswrapper[4764]: I0202 09:07:48.701777 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:48 crc kubenswrapper[4764]: I0202 09:07:48.701873 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:48 crc kubenswrapper[4764]: I0202 09:07:48.701895 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:48 crc kubenswrapper[4764]: I0202 09:07:48.701917 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:48 crc kubenswrapper[4764]: I0202 09:07:48.701985 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:48Z","lastTransitionTime":"2026-02-02T09:07:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:48 crc kubenswrapper[4764]: I0202 09:07:48.709770 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:48Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:48 crc kubenswrapper[4764]: I0202 09:07:48.750706 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nc7qt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1a2c7b22068b7165927fc158885d48bf9952b9eba521d8ae9a599eee53dc305\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2hz4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nc7qt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:48Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:48 crc kubenswrapper[4764]: I0202 09:07:48.787641 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-30 14:09:02.173091118 +0000 UTC Feb 02 09:07:48 crc kubenswrapper[4764]: I0202 09:07:48.804547 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:48 crc kubenswrapper[4764]: I0202 09:07:48.804600 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:48 crc kubenswrapper[4764]: I0202 09:07:48.804612 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:48 crc kubenswrapper[4764]: I0202 09:07:48.804625 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:48 crc kubenswrapper[4764]: I0202 09:07:48.804635 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:48Z","lastTransitionTime":"2026-02-02T09:07:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:48 crc kubenswrapper[4764]: I0202 09:07:48.824886 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 09:07:48 crc kubenswrapper[4764]: I0202 09:07:48.824948 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 09:07:48 crc kubenswrapper[4764]: I0202 09:07:48.824887 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 09:07:48 crc kubenswrapper[4764]: E0202 09:07:48.825018 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 09:07:48 crc kubenswrapper[4764]: E0202 09:07:48.825115 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 09:07:48 crc kubenswrapper[4764]: E0202 09:07:48.825211 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 09:07:48 crc kubenswrapper[4764]: I0202 09:07:48.907156 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:48 crc kubenswrapper[4764]: I0202 09:07:48.907202 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:48 crc kubenswrapper[4764]: I0202 09:07:48.907215 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:48 crc kubenswrapper[4764]: I0202 09:07:48.907230 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:48 crc kubenswrapper[4764]: I0202 09:07:48.907239 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:48Z","lastTransitionTime":"2026-02-02T09:07:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:49 crc kubenswrapper[4764]: I0202 09:07:49.009521 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:49 crc kubenswrapper[4764]: I0202 09:07:49.009577 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:49 crc kubenswrapper[4764]: I0202 09:07:49.009599 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:49 crc kubenswrapper[4764]: I0202 09:07:49.009626 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:49 crc kubenswrapper[4764]: I0202 09:07:49.009646 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:49Z","lastTransitionTime":"2026-02-02T09:07:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:49 crc kubenswrapper[4764]: I0202 09:07:49.047743 4764 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 02 09:07:49 crc kubenswrapper[4764]: I0202 09:07:49.048474 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" Feb 02 09:07:49 crc kubenswrapper[4764]: I0202 09:07:49.048541 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" Feb 02 09:07:49 crc kubenswrapper[4764]: I0202 09:07:49.091179 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" Feb 02 09:07:49 crc kubenswrapper[4764]: I0202 09:07:49.092471 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" Feb 02 09:07:49 crc kubenswrapper[4764]: I0202 09:07:49.109117 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed747fd8-e3ce-4319-a6da-bda0fd51b85a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6f064bdc66470aace3180d4988244026b54009a63cff199e41c5d02a1d18479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://60258f84a1d8dcfe48b7e9688ed50f077e201971f20f3f6663fbf1ba58a4611a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f84be2e08217a2ac3edde858d78b4410376ed9bae3bcabe2a7f9b2fbe2f69c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a69d25d352d7ceac5c735bc1fcf446a1fea6dc22c65a5296cfab828cea17377\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:49Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:49 crc kubenswrapper[4764]: I0202 09:07:49.111642 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:49 crc kubenswrapper[4764]: I0202 09:07:49.111667 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:49 crc kubenswrapper[4764]: I0202 09:07:49.111677 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:49 crc kubenswrapper[4764]: I0202 09:07:49.111689 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:49 crc kubenswrapper[4764]: I0202 09:07:49.111698 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:49Z","lastTransitionTime":"2026-02-02T09:07:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:49 crc kubenswrapper[4764]: I0202 09:07:49.130845 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f6904228e052fff69ad1fd66df97758b6b60ba4351e97a7fbd03eee4df1b653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:49Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:49 crc kubenswrapper[4764]: I0202 09:07:49.145064 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6zxff" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aaba93da-12a5-43df-b7c7-4c8b988163c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3db293678a907cd55b31b1d0c0bd7c30e1ba75ecd5e13cbc68ab98fbf0fbd9d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqwb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6zxff\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:49Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:49 crc kubenswrapper[4764]: I0202 09:07:49.157404 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d192f670-9f9d-4539-9641-e4bed73acdd4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://040c5e8a5659f85f0d52d1a1e064154d998ae104f711101d2204c0b1caebbb30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbksf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fdf06410349b7c99aa166e87fbe80bf68e2ae923635a6a978b7e2cd50383f7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbksf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4ndm4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:49Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:49 crc kubenswrapper[4764]: I0202 09:07:49.168759 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pcjmn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6e5ccd3b-b657-47a2-9c4d-bc9adb126533\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72863eef568d92a1abeb3cde2e89630a4f5fb7626c6949448ff5195774c98eab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2f7zl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pcjmn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:49Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:49 crc kubenswrapper[4764]: I0202 09:07:49.188418 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:49Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:49 crc kubenswrapper[4764]: I0202 09:07:49.204149 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9p5dc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"660ac006-2495-45c4-9fb3-e9c1dddcf7a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da812b124c97aa8ad1563f4bcfac8f3bfa025096fa4a2561e5847b1bf9cb30b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23e57e65a6f5c3059506b9db21e9ac2492c30bd9191a98db0281915e9ce39797\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23e57e65a6f5c3059506b9db21e9ac2492c30bd9191a98db0281915e9ce39797\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2417b66816da1a6cd1791a4b91eca4a3c4f3694462e066e7f0722f1b9c2105c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2417b66816da1a6cd1791a4b91eca4a3c4f3694462e066e7f0722f1b9c2105c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://883641f29e9120c2cea6b272bacf1dc63c752f55ab0408cdaf73db85c85ac369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://883641f29e9120c2cea6b272bacf1dc63c752f55ab0408cdaf73db85c85ac369\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2745baba1188ba5e8f820616e4117a12c50ecee5b68e71ad9cabbb621b6e131f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2745baba1188ba5e8f820616e4117a12c50ecee5b68e71ad9cabbb621b6e131f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8925e435be4d48d86824fc35f8b8819fb35d63e385164c4a20edb3bf3035c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a8925e435be4d48d86824fc35f8b8819fb35d63e385164c4a20edb3bf3035c3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d3fbf668ae7d04c8d9e574f43c2fe90c5604204ba38a3c60847afdfbceb7a57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d3fbf668ae7d04c8d9e574f43c2fe90c5604204ba38a3c60847afdfbceb7a57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9p5dc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:49Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:49 crc kubenswrapper[4764]: I0202 09:07:49.213531 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:49 crc kubenswrapper[4764]: I0202 09:07:49.213570 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:49 crc kubenswrapper[4764]: I0202 09:07:49.213583 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:49 crc kubenswrapper[4764]: I0202 09:07:49.213599 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:49 crc kubenswrapper[4764]: I0202 09:07:49.213611 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:49Z","lastTransitionTime":"2026-02-02T09:07:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:49 crc kubenswrapper[4764]: I0202 09:07:49.220524 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87d61aea-46f6-4ee6-a7a4-767a6a67fea4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e78b6d08e4d616cefbb73e3fd43b3c6886dcd8aa997c7b4948bf1ff63401293\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d468715c2da33996561c40043216233fd12e21539d0c58dd1f3b892fd9f91a9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b184f7c936dc4046f74c0018269cbc2c4b7ceddb650b4e871d3f8a707e0df88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://269c5074371bd00f8363fb3888c11f1df383081a61497aea6578f08700881d50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b25bc9395831524c53f9e565b4efbe74f21792efafeb1dd507161bc91aa0ecdf\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T09:07:22Z\\\",\\\"message\\\":\\\"W0202 09:07:21.597488 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0202 09:07:21.598117 1 crypto.go:601] Generating new CA for check-endpoints-signer@1770023241 cert, and key in /tmp/serving-cert-3856183724/serving-signer.crt, /tmp/serving-cert-3856183724/serving-signer.key\\\\nI0202 09:07:22.012184 1 observer_polling.go:159] Starting file observer\\\\nW0202 09:07:22.015493 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0202 09:07:22.015797 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 09:07:22.016596 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3856183724/tls.crt::/tmp/serving-cert-3856183724/tls.key\\\\\\\"\\\\nF0202 09:07:22.270605 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b44e17b8ca01f549bc10081dd5f109cb1b9fff364bf5bd4fd90d9f0e3eeb04db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfb06a5ec9262cf39b108a66014d15e325cdbd5441a2d9e7a811dd57265c99a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfb06a5ec9262cf39b108a66014d15e325cdbd5441a2d9e7a811dd57265c99a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:49Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:49 crc kubenswrapper[4764]: I0202 09:07:49.235516 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5528225d91ac2778bef4554abb0bd036ac5d3b6cdac7357bab69e70c3894b6f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://071bbb80d0747cf3e109ba626d47f7d05f85f4cc0b3aba77276c795741f47ac6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:49Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:49 crc kubenswrapper[4764]: I0202 09:07:49.253589 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:49Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:49 crc kubenswrapper[4764]: I0202 09:07:49.267774 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nc7qt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1a2c7b22068b7165927fc158885d48bf9952b9eba521d8ae9a599eee53dc305\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2hz4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nc7qt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:49Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:49 crc kubenswrapper[4764]: I0202 09:07:49.293436 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d66ab3c0-9b5b-41d5-b6c0-91ccddccd7e9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d92ac486f5dade46bc75de32099859a71fc2ed25a7abbe6ae8937df128e7df30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a4ac273a4f127d79e5fad7d4c30cd2044a346f5734dba5d33d68cd092c0c2d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d38c7cb5dcd22fb07524494ebe77aa1a7f4f808f566043563817e47e7ec12a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://408722c7f83d28919431532f8c797a1c795fd5d96736d14e7f06a05744899702\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f473b18cd8003f75bb1dafb82bcec45745a73ccd10434297356dac185a3d15c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e229f484d19cbdabedf53d90a822a4434db9e00b1301538e885e26acc19d25f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e229f484d19cbdabedf53d90a822a4434db9e00b1301538e885e26acc19d25f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a9cc673be8ff245dba81278e8cd34cfe45ed36ff0b8bd5b30a0ddf333ec6371\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a9cc673be8ff245dba81278e8cd34cfe45ed36ff0b8bd5b30a0ddf333ec6371\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bc7e49c7a5ae66ea939060192fa6a346dc0b0b9aa9750b29879d3c426cc0d276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc7e49c7a5ae66ea939060192fa6a346dc0b0b9aa9750b29879d3c426cc0d276\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:49Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:49 crc kubenswrapper[4764]: I0202 09:07:49.308303 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa9d037e66cb53b9dd698eff46284c9e718fd842cda6f7b525f0c79062d70bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:49Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:49 crc kubenswrapper[4764]: I0202 09:07:49.316354 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:49 crc kubenswrapper[4764]: I0202 09:07:49.316396 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:49 crc kubenswrapper[4764]: I0202 09:07:49.316404 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:49 crc kubenswrapper[4764]: I0202 09:07:49.316420 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:49 crc kubenswrapper[4764]: I0202 09:07:49.316428 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:49Z","lastTransitionTime":"2026-02-02T09:07:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:49 crc kubenswrapper[4764]: I0202 09:07:49.324884 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:49Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:49 crc kubenswrapper[4764]: I0202 09:07:49.359149 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24632dda-6100-4ab6-a28e-214ddae4360c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a6c67e9b7708cc0d85b0b6ffbbade0d34867e159a9c5487ca2606475c667a50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bffda48d7c881c1470251b501661324859373246cc1a21811bd36ca24887c7aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a989645cf485cd8ff80ffff1a86cd7808e1564eba5667535e161cc68db13b26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20b71d3ee7a381baa51b0b0d0470283022cfede30f491135737b5e82d631d763\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c380383b6a014d1b694a13ddc22f54b20b9e4f9d5c9b9000d76187621dce0a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5529f2bff87753428e9173826fc1880acf96a4c134fa9ecad195b15697cad73a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://469f42001223aec6a9629296820e4925288790b1a17d6c7ac106dab6989bfc77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b4cfbd4e682bcdcaff785013bdb72580ce36fced733e235538536827036b93d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://579903eadb183466a2564737e6b678e04308239c0cd75b6808892053b97afbc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://579903eadb183466a2564737e6b678e04308239c0cd75b6808892053b97afbc0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zhn7j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:49Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:49 crc kubenswrapper[4764]: I0202 09:07:49.398822 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5528225d91ac2778bef4554abb0bd036ac5d3b6cdac7357bab69e70c3894b6f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://071bbb80d0747cf3e109ba626d47f7d05f85f4cc0b3aba77276c795741f47ac6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:49Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:49 crc kubenswrapper[4764]: I0202 09:07:49.418186 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:49 crc kubenswrapper[4764]: I0202 09:07:49.418244 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:49 crc kubenswrapper[4764]: I0202 09:07:49.418260 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:49 crc kubenswrapper[4764]: I0202 09:07:49.418284 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:49 crc kubenswrapper[4764]: I0202 09:07:49.418304 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:49Z","lastTransitionTime":"2026-02-02T09:07:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:49 crc kubenswrapper[4764]: I0202 09:07:49.434413 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:49Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:49 crc kubenswrapper[4764]: I0202 09:07:49.470812 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nc7qt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1a2c7b22068b7165927fc158885d48bf9952b9eba521d8ae9a599eee53dc305\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2hz4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nc7qt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:49Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:49 crc kubenswrapper[4764]: I0202 09:07:49.521382 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:49 crc kubenswrapper[4764]: I0202 09:07:49.521418 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:49 crc kubenswrapper[4764]: I0202 09:07:49.521430 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:49 crc kubenswrapper[4764]: I0202 09:07:49.521446 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:49 crc kubenswrapper[4764]: I0202 09:07:49.521458 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:49Z","lastTransitionTime":"2026-02-02T09:07:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:49 crc kubenswrapper[4764]: I0202 09:07:49.521970 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9p5dc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"660ac006-2495-45c4-9fb3-e9c1dddcf7a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da812b124c97aa8ad1563f4bcfac8f3bfa025096fa4a2561e5847b1bf9cb30b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23e57e65a6f5c3059506b9db21e9ac2492c30bd9191a98db0281915e9ce39797\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23e57e65a6f5c3059506b9db21e9ac2492c30bd9191a98db0281915e9ce39797\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2417b66816da1a6cd1791a4b91eca4a3c4f3694462e066e7f0722f1b9c2105c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2417b66816da1a6cd1791a4b91eca4a3c4f3694462e066e7f0722f1b9c2105c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://883641f29e9120c2cea6b272bacf1dc63c752f55ab0408cdaf73db85c85ac369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://883641f29e9120c2cea6b272bacf1dc63c752f55ab0408cdaf73db85c85ac369\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2745baba1188ba5e8f820616e4117a12c50ecee5b68e71ad9cabbb621b6e131f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2745baba1188ba5e8f820616e4117a12c50ecee5b68e71ad9cabbb621b6e131f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8925e435be4d48d86824fc35f8b8819fb35d63e385164c4a20edb3bf3035c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a8925e435be4d48d86824fc35f8b8819fb35d63e385164c4a20edb3bf3035c3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d3fbf668ae7d04c8d9e574f43c2fe90c5604204ba38a3c60847afdfbceb7a57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d3fbf668ae7d04c8d9e574f43c2fe90c5604204ba38a3c60847afdfbceb7a57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9p5dc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:49Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:49 crc kubenswrapper[4764]: I0202 09:07:49.549520 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87d61aea-46f6-4ee6-a7a4-767a6a67fea4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e78b6d08e4d616cefbb73e3fd43b3c6886dcd8aa997c7b4948bf1ff63401293\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d468715c2da33996561c40043216233fd12e21539d0c58dd1f3b892fd9f91a9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b184f7c936dc4046f74c0018269cbc2c4b7ceddb650b4e871d3f8a707e0df88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://269c5074371bd00f8363fb3888c11f1df383081a61497aea6578f08700881d50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b25bc9395831524c53f9e565b4efbe74f21792efafeb1dd507161bc91aa0ecdf\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T09:07:22Z\\\",\\\"message\\\":\\\"W0202 09:07:21.597488 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0202 09:07:21.598117 1 crypto.go:601] Generating new CA for check-endpoints-signer@1770023241 cert, and key in /tmp/serving-cert-3856183724/serving-signer.crt, /tmp/serving-cert-3856183724/serving-signer.key\\\\nI0202 09:07:22.012184 1 observer_polling.go:159] Starting file observer\\\\nW0202 09:07:22.015493 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0202 09:07:22.015797 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 09:07:22.016596 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3856183724/tls.crt::/tmp/serving-cert-3856183724/tls.key\\\\\\\"\\\\nF0202 09:07:22.270605 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b44e17b8ca01f549bc10081dd5f109cb1b9fff364bf5bd4fd90d9f0e3eeb04db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfb06a5ec9262cf39b108a66014d15e325cdbd5441a2d9e7a811dd57265c99a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfb06a5ec9262cf39b108a66014d15e325cdbd5441a2d9e7a811dd57265c99a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:49Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:49 crc kubenswrapper[4764]: I0202 09:07:49.591949 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa9d037e66cb53b9dd698eff46284c9e718fd842cda6f7b525f0c79062d70bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:49Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:49 crc kubenswrapper[4764]: I0202 09:07:49.624745 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:49 crc kubenswrapper[4764]: I0202 09:07:49.624825 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:49 crc kubenswrapper[4764]: I0202 09:07:49.624850 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:49 crc kubenswrapper[4764]: I0202 09:07:49.624882 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:49 crc kubenswrapper[4764]: I0202 09:07:49.624905 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:49Z","lastTransitionTime":"2026-02-02T09:07:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:49 crc kubenswrapper[4764]: I0202 09:07:49.630925 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:49Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:49 crc kubenswrapper[4764]: I0202 09:07:49.678306 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24632dda-6100-4ab6-a28e-214ddae4360c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a6c67e9b7708cc0d85b0b6ffbbade0d34867e159a9c5487ca2606475c667a50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bffda48d7c881c1470251b501661324859373246cc1a21811bd36ca24887c7aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a989645cf485cd8ff80ffff1a86cd7808e1564eba5667535e161cc68db13b26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20b71d3ee7a381baa51b0b0d0470283022cfede30f491135737b5e82d631d763\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c380383b6a014d1b694a13ddc22f54b20b9e4f9d5c9b9000d76187621dce0a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5529f2bff87753428e9173826fc1880acf96a4c134fa9ecad195b15697cad73a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://469f42001223aec6a9629296820e4925288790b1a17d6c7ac106dab6989bfc77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b4cfbd4e682bcdcaff785013bdb72580ce36fced733e235538536827036b93d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://579903eadb183466a2564737e6b678e04308239c0cd75b6808892053b97afbc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://579903eadb183466a2564737e6b678e04308239c0cd75b6808892053b97afbc0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zhn7j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:49Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:49 crc kubenswrapper[4764]: I0202 09:07:49.716877 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d66ab3c0-9b5b-41d5-b6c0-91ccddccd7e9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d92ac486f5dade46bc75de32099859a71fc2ed25a7abbe6ae8937df128e7df30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a4ac273a4f127d79e5fad7d4c30cd2044a346f5734dba5d33d68cd092c0c2d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d38c7cb5dcd22fb07524494ebe77aa1a7f4f808f566043563817e47e7ec12a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://408722c7f83d28919431532f8c797a1c795fd5d96736d14e7f06a05744899702\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f473b18cd8003f75bb1dafb82bcec45745a73ccd10434297356dac185a3d15c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e229f484d19cbdabedf53d90a822a4434db9e00b1301538e885e26acc19d25f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e229f484d19cbdabedf53d90a822a4434db9e00b1301538e885e26acc19d25f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a9cc673be8ff245dba81278e8cd34cfe45ed36ff0b8bd5b30a0ddf333ec6371\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a9cc673be8ff245dba81278e8cd34cfe45ed36ff0b8bd5b30a0ddf333ec6371\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bc7e49c7a5ae66ea939060192fa6a346dc0b0b9aa9750b29879d3c426cc0d276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc7e49c7a5ae66ea939060192fa6a346dc0b0b9aa9750b29879d3c426cc0d276\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:49Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:49 crc kubenswrapper[4764]: I0202 09:07:49.728118 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:49 crc kubenswrapper[4764]: I0202 09:07:49.728159 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:49 crc kubenswrapper[4764]: I0202 09:07:49.728169 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:49 crc kubenswrapper[4764]: I0202 09:07:49.728205 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:49 crc kubenswrapper[4764]: I0202 09:07:49.728214 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:49Z","lastTransitionTime":"2026-02-02T09:07:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:49 crc kubenswrapper[4764]: I0202 09:07:49.749012 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f6904228e052fff69ad1fd66df97758b6b60ba4351e97a7fbd03eee4df1b653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:49Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:49 crc kubenswrapper[4764]: I0202 09:07:49.788317 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-10 03:00:56.569663539 +0000 UTC Feb 02 09:07:49 crc kubenswrapper[4764]: I0202 09:07:49.789752 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6zxff" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aaba93da-12a5-43df-b7c7-4c8b988163c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3db293678a907cd55b31b1d0c0bd7c30e1ba75ecd5e13cbc68ab98fbf0fbd9d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqwb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6zxff\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:49Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:49 crc kubenswrapper[4764]: I0202 09:07:49.831806 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:49 crc kubenswrapper[4764]: I0202 09:07:49.831847 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:49 crc kubenswrapper[4764]: I0202 09:07:49.831860 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:49 crc kubenswrapper[4764]: I0202 09:07:49.831879 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:49 crc kubenswrapper[4764]: I0202 09:07:49.831891 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:49Z","lastTransitionTime":"2026-02-02T09:07:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:49 crc kubenswrapper[4764]: I0202 09:07:49.832983 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d192f670-9f9d-4539-9641-e4bed73acdd4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://040c5e8a5659f85f0d52d1a1e064154d998ae104f711101d2204c0b1caebbb30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbksf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fdf06410349b7c99aa166e87fbe80bf68e2ae923635a6a978b7e2cd50383f7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbksf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4ndm4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:49Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:49 crc kubenswrapper[4764]: I0202 09:07:49.871200 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pcjmn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6e5ccd3b-b657-47a2-9c4d-bc9adb126533\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72863eef568d92a1abeb3cde2e89630a4f5fb7626c6949448ff5195774c98eab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2f7zl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pcjmn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:49Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:49 crc kubenswrapper[4764]: I0202 09:07:49.910003 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed747fd8-e3ce-4319-a6da-bda0fd51b85a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6f064bdc66470aace3180d4988244026b54009a63cff199e41c5d02a1d18479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://60258f84a1d8dcfe48b7e9688ed50f077e201971f20f3f6663fbf1ba58a4611a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f84be2e08217a2ac3edde858d78b4410376ed9bae3bcabe2a7f9b2fbe2f69c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a69d25d352d7ceac5c735bc1fcf446a1fea6dc22c65a5296cfab828cea17377\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:49Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:49 crc kubenswrapper[4764]: I0202 09:07:49.934031 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:49 crc kubenswrapper[4764]: I0202 09:07:49.934068 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:49 crc kubenswrapper[4764]: I0202 09:07:49.934079 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:49 crc kubenswrapper[4764]: I0202 09:07:49.934098 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:49 crc kubenswrapper[4764]: I0202 09:07:49.934111 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:49Z","lastTransitionTime":"2026-02-02T09:07:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:49 crc kubenswrapper[4764]: I0202 09:07:49.948703 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:49Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:50 crc kubenswrapper[4764]: I0202 09:07:50.036734 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:50 crc kubenswrapper[4764]: I0202 09:07:50.036795 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:50 crc kubenswrapper[4764]: I0202 09:07:50.036816 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:50 crc kubenswrapper[4764]: I0202 09:07:50.036842 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:50 crc kubenswrapper[4764]: I0202 09:07:50.036860 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:50Z","lastTransitionTime":"2026-02-02T09:07:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:50 crc kubenswrapper[4764]: I0202 09:07:50.049682 4764 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 02 09:07:50 crc kubenswrapper[4764]: I0202 09:07:50.138849 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:50 crc kubenswrapper[4764]: I0202 09:07:50.138879 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:50 crc kubenswrapper[4764]: I0202 09:07:50.138888 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:50 crc kubenswrapper[4764]: I0202 09:07:50.138903 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:50 crc kubenswrapper[4764]: I0202 09:07:50.138913 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:50Z","lastTransitionTime":"2026-02-02T09:07:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:50 crc kubenswrapper[4764]: I0202 09:07:50.241814 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:50 crc kubenswrapper[4764]: I0202 09:07:50.241861 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:50 crc kubenswrapper[4764]: I0202 09:07:50.241878 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:50 crc kubenswrapper[4764]: I0202 09:07:50.241908 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:50 crc kubenswrapper[4764]: I0202 09:07:50.241923 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:50Z","lastTransitionTime":"2026-02-02T09:07:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:50 crc kubenswrapper[4764]: I0202 09:07:50.344826 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:50 crc kubenswrapper[4764]: I0202 09:07:50.344856 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:50 crc kubenswrapper[4764]: I0202 09:07:50.344864 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:50 crc kubenswrapper[4764]: I0202 09:07:50.344876 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:50 crc kubenswrapper[4764]: I0202 09:07:50.344884 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:50Z","lastTransitionTime":"2026-02-02T09:07:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:50 crc kubenswrapper[4764]: I0202 09:07:50.446985 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:50 crc kubenswrapper[4764]: I0202 09:07:50.447009 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:50 crc kubenswrapper[4764]: I0202 09:07:50.447017 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:50 crc kubenswrapper[4764]: I0202 09:07:50.447029 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:50 crc kubenswrapper[4764]: I0202 09:07:50.447037 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:50Z","lastTransitionTime":"2026-02-02T09:07:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:50 crc kubenswrapper[4764]: I0202 09:07:50.550301 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:50 crc kubenswrapper[4764]: I0202 09:07:50.550368 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:50 crc kubenswrapper[4764]: I0202 09:07:50.550393 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:50 crc kubenswrapper[4764]: I0202 09:07:50.550419 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:50 crc kubenswrapper[4764]: I0202 09:07:50.550438 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:50Z","lastTransitionTime":"2026-02-02T09:07:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:50 crc kubenswrapper[4764]: I0202 09:07:50.653354 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:50 crc kubenswrapper[4764]: I0202 09:07:50.653410 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:50 crc kubenswrapper[4764]: I0202 09:07:50.653422 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:50 crc kubenswrapper[4764]: I0202 09:07:50.653440 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:50 crc kubenswrapper[4764]: I0202 09:07:50.653452 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:50Z","lastTransitionTime":"2026-02-02T09:07:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:50 crc kubenswrapper[4764]: I0202 09:07:50.756280 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:50 crc kubenswrapper[4764]: I0202 09:07:50.756327 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:50 crc kubenswrapper[4764]: I0202 09:07:50.756336 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:50 crc kubenswrapper[4764]: I0202 09:07:50.756356 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:50 crc kubenswrapper[4764]: I0202 09:07:50.756368 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:50Z","lastTransitionTime":"2026-02-02T09:07:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:50 crc kubenswrapper[4764]: I0202 09:07:50.789912 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-06 12:18:50.130854625 +0000 UTC Feb 02 09:07:50 crc kubenswrapper[4764]: I0202 09:07:50.825192 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 09:07:50 crc kubenswrapper[4764]: I0202 09:07:50.825214 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 09:07:50 crc kubenswrapper[4764]: I0202 09:07:50.825270 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 09:07:50 crc kubenswrapper[4764]: E0202 09:07:50.825330 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 09:07:50 crc kubenswrapper[4764]: E0202 09:07:50.825486 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 09:07:50 crc kubenswrapper[4764]: E0202 09:07:50.825582 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 09:07:50 crc kubenswrapper[4764]: I0202 09:07:50.858104 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:50 crc kubenswrapper[4764]: I0202 09:07:50.858167 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:50 crc kubenswrapper[4764]: I0202 09:07:50.858182 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:50 crc kubenswrapper[4764]: I0202 09:07:50.858197 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:50 crc kubenswrapper[4764]: I0202 09:07:50.858207 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:50Z","lastTransitionTime":"2026-02-02T09:07:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:50 crc kubenswrapper[4764]: I0202 09:07:50.960035 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:50 crc kubenswrapper[4764]: I0202 09:07:50.960202 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:50 crc kubenswrapper[4764]: I0202 09:07:50.960291 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:50 crc kubenswrapper[4764]: I0202 09:07:50.960378 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:50 crc kubenswrapper[4764]: I0202 09:07:50.960486 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:50Z","lastTransitionTime":"2026-02-02T09:07:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:51 crc kubenswrapper[4764]: I0202 09:07:51.054727 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zhn7j_24632dda-6100-4ab6-a28e-214ddae4360c/ovnkube-controller/0.log" Feb 02 09:07:51 crc kubenswrapper[4764]: I0202 09:07:51.057402 4764 generic.go:334] "Generic (PLEG): container finished" podID="24632dda-6100-4ab6-a28e-214ddae4360c" containerID="469f42001223aec6a9629296820e4925288790b1a17d6c7ac106dab6989bfc77" exitCode=1 Feb 02 09:07:51 crc kubenswrapper[4764]: I0202 09:07:51.057569 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" event={"ID":"24632dda-6100-4ab6-a28e-214ddae4360c","Type":"ContainerDied","Data":"469f42001223aec6a9629296820e4925288790b1a17d6c7ac106dab6989bfc77"} Feb 02 09:07:51 crc kubenswrapper[4764]: I0202 09:07:51.058020 4764 scope.go:117] "RemoveContainer" containerID="469f42001223aec6a9629296820e4925288790b1a17d6c7ac106dab6989bfc77" Feb 02 09:07:51 crc kubenswrapper[4764]: I0202 09:07:51.062737 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:51 crc kubenswrapper[4764]: I0202 09:07:51.062766 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:51 crc kubenswrapper[4764]: I0202 09:07:51.062775 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:51 crc kubenswrapper[4764]: I0202 09:07:51.062788 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:51 crc kubenswrapper[4764]: I0202 09:07:51.062797 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:51Z","lastTransitionTime":"2026-02-02T09:07:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:51 crc kubenswrapper[4764]: I0202 09:07:51.076709 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d66ab3c0-9b5b-41d5-b6c0-91ccddccd7e9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d92ac486f5dade46bc75de32099859a71fc2ed25a7abbe6ae8937df128e7df30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a4ac273a4f127d79e5fad7d4c30cd2044a346f5734dba5d33d68cd092c0c2d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d38c7cb5dcd22fb07524494ebe77aa1a7f4f808f566043563817e47e7ec12a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://408722c7f83d28919431532f8c797a1c795fd5d96736d14e7f06a05744899702\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f473b18cd8003f75bb1dafb82bcec45745a73ccd10434297356dac185a3d15c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e229f484d19cbdabedf53d90a822a4434db9e00b1301538e885e26acc19d25f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e229f484d19cbdabedf53d90a822a4434db9e00b1301538e885e26acc19d25f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a9cc673be8ff245dba81278e8cd34cfe45ed36ff0b8bd5b30a0ddf333ec6371\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a9cc673be8ff245dba81278e8cd34cfe45ed36ff0b8bd5b30a0ddf333ec6371\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bc7e49c7a5ae66ea939060192fa6a346dc0b0b9aa9750b29879d3c426cc0d276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc7e49c7a5ae66ea939060192fa6a346dc0b0b9aa9750b29879d3c426cc0d276\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:51Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:51 crc kubenswrapper[4764]: I0202 09:07:51.094450 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa9d037e66cb53b9dd698eff46284c9e718fd842cda6f7b525f0c79062d70bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:51Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:51 crc kubenswrapper[4764]: I0202 09:07:51.108436 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:51Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:51 crc kubenswrapper[4764]: I0202 09:07:51.125804 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24632dda-6100-4ab6-a28e-214ddae4360c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a6c67e9b7708cc0d85b0b6ffbbade0d34867e159a9c5487ca2606475c667a50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bffda48d7c881c1470251b501661324859373246cc1a21811bd36ca24887c7aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a989645cf485cd8ff80ffff1a86cd7808e1564eba5667535e161cc68db13b26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20b71d3ee7a381baa51b0b0d0470283022cfede30f491135737b5e82d631d763\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c380383b6a014d1b694a13ddc22f54b20b9e4f9d5c9b9000d76187621dce0a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5529f2bff87753428e9173826fc1880acf96a4c134fa9ecad195b15697cad73a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://469f42001223aec6a9629296820e4925288790b1a17d6c7ac106dab6989bfc77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://469f42001223aec6a9629296820e4925288790b1a17d6c7ac106dab6989bfc77\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-02T09:07:50Z\\\",\\\"message\\\":\\\"5993 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0202 09:07:50.920631 5993 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0202 09:07:50.920653 5993 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0202 09:07:50.920876 5993 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0202 09:07:50.920634 5993 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0202 09:07:50.921026 5993 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0202 09:07:50.921057 5993 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0202 09:07:50.921081 5993 factory.go:656] Stopping watch factory\\\\nI0202 09:07:50.921098 5993 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0202 09:07:50.921161 5993 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0202 09:07:50.921186 5993 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0202 09:07:50.921195 5993 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0202 09:07:50.921259 5993 handler.go:208] Removed *v1.Node event handler 2\\\\nI0202 09:07:50.921280 5993 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b4cfbd4e682bcdcaff785013bdb72580ce36fced733e235538536827036b93d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://579903eadb183466a2564737e6b678e04308239c0cd75b6808892053b97afbc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://579903eadb183466a2564737e6b678e04308239c0cd75b6808892053b97afbc0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zhn7j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:51Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:51 crc kubenswrapper[4764]: I0202 09:07:51.137691 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed747fd8-e3ce-4319-a6da-bda0fd51b85a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6f064bdc66470aace3180d4988244026b54009a63cff199e41c5d02a1d18479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://60258f84a1d8dcfe48b7e9688ed50f077e201971f20f3f6663fbf1ba58a4611a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f84be2e08217a2ac3edde858d78b4410376ed9bae3bcabe2a7f9b2fbe2f69c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a69d25d352d7ceac5c735bc1fcf446a1fea6dc22c65a5296cfab828cea17377\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:51Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:51 crc kubenswrapper[4764]: I0202 09:07:51.150982 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f6904228e052fff69ad1fd66df97758b6b60ba4351e97a7fbd03eee4df1b653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:51Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:51 crc kubenswrapper[4764]: I0202 09:07:51.160643 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6zxff" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aaba93da-12a5-43df-b7c7-4c8b988163c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3db293678a907cd55b31b1d0c0bd7c30e1ba75ecd5e13cbc68ab98fbf0fbd9d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqwb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6zxff\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:51Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:51 crc kubenswrapper[4764]: I0202 09:07:51.165511 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:51 crc kubenswrapper[4764]: I0202 09:07:51.165551 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:51 crc kubenswrapper[4764]: I0202 09:07:51.165562 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:51 crc kubenswrapper[4764]: I0202 09:07:51.165578 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:51 crc kubenswrapper[4764]: I0202 09:07:51.165590 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:51Z","lastTransitionTime":"2026-02-02T09:07:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:51 crc kubenswrapper[4764]: I0202 09:07:51.175997 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d192f670-9f9d-4539-9641-e4bed73acdd4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://040c5e8a5659f85f0d52d1a1e064154d998ae104f711101d2204c0b1caebbb30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbksf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fdf06410349b7c99aa166e87fbe80bf68e2ae923635a6a978b7e2cd50383f7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbksf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4ndm4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:51Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:51 crc kubenswrapper[4764]: I0202 09:07:51.191149 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pcjmn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6e5ccd3b-b657-47a2-9c4d-bc9adb126533\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72863eef568d92a1abeb3cde2e89630a4f5fb7626c6949448ff5195774c98eab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2f7zl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pcjmn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:51Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:51 crc kubenswrapper[4764]: I0202 09:07:51.204249 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:51Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:51 crc kubenswrapper[4764]: I0202 09:07:51.222348 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87d61aea-46f6-4ee6-a7a4-767a6a67fea4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e78b6d08e4d616cefbb73e3fd43b3c6886dcd8aa997c7b4948bf1ff63401293\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d468715c2da33996561c40043216233fd12e21539d0c58dd1f3b892fd9f91a9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b184f7c936dc4046f74c0018269cbc2c4b7ceddb650b4e871d3f8a707e0df88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://269c5074371bd00f8363fb3888c11f1df383081a61497aea6578f08700881d50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b25bc9395831524c53f9e565b4efbe74f21792efafeb1dd507161bc91aa0ecdf\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T09:07:22Z\\\",\\\"message\\\":\\\"W0202 09:07:21.597488 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0202 09:07:21.598117 1 crypto.go:601] Generating new CA for check-endpoints-signer@1770023241 cert, and key in /tmp/serving-cert-3856183724/serving-signer.crt, /tmp/serving-cert-3856183724/serving-signer.key\\\\nI0202 09:07:22.012184 1 observer_polling.go:159] Starting file observer\\\\nW0202 09:07:22.015493 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0202 09:07:22.015797 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 09:07:22.016596 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3856183724/tls.crt::/tmp/serving-cert-3856183724/tls.key\\\\\\\"\\\\nF0202 09:07:22.270605 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b44e17b8ca01f549bc10081dd5f109cb1b9fff364bf5bd4fd90d9f0e3eeb04db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfb06a5ec9262cf39b108a66014d15e325cdbd5441a2d9e7a811dd57265c99a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfb06a5ec9262cf39b108a66014d15e325cdbd5441a2d9e7a811dd57265c99a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:51Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:51 crc kubenswrapper[4764]: I0202 09:07:51.238805 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5528225d91ac2778bef4554abb0bd036ac5d3b6cdac7357bab69e70c3894b6f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://071bbb80d0747cf3e109ba626d47f7d05f85f4cc0b3aba77276c795741f47ac6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:51Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:51 crc kubenswrapper[4764]: I0202 09:07:51.256453 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:51Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:51 crc kubenswrapper[4764]: I0202 09:07:51.268167 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:51 crc kubenswrapper[4764]: I0202 09:07:51.268193 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:51 crc kubenswrapper[4764]: I0202 09:07:51.268201 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:51 crc kubenswrapper[4764]: I0202 09:07:51.268212 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:51 crc kubenswrapper[4764]: I0202 09:07:51.268222 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:51Z","lastTransitionTime":"2026-02-02T09:07:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:51 crc kubenswrapper[4764]: I0202 09:07:51.274966 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nc7qt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1a2c7b22068b7165927fc158885d48bf9952b9eba521d8ae9a599eee53dc305\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2hz4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nc7qt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:51Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:51 crc kubenswrapper[4764]: I0202 09:07:51.290363 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9p5dc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"660ac006-2495-45c4-9fb3-e9c1dddcf7a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da812b124c97aa8ad1563f4bcfac8f3bfa025096fa4a2561e5847b1bf9cb30b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23e57e65a6f5c3059506b9db21e9ac2492c30bd9191a98db0281915e9ce39797\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23e57e65a6f5c3059506b9db21e9ac2492c30bd9191a98db0281915e9ce39797\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2417b66816da1a6cd1791a4b91eca4a3c4f3694462e066e7f0722f1b9c2105c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2417b66816da1a6cd1791a4b91eca4a3c4f3694462e066e7f0722f1b9c2105c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://883641f29e9120c2cea6b272bacf1dc63c752f55ab0408cdaf73db85c85ac369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://883641f29e9120c2cea6b272bacf1dc63c752f55ab0408cdaf73db85c85ac369\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2745baba1188ba5e8f820616e4117a12c50ecee5b68e71ad9cabbb621b6e131f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2745baba1188ba5e8f820616e4117a12c50ecee5b68e71ad9cabbb621b6e131f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8925e435be4d48d86824fc35f8b8819fb35d63e385164c4a20edb3bf3035c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a8925e435be4d48d86824fc35f8b8819fb35d63e385164c4a20edb3bf3035c3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d3fbf668ae7d04c8d9e574f43c2fe90c5604204ba38a3c60847afdfbceb7a57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d3fbf668ae7d04c8d9e574f43c2fe90c5604204ba38a3c60847afdfbceb7a57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9p5dc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:51Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:51 crc kubenswrapper[4764]: I0202 09:07:51.370499 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:51 crc kubenswrapper[4764]: I0202 09:07:51.370781 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:51 crc kubenswrapper[4764]: I0202 09:07:51.370908 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:51 crc kubenswrapper[4764]: I0202 09:07:51.371029 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:51 crc kubenswrapper[4764]: I0202 09:07:51.371125 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:51Z","lastTransitionTime":"2026-02-02T09:07:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:51 crc kubenswrapper[4764]: I0202 09:07:51.474958 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:51 crc kubenswrapper[4764]: I0202 09:07:51.475007 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:51 crc kubenswrapper[4764]: I0202 09:07:51.475024 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:51 crc kubenswrapper[4764]: I0202 09:07:51.475078 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:51 crc kubenswrapper[4764]: I0202 09:07:51.475099 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:51Z","lastTransitionTime":"2026-02-02T09:07:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:51 crc kubenswrapper[4764]: I0202 09:07:51.577990 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:51 crc kubenswrapper[4764]: I0202 09:07:51.578029 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:51 crc kubenswrapper[4764]: I0202 09:07:51.578040 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:51 crc kubenswrapper[4764]: I0202 09:07:51.578055 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:51 crc kubenswrapper[4764]: I0202 09:07:51.578065 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:51Z","lastTransitionTime":"2026-02-02T09:07:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:51 crc kubenswrapper[4764]: I0202 09:07:51.679922 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:51 crc kubenswrapper[4764]: I0202 09:07:51.679981 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:51 crc kubenswrapper[4764]: I0202 09:07:51.679992 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:51 crc kubenswrapper[4764]: I0202 09:07:51.680015 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:51 crc kubenswrapper[4764]: I0202 09:07:51.680029 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:51Z","lastTransitionTime":"2026-02-02T09:07:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:51 crc kubenswrapper[4764]: I0202 09:07:51.783156 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:51 crc kubenswrapper[4764]: I0202 09:07:51.783212 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:51 crc kubenswrapper[4764]: I0202 09:07:51.783228 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:51 crc kubenswrapper[4764]: I0202 09:07:51.783251 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:51 crc kubenswrapper[4764]: I0202 09:07:51.783269 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:51Z","lastTransitionTime":"2026-02-02T09:07:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:51 crc kubenswrapper[4764]: I0202 09:07:51.790273 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-07 09:13:08.284918068 +0000 UTC Feb 02 09:07:51 crc kubenswrapper[4764]: I0202 09:07:51.885815 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:51 crc kubenswrapper[4764]: I0202 09:07:51.885857 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:51 crc kubenswrapper[4764]: I0202 09:07:51.885865 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:51 crc kubenswrapper[4764]: I0202 09:07:51.885879 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:51 crc kubenswrapper[4764]: I0202 09:07:51.885887 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:51Z","lastTransitionTime":"2026-02-02T09:07:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:51 crc kubenswrapper[4764]: I0202 09:07:51.988824 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:51 crc kubenswrapper[4764]: I0202 09:07:51.988859 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:51 crc kubenswrapper[4764]: I0202 09:07:51.988870 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:51 crc kubenswrapper[4764]: I0202 09:07:51.988885 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:51 crc kubenswrapper[4764]: I0202 09:07:51.988898 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:51Z","lastTransitionTime":"2026-02-02T09:07:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:52 crc kubenswrapper[4764]: I0202 09:07:52.064624 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zhn7j_24632dda-6100-4ab6-a28e-214ddae4360c/ovnkube-controller/0.log" Feb 02 09:07:52 crc kubenswrapper[4764]: I0202 09:07:52.078822 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" event={"ID":"24632dda-6100-4ab6-a28e-214ddae4360c","Type":"ContainerStarted","Data":"12244061eabaae7fa9057ac5cee894c3b308b466411bcb910682e7f65d240dd7"} Feb 02 09:07:52 crc kubenswrapper[4764]: I0202 09:07:52.078971 4764 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 02 09:07:52 crc kubenswrapper[4764]: I0202 09:07:52.093279 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:52 crc kubenswrapper[4764]: I0202 09:07:52.093343 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:52 crc kubenswrapper[4764]: I0202 09:07:52.093361 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:52 crc kubenswrapper[4764]: I0202 09:07:52.093387 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:52 crc kubenswrapper[4764]: I0202 09:07:52.093408 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:52Z","lastTransitionTime":"2026-02-02T09:07:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:52 crc kubenswrapper[4764]: I0202 09:07:52.101777 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d66ab3c0-9b5b-41d5-b6c0-91ccddccd7e9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d92ac486f5dade46bc75de32099859a71fc2ed25a7abbe6ae8937df128e7df30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a4ac273a4f127d79e5fad7d4c30cd2044a346f5734dba5d33d68cd092c0c2d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d38c7cb5dcd22fb07524494ebe77aa1a7f4f808f566043563817e47e7ec12a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://408722c7f83d28919431532f8c797a1c795fd5d96736d14e7f06a05744899702\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f473b18cd8003f75bb1dafb82bcec45745a73ccd10434297356dac185a3d15c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e229f484d19cbdabedf53d90a822a4434db9e00b1301538e885e26acc19d25f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e229f484d19cbdabedf53d90a822a4434db9e00b1301538e885e26acc19d25f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a9cc673be8ff245dba81278e8cd34cfe45ed36ff0b8bd5b30a0ddf333ec6371\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a9cc673be8ff245dba81278e8cd34cfe45ed36ff0b8bd5b30a0ddf333ec6371\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bc7e49c7a5ae66ea939060192fa6a346dc0b0b9aa9750b29879d3c426cc0d276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc7e49c7a5ae66ea939060192fa6a346dc0b0b9aa9750b29879d3c426cc0d276\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:52Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:52 crc kubenswrapper[4764]: I0202 09:07:52.114776 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa9d037e66cb53b9dd698eff46284c9e718fd842cda6f7b525f0c79062d70bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:52Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:52 crc kubenswrapper[4764]: I0202 09:07:52.132518 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:52Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:52 crc kubenswrapper[4764]: I0202 09:07:52.163663 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24632dda-6100-4ab6-a28e-214ddae4360c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a6c67e9b7708cc0d85b0b6ffbbade0d34867e159a9c5487ca2606475c667a50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bffda48d7c881c1470251b501661324859373246cc1a21811bd36ca24887c7aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a989645cf485cd8ff80ffff1a86cd7808e1564eba5667535e161cc68db13b26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20b71d3ee7a381baa51b0b0d0470283022cfede30f491135737b5e82d631d763\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c380383b6a014d1b694a13ddc22f54b20b9e4f9d5c9b9000d76187621dce0a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5529f2bff87753428e9173826fc1880acf96a4c134fa9ecad195b15697cad73a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12244061eabaae7fa9057ac5cee894c3b308b466411bcb910682e7f65d240dd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://469f42001223aec6a9629296820e4925288790b1a17d6c7ac106dab6989bfc77\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-02T09:07:50Z\\\",\\\"message\\\":\\\"5993 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0202 09:07:50.920631 5993 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0202 09:07:50.920653 5993 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0202 09:07:50.920876 5993 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0202 09:07:50.920634 5993 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0202 09:07:50.921026 5993 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0202 09:07:50.921057 5993 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0202 09:07:50.921081 5993 factory.go:656] Stopping watch factory\\\\nI0202 09:07:50.921098 5993 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0202 09:07:50.921161 5993 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0202 09:07:50.921186 5993 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0202 09:07:50.921195 5993 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0202 09:07:50.921259 5993 handler.go:208] Removed *v1.Node event handler 2\\\\nI0202 09:07:50.921280 5993 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:47Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b4cfbd4e682bcdcaff785013bdb72580ce36fced733e235538536827036b93d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://579903eadb183466a2564737e6b678e04308239c0cd75b6808892053b97afbc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://579903eadb183466a2564737e6b678e04308239c0cd75b6808892053b97afbc0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zhn7j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:52Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:52 crc kubenswrapper[4764]: I0202 09:07:52.181635 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed747fd8-e3ce-4319-a6da-bda0fd51b85a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6f064bdc66470aace3180d4988244026b54009a63cff199e41c5d02a1d18479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://60258f84a1d8dcfe48b7e9688ed50f077e201971f20f3f6663fbf1ba58a4611a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f84be2e08217a2ac3edde858d78b4410376ed9bae3bcabe2a7f9b2fbe2f69c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a69d25d352d7ceac5c735bc1fcf446a1fea6dc22c65a5296cfab828cea17377\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:52Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:52 crc kubenswrapper[4764]: I0202 09:07:52.196773 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:52 crc kubenswrapper[4764]: I0202 09:07:52.196825 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:52 crc kubenswrapper[4764]: I0202 09:07:52.196835 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:52 crc kubenswrapper[4764]: I0202 09:07:52.196849 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:52 crc kubenswrapper[4764]: I0202 09:07:52.196858 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:52Z","lastTransitionTime":"2026-02-02T09:07:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:52 crc kubenswrapper[4764]: I0202 09:07:52.202026 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f6904228e052fff69ad1fd66df97758b6b60ba4351e97a7fbd03eee4df1b653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:52Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:52 crc kubenswrapper[4764]: I0202 09:07:52.227049 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6zxff" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aaba93da-12a5-43df-b7c7-4c8b988163c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3db293678a907cd55b31b1d0c0bd7c30e1ba75ecd5e13cbc68ab98fbf0fbd9d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqwb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6zxff\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:52Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:52 crc kubenswrapper[4764]: I0202 09:07:52.240195 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d192f670-9f9d-4539-9641-e4bed73acdd4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://040c5e8a5659f85f0d52d1a1e064154d998ae104f711101d2204c0b1caebbb30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbksf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fdf06410349b7c99aa166e87fbe80bf68e2ae923635a6a978b7e2cd50383f7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbksf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4ndm4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:52Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:52 crc kubenswrapper[4764]: I0202 09:07:52.253910 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pcjmn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6e5ccd3b-b657-47a2-9c4d-bc9adb126533\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72863eef568d92a1abeb3cde2e89630a4f5fb7626c6949448ff5195774c98eab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2f7zl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pcjmn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:52Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:52 crc kubenswrapper[4764]: I0202 09:07:52.277607 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:52Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:52 crc kubenswrapper[4764]: I0202 09:07:52.294074 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87d61aea-46f6-4ee6-a7a4-767a6a67fea4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e78b6d08e4d616cefbb73e3fd43b3c6886dcd8aa997c7b4948bf1ff63401293\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d468715c2da33996561c40043216233fd12e21539d0c58dd1f3b892fd9f91a9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b184f7c936dc4046f74c0018269cbc2c4b7ceddb650b4e871d3f8a707e0df88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://269c5074371bd00f8363fb3888c11f1df383081a61497aea6578f08700881d50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b25bc9395831524c53f9e565b4efbe74f21792efafeb1dd507161bc91aa0ecdf\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T09:07:22Z\\\",\\\"message\\\":\\\"W0202 09:07:21.597488 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0202 09:07:21.598117 1 crypto.go:601] Generating new CA for check-endpoints-signer@1770023241 cert, and key in /tmp/serving-cert-3856183724/serving-signer.crt, /tmp/serving-cert-3856183724/serving-signer.key\\\\nI0202 09:07:22.012184 1 observer_polling.go:159] Starting file observer\\\\nW0202 09:07:22.015493 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0202 09:07:22.015797 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 09:07:22.016596 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3856183724/tls.crt::/tmp/serving-cert-3856183724/tls.key\\\\\\\"\\\\nF0202 09:07:22.270605 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b44e17b8ca01f549bc10081dd5f109cb1b9fff364bf5bd4fd90d9f0e3eeb04db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfb06a5ec9262cf39b108a66014d15e325cdbd5441a2d9e7a811dd57265c99a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfb06a5ec9262cf39b108a66014d15e325cdbd5441a2d9e7a811dd57265c99a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:52Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:52 crc kubenswrapper[4764]: I0202 09:07:52.298924 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:52 crc kubenswrapper[4764]: I0202 09:07:52.298977 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:52 crc kubenswrapper[4764]: I0202 09:07:52.298992 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:52 crc kubenswrapper[4764]: I0202 09:07:52.299032 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:52 crc kubenswrapper[4764]: I0202 09:07:52.299044 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:52Z","lastTransitionTime":"2026-02-02T09:07:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:52 crc kubenswrapper[4764]: I0202 09:07:52.307337 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5528225d91ac2778bef4554abb0bd036ac5d3b6cdac7357bab69e70c3894b6f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://071bbb80d0747cf3e109ba626d47f7d05f85f4cc0b3aba77276c795741f47ac6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:52Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:52 crc kubenswrapper[4764]: I0202 09:07:52.319169 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:52Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:52 crc kubenswrapper[4764]: I0202 09:07:52.330532 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nc7qt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1a2c7b22068b7165927fc158885d48bf9952b9eba521d8ae9a599eee53dc305\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2hz4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nc7qt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:52Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:52 crc kubenswrapper[4764]: I0202 09:07:52.343463 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9p5dc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"660ac006-2495-45c4-9fb3-e9c1dddcf7a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da812b124c97aa8ad1563f4bcfac8f3bfa025096fa4a2561e5847b1bf9cb30b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23e57e65a6f5c3059506b9db21e9ac2492c30bd9191a98db0281915e9ce39797\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23e57e65a6f5c3059506b9db21e9ac2492c30bd9191a98db0281915e9ce39797\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2417b66816da1a6cd1791a4b91eca4a3c4f3694462e066e7f0722f1b9c2105c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2417b66816da1a6cd1791a4b91eca4a3c4f3694462e066e7f0722f1b9c2105c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://883641f29e9120c2cea6b272bacf1dc63c752f55ab0408cdaf73db85c85ac369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://883641f29e9120c2cea6b272bacf1dc63c752f55ab0408cdaf73db85c85ac369\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2745baba1188ba5e8f820616e4117a12c50ecee5b68e71ad9cabbb621b6e131f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2745baba1188ba5e8f820616e4117a12c50ecee5b68e71ad9cabbb621b6e131f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8925e435be4d48d86824fc35f8b8819fb35d63e385164c4a20edb3bf3035c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a8925e435be4d48d86824fc35f8b8819fb35d63e385164c4a20edb3bf3035c3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d3fbf668ae7d04c8d9e574f43c2fe90c5604204ba38a3c60847afdfbceb7a57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d3fbf668ae7d04c8d9e574f43c2fe90c5604204ba38a3c60847afdfbceb7a57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9p5dc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:52Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:52 crc kubenswrapper[4764]: I0202 09:07:52.401355 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:52 crc kubenswrapper[4764]: I0202 09:07:52.401409 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:52 crc kubenswrapper[4764]: I0202 09:07:52.401419 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:52 crc kubenswrapper[4764]: I0202 09:07:52.401432 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:52 crc kubenswrapper[4764]: I0202 09:07:52.401441 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:52Z","lastTransitionTime":"2026-02-02T09:07:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:52 crc kubenswrapper[4764]: I0202 09:07:52.494982 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lh6tm"] Feb 02 09:07:52 crc kubenswrapper[4764]: I0202 09:07:52.495391 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lh6tm" Feb 02 09:07:52 crc kubenswrapper[4764]: I0202 09:07:52.499025 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Feb 02 09:07:52 crc kubenswrapper[4764]: I0202 09:07:52.499505 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Feb 02 09:07:52 crc kubenswrapper[4764]: I0202 09:07:52.503560 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:52 crc kubenswrapper[4764]: I0202 09:07:52.503590 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:52 crc kubenswrapper[4764]: I0202 09:07:52.503599 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:52 crc kubenswrapper[4764]: I0202 09:07:52.503613 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:52 crc kubenswrapper[4764]: I0202 09:07:52.503622 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:52Z","lastTransitionTime":"2026-02-02T09:07:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:52 crc kubenswrapper[4764]: I0202 09:07:52.516900 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87d61aea-46f6-4ee6-a7a4-767a6a67fea4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e78b6d08e4d616cefbb73e3fd43b3c6886dcd8aa997c7b4948bf1ff63401293\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d468715c2da33996561c40043216233fd12e21539d0c58dd1f3b892fd9f91a9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b184f7c936dc4046f74c0018269cbc2c4b7ceddb650b4e871d3f8a707e0df88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://269c5074371bd00f8363fb3888c11f1df383081a61497aea6578f08700881d50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b25bc9395831524c53f9e565b4efbe74f21792efafeb1dd507161bc91aa0ecdf\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T09:07:22Z\\\",\\\"message\\\":\\\"W0202 09:07:21.597488 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0202 09:07:21.598117 1 crypto.go:601] Generating new CA for check-endpoints-signer@1770023241 cert, and key in /tmp/serving-cert-3856183724/serving-signer.crt, /tmp/serving-cert-3856183724/serving-signer.key\\\\nI0202 09:07:22.012184 1 observer_polling.go:159] Starting file observer\\\\nW0202 09:07:22.015493 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0202 09:07:22.015797 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 09:07:22.016596 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3856183724/tls.crt::/tmp/serving-cert-3856183724/tls.key\\\\\\\"\\\\nF0202 09:07:22.270605 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b44e17b8ca01f549bc10081dd5f109cb1b9fff364bf5bd4fd90d9f0e3eeb04db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfb06a5ec9262cf39b108a66014d15e325cdbd5441a2d9e7a811dd57265c99a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfb06a5ec9262cf39b108a66014d15e325cdbd5441a2d9e7a811dd57265c99a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:52Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:52 crc kubenswrapper[4764]: I0202 09:07:52.533437 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bb2sk\" (UniqueName: \"kubernetes.io/projected/a2152e8a-f456-47b5-a547-b65ec837ecbc-kube-api-access-bb2sk\") pod \"ovnkube-control-plane-749d76644c-lh6tm\" (UID: \"a2152e8a-f456-47b5-a547-b65ec837ecbc\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lh6tm" Feb 02 09:07:52 crc kubenswrapper[4764]: I0202 09:07:52.533569 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/a2152e8a-f456-47b5-a547-b65ec837ecbc-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-lh6tm\" (UID: \"a2152e8a-f456-47b5-a547-b65ec837ecbc\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lh6tm" Feb 02 09:07:52 crc kubenswrapper[4764]: I0202 09:07:52.533618 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/a2152e8a-f456-47b5-a547-b65ec837ecbc-env-overrides\") pod \"ovnkube-control-plane-749d76644c-lh6tm\" (UID: \"a2152e8a-f456-47b5-a547-b65ec837ecbc\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lh6tm" Feb 02 09:07:52 crc kubenswrapper[4764]: I0202 09:07:52.533683 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/a2152e8a-f456-47b5-a547-b65ec837ecbc-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-lh6tm\" (UID: \"a2152e8a-f456-47b5-a547-b65ec837ecbc\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lh6tm" Feb 02 09:07:52 crc kubenswrapper[4764]: I0202 09:07:52.534106 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5528225d91ac2778bef4554abb0bd036ac5d3b6cdac7357bab69e70c3894b6f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://071bbb80d0747cf3e109ba626d47f7d05f85f4cc0b3aba77276c795741f47ac6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:52Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:52 crc kubenswrapper[4764]: I0202 09:07:52.548093 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:52Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:52 crc kubenswrapper[4764]: I0202 09:07:52.565548 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nc7qt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1a2c7b22068b7165927fc158885d48bf9952b9eba521d8ae9a599eee53dc305\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2hz4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nc7qt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:52Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:52 crc kubenswrapper[4764]: I0202 09:07:52.590076 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9p5dc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"660ac006-2495-45c4-9fb3-e9c1dddcf7a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da812b124c97aa8ad1563f4bcfac8f3bfa025096fa4a2561e5847b1bf9cb30b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23e57e65a6f5c3059506b9db21e9ac2492c30bd9191a98db0281915e9ce39797\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23e57e65a6f5c3059506b9db21e9ac2492c30bd9191a98db0281915e9ce39797\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2417b66816da1a6cd1791a4b91eca4a3c4f3694462e066e7f0722f1b9c2105c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2417b66816da1a6cd1791a4b91eca4a3c4f3694462e066e7f0722f1b9c2105c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://883641f29e9120c2cea6b272bacf1dc63c752f55ab0408cdaf73db85c85ac369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://883641f29e9120c2cea6b272bacf1dc63c752f55ab0408cdaf73db85c85ac369\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2745baba1188ba5e8f820616e4117a12c50ecee5b68e71ad9cabbb621b6e131f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2745baba1188ba5e8f820616e4117a12c50ecee5b68e71ad9cabbb621b6e131f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8925e435be4d48d86824fc35f8b8819fb35d63e385164c4a20edb3bf3035c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a8925e435be4d48d86824fc35f8b8819fb35d63e385164c4a20edb3bf3035c3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d3fbf668ae7d04c8d9e574f43c2fe90c5604204ba38a3c60847afdfbceb7a57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d3fbf668ae7d04c8d9e574f43c2fe90c5604204ba38a3c60847afdfbceb7a57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9p5dc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:52Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:52 crc kubenswrapper[4764]: I0202 09:07:52.606482 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:52 crc kubenswrapper[4764]: I0202 09:07:52.606525 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:52 crc kubenswrapper[4764]: I0202 09:07:52.606535 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:52 crc kubenswrapper[4764]: I0202 09:07:52.606550 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:52 crc kubenswrapper[4764]: I0202 09:07:52.606562 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:52Z","lastTransitionTime":"2026-02-02T09:07:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:52 crc kubenswrapper[4764]: I0202 09:07:52.614210 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d66ab3c0-9b5b-41d5-b6c0-91ccddccd7e9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d92ac486f5dade46bc75de32099859a71fc2ed25a7abbe6ae8937df128e7df30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a4ac273a4f127d79e5fad7d4c30cd2044a346f5734dba5d33d68cd092c0c2d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d38c7cb5dcd22fb07524494ebe77aa1a7f4f808f566043563817e47e7ec12a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://408722c7f83d28919431532f8c797a1c795fd5d96736d14e7f06a05744899702\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f473b18cd8003f75bb1dafb82bcec45745a73ccd10434297356dac185a3d15c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e229f484d19cbdabedf53d90a822a4434db9e00b1301538e885e26acc19d25f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e229f484d19cbdabedf53d90a822a4434db9e00b1301538e885e26acc19d25f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a9cc673be8ff245dba81278e8cd34cfe45ed36ff0b8bd5b30a0ddf333ec6371\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a9cc673be8ff245dba81278e8cd34cfe45ed36ff0b8bd5b30a0ddf333ec6371\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bc7e49c7a5ae66ea939060192fa6a346dc0b0b9aa9750b29879d3c426cc0d276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc7e49c7a5ae66ea939060192fa6a346dc0b0b9aa9750b29879d3c426cc0d276\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:52Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:52 crc kubenswrapper[4764]: I0202 09:07:52.630131 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa9d037e66cb53b9dd698eff46284c9e718fd842cda6f7b525f0c79062d70bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:52Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:52 crc kubenswrapper[4764]: I0202 09:07:52.634177 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/a2152e8a-f456-47b5-a547-b65ec837ecbc-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-lh6tm\" (UID: \"a2152e8a-f456-47b5-a547-b65ec837ecbc\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lh6tm" Feb 02 09:07:52 crc kubenswrapper[4764]: I0202 09:07:52.634216 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/a2152e8a-f456-47b5-a547-b65ec837ecbc-env-overrides\") pod \"ovnkube-control-plane-749d76644c-lh6tm\" (UID: \"a2152e8a-f456-47b5-a547-b65ec837ecbc\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lh6tm" Feb 02 09:07:52 crc kubenswrapper[4764]: I0202 09:07:52.634257 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/a2152e8a-f456-47b5-a547-b65ec837ecbc-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-lh6tm\" (UID: \"a2152e8a-f456-47b5-a547-b65ec837ecbc\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lh6tm" Feb 02 09:07:52 crc kubenswrapper[4764]: I0202 09:07:52.634293 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bb2sk\" (UniqueName: \"kubernetes.io/projected/a2152e8a-f456-47b5-a547-b65ec837ecbc-kube-api-access-bb2sk\") pod \"ovnkube-control-plane-749d76644c-lh6tm\" (UID: \"a2152e8a-f456-47b5-a547-b65ec837ecbc\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lh6tm" Feb 02 09:07:52 crc kubenswrapper[4764]: I0202 09:07:52.635018 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/a2152e8a-f456-47b5-a547-b65ec837ecbc-env-overrides\") pod \"ovnkube-control-plane-749d76644c-lh6tm\" (UID: \"a2152e8a-f456-47b5-a547-b65ec837ecbc\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lh6tm" Feb 02 09:07:52 crc kubenswrapper[4764]: I0202 09:07:52.635197 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/a2152e8a-f456-47b5-a547-b65ec837ecbc-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-lh6tm\" (UID: \"a2152e8a-f456-47b5-a547-b65ec837ecbc\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lh6tm" Feb 02 09:07:52 crc kubenswrapper[4764]: I0202 09:07:52.641662 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/a2152e8a-f456-47b5-a547-b65ec837ecbc-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-lh6tm\" (UID: \"a2152e8a-f456-47b5-a547-b65ec837ecbc\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lh6tm" Feb 02 09:07:52 crc kubenswrapper[4764]: I0202 09:07:52.656002 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:52Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:52 crc kubenswrapper[4764]: I0202 09:07:52.656393 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bb2sk\" (UniqueName: \"kubernetes.io/projected/a2152e8a-f456-47b5-a547-b65ec837ecbc-kube-api-access-bb2sk\") pod \"ovnkube-control-plane-749d76644c-lh6tm\" (UID: \"a2152e8a-f456-47b5-a547-b65ec837ecbc\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lh6tm" Feb 02 09:07:52 crc kubenswrapper[4764]: I0202 09:07:52.677380 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24632dda-6100-4ab6-a28e-214ddae4360c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a6c67e9b7708cc0d85b0b6ffbbade0d34867e159a9c5487ca2606475c667a50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bffda48d7c881c1470251b501661324859373246cc1a21811bd36ca24887c7aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a989645cf485cd8ff80ffff1a86cd7808e1564eba5667535e161cc68db13b26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20b71d3ee7a381baa51b0b0d0470283022cfede30f491135737b5e82d631d763\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c380383b6a014d1b694a13ddc22f54b20b9e4f9d5c9b9000d76187621dce0a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5529f2bff87753428e9173826fc1880acf96a4c134fa9ecad195b15697cad73a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12244061eabaae7fa9057ac5cee894c3b308b466411bcb910682e7f65d240dd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://469f42001223aec6a9629296820e4925288790b1a17d6c7ac106dab6989bfc77\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-02T09:07:50Z\\\",\\\"message\\\":\\\"5993 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0202 09:07:50.920631 5993 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0202 09:07:50.920653 5993 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0202 09:07:50.920876 5993 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0202 09:07:50.920634 5993 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0202 09:07:50.921026 5993 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0202 09:07:50.921057 5993 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0202 09:07:50.921081 5993 factory.go:656] Stopping watch factory\\\\nI0202 09:07:50.921098 5993 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0202 09:07:50.921161 5993 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0202 09:07:50.921186 5993 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0202 09:07:50.921195 5993 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0202 09:07:50.921259 5993 handler.go:208] Removed *v1.Node event handler 2\\\\nI0202 09:07:50.921280 5993 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:47Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b4cfbd4e682bcdcaff785013bdb72580ce36fced733e235538536827036b93d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://579903eadb183466a2564737e6b678e04308239c0cd75b6808892053b97afbc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://579903eadb183466a2564737e6b678e04308239c0cd75b6808892053b97afbc0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zhn7j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:52Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:52 crc kubenswrapper[4764]: I0202 09:07:52.691368 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed747fd8-e3ce-4319-a6da-bda0fd51b85a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6f064bdc66470aace3180d4988244026b54009a63cff199e41c5d02a1d18479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://60258f84a1d8dcfe48b7e9688ed50f077e201971f20f3f6663fbf1ba58a4611a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f84be2e08217a2ac3edde858d78b4410376ed9bae3bcabe2a7f9b2fbe2f69c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a69d25d352d7ceac5c735bc1fcf446a1fea6dc22c65a5296cfab828cea17377\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:52Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:52 crc kubenswrapper[4764]: I0202 09:07:52.703327 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f6904228e052fff69ad1fd66df97758b6b60ba4351e97a7fbd03eee4df1b653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:52Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:52 crc kubenswrapper[4764]: I0202 09:07:52.709303 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:52 crc kubenswrapper[4764]: I0202 09:07:52.709340 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:52 crc kubenswrapper[4764]: I0202 09:07:52.709350 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:52 crc kubenswrapper[4764]: I0202 09:07:52.709366 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:52 crc kubenswrapper[4764]: I0202 09:07:52.709377 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:52Z","lastTransitionTime":"2026-02-02T09:07:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:52 crc kubenswrapper[4764]: I0202 09:07:52.717550 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6zxff" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aaba93da-12a5-43df-b7c7-4c8b988163c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3db293678a907cd55b31b1d0c0bd7c30e1ba75ecd5e13cbc68ab98fbf0fbd9d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqwb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6zxff\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:52Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:52 crc kubenswrapper[4764]: I0202 09:07:52.730915 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d192f670-9f9d-4539-9641-e4bed73acdd4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://040c5e8a5659f85f0d52d1a1e064154d998ae104f711101d2204c0b1caebbb30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbksf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fdf06410349b7c99aa166e87fbe80bf68e2ae923635a6a978b7e2cd50383f7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbksf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4ndm4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:52Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:52 crc kubenswrapper[4764]: I0202 09:07:52.742420 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pcjmn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6e5ccd3b-b657-47a2-9c4d-bc9adb126533\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72863eef568d92a1abeb3cde2e89630a4f5fb7626c6949448ff5195774c98eab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2f7zl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pcjmn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:52Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:52 crc kubenswrapper[4764]: I0202 09:07:52.753897 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:52Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:52 crc kubenswrapper[4764]: I0202 09:07:52.763547 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lh6tm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2152e8a-f456-47b5-a547-b65ec837ecbc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bb2sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bb2sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lh6tm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:52Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:52 crc kubenswrapper[4764]: I0202 09:07:52.791158 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-27 15:13:55.962626693 +0000 UTC Feb 02 09:07:52 crc kubenswrapper[4764]: I0202 09:07:52.809478 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lh6tm" Feb 02 09:07:52 crc kubenswrapper[4764]: I0202 09:07:52.812040 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:52 crc kubenswrapper[4764]: I0202 09:07:52.812085 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:52 crc kubenswrapper[4764]: I0202 09:07:52.812095 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:52 crc kubenswrapper[4764]: I0202 09:07:52.812113 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:52 crc kubenswrapper[4764]: I0202 09:07:52.812125 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:52Z","lastTransitionTime":"2026-02-02T09:07:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:52 crc kubenswrapper[4764]: W0202 09:07:52.823071 4764 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda2152e8a_f456_47b5_a547_b65ec837ecbc.slice/crio-28f42d072678d4f1cc1fdc5759ca9072462e06746827c54847b16abb94581902 WatchSource:0}: Error finding container 28f42d072678d4f1cc1fdc5759ca9072462e06746827c54847b16abb94581902: Status 404 returned error can't find the container with id 28f42d072678d4f1cc1fdc5759ca9072462e06746827c54847b16abb94581902 Feb 02 09:07:52 crc kubenswrapper[4764]: I0202 09:07:52.824513 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 09:07:52 crc kubenswrapper[4764]: I0202 09:07:52.824524 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 09:07:52 crc kubenswrapper[4764]: I0202 09:07:52.824541 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 09:07:52 crc kubenswrapper[4764]: E0202 09:07:52.824781 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 09:07:52 crc kubenswrapper[4764]: E0202 09:07:52.824865 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 09:07:52 crc kubenswrapper[4764]: E0202 09:07:52.824988 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 09:07:52 crc kubenswrapper[4764]: I0202 09:07:52.914321 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:52 crc kubenswrapper[4764]: I0202 09:07:52.914378 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:52 crc kubenswrapper[4764]: I0202 09:07:52.914395 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:52 crc kubenswrapper[4764]: I0202 09:07:52.914416 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:52 crc kubenswrapper[4764]: I0202 09:07:52.914432 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:52Z","lastTransitionTime":"2026-02-02T09:07:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:53 crc kubenswrapper[4764]: I0202 09:07:53.016272 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:53 crc kubenswrapper[4764]: I0202 09:07:53.016305 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:53 crc kubenswrapper[4764]: I0202 09:07:53.016315 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:53 crc kubenswrapper[4764]: I0202 09:07:53.016329 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:53 crc kubenswrapper[4764]: I0202 09:07:53.016340 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:53Z","lastTransitionTime":"2026-02-02T09:07:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:53 crc kubenswrapper[4764]: I0202 09:07:53.082262 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lh6tm" event={"ID":"a2152e8a-f456-47b5-a547-b65ec837ecbc","Type":"ContainerStarted","Data":"38e176a00b07f2960fdd7ba2aadd1468a5e083b8d154e9d118d13f2585da81e8"} Feb 02 09:07:53 crc kubenswrapper[4764]: I0202 09:07:53.082350 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lh6tm" event={"ID":"a2152e8a-f456-47b5-a547-b65ec837ecbc","Type":"ContainerStarted","Data":"28f42d072678d4f1cc1fdc5759ca9072462e06746827c54847b16abb94581902"} Feb 02 09:07:53 crc kubenswrapper[4764]: I0202 09:07:53.084135 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zhn7j_24632dda-6100-4ab6-a28e-214ddae4360c/ovnkube-controller/1.log" Feb 02 09:07:53 crc kubenswrapper[4764]: I0202 09:07:53.084920 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zhn7j_24632dda-6100-4ab6-a28e-214ddae4360c/ovnkube-controller/0.log" Feb 02 09:07:53 crc kubenswrapper[4764]: I0202 09:07:53.091306 4764 generic.go:334] "Generic (PLEG): container finished" podID="24632dda-6100-4ab6-a28e-214ddae4360c" containerID="12244061eabaae7fa9057ac5cee894c3b308b466411bcb910682e7f65d240dd7" exitCode=1 Feb 02 09:07:53 crc kubenswrapper[4764]: I0202 09:07:53.091352 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" event={"ID":"24632dda-6100-4ab6-a28e-214ddae4360c","Type":"ContainerDied","Data":"12244061eabaae7fa9057ac5cee894c3b308b466411bcb910682e7f65d240dd7"} Feb 02 09:07:53 crc kubenswrapper[4764]: I0202 09:07:53.091388 4764 scope.go:117] "RemoveContainer" containerID="469f42001223aec6a9629296820e4925288790b1a17d6c7ac106dab6989bfc77" Feb 02 09:07:53 crc kubenswrapper[4764]: I0202 09:07:53.092021 4764 scope.go:117] "RemoveContainer" containerID="12244061eabaae7fa9057ac5cee894c3b308b466411bcb910682e7f65d240dd7" Feb 02 09:07:53 crc kubenswrapper[4764]: E0202 09:07:53.092203 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-zhn7j_openshift-ovn-kubernetes(24632dda-6100-4ab6-a28e-214ddae4360c)\"" pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" podUID="24632dda-6100-4ab6-a28e-214ddae4360c" Feb 02 09:07:53 crc kubenswrapper[4764]: I0202 09:07:53.118792 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:53 crc kubenswrapper[4764]: I0202 09:07:53.118834 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:53 crc kubenswrapper[4764]: I0202 09:07:53.118843 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:53 crc kubenswrapper[4764]: I0202 09:07:53.118859 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:53 crc kubenswrapper[4764]: I0202 09:07:53.118869 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:53Z","lastTransitionTime":"2026-02-02T09:07:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:53 crc kubenswrapper[4764]: I0202 09:07:53.119314 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d66ab3c0-9b5b-41d5-b6c0-91ccddccd7e9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d92ac486f5dade46bc75de32099859a71fc2ed25a7abbe6ae8937df128e7df30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a4ac273a4f127d79e5fad7d4c30cd2044a346f5734dba5d33d68cd092c0c2d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d38c7cb5dcd22fb07524494ebe77aa1a7f4f808f566043563817e47e7ec12a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://408722c7f83d28919431532f8c797a1c795fd5d96736d14e7f06a05744899702\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f473b18cd8003f75bb1dafb82bcec45745a73ccd10434297356dac185a3d15c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e229f484d19cbdabedf53d90a822a4434db9e00b1301538e885e26acc19d25f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e229f484d19cbdabedf53d90a822a4434db9e00b1301538e885e26acc19d25f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a9cc673be8ff245dba81278e8cd34cfe45ed36ff0b8bd5b30a0ddf333ec6371\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a9cc673be8ff245dba81278e8cd34cfe45ed36ff0b8bd5b30a0ddf333ec6371\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bc7e49c7a5ae66ea939060192fa6a346dc0b0b9aa9750b29879d3c426cc0d276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc7e49c7a5ae66ea939060192fa6a346dc0b0b9aa9750b29879d3c426cc0d276\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:53Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:53 crc kubenswrapper[4764]: I0202 09:07:53.130785 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa9d037e66cb53b9dd698eff46284c9e718fd842cda6f7b525f0c79062d70bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:53Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:53 crc kubenswrapper[4764]: I0202 09:07:53.145752 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:53Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:53 crc kubenswrapper[4764]: I0202 09:07:53.166255 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24632dda-6100-4ab6-a28e-214ddae4360c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a6c67e9b7708cc0d85b0b6ffbbade0d34867e159a9c5487ca2606475c667a50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bffda48d7c881c1470251b501661324859373246cc1a21811bd36ca24887c7aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a989645cf485cd8ff80ffff1a86cd7808e1564eba5667535e161cc68db13b26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20b71d3ee7a381baa51b0b0d0470283022cfede30f491135737b5e82d631d763\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c380383b6a014d1b694a13ddc22f54b20b9e4f9d5c9b9000d76187621dce0a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5529f2bff87753428e9173826fc1880acf96a4c134fa9ecad195b15697cad73a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12244061eabaae7fa9057ac5cee894c3b308b466411bcb910682e7f65d240dd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://469f42001223aec6a9629296820e4925288790b1a17d6c7ac106dab6989bfc77\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-02T09:07:50Z\\\",\\\"message\\\":\\\"5993 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0202 09:07:50.920631 5993 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0202 09:07:50.920653 5993 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0202 09:07:50.920876 5993 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0202 09:07:50.920634 5993 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0202 09:07:50.921026 5993 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0202 09:07:50.921057 5993 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0202 09:07:50.921081 5993 factory.go:656] Stopping watch factory\\\\nI0202 09:07:50.921098 5993 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0202 09:07:50.921161 5993 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0202 09:07:50.921186 5993 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0202 09:07:50.921195 5993 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0202 09:07:50.921259 5993 handler.go:208] Removed *v1.Node event handler 2\\\\nI0202 09:07:50.921280 5993 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:47Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://12244061eabaae7fa9057ac5cee894c3b308b466411bcb910682e7f65d240dd7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-02T09:07:52Z\\\",\\\"message\\\":\\\"l_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-kube-apiserver/apiserver]} name:Service_openshift-kube-apiserver/apiserver_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.93:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {d71b38eb-32af-4c0f-9490-7c317c111e3a}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0202 09:07:51.914338 6129 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-kube-apiserver/apiserver]} name:Service_openshift-kube-apiserver/apiserver_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.93:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {d71b38eb-32af-4c0f-9490-7c317c111e3a}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF0202 09:07:51.914404 6129 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b4cfbd4e682bcdcaff785013bdb72580ce36fced733e235538536827036b93d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://579903eadb183466a2564737e6b678e04308239c0cd75b6808892053b97afbc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://579903eadb183466a2564737e6b678e04308239c0cd75b6808892053b97afbc0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zhn7j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:53Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:53 crc kubenswrapper[4764]: I0202 09:07:53.178369 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed747fd8-e3ce-4319-a6da-bda0fd51b85a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6f064bdc66470aace3180d4988244026b54009a63cff199e41c5d02a1d18479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://60258f84a1d8dcfe48b7e9688ed50f077e201971f20f3f6663fbf1ba58a4611a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f84be2e08217a2ac3edde858d78b4410376ed9bae3bcabe2a7f9b2fbe2f69c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a69d25d352d7ceac5c735bc1fcf446a1fea6dc22c65a5296cfab828cea17377\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:53Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:53 crc kubenswrapper[4764]: I0202 09:07:53.188099 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f6904228e052fff69ad1fd66df97758b6b60ba4351e97a7fbd03eee4df1b653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:53Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:53 crc kubenswrapper[4764]: I0202 09:07:53.197537 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6zxff" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aaba93da-12a5-43df-b7c7-4c8b988163c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3db293678a907cd55b31b1d0c0bd7c30e1ba75ecd5e13cbc68ab98fbf0fbd9d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqwb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6zxff\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:53Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:53 crc kubenswrapper[4764]: I0202 09:07:53.207443 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d192f670-9f9d-4539-9641-e4bed73acdd4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://040c5e8a5659f85f0d52d1a1e064154d998ae104f711101d2204c0b1caebbb30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbksf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fdf06410349b7c99aa166e87fbe80bf68e2ae923635a6a978b7e2cd50383f7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbksf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4ndm4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:53Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:53 crc kubenswrapper[4764]: I0202 09:07:53.216498 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pcjmn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6e5ccd3b-b657-47a2-9c4d-bc9adb126533\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72863eef568d92a1abeb3cde2e89630a4f5fb7626c6949448ff5195774c98eab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2f7zl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pcjmn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:53Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:53 crc kubenswrapper[4764]: I0202 09:07:53.220536 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:53 crc kubenswrapper[4764]: I0202 09:07:53.220572 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:53 crc kubenswrapper[4764]: I0202 09:07:53.220585 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:53 crc kubenswrapper[4764]: I0202 09:07:53.220602 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:53 crc kubenswrapper[4764]: I0202 09:07:53.220613 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:53Z","lastTransitionTime":"2026-02-02T09:07:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:53 crc kubenswrapper[4764]: I0202 09:07:53.228354 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:53Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:53 crc kubenswrapper[4764]: I0202 09:07:53.238304 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lh6tm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2152e8a-f456-47b5-a547-b65ec837ecbc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bb2sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bb2sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lh6tm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:53Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:53 crc kubenswrapper[4764]: I0202 09:07:53.253925 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9p5dc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"660ac006-2495-45c4-9fb3-e9c1dddcf7a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da812b124c97aa8ad1563f4bcfac8f3bfa025096fa4a2561e5847b1bf9cb30b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23e57e65a6f5c3059506b9db21e9ac2492c30bd9191a98db0281915e9ce39797\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23e57e65a6f5c3059506b9db21e9ac2492c30bd9191a98db0281915e9ce39797\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2417b66816da1a6cd1791a4b91eca4a3c4f3694462e066e7f0722f1b9c2105c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2417b66816da1a6cd1791a4b91eca4a3c4f3694462e066e7f0722f1b9c2105c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://883641f29e9120c2cea6b272bacf1dc63c752f55ab0408cdaf73db85c85ac369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://883641f29e9120c2cea6b272bacf1dc63c752f55ab0408cdaf73db85c85ac369\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2745baba1188ba5e8f820616e4117a12c50ecee5b68e71ad9cabbb621b6e131f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2745baba1188ba5e8f820616e4117a12c50ecee5b68e71ad9cabbb621b6e131f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8925e435be4d48d86824fc35f8b8819fb35d63e385164c4a20edb3bf3035c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a8925e435be4d48d86824fc35f8b8819fb35d63e385164c4a20edb3bf3035c3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d3fbf668ae7d04c8d9e574f43c2fe90c5604204ba38a3c60847afdfbceb7a57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d3fbf668ae7d04c8d9e574f43c2fe90c5604204ba38a3c60847afdfbceb7a57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9p5dc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:53Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:53 crc kubenswrapper[4764]: I0202 09:07:53.272640 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87d61aea-46f6-4ee6-a7a4-767a6a67fea4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e78b6d08e4d616cefbb73e3fd43b3c6886dcd8aa997c7b4948bf1ff63401293\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d468715c2da33996561c40043216233fd12e21539d0c58dd1f3b892fd9f91a9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b184f7c936dc4046f74c0018269cbc2c4b7ceddb650b4e871d3f8a707e0df88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://269c5074371bd00f8363fb3888c11f1df383081a61497aea6578f08700881d50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b25bc9395831524c53f9e565b4efbe74f21792efafeb1dd507161bc91aa0ecdf\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T09:07:22Z\\\",\\\"message\\\":\\\"W0202 09:07:21.597488 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0202 09:07:21.598117 1 crypto.go:601] Generating new CA for check-endpoints-signer@1770023241 cert, and key in /tmp/serving-cert-3856183724/serving-signer.crt, /tmp/serving-cert-3856183724/serving-signer.key\\\\nI0202 09:07:22.012184 1 observer_polling.go:159] Starting file observer\\\\nW0202 09:07:22.015493 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0202 09:07:22.015797 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 09:07:22.016596 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3856183724/tls.crt::/tmp/serving-cert-3856183724/tls.key\\\\\\\"\\\\nF0202 09:07:22.270605 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b44e17b8ca01f549bc10081dd5f109cb1b9fff364bf5bd4fd90d9f0e3eeb04db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfb06a5ec9262cf39b108a66014d15e325cdbd5441a2d9e7a811dd57265c99a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfb06a5ec9262cf39b108a66014d15e325cdbd5441a2d9e7a811dd57265c99a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:53Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:53 crc kubenswrapper[4764]: I0202 09:07:53.291489 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5528225d91ac2778bef4554abb0bd036ac5d3b6cdac7357bab69e70c3894b6f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://071bbb80d0747cf3e109ba626d47f7d05f85f4cc0b3aba77276c795741f47ac6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:53Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:53 crc kubenswrapper[4764]: I0202 09:07:53.306383 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:53Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:53 crc kubenswrapper[4764]: I0202 09:07:53.318871 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nc7qt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1a2c7b22068b7165927fc158885d48bf9952b9eba521d8ae9a599eee53dc305\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2hz4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nc7qt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:53Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:53 crc kubenswrapper[4764]: I0202 09:07:53.322499 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:53 crc kubenswrapper[4764]: I0202 09:07:53.322553 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:53 crc kubenswrapper[4764]: I0202 09:07:53.322565 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:53 crc kubenswrapper[4764]: I0202 09:07:53.322580 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:53 crc kubenswrapper[4764]: I0202 09:07:53.322591 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:53Z","lastTransitionTime":"2026-02-02T09:07:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:53 crc kubenswrapper[4764]: I0202 09:07:53.425286 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:53 crc kubenswrapper[4764]: I0202 09:07:53.425328 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:53 crc kubenswrapper[4764]: I0202 09:07:53.425341 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:53 crc kubenswrapper[4764]: I0202 09:07:53.425359 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:53 crc kubenswrapper[4764]: I0202 09:07:53.425370 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:53Z","lastTransitionTime":"2026-02-02T09:07:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:53 crc kubenswrapper[4764]: I0202 09:07:53.528130 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:53 crc kubenswrapper[4764]: I0202 09:07:53.528185 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:53 crc kubenswrapper[4764]: I0202 09:07:53.528202 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:53 crc kubenswrapper[4764]: I0202 09:07:53.528222 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:53 crc kubenswrapper[4764]: I0202 09:07:53.528237 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:53Z","lastTransitionTime":"2026-02-02T09:07:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:53 crc kubenswrapper[4764]: I0202 09:07:53.630988 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:53 crc kubenswrapper[4764]: I0202 09:07:53.631020 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:53 crc kubenswrapper[4764]: I0202 09:07:53.631030 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:53 crc kubenswrapper[4764]: I0202 09:07:53.631044 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:53 crc kubenswrapper[4764]: I0202 09:07:53.631055 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:53Z","lastTransitionTime":"2026-02-02T09:07:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:53 crc kubenswrapper[4764]: I0202 09:07:53.733383 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:53 crc kubenswrapper[4764]: I0202 09:07:53.733414 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:53 crc kubenswrapper[4764]: I0202 09:07:53.733423 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:53 crc kubenswrapper[4764]: I0202 09:07:53.733471 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:53 crc kubenswrapper[4764]: I0202 09:07:53.733481 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:53Z","lastTransitionTime":"2026-02-02T09:07:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:53 crc kubenswrapper[4764]: I0202 09:07:53.792213 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-06 13:18:42.638884567 +0000 UTC Feb 02 09:07:53 crc kubenswrapper[4764]: I0202 09:07:53.835874 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:53 crc kubenswrapper[4764]: I0202 09:07:53.835911 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:53 crc kubenswrapper[4764]: I0202 09:07:53.835921 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:53 crc kubenswrapper[4764]: I0202 09:07:53.835961 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:53 crc kubenswrapper[4764]: I0202 09:07:53.835972 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:53Z","lastTransitionTime":"2026-02-02T09:07:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:53 crc kubenswrapper[4764]: I0202 09:07:53.938786 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:53 crc kubenswrapper[4764]: I0202 09:07:53.938821 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:53 crc kubenswrapper[4764]: I0202 09:07:53.938830 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:53 crc kubenswrapper[4764]: I0202 09:07:53.938846 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:53 crc kubenswrapper[4764]: I0202 09:07:53.938858 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:53Z","lastTransitionTime":"2026-02-02T09:07:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:53 crc kubenswrapper[4764]: I0202 09:07:53.976016 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-kwtmr"] Feb 02 09:07:53 crc kubenswrapper[4764]: I0202 09:07:53.976595 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kwtmr" Feb 02 09:07:53 crc kubenswrapper[4764]: E0202 09:07:53.976685 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kwtmr" podUID="991faa9a-dd25-4f49-82bd-ce60cefd4af2" Feb 02 09:07:53 crc kubenswrapper[4764]: I0202 09:07:53.991493 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f6904228e052fff69ad1fd66df97758b6b60ba4351e97a7fbd03eee4df1b653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:53Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:54 crc kubenswrapper[4764]: I0202 09:07:54.002577 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6zxff" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aaba93da-12a5-43df-b7c7-4c8b988163c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3db293678a907cd55b31b1d0c0bd7c30e1ba75ecd5e13cbc68ab98fbf0fbd9d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqwb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6zxff\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:54Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:54 crc kubenswrapper[4764]: I0202 09:07:54.012721 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d192f670-9f9d-4539-9641-e4bed73acdd4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://040c5e8a5659f85f0d52d1a1e064154d998ae104f711101d2204c0b1caebbb30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbksf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fdf06410349b7c99aa166e87fbe80bf68e2ae923635a6a978b7e2cd50383f7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbksf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4ndm4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:54Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:54 crc kubenswrapper[4764]: I0202 09:07:54.023486 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pcjmn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6e5ccd3b-b657-47a2-9c4d-bc9adb126533\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72863eef568d92a1abeb3cde2e89630a4f5fb7626c6949448ff5195774c98eab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2f7zl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pcjmn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:54Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:54 crc kubenswrapper[4764]: I0202 09:07:54.036784 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed747fd8-e3ce-4319-a6da-bda0fd51b85a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6f064bdc66470aace3180d4988244026b54009a63cff199e41c5d02a1d18479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://60258f84a1d8dcfe48b7e9688ed50f077e201971f20f3f6663fbf1ba58a4611a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f84be2e08217a2ac3edde858d78b4410376ed9bae3bcabe2a7f9b2fbe2f69c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a69d25d352d7ceac5c735bc1fcf446a1fea6dc22c65a5296cfab828cea17377\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:54Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:54 crc kubenswrapper[4764]: I0202 09:07:54.041218 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:54 crc kubenswrapper[4764]: I0202 09:07:54.041284 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:54 crc kubenswrapper[4764]: I0202 09:07:54.041300 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:54 crc kubenswrapper[4764]: I0202 09:07:54.041322 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:54 crc kubenswrapper[4764]: I0202 09:07:54.041339 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:54Z","lastTransitionTime":"2026-02-02T09:07:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:54 crc kubenswrapper[4764]: I0202 09:07:54.046747 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h865b\" (UniqueName: \"kubernetes.io/projected/991faa9a-dd25-4f49-82bd-ce60cefd4af2-kube-api-access-h865b\") pod \"network-metrics-daemon-kwtmr\" (UID: \"991faa9a-dd25-4f49-82bd-ce60cefd4af2\") " pod="openshift-multus/network-metrics-daemon-kwtmr" Feb 02 09:07:54 crc kubenswrapper[4764]: I0202 09:07:54.046895 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/991faa9a-dd25-4f49-82bd-ce60cefd4af2-metrics-certs\") pod \"network-metrics-daemon-kwtmr\" (UID: \"991faa9a-dd25-4f49-82bd-ce60cefd4af2\") " pod="openshift-multus/network-metrics-daemon-kwtmr" Feb 02 09:07:54 crc kubenswrapper[4764]: I0202 09:07:54.051584 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:54Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:54 crc kubenswrapper[4764]: I0202 09:07:54.070359 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lh6tm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2152e8a-f456-47b5-a547-b65ec837ecbc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bb2sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bb2sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lh6tm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:54Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:54 crc kubenswrapper[4764]: I0202 09:07:54.085527 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5528225d91ac2778bef4554abb0bd036ac5d3b6cdac7357bab69e70c3894b6f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://071bbb80d0747cf3e109ba626d47f7d05f85f4cc0b3aba77276c795741f47ac6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:54Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:54 crc kubenswrapper[4764]: I0202 09:07:54.096375 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lh6tm" event={"ID":"a2152e8a-f456-47b5-a547-b65ec837ecbc","Type":"ContainerStarted","Data":"996e649011d5255783284421ddea8e0a5c383ef675b0920c3fdc0af6503d80e4"} Feb 02 09:07:54 crc kubenswrapper[4764]: I0202 09:07:54.100381 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zhn7j_24632dda-6100-4ab6-a28e-214ddae4360c/ovnkube-controller/1.log" Feb 02 09:07:54 crc kubenswrapper[4764]: I0202 09:07:54.106907 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:54Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:54 crc kubenswrapper[4764]: I0202 09:07:54.107262 4764 scope.go:117] "RemoveContainer" containerID="12244061eabaae7fa9057ac5cee894c3b308b466411bcb910682e7f65d240dd7" Feb 02 09:07:54 crc kubenswrapper[4764]: E0202 09:07:54.107709 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-zhn7j_openshift-ovn-kubernetes(24632dda-6100-4ab6-a28e-214ddae4360c)\"" pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" podUID="24632dda-6100-4ab6-a28e-214ddae4360c" Feb 02 09:07:54 crc kubenswrapper[4764]: I0202 09:07:54.121174 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nc7qt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1a2c7b22068b7165927fc158885d48bf9952b9eba521d8ae9a599eee53dc305\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2hz4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nc7qt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:54Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:54 crc kubenswrapper[4764]: I0202 09:07:54.135376 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9p5dc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"660ac006-2495-45c4-9fb3-e9c1dddcf7a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da812b124c97aa8ad1563f4bcfac8f3bfa025096fa4a2561e5847b1bf9cb30b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23e57e65a6f5c3059506b9db21e9ac2492c30bd9191a98db0281915e9ce39797\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23e57e65a6f5c3059506b9db21e9ac2492c30bd9191a98db0281915e9ce39797\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2417b66816da1a6cd1791a4b91eca4a3c4f3694462e066e7f0722f1b9c2105c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2417b66816da1a6cd1791a4b91eca4a3c4f3694462e066e7f0722f1b9c2105c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://883641f29e9120c2cea6b272bacf1dc63c752f55ab0408cdaf73db85c85ac369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://883641f29e9120c2cea6b272bacf1dc63c752f55ab0408cdaf73db85c85ac369\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2745baba1188ba5e8f820616e4117a12c50ecee5b68e71ad9cabbb621b6e131f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2745baba1188ba5e8f820616e4117a12c50ecee5b68e71ad9cabbb621b6e131f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8925e435be4d48d86824fc35f8b8819fb35d63e385164c4a20edb3bf3035c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a8925e435be4d48d86824fc35f8b8819fb35d63e385164c4a20edb3bf3035c3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d3fbf668ae7d04c8d9e574f43c2fe90c5604204ba38a3c60847afdfbceb7a57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d3fbf668ae7d04c8d9e574f43c2fe90c5604204ba38a3c60847afdfbceb7a57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9p5dc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:54Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:54 crc kubenswrapper[4764]: I0202 09:07:54.144398 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:54 crc kubenswrapper[4764]: I0202 09:07:54.144442 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:54 crc kubenswrapper[4764]: I0202 09:07:54.144456 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:54 crc kubenswrapper[4764]: I0202 09:07:54.144479 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:54 crc kubenswrapper[4764]: I0202 09:07:54.144495 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:54Z","lastTransitionTime":"2026-02-02T09:07:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:54 crc kubenswrapper[4764]: I0202 09:07:54.148173 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h865b\" (UniqueName: \"kubernetes.io/projected/991faa9a-dd25-4f49-82bd-ce60cefd4af2-kube-api-access-h865b\") pod \"network-metrics-daemon-kwtmr\" (UID: \"991faa9a-dd25-4f49-82bd-ce60cefd4af2\") " pod="openshift-multus/network-metrics-daemon-kwtmr" Feb 02 09:07:54 crc kubenswrapper[4764]: I0202 09:07:54.148393 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/991faa9a-dd25-4f49-82bd-ce60cefd4af2-metrics-certs\") pod \"network-metrics-daemon-kwtmr\" (UID: \"991faa9a-dd25-4f49-82bd-ce60cefd4af2\") " pod="openshift-multus/network-metrics-daemon-kwtmr" Feb 02 09:07:54 crc kubenswrapper[4764]: E0202 09:07:54.149263 4764 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 02 09:07:54 crc kubenswrapper[4764]: E0202 09:07:54.149342 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/991faa9a-dd25-4f49-82bd-ce60cefd4af2-metrics-certs podName:991faa9a-dd25-4f49-82bd-ce60cefd4af2 nodeName:}" failed. No retries permitted until 2026-02-02 09:07:54.649318162 +0000 UTC m=+37.583042250 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/991faa9a-dd25-4f49-82bd-ce60cefd4af2-metrics-certs") pod "network-metrics-daemon-kwtmr" (UID: "991faa9a-dd25-4f49-82bd-ce60cefd4af2") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 02 09:07:54 crc kubenswrapper[4764]: I0202 09:07:54.149973 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kwtmr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"991faa9a-dd25-4f49-82bd-ce60cefd4af2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h865b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h865b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:53Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kwtmr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:54Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:54 crc kubenswrapper[4764]: I0202 09:07:54.164361 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87d61aea-46f6-4ee6-a7a4-767a6a67fea4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e78b6d08e4d616cefbb73e3fd43b3c6886dcd8aa997c7b4948bf1ff63401293\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d468715c2da33996561c40043216233fd12e21539d0c58dd1f3b892fd9f91a9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b184f7c936dc4046f74c0018269cbc2c4b7ceddb650b4e871d3f8a707e0df88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://269c5074371bd00f8363fb3888c11f1df383081a61497aea6578f08700881d50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b25bc9395831524c53f9e565b4efbe74f21792efafeb1dd507161bc91aa0ecdf\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T09:07:22Z\\\",\\\"message\\\":\\\"W0202 09:07:21.597488 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0202 09:07:21.598117 1 crypto.go:601] Generating new CA for check-endpoints-signer@1770023241 cert, and key in /tmp/serving-cert-3856183724/serving-signer.crt, /tmp/serving-cert-3856183724/serving-signer.key\\\\nI0202 09:07:22.012184 1 observer_polling.go:159] Starting file observer\\\\nW0202 09:07:22.015493 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0202 09:07:22.015797 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 09:07:22.016596 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3856183724/tls.crt::/tmp/serving-cert-3856183724/tls.key\\\\\\\"\\\\nF0202 09:07:22.270605 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b44e17b8ca01f549bc10081dd5f109cb1b9fff364bf5bd4fd90d9f0e3eeb04db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfb06a5ec9262cf39b108a66014d15e325cdbd5441a2d9e7a811dd57265c99a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfb06a5ec9262cf39b108a66014d15e325cdbd5441a2d9e7a811dd57265c99a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:54Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:54 crc kubenswrapper[4764]: I0202 09:07:54.171115 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h865b\" (UniqueName: \"kubernetes.io/projected/991faa9a-dd25-4f49-82bd-ce60cefd4af2-kube-api-access-h865b\") pod \"network-metrics-daemon-kwtmr\" (UID: \"991faa9a-dd25-4f49-82bd-ce60cefd4af2\") " pod="openshift-multus/network-metrics-daemon-kwtmr" Feb 02 09:07:54 crc kubenswrapper[4764]: I0202 09:07:54.178402 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa9d037e66cb53b9dd698eff46284c9e718fd842cda6f7b525f0c79062d70bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:54Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:54 crc kubenswrapper[4764]: I0202 09:07:54.192487 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:54Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:54 crc kubenswrapper[4764]: I0202 09:07:54.212330 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24632dda-6100-4ab6-a28e-214ddae4360c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a6c67e9b7708cc0d85b0b6ffbbade0d34867e159a9c5487ca2606475c667a50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bffda48d7c881c1470251b501661324859373246cc1a21811bd36ca24887c7aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a989645cf485cd8ff80ffff1a86cd7808e1564eba5667535e161cc68db13b26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20b71d3ee7a381baa51b0b0d0470283022cfede30f491135737b5e82d631d763\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c380383b6a014d1b694a13ddc22f54b20b9e4f9d5c9b9000d76187621dce0a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5529f2bff87753428e9173826fc1880acf96a4c134fa9ecad195b15697cad73a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12244061eabaae7fa9057ac5cee894c3b308b466411bcb910682e7f65d240dd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://469f42001223aec6a9629296820e4925288790b1a17d6c7ac106dab6989bfc77\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-02T09:07:50Z\\\",\\\"message\\\":\\\"5993 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0202 09:07:50.920631 5993 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0202 09:07:50.920653 5993 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0202 09:07:50.920876 5993 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0202 09:07:50.920634 5993 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0202 09:07:50.921026 5993 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0202 09:07:50.921057 5993 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0202 09:07:50.921081 5993 factory.go:656] Stopping watch factory\\\\nI0202 09:07:50.921098 5993 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0202 09:07:50.921161 5993 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0202 09:07:50.921186 5993 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0202 09:07:50.921195 5993 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0202 09:07:50.921259 5993 handler.go:208] Removed *v1.Node event handler 2\\\\nI0202 09:07:50.921280 5993 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:47Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://12244061eabaae7fa9057ac5cee894c3b308b466411bcb910682e7f65d240dd7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-02T09:07:52Z\\\",\\\"message\\\":\\\"l_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-kube-apiserver/apiserver]} name:Service_openshift-kube-apiserver/apiserver_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.93:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {d71b38eb-32af-4c0f-9490-7c317c111e3a}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0202 09:07:51.914338 6129 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-kube-apiserver/apiserver]} name:Service_openshift-kube-apiserver/apiserver_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.93:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {d71b38eb-32af-4c0f-9490-7c317c111e3a}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF0202 09:07:51.914404 6129 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b4cfbd4e682bcdcaff785013bdb72580ce36fced733e235538536827036b93d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://579903eadb183466a2564737e6b678e04308239c0cd75b6808892053b97afbc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://579903eadb183466a2564737e6b678e04308239c0cd75b6808892053b97afbc0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zhn7j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:54Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:54 crc kubenswrapper[4764]: I0202 09:07:54.247218 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:54 crc kubenswrapper[4764]: I0202 09:07:54.247257 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:54 crc kubenswrapper[4764]: I0202 09:07:54.247269 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:54 crc kubenswrapper[4764]: I0202 09:07:54.247286 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:54 crc kubenswrapper[4764]: I0202 09:07:54.247296 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:54Z","lastTransitionTime":"2026-02-02T09:07:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:54 crc kubenswrapper[4764]: I0202 09:07:54.255162 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d66ab3c0-9b5b-41d5-b6c0-91ccddccd7e9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d92ac486f5dade46bc75de32099859a71fc2ed25a7abbe6ae8937df128e7df30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a4ac273a4f127d79e5fad7d4c30cd2044a346f5734dba5d33d68cd092c0c2d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d38c7cb5dcd22fb07524494ebe77aa1a7f4f808f566043563817e47e7ec12a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://408722c7f83d28919431532f8c797a1c795fd5d96736d14e7f06a05744899702\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f473b18cd8003f75bb1dafb82bcec45745a73ccd10434297356dac185a3d15c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e229f484d19cbdabedf53d90a822a4434db9e00b1301538e885e26acc19d25f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e229f484d19cbdabedf53d90a822a4434db9e00b1301538e885e26acc19d25f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a9cc673be8ff245dba81278e8cd34cfe45ed36ff0b8bd5b30a0ddf333ec6371\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a9cc673be8ff245dba81278e8cd34cfe45ed36ff0b8bd5b30a0ddf333ec6371\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bc7e49c7a5ae66ea939060192fa6a346dc0b0b9aa9750b29879d3c426cc0d276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc7e49c7a5ae66ea939060192fa6a346dc0b0b9aa9750b29879d3c426cc0d276\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:54Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:54 crc kubenswrapper[4764]: I0202 09:07:54.278046 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:54Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:54 crc kubenswrapper[4764]: I0202 09:07:54.290698 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lh6tm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2152e8a-f456-47b5-a547-b65ec837ecbc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38e176a00b07f2960fdd7ba2aadd1468a5e083b8d154e9d118d13f2585da81e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bb2sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://996e649011d5255783284421ddea8e0a5c383ef675b0920c3fdc0af6503d80e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bb2sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lh6tm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:54Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:54 crc kubenswrapper[4764]: I0202 09:07:54.301722 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nc7qt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1a2c7b22068b7165927fc158885d48bf9952b9eba521d8ae9a599eee53dc305\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2hz4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nc7qt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:54Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:54 crc kubenswrapper[4764]: I0202 09:07:54.314323 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9p5dc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"660ac006-2495-45c4-9fb3-e9c1dddcf7a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da812b124c97aa8ad1563f4bcfac8f3bfa025096fa4a2561e5847b1bf9cb30b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23e57e65a6f5c3059506b9db21e9ac2492c30bd9191a98db0281915e9ce39797\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23e57e65a6f5c3059506b9db21e9ac2492c30bd9191a98db0281915e9ce39797\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2417b66816da1a6cd1791a4b91eca4a3c4f3694462e066e7f0722f1b9c2105c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2417b66816da1a6cd1791a4b91eca4a3c4f3694462e066e7f0722f1b9c2105c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://883641f29e9120c2cea6b272bacf1dc63c752f55ab0408cdaf73db85c85ac369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://883641f29e9120c2cea6b272bacf1dc63c752f55ab0408cdaf73db85c85ac369\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2745baba1188ba5e8f820616e4117a12c50ecee5b68e71ad9cabbb621b6e131f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2745baba1188ba5e8f820616e4117a12c50ecee5b68e71ad9cabbb621b6e131f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8925e435be4d48d86824fc35f8b8819fb35d63e385164c4a20edb3bf3035c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a8925e435be4d48d86824fc35f8b8819fb35d63e385164c4a20edb3bf3035c3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d3fbf668ae7d04c8d9e574f43c2fe90c5604204ba38a3c60847afdfbceb7a57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d3fbf668ae7d04c8d9e574f43c2fe90c5604204ba38a3c60847afdfbceb7a57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9p5dc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:54Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:54 crc kubenswrapper[4764]: I0202 09:07:54.324731 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kwtmr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"991faa9a-dd25-4f49-82bd-ce60cefd4af2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h865b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h865b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:53Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kwtmr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:54Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:54 crc kubenswrapper[4764]: I0202 09:07:54.339117 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87d61aea-46f6-4ee6-a7a4-767a6a67fea4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e78b6d08e4d616cefbb73e3fd43b3c6886dcd8aa997c7b4948bf1ff63401293\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d468715c2da33996561c40043216233fd12e21539d0c58dd1f3b892fd9f91a9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b184f7c936dc4046f74c0018269cbc2c4b7ceddb650b4e871d3f8a707e0df88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://269c5074371bd00f8363fb3888c11f1df383081a61497aea6578f08700881d50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b25bc9395831524c53f9e565b4efbe74f21792efafeb1dd507161bc91aa0ecdf\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T09:07:22Z\\\",\\\"message\\\":\\\"W0202 09:07:21.597488 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0202 09:07:21.598117 1 crypto.go:601] Generating new CA for check-endpoints-signer@1770023241 cert, and key in /tmp/serving-cert-3856183724/serving-signer.crt, /tmp/serving-cert-3856183724/serving-signer.key\\\\nI0202 09:07:22.012184 1 observer_polling.go:159] Starting file observer\\\\nW0202 09:07:22.015493 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0202 09:07:22.015797 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 09:07:22.016596 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3856183724/tls.crt::/tmp/serving-cert-3856183724/tls.key\\\\\\\"\\\\nF0202 09:07:22.270605 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b44e17b8ca01f549bc10081dd5f109cb1b9fff364bf5bd4fd90d9f0e3eeb04db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfb06a5ec9262cf39b108a66014d15e325cdbd5441a2d9e7a811dd57265c99a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfb06a5ec9262cf39b108a66014d15e325cdbd5441a2d9e7a811dd57265c99a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:54Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:54 crc kubenswrapper[4764]: I0202 09:07:54.349229 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:54 crc kubenswrapper[4764]: I0202 09:07:54.349306 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:54 crc kubenswrapper[4764]: I0202 09:07:54.349333 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:54 crc kubenswrapper[4764]: I0202 09:07:54.349363 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:54 crc kubenswrapper[4764]: I0202 09:07:54.349384 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:54Z","lastTransitionTime":"2026-02-02T09:07:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:54 crc kubenswrapper[4764]: I0202 09:07:54.352331 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5528225d91ac2778bef4554abb0bd036ac5d3b6cdac7357bab69e70c3894b6f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://071bbb80d0747cf3e109ba626d47f7d05f85f4cc0b3aba77276c795741f47ac6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:54Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:54 crc kubenswrapper[4764]: I0202 09:07:54.363241 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:54Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:54 crc kubenswrapper[4764]: I0202 09:07:54.380099 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d66ab3c0-9b5b-41d5-b6c0-91ccddccd7e9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d92ac486f5dade46bc75de32099859a71fc2ed25a7abbe6ae8937df128e7df30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a4ac273a4f127d79e5fad7d4c30cd2044a346f5734dba5d33d68cd092c0c2d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d38c7cb5dcd22fb07524494ebe77aa1a7f4f808f566043563817e47e7ec12a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://408722c7f83d28919431532f8c797a1c795fd5d96736d14e7f06a05744899702\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f473b18cd8003f75bb1dafb82bcec45745a73ccd10434297356dac185a3d15c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e229f484d19cbdabedf53d90a822a4434db9e00b1301538e885e26acc19d25f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e229f484d19cbdabedf53d90a822a4434db9e00b1301538e885e26acc19d25f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a9cc673be8ff245dba81278e8cd34cfe45ed36ff0b8bd5b30a0ddf333ec6371\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a9cc673be8ff245dba81278e8cd34cfe45ed36ff0b8bd5b30a0ddf333ec6371\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bc7e49c7a5ae66ea939060192fa6a346dc0b0b9aa9750b29879d3c426cc0d276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc7e49c7a5ae66ea939060192fa6a346dc0b0b9aa9750b29879d3c426cc0d276\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:54Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:54 crc kubenswrapper[4764]: I0202 09:07:54.391771 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa9d037e66cb53b9dd698eff46284c9e718fd842cda6f7b525f0c79062d70bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:54Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:54 crc kubenswrapper[4764]: I0202 09:07:54.402136 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:54Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:54 crc kubenswrapper[4764]: I0202 09:07:54.418132 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24632dda-6100-4ab6-a28e-214ddae4360c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a6c67e9b7708cc0d85b0b6ffbbade0d34867e159a9c5487ca2606475c667a50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bffda48d7c881c1470251b501661324859373246cc1a21811bd36ca24887c7aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a989645cf485cd8ff80ffff1a86cd7808e1564eba5667535e161cc68db13b26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20b71d3ee7a381baa51b0b0d0470283022cfede30f491135737b5e82d631d763\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c380383b6a014d1b694a13ddc22f54b20b9e4f9d5c9b9000d76187621dce0a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5529f2bff87753428e9173826fc1880acf96a4c134fa9ecad195b15697cad73a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12244061eabaae7fa9057ac5cee894c3b308b466411bcb910682e7f65d240dd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://12244061eabaae7fa9057ac5cee894c3b308b466411bcb910682e7f65d240dd7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-02T09:07:52Z\\\",\\\"message\\\":\\\"l_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-kube-apiserver/apiserver]} name:Service_openshift-kube-apiserver/apiserver_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.93:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {d71b38eb-32af-4c0f-9490-7c317c111e3a}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0202 09:07:51.914338 6129 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-kube-apiserver/apiserver]} name:Service_openshift-kube-apiserver/apiserver_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.93:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {d71b38eb-32af-4c0f-9490-7c317c111e3a}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF0202 09:07:51.914404 6129 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:51Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-zhn7j_openshift-ovn-kubernetes(24632dda-6100-4ab6-a28e-214ddae4360c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b4cfbd4e682bcdcaff785013bdb72580ce36fced733e235538536827036b93d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://579903eadb183466a2564737e6b678e04308239c0cd75b6808892053b97afbc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://579903eadb183466a2564737e6b678e04308239c0cd75b6808892053b97afbc0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zhn7j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:54Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:54 crc kubenswrapper[4764]: I0202 09:07:54.427118 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pcjmn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6e5ccd3b-b657-47a2-9c4d-bc9adb126533\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72863eef568d92a1abeb3cde2e89630a4f5fb7626c6949448ff5195774c98eab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2f7zl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pcjmn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:54Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:54 crc kubenswrapper[4764]: I0202 09:07:54.437571 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed747fd8-e3ce-4319-a6da-bda0fd51b85a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6f064bdc66470aace3180d4988244026b54009a63cff199e41c5d02a1d18479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://60258f84a1d8dcfe48b7e9688ed50f077e201971f20f3f6663fbf1ba58a4611a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f84be2e08217a2ac3edde858d78b4410376ed9bae3bcabe2a7f9b2fbe2f69c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a69d25d352d7ceac5c735bc1fcf446a1fea6dc22c65a5296cfab828cea17377\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:54Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:54 crc kubenswrapper[4764]: I0202 09:07:54.450367 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 09:07:54 crc kubenswrapper[4764]: E0202 09:07:54.450715 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 09:08:10.450685258 +0000 UTC m=+53.384409376 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:07:54 crc kubenswrapper[4764]: I0202 09:07:54.451687 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:54 crc kubenswrapper[4764]: I0202 09:07:54.451950 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:54 crc kubenswrapper[4764]: I0202 09:07:54.452028 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:54 crc kubenswrapper[4764]: I0202 09:07:54.452102 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:54 crc kubenswrapper[4764]: I0202 09:07:54.452166 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:54Z","lastTransitionTime":"2026-02-02T09:07:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:54 crc kubenswrapper[4764]: I0202 09:07:54.451742 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f6904228e052fff69ad1fd66df97758b6b60ba4351e97a7fbd03eee4df1b653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:54Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:54 crc kubenswrapper[4764]: I0202 09:07:54.461389 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6zxff" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aaba93da-12a5-43df-b7c7-4c8b988163c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3db293678a907cd55b31b1d0c0bd7c30e1ba75ecd5e13cbc68ab98fbf0fbd9d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqwb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6zxff\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:54Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:54 crc kubenswrapper[4764]: I0202 09:07:54.471800 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d192f670-9f9d-4539-9641-e4bed73acdd4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://040c5e8a5659f85f0d52d1a1e064154d998ae104f711101d2204c0b1caebbb30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbksf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fdf06410349b7c99aa166e87fbe80bf68e2ae923635a6a978b7e2cd50383f7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbksf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4ndm4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:54Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:54 crc kubenswrapper[4764]: I0202 09:07:54.551481 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 09:07:54 crc kubenswrapper[4764]: I0202 09:07:54.551539 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 09:07:54 crc kubenswrapper[4764]: I0202 09:07:54.551556 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 09:07:54 crc kubenswrapper[4764]: I0202 09:07:54.551575 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 09:07:54 crc kubenswrapper[4764]: E0202 09:07:54.551696 4764 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 02 09:07:54 crc kubenswrapper[4764]: E0202 09:07:54.551711 4764 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 02 09:07:54 crc kubenswrapper[4764]: E0202 09:07:54.551721 4764 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 02 09:07:54 crc kubenswrapper[4764]: E0202 09:07:54.551771 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-02 09:08:10.551758482 +0000 UTC m=+53.485482570 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 02 09:07:54 crc kubenswrapper[4764]: E0202 09:07:54.551819 4764 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 02 09:07:54 crc kubenswrapper[4764]: E0202 09:07:54.551826 4764 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 02 09:07:54 crc kubenswrapper[4764]: E0202 09:07:54.551835 4764 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 02 09:07:54 crc kubenswrapper[4764]: E0202 09:07:54.551859 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-02 09:08:10.551851174 +0000 UTC m=+53.485575262 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 02 09:07:54 crc kubenswrapper[4764]: E0202 09:07:54.551888 4764 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 02 09:07:54 crc kubenswrapper[4764]: E0202 09:07:54.551907 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-02 09:08:10.551902316 +0000 UTC m=+53.485626404 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 02 09:07:54 crc kubenswrapper[4764]: E0202 09:07:54.551959 4764 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 02 09:07:54 crc kubenswrapper[4764]: E0202 09:07:54.551981 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-02 09:08:10.551975547 +0000 UTC m=+53.485699635 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 02 09:07:54 crc kubenswrapper[4764]: I0202 09:07:54.554977 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:54 crc kubenswrapper[4764]: I0202 09:07:54.555007 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:54 crc kubenswrapper[4764]: I0202 09:07:54.555015 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:54 crc kubenswrapper[4764]: I0202 09:07:54.555030 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:54 crc kubenswrapper[4764]: I0202 09:07:54.555039 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:54Z","lastTransitionTime":"2026-02-02T09:07:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:54 crc kubenswrapper[4764]: I0202 09:07:54.652632 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/991faa9a-dd25-4f49-82bd-ce60cefd4af2-metrics-certs\") pod \"network-metrics-daemon-kwtmr\" (UID: \"991faa9a-dd25-4f49-82bd-ce60cefd4af2\") " pod="openshift-multus/network-metrics-daemon-kwtmr" Feb 02 09:07:54 crc kubenswrapper[4764]: E0202 09:07:54.652873 4764 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 02 09:07:54 crc kubenswrapper[4764]: E0202 09:07:54.653046 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/991faa9a-dd25-4f49-82bd-ce60cefd4af2-metrics-certs podName:991faa9a-dd25-4f49-82bd-ce60cefd4af2 nodeName:}" failed. No retries permitted until 2026-02-02 09:07:55.653015151 +0000 UTC m=+38.586739269 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/991faa9a-dd25-4f49-82bd-ce60cefd4af2-metrics-certs") pod "network-metrics-daemon-kwtmr" (UID: "991faa9a-dd25-4f49-82bd-ce60cefd4af2") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 02 09:07:54 crc kubenswrapper[4764]: I0202 09:07:54.659051 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:54 crc kubenswrapper[4764]: I0202 09:07:54.659094 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:54 crc kubenswrapper[4764]: I0202 09:07:54.659106 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:54 crc kubenswrapper[4764]: I0202 09:07:54.659124 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:54 crc kubenswrapper[4764]: I0202 09:07:54.659142 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:54Z","lastTransitionTime":"2026-02-02T09:07:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:54 crc kubenswrapper[4764]: I0202 09:07:54.762740 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:54 crc kubenswrapper[4764]: I0202 09:07:54.762792 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:54 crc kubenswrapper[4764]: I0202 09:07:54.762803 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:54 crc kubenswrapper[4764]: I0202 09:07:54.762822 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:54 crc kubenswrapper[4764]: I0202 09:07:54.762835 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:54Z","lastTransitionTime":"2026-02-02T09:07:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:54 crc kubenswrapper[4764]: I0202 09:07:54.792975 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-19 10:49:05.220348786 +0000 UTC Feb 02 09:07:54 crc kubenswrapper[4764]: I0202 09:07:54.817468 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" Feb 02 09:07:54 crc kubenswrapper[4764]: I0202 09:07:54.824695 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 09:07:54 crc kubenswrapper[4764]: I0202 09:07:54.824759 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 09:07:54 crc kubenswrapper[4764]: E0202 09:07:54.824830 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 09:07:54 crc kubenswrapper[4764]: E0202 09:07:54.824914 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 09:07:54 crc kubenswrapper[4764]: I0202 09:07:54.824913 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 09:07:54 crc kubenswrapper[4764]: E0202 09:07:54.825241 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 09:07:54 crc kubenswrapper[4764]: I0202 09:07:54.865370 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:54 crc kubenswrapper[4764]: I0202 09:07:54.865421 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:54 crc kubenswrapper[4764]: I0202 09:07:54.865433 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:54 crc kubenswrapper[4764]: I0202 09:07:54.865451 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:54 crc kubenswrapper[4764]: I0202 09:07:54.865465 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:54Z","lastTransitionTime":"2026-02-02T09:07:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:54 crc kubenswrapper[4764]: I0202 09:07:54.967734 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:54 crc kubenswrapper[4764]: I0202 09:07:54.967978 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:54 crc kubenswrapper[4764]: I0202 09:07:54.968051 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:54 crc kubenswrapper[4764]: I0202 09:07:54.968120 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:54 crc kubenswrapper[4764]: I0202 09:07:54.968174 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:54Z","lastTransitionTime":"2026-02-02T09:07:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:55 crc kubenswrapper[4764]: I0202 09:07:55.070994 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:55 crc kubenswrapper[4764]: I0202 09:07:55.071278 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:55 crc kubenswrapper[4764]: I0202 09:07:55.071361 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:55 crc kubenswrapper[4764]: I0202 09:07:55.071451 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:55 crc kubenswrapper[4764]: I0202 09:07:55.071518 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:55Z","lastTransitionTime":"2026-02-02T09:07:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:55 crc kubenswrapper[4764]: I0202 09:07:55.109122 4764 scope.go:117] "RemoveContainer" containerID="12244061eabaae7fa9057ac5cee894c3b308b466411bcb910682e7f65d240dd7" Feb 02 09:07:55 crc kubenswrapper[4764]: E0202 09:07:55.109386 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-zhn7j_openshift-ovn-kubernetes(24632dda-6100-4ab6-a28e-214ddae4360c)\"" pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" podUID="24632dda-6100-4ab6-a28e-214ddae4360c" Feb 02 09:07:55 crc kubenswrapper[4764]: I0202 09:07:55.174444 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:55 crc kubenswrapper[4764]: I0202 09:07:55.174803 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:55 crc kubenswrapper[4764]: I0202 09:07:55.175000 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:55 crc kubenswrapper[4764]: I0202 09:07:55.175183 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:55 crc kubenswrapper[4764]: I0202 09:07:55.175353 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:55Z","lastTransitionTime":"2026-02-02T09:07:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:55 crc kubenswrapper[4764]: I0202 09:07:55.257736 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:55 crc kubenswrapper[4764]: I0202 09:07:55.258081 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:55 crc kubenswrapper[4764]: I0202 09:07:55.258197 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:55 crc kubenswrapper[4764]: I0202 09:07:55.258330 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:55 crc kubenswrapper[4764]: I0202 09:07:55.258424 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:55Z","lastTransitionTime":"2026-02-02T09:07:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:55 crc kubenswrapper[4764]: E0202 09:07:55.279729 4764 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:07:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:07:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:07:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:07:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"31973a70-ce17-4811-9a14-f5476c6979db\\\",\\\"systemUUID\\\":\\\"ca92679c-8580-43c3-be51-62441e93d22b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:55Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:55 crc kubenswrapper[4764]: I0202 09:07:55.285575 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:55 crc kubenswrapper[4764]: I0202 09:07:55.285665 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:55 crc kubenswrapper[4764]: I0202 09:07:55.285690 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:55 crc kubenswrapper[4764]: I0202 09:07:55.285721 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:55 crc kubenswrapper[4764]: I0202 09:07:55.285739 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:55Z","lastTransitionTime":"2026-02-02T09:07:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:55 crc kubenswrapper[4764]: E0202 09:07:55.309291 4764 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:07:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:07:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:07:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:07:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"31973a70-ce17-4811-9a14-f5476c6979db\\\",\\\"systemUUID\\\":\\\"ca92679c-8580-43c3-be51-62441e93d22b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:55Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:55 crc kubenswrapper[4764]: I0202 09:07:55.314984 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:55 crc kubenswrapper[4764]: I0202 09:07:55.315044 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:55 crc kubenswrapper[4764]: I0202 09:07:55.315063 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:55 crc kubenswrapper[4764]: I0202 09:07:55.315089 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:55 crc kubenswrapper[4764]: I0202 09:07:55.315108 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:55Z","lastTransitionTime":"2026-02-02T09:07:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:55 crc kubenswrapper[4764]: E0202 09:07:55.333332 4764 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:07:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:07:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:07:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:07:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"31973a70-ce17-4811-9a14-f5476c6979db\\\",\\\"systemUUID\\\":\\\"ca92679c-8580-43c3-be51-62441e93d22b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:55Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:55 crc kubenswrapper[4764]: I0202 09:07:55.339225 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:55 crc kubenswrapper[4764]: I0202 09:07:55.339264 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:55 crc kubenswrapper[4764]: I0202 09:07:55.339305 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:55 crc kubenswrapper[4764]: I0202 09:07:55.339322 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:55 crc kubenswrapper[4764]: I0202 09:07:55.339334 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:55Z","lastTransitionTime":"2026-02-02T09:07:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:55 crc kubenswrapper[4764]: E0202 09:07:55.356593 4764 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:07:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:07:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:07:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:07:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"31973a70-ce17-4811-9a14-f5476c6979db\\\",\\\"systemUUID\\\":\\\"ca92679c-8580-43c3-be51-62441e93d22b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:55Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:55 crc kubenswrapper[4764]: I0202 09:07:55.361272 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:55 crc kubenswrapper[4764]: I0202 09:07:55.361352 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:55 crc kubenswrapper[4764]: I0202 09:07:55.361380 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:55 crc kubenswrapper[4764]: I0202 09:07:55.361407 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:55 crc kubenswrapper[4764]: I0202 09:07:55.361426 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:55Z","lastTransitionTime":"2026-02-02T09:07:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:55 crc kubenswrapper[4764]: E0202 09:07:55.377662 4764 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:07:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:07:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:07:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:07:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"31973a70-ce17-4811-9a14-f5476c6979db\\\",\\\"systemUUID\\\":\\\"ca92679c-8580-43c3-be51-62441e93d22b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:55Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:55 crc kubenswrapper[4764]: E0202 09:07:55.377814 4764 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 02 09:07:55 crc kubenswrapper[4764]: I0202 09:07:55.379896 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:55 crc kubenswrapper[4764]: I0202 09:07:55.379978 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:55 crc kubenswrapper[4764]: I0202 09:07:55.379993 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:55 crc kubenswrapper[4764]: I0202 09:07:55.380011 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:55 crc kubenswrapper[4764]: I0202 09:07:55.380023 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:55Z","lastTransitionTime":"2026-02-02T09:07:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:55 crc kubenswrapper[4764]: I0202 09:07:55.482247 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:55 crc kubenswrapper[4764]: I0202 09:07:55.482288 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:55 crc kubenswrapper[4764]: I0202 09:07:55.482299 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:55 crc kubenswrapper[4764]: I0202 09:07:55.482315 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:55 crc kubenswrapper[4764]: I0202 09:07:55.482326 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:55Z","lastTransitionTime":"2026-02-02T09:07:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:55 crc kubenswrapper[4764]: I0202 09:07:55.584953 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:55 crc kubenswrapper[4764]: I0202 09:07:55.585008 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:55 crc kubenswrapper[4764]: I0202 09:07:55.585034 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:55 crc kubenswrapper[4764]: I0202 09:07:55.585060 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:55 crc kubenswrapper[4764]: I0202 09:07:55.585076 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:55Z","lastTransitionTime":"2026-02-02T09:07:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:55 crc kubenswrapper[4764]: I0202 09:07:55.664213 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/991faa9a-dd25-4f49-82bd-ce60cefd4af2-metrics-certs\") pod \"network-metrics-daemon-kwtmr\" (UID: \"991faa9a-dd25-4f49-82bd-ce60cefd4af2\") " pod="openshift-multus/network-metrics-daemon-kwtmr" Feb 02 09:07:55 crc kubenswrapper[4764]: E0202 09:07:55.664418 4764 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 02 09:07:55 crc kubenswrapper[4764]: E0202 09:07:55.664509 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/991faa9a-dd25-4f49-82bd-ce60cefd4af2-metrics-certs podName:991faa9a-dd25-4f49-82bd-ce60cefd4af2 nodeName:}" failed. No retries permitted until 2026-02-02 09:07:57.664480719 +0000 UTC m=+40.598204817 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/991faa9a-dd25-4f49-82bd-ce60cefd4af2-metrics-certs") pod "network-metrics-daemon-kwtmr" (UID: "991faa9a-dd25-4f49-82bd-ce60cefd4af2") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 02 09:07:55 crc kubenswrapper[4764]: I0202 09:07:55.687434 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:55 crc kubenswrapper[4764]: I0202 09:07:55.687492 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:55 crc kubenswrapper[4764]: I0202 09:07:55.687506 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:55 crc kubenswrapper[4764]: I0202 09:07:55.687531 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:55 crc kubenswrapper[4764]: I0202 09:07:55.687547 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:55Z","lastTransitionTime":"2026-02-02T09:07:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:55 crc kubenswrapper[4764]: I0202 09:07:55.790862 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:55 crc kubenswrapper[4764]: I0202 09:07:55.790904 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:55 crc kubenswrapper[4764]: I0202 09:07:55.790915 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:55 crc kubenswrapper[4764]: I0202 09:07:55.790928 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:55 crc kubenswrapper[4764]: I0202 09:07:55.790952 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:55Z","lastTransitionTime":"2026-02-02T09:07:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:55 crc kubenswrapper[4764]: I0202 09:07:55.794652 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-16 15:27:56.310849344 +0000 UTC Feb 02 09:07:55 crc kubenswrapper[4764]: I0202 09:07:55.824904 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kwtmr" Feb 02 09:07:55 crc kubenswrapper[4764]: E0202 09:07:55.825076 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kwtmr" podUID="991faa9a-dd25-4f49-82bd-ce60cefd4af2" Feb 02 09:07:55 crc kubenswrapper[4764]: I0202 09:07:55.893959 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:55 crc kubenswrapper[4764]: I0202 09:07:55.894002 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:55 crc kubenswrapper[4764]: I0202 09:07:55.894016 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:55 crc kubenswrapper[4764]: I0202 09:07:55.894037 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:55 crc kubenswrapper[4764]: I0202 09:07:55.894055 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:55Z","lastTransitionTime":"2026-02-02T09:07:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:55 crc kubenswrapper[4764]: I0202 09:07:55.997039 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:55 crc kubenswrapper[4764]: I0202 09:07:55.997108 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:55 crc kubenswrapper[4764]: I0202 09:07:55.997127 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:55 crc kubenswrapper[4764]: I0202 09:07:55.997151 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:55 crc kubenswrapper[4764]: I0202 09:07:55.997173 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:55Z","lastTransitionTime":"2026-02-02T09:07:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:56 crc kubenswrapper[4764]: I0202 09:07:56.100350 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:56 crc kubenswrapper[4764]: I0202 09:07:56.100407 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:56 crc kubenswrapper[4764]: I0202 09:07:56.100419 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:56 crc kubenswrapper[4764]: I0202 09:07:56.100439 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:56 crc kubenswrapper[4764]: I0202 09:07:56.100451 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:56Z","lastTransitionTime":"2026-02-02T09:07:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:56 crc kubenswrapper[4764]: I0202 09:07:56.202601 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:56 crc kubenswrapper[4764]: I0202 09:07:56.202648 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:56 crc kubenswrapper[4764]: I0202 09:07:56.202659 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:56 crc kubenswrapper[4764]: I0202 09:07:56.202673 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:56 crc kubenswrapper[4764]: I0202 09:07:56.202683 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:56Z","lastTransitionTime":"2026-02-02T09:07:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:56 crc kubenswrapper[4764]: I0202 09:07:56.305809 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:56 crc kubenswrapper[4764]: I0202 09:07:56.305870 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:56 crc kubenswrapper[4764]: I0202 09:07:56.305887 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:56 crc kubenswrapper[4764]: I0202 09:07:56.305912 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:56 crc kubenswrapper[4764]: I0202 09:07:56.305960 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:56Z","lastTransitionTime":"2026-02-02T09:07:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:56 crc kubenswrapper[4764]: I0202 09:07:56.408282 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:56 crc kubenswrapper[4764]: I0202 09:07:56.408349 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:56 crc kubenswrapper[4764]: I0202 09:07:56.408360 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:56 crc kubenswrapper[4764]: I0202 09:07:56.408375 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:56 crc kubenswrapper[4764]: I0202 09:07:56.408386 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:56Z","lastTransitionTime":"2026-02-02T09:07:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:56 crc kubenswrapper[4764]: I0202 09:07:56.510373 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:56 crc kubenswrapper[4764]: I0202 09:07:56.510435 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:56 crc kubenswrapper[4764]: I0202 09:07:56.510444 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:56 crc kubenswrapper[4764]: I0202 09:07:56.510476 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:56 crc kubenswrapper[4764]: I0202 09:07:56.510486 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:56Z","lastTransitionTime":"2026-02-02T09:07:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:56 crc kubenswrapper[4764]: I0202 09:07:56.613245 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:56 crc kubenswrapper[4764]: I0202 09:07:56.613282 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:56 crc kubenswrapper[4764]: I0202 09:07:56.613293 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:56 crc kubenswrapper[4764]: I0202 09:07:56.613308 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:56 crc kubenswrapper[4764]: I0202 09:07:56.613319 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:56Z","lastTransitionTime":"2026-02-02T09:07:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:56 crc kubenswrapper[4764]: I0202 09:07:56.717341 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:56 crc kubenswrapper[4764]: I0202 09:07:56.717378 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:56 crc kubenswrapper[4764]: I0202 09:07:56.717390 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:56 crc kubenswrapper[4764]: I0202 09:07:56.717405 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:56 crc kubenswrapper[4764]: I0202 09:07:56.717443 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:56Z","lastTransitionTime":"2026-02-02T09:07:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:56 crc kubenswrapper[4764]: I0202 09:07:56.795008 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-17 17:24:41.610005955 +0000 UTC Feb 02 09:07:56 crc kubenswrapper[4764]: I0202 09:07:56.820474 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:56 crc kubenswrapper[4764]: I0202 09:07:56.820531 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:56 crc kubenswrapper[4764]: I0202 09:07:56.820548 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:56 crc kubenswrapper[4764]: I0202 09:07:56.820572 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:56 crc kubenswrapper[4764]: I0202 09:07:56.820588 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:56Z","lastTransitionTime":"2026-02-02T09:07:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:56 crc kubenswrapper[4764]: I0202 09:07:56.824879 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 09:07:56 crc kubenswrapper[4764]: E0202 09:07:56.825104 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 09:07:56 crc kubenswrapper[4764]: I0202 09:07:56.825403 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 09:07:56 crc kubenswrapper[4764]: I0202 09:07:56.825459 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 09:07:56 crc kubenswrapper[4764]: E0202 09:07:56.825644 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 09:07:56 crc kubenswrapper[4764]: E0202 09:07:56.825742 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 09:07:56 crc kubenswrapper[4764]: I0202 09:07:56.922980 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:56 crc kubenswrapper[4764]: I0202 09:07:56.923037 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:56 crc kubenswrapper[4764]: I0202 09:07:56.923056 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:56 crc kubenswrapper[4764]: I0202 09:07:56.923080 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:56 crc kubenswrapper[4764]: I0202 09:07:56.923099 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:56Z","lastTransitionTime":"2026-02-02T09:07:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:57 crc kubenswrapper[4764]: I0202 09:07:57.026705 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:57 crc kubenswrapper[4764]: I0202 09:07:57.026771 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:57 crc kubenswrapper[4764]: I0202 09:07:57.026783 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:57 crc kubenswrapper[4764]: I0202 09:07:57.026799 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:57 crc kubenswrapper[4764]: I0202 09:07:57.026810 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:57Z","lastTransitionTime":"2026-02-02T09:07:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:57 crc kubenswrapper[4764]: I0202 09:07:57.128858 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:57 crc kubenswrapper[4764]: I0202 09:07:57.128899 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:57 crc kubenswrapper[4764]: I0202 09:07:57.128911 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:57 crc kubenswrapper[4764]: I0202 09:07:57.128927 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:57 crc kubenswrapper[4764]: I0202 09:07:57.128955 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:57Z","lastTransitionTime":"2026-02-02T09:07:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:57 crc kubenswrapper[4764]: I0202 09:07:57.231780 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:57 crc kubenswrapper[4764]: I0202 09:07:57.231828 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:57 crc kubenswrapper[4764]: I0202 09:07:57.231840 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:57 crc kubenswrapper[4764]: I0202 09:07:57.231858 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:57 crc kubenswrapper[4764]: I0202 09:07:57.231871 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:57Z","lastTransitionTime":"2026-02-02T09:07:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:57 crc kubenswrapper[4764]: I0202 09:07:57.334521 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:57 crc kubenswrapper[4764]: I0202 09:07:57.334571 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:57 crc kubenswrapper[4764]: I0202 09:07:57.334582 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:57 crc kubenswrapper[4764]: I0202 09:07:57.334600 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:57 crc kubenswrapper[4764]: I0202 09:07:57.334611 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:57Z","lastTransitionTime":"2026-02-02T09:07:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:57 crc kubenswrapper[4764]: I0202 09:07:57.437299 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:57 crc kubenswrapper[4764]: I0202 09:07:57.437340 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:57 crc kubenswrapper[4764]: I0202 09:07:57.437348 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:57 crc kubenswrapper[4764]: I0202 09:07:57.437362 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:57 crc kubenswrapper[4764]: I0202 09:07:57.437373 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:57Z","lastTransitionTime":"2026-02-02T09:07:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:57 crc kubenswrapper[4764]: I0202 09:07:57.540403 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:57 crc kubenswrapper[4764]: I0202 09:07:57.540454 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:57 crc kubenswrapper[4764]: I0202 09:07:57.540464 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:57 crc kubenswrapper[4764]: I0202 09:07:57.540483 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:57 crc kubenswrapper[4764]: I0202 09:07:57.540495 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:57Z","lastTransitionTime":"2026-02-02T09:07:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:57 crc kubenswrapper[4764]: I0202 09:07:57.643119 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:57 crc kubenswrapper[4764]: I0202 09:07:57.643182 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:57 crc kubenswrapper[4764]: I0202 09:07:57.643201 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:57 crc kubenswrapper[4764]: I0202 09:07:57.643226 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:57 crc kubenswrapper[4764]: I0202 09:07:57.643244 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:57Z","lastTransitionTime":"2026-02-02T09:07:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:57 crc kubenswrapper[4764]: I0202 09:07:57.688055 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/991faa9a-dd25-4f49-82bd-ce60cefd4af2-metrics-certs\") pod \"network-metrics-daemon-kwtmr\" (UID: \"991faa9a-dd25-4f49-82bd-ce60cefd4af2\") " pod="openshift-multus/network-metrics-daemon-kwtmr" Feb 02 09:07:57 crc kubenswrapper[4764]: E0202 09:07:57.688247 4764 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 02 09:07:57 crc kubenswrapper[4764]: E0202 09:07:57.688320 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/991faa9a-dd25-4f49-82bd-ce60cefd4af2-metrics-certs podName:991faa9a-dd25-4f49-82bd-ce60cefd4af2 nodeName:}" failed. No retries permitted until 2026-02-02 09:08:01.688302147 +0000 UTC m=+44.622026255 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/991faa9a-dd25-4f49-82bd-ce60cefd4af2-metrics-certs") pod "network-metrics-daemon-kwtmr" (UID: "991faa9a-dd25-4f49-82bd-ce60cefd4af2") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 02 09:07:57 crc kubenswrapper[4764]: I0202 09:07:57.746178 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:57 crc kubenswrapper[4764]: I0202 09:07:57.746300 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:57 crc kubenswrapper[4764]: I0202 09:07:57.746321 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:57 crc kubenswrapper[4764]: I0202 09:07:57.746397 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:57 crc kubenswrapper[4764]: I0202 09:07:57.746419 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:57Z","lastTransitionTime":"2026-02-02T09:07:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:57 crc kubenswrapper[4764]: I0202 09:07:57.796059 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-29 22:47:28.747869011 +0000 UTC Feb 02 09:07:57 crc kubenswrapper[4764]: I0202 09:07:57.825614 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kwtmr" Feb 02 09:07:57 crc kubenswrapper[4764]: E0202 09:07:57.826033 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kwtmr" podUID="991faa9a-dd25-4f49-82bd-ce60cefd4af2" Feb 02 09:07:57 crc kubenswrapper[4764]: I0202 09:07:57.843979 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6zxff" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aaba93da-12a5-43df-b7c7-4c8b988163c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3db293678a907cd55b31b1d0c0bd7c30e1ba75ecd5e13cbc68ab98fbf0fbd9d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqwb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6zxff\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:57Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:57 crc kubenswrapper[4764]: I0202 09:07:57.849248 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:57 crc kubenswrapper[4764]: I0202 09:07:57.849306 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:57 crc kubenswrapper[4764]: I0202 09:07:57.849324 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:57 crc kubenswrapper[4764]: I0202 09:07:57.849347 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:57 crc kubenswrapper[4764]: I0202 09:07:57.849364 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:57Z","lastTransitionTime":"2026-02-02T09:07:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:57 crc kubenswrapper[4764]: I0202 09:07:57.857702 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d192f670-9f9d-4539-9641-e4bed73acdd4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://040c5e8a5659f85f0d52d1a1e064154d998ae104f711101d2204c0b1caebbb30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbksf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fdf06410349b7c99aa166e87fbe80bf68e2ae923635a6a978b7e2cd50383f7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbksf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4ndm4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:57Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:57 crc kubenswrapper[4764]: I0202 09:07:57.869256 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pcjmn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6e5ccd3b-b657-47a2-9c4d-bc9adb126533\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72863eef568d92a1abeb3cde2e89630a4f5fb7626c6949448ff5195774c98eab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2f7zl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pcjmn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:57Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:57 crc kubenswrapper[4764]: I0202 09:07:57.883254 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed747fd8-e3ce-4319-a6da-bda0fd51b85a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6f064bdc66470aace3180d4988244026b54009a63cff199e41c5d02a1d18479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://60258f84a1d8dcfe48b7e9688ed50f077e201971f20f3f6663fbf1ba58a4611a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f84be2e08217a2ac3edde858d78b4410376ed9bae3bcabe2a7f9b2fbe2f69c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a69d25d352d7ceac5c735bc1fcf446a1fea6dc22c65a5296cfab828cea17377\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:57Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:57 crc kubenswrapper[4764]: I0202 09:07:57.898254 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f6904228e052fff69ad1fd66df97758b6b60ba4351e97a7fbd03eee4df1b653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:57Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:57 crc kubenswrapper[4764]: I0202 09:07:57.917628 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lh6tm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2152e8a-f456-47b5-a547-b65ec837ecbc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38e176a00b07f2960fdd7ba2aadd1468a5e083b8d154e9d118d13f2585da81e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bb2sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://996e649011d5255783284421ddea8e0a5c383ef675b0920c3fdc0af6503d80e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bb2sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lh6tm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:57Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:57 crc kubenswrapper[4764]: I0202 09:07:57.931807 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:57Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:57 crc kubenswrapper[4764]: I0202 09:07:57.949083 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5528225d91ac2778bef4554abb0bd036ac5d3b6cdac7357bab69e70c3894b6f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://071bbb80d0747cf3e109ba626d47f7d05f85f4cc0b3aba77276c795741f47ac6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:57Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:57 crc kubenswrapper[4764]: I0202 09:07:57.952128 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:57 crc kubenswrapper[4764]: I0202 09:07:57.952174 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:57 crc kubenswrapper[4764]: I0202 09:07:57.952186 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:57 crc kubenswrapper[4764]: I0202 09:07:57.952202 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:57 crc kubenswrapper[4764]: I0202 09:07:57.952217 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:57Z","lastTransitionTime":"2026-02-02T09:07:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:57 crc kubenswrapper[4764]: I0202 09:07:57.968190 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:57Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:57 crc kubenswrapper[4764]: I0202 09:07:57.984844 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nc7qt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1a2c7b22068b7165927fc158885d48bf9952b9eba521d8ae9a599eee53dc305\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2hz4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nc7qt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:57Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:58 crc kubenswrapper[4764]: I0202 09:07:58.002916 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9p5dc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"660ac006-2495-45c4-9fb3-e9c1dddcf7a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da812b124c97aa8ad1563f4bcfac8f3bfa025096fa4a2561e5847b1bf9cb30b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23e57e65a6f5c3059506b9db21e9ac2492c30bd9191a98db0281915e9ce39797\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23e57e65a6f5c3059506b9db21e9ac2492c30bd9191a98db0281915e9ce39797\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2417b66816da1a6cd1791a4b91eca4a3c4f3694462e066e7f0722f1b9c2105c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2417b66816da1a6cd1791a4b91eca4a3c4f3694462e066e7f0722f1b9c2105c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://883641f29e9120c2cea6b272bacf1dc63c752f55ab0408cdaf73db85c85ac369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://883641f29e9120c2cea6b272bacf1dc63c752f55ab0408cdaf73db85c85ac369\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2745baba1188ba5e8f820616e4117a12c50ecee5b68e71ad9cabbb621b6e131f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2745baba1188ba5e8f820616e4117a12c50ecee5b68e71ad9cabbb621b6e131f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8925e435be4d48d86824fc35f8b8819fb35d63e385164c4a20edb3bf3035c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a8925e435be4d48d86824fc35f8b8819fb35d63e385164c4a20edb3bf3035c3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d3fbf668ae7d04c8d9e574f43c2fe90c5604204ba38a3c60847afdfbceb7a57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d3fbf668ae7d04c8d9e574f43c2fe90c5604204ba38a3c60847afdfbceb7a57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9p5dc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:57Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:58 crc kubenswrapper[4764]: I0202 09:07:58.016570 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kwtmr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"991faa9a-dd25-4f49-82bd-ce60cefd4af2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h865b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h865b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:53Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kwtmr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:58Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:58 crc kubenswrapper[4764]: I0202 09:07:58.032222 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87d61aea-46f6-4ee6-a7a4-767a6a67fea4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e78b6d08e4d616cefbb73e3fd43b3c6886dcd8aa997c7b4948bf1ff63401293\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d468715c2da33996561c40043216233fd12e21539d0c58dd1f3b892fd9f91a9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b184f7c936dc4046f74c0018269cbc2c4b7ceddb650b4e871d3f8a707e0df88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://269c5074371bd00f8363fb3888c11f1df383081a61497aea6578f08700881d50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b25bc9395831524c53f9e565b4efbe74f21792efafeb1dd507161bc91aa0ecdf\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T09:07:22Z\\\",\\\"message\\\":\\\"W0202 09:07:21.597488 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0202 09:07:21.598117 1 crypto.go:601] Generating new CA for check-endpoints-signer@1770023241 cert, and key in /tmp/serving-cert-3856183724/serving-signer.crt, /tmp/serving-cert-3856183724/serving-signer.key\\\\nI0202 09:07:22.012184 1 observer_polling.go:159] Starting file observer\\\\nW0202 09:07:22.015493 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0202 09:07:22.015797 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 09:07:22.016596 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3856183724/tls.crt::/tmp/serving-cert-3856183724/tls.key\\\\\\\"\\\\nF0202 09:07:22.270605 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b44e17b8ca01f549bc10081dd5f109cb1b9fff364bf5bd4fd90d9f0e3eeb04db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfb06a5ec9262cf39b108a66014d15e325cdbd5441a2d9e7a811dd57265c99a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfb06a5ec9262cf39b108a66014d15e325cdbd5441a2d9e7a811dd57265c99a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:58Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:58 crc kubenswrapper[4764]: I0202 09:07:58.044021 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:58Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:58 crc kubenswrapper[4764]: I0202 09:07:58.054445 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:58 crc kubenswrapper[4764]: I0202 09:07:58.054475 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:58 crc kubenswrapper[4764]: I0202 09:07:58.054484 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:58 crc kubenswrapper[4764]: I0202 09:07:58.054496 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:58 crc kubenswrapper[4764]: I0202 09:07:58.054508 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:58Z","lastTransitionTime":"2026-02-02T09:07:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:58 crc kubenswrapper[4764]: I0202 09:07:58.063437 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24632dda-6100-4ab6-a28e-214ddae4360c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a6c67e9b7708cc0d85b0b6ffbbade0d34867e159a9c5487ca2606475c667a50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bffda48d7c881c1470251b501661324859373246cc1a21811bd36ca24887c7aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a989645cf485cd8ff80ffff1a86cd7808e1564eba5667535e161cc68db13b26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20b71d3ee7a381baa51b0b0d0470283022cfede30f491135737b5e82d631d763\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c380383b6a014d1b694a13ddc22f54b20b9e4f9d5c9b9000d76187621dce0a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5529f2bff87753428e9173826fc1880acf96a4c134fa9ecad195b15697cad73a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12244061eabaae7fa9057ac5cee894c3b308b466411bcb910682e7f65d240dd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://12244061eabaae7fa9057ac5cee894c3b308b466411bcb910682e7f65d240dd7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-02T09:07:52Z\\\",\\\"message\\\":\\\"l_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-kube-apiserver/apiserver]} name:Service_openshift-kube-apiserver/apiserver_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.93:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {d71b38eb-32af-4c0f-9490-7c317c111e3a}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0202 09:07:51.914338 6129 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-kube-apiserver/apiserver]} name:Service_openshift-kube-apiserver/apiserver_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.93:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {d71b38eb-32af-4c0f-9490-7c317c111e3a}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF0202 09:07:51.914404 6129 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:51Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-zhn7j_openshift-ovn-kubernetes(24632dda-6100-4ab6-a28e-214ddae4360c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b4cfbd4e682bcdcaff785013bdb72580ce36fced733e235538536827036b93d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://579903eadb183466a2564737e6b678e04308239c0cd75b6808892053b97afbc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://579903eadb183466a2564737e6b678e04308239c0cd75b6808892053b97afbc0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zhn7j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:58Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:58 crc kubenswrapper[4764]: I0202 09:07:58.082490 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d66ab3c0-9b5b-41d5-b6c0-91ccddccd7e9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d92ac486f5dade46bc75de32099859a71fc2ed25a7abbe6ae8937df128e7df30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a4ac273a4f127d79e5fad7d4c30cd2044a346f5734dba5d33d68cd092c0c2d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d38c7cb5dcd22fb07524494ebe77aa1a7f4f808f566043563817e47e7ec12a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://408722c7f83d28919431532f8c797a1c795fd5d96736d14e7f06a05744899702\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f473b18cd8003f75bb1dafb82bcec45745a73ccd10434297356dac185a3d15c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e229f484d19cbdabedf53d90a822a4434db9e00b1301538e885e26acc19d25f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e229f484d19cbdabedf53d90a822a4434db9e00b1301538e885e26acc19d25f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a9cc673be8ff245dba81278e8cd34cfe45ed36ff0b8bd5b30a0ddf333ec6371\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a9cc673be8ff245dba81278e8cd34cfe45ed36ff0b8bd5b30a0ddf333ec6371\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bc7e49c7a5ae66ea939060192fa6a346dc0b0b9aa9750b29879d3c426cc0d276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc7e49c7a5ae66ea939060192fa6a346dc0b0b9aa9750b29879d3c426cc0d276\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:58Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:58 crc kubenswrapper[4764]: I0202 09:07:58.094944 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa9d037e66cb53b9dd698eff46284c9e718fd842cda6f7b525f0c79062d70bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:07:58Z is after 2025-08-24T17:21:41Z" Feb 02 09:07:58 crc kubenswrapper[4764]: I0202 09:07:58.156801 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:58 crc kubenswrapper[4764]: I0202 09:07:58.156844 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:58 crc kubenswrapper[4764]: I0202 09:07:58.156853 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:58 crc kubenswrapper[4764]: I0202 09:07:58.156884 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:58 crc kubenswrapper[4764]: I0202 09:07:58.156896 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:58Z","lastTransitionTime":"2026-02-02T09:07:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:58 crc kubenswrapper[4764]: I0202 09:07:58.259376 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:58 crc kubenswrapper[4764]: I0202 09:07:58.259421 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:58 crc kubenswrapper[4764]: I0202 09:07:58.259433 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:58 crc kubenswrapper[4764]: I0202 09:07:58.259447 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:58 crc kubenswrapper[4764]: I0202 09:07:58.259457 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:58Z","lastTransitionTime":"2026-02-02T09:07:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:58 crc kubenswrapper[4764]: I0202 09:07:58.363015 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:58 crc kubenswrapper[4764]: I0202 09:07:58.363305 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:58 crc kubenswrapper[4764]: I0202 09:07:58.363473 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:58 crc kubenswrapper[4764]: I0202 09:07:58.363656 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:58 crc kubenswrapper[4764]: I0202 09:07:58.363913 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:58Z","lastTransitionTime":"2026-02-02T09:07:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:58 crc kubenswrapper[4764]: I0202 09:07:58.470325 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:58 crc kubenswrapper[4764]: I0202 09:07:58.470410 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:58 crc kubenswrapper[4764]: I0202 09:07:58.470468 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:58 crc kubenswrapper[4764]: I0202 09:07:58.470499 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:58 crc kubenswrapper[4764]: I0202 09:07:58.470533 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:58Z","lastTransitionTime":"2026-02-02T09:07:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:58 crc kubenswrapper[4764]: I0202 09:07:58.573556 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:58 crc kubenswrapper[4764]: I0202 09:07:58.573898 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:58 crc kubenswrapper[4764]: I0202 09:07:58.574084 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:58 crc kubenswrapper[4764]: I0202 09:07:58.574302 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:58 crc kubenswrapper[4764]: I0202 09:07:58.574412 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:58Z","lastTransitionTime":"2026-02-02T09:07:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:58 crc kubenswrapper[4764]: I0202 09:07:58.677355 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:58 crc kubenswrapper[4764]: I0202 09:07:58.677430 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:58 crc kubenswrapper[4764]: I0202 09:07:58.677448 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:58 crc kubenswrapper[4764]: I0202 09:07:58.677471 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:58 crc kubenswrapper[4764]: I0202 09:07:58.677487 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:58Z","lastTransitionTime":"2026-02-02T09:07:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:58 crc kubenswrapper[4764]: I0202 09:07:58.780085 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:58 crc kubenswrapper[4764]: I0202 09:07:58.780128 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:58 crc kubenswrapper[4764]: I0202 09:07:58.780138 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:58 crc kubenswrapper[4764]: I0202 09:07:58.780151 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:58 crc kubenswrapper[4764]: I0202 09:07:58.780175 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:58Z","lastTransitionTime":"2026-02-02T09:07:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:58 crc kubenswrapper[4764]: I0202 09:07:58.796570 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-07 13:46:11.195039687 +0000 UTC Feb 02 09:07:58 crc kubenswrapper[4764]: I0202 09:07:58.825063 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 09:07:58 crc kubenswrapper[4764]: I0202 09:07:58.825079 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 09:07:58 crc kubenswrapper[4764]: E0202 09:07:58.825201 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 09:07:58 crc kubenswrapper[4764]: I0202 09:07:58.825083 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 09:07:58 crc kubenswrapper[4764]: E0202 09:07:58.825299 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 09:07:58 crc kubenswrapper[4764]: E0202 09:07:58.825360 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 09:07:58 crc kubenswrapper[4764]: I0202 09:07:58.883182 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:58 crc kubenswrapper[4764]: I0202 09:07:58.883271 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:58 crc kubenswrapper[4764]: I0202 09:07:58.883299 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:58 crc kubenswrapper[4764]: I0202 09:07:58.883328 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:58 crc kubenswrapper[4764]: I0202 09:07:58.883349 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:58Z","lastTransitionTime":"2026-02-02T09:07:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:58 crc kubenswrapper[4764]: I0202 09:07:58.986586 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:58 crc kubenswrapper[4764]: I0202 09:07:58.986711 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:58 crc kubenswrapper[4764]: I0202 09:07:58.986729 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:58 crc kubenswrapper[4764]: I0202 09:07:58.986752 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:58 crc kubenswrapper[4764]: I0202 09:07:58.986769 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:58Z","lastTransitionTime":"2026-02-02T09:07:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:59 crc kubenswrapper[4764]: I0202 09:07:59.089021 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:59 crc kubenswrapper[4764]: I0202 09:07:59.089066 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:59 crc kubenswrapper[4764]: I0202 09:07:59.089076 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:59 crc kubenswrapper[4764]: I0202 09:07:59.089089 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:59 crc kubenswrapper[4764]: I0202 09:07:59.089098 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:59Z","lastTransitionTime":"2026-02-02T09:07:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:59 crc kubenswrapper[4764]: I0202 09:07:59.191861 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:59 crc kubenswrapper[4764]: I0202 09:07:59.191894 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:59 crc kubenswrapper[4764]: I0202 09:07:59.191903 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:59 crc kubenswrapper[4764]: I0202 09:07:59.191915 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:59 crc kubenswrapper[4764]: I0202 09:07:59.191923 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:59Z","lastTransitionTime":"2026-02-02T09:07:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:59 crc kubenswrapper[4764]: I0202 09:07:59.294764 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:59 crc kubenswrapper[4764]: I0202 09:07:59.294843 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:59 crc kubenswrapper[4764]: I0202 09:07:59.294860 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:59 crc kubenswrapper[4764]: I0202 09:07:59.294885 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:59 crc kubenswrapper[4764]: I0202 09:07:59.294901 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:59Z","lastTransitionTime":"2026-02-02T09:07:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:59 crc kubenswrapper[4764]: I0202 09:07:59.397607 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:59 crc kubenswrapper[4764]: I0202 09:07:59.397833 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:59 crc kubenswrapper[4764]: I0202 09:07:59.397852 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:59 crc kubenswrapper[4764]: I0202 09:07:59.397874 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:59 crc kubenswrapper[4764]: I0202 09:07:59.397893 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:59Z","lastTransitionTime":"2026-02-02T09:07:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:59 crc kubenswrapper[4764]: I0202 09:07:59.500551 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:59 crc kubenswrapper[4764]: I0202 09:07:59.500627 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:59 crc kubenswrapper[4764]: I0202 09:07:59.500647 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:59 crc kubenswrapper[4764]: I0202 09:07:59.500680 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:59 crc kubenswrapper[4764]: I0202 09:07:59.500702 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:59Z","lastTransitionTime":"2026-02-02T09:07:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:59 crc kubenswrapper[4764]: I0202 09:07:59.603324 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:59 crc kubenswrapper[4764]: I0202 09:07:59.603383 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:59 crc kubenswrapper[4764]: I0202 09:07:59.603400 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:59 crc kubenswrapper[4764]: I0202 09:07:59.603424 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:59 crc kubenswrapper[4764]: I0202 09:07:59.603442 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:59Z","lastTransitionTime":"2026-02-02T09:07:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:59 crc kubenswrapper[4764]: I0202 09:07:59.705827 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:59 crc kubenswrapper[4764]: I0202 09:07:59.705907 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:59 crc kubenswrapper[4764]: I0202 09:07:59.705930 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:59 crc kubenswrapper[4764]: I0202 09:07:59.706005 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:59 crc kubenswrapper[4764]: I0202 09:07:59.706029 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:59Z","lastTransitionTime":"2026-02-02T09:07:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:59 crc kubenswrapper[4764]: I0202 09:07:59.797294 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-01 22:12:04.772273935 +0000 UTC Feb 02 09:07:59 crc kubenswrapper[4764]: I0202 09:07:59.809522 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:59 crc kubenswrapper[4764]: I0202 09:07:59.809620 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:59 crc kubenswrapper[4764]: I0202 09:07:59.809645 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:59 crc kubenswrapper[4764]: I0202 09:07:59.809679 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:59 crc kubenswrapper[4764]: I0202 09:07:59.809725 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:59Z","lastTransitionTime":"2026-02-02T09:07:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:07:59 crc kubenswrapper[4764]: I0202 09:07:59.825065 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kwtmr" Feb 02 09:07:59 crc kubenswrapper[4764]: E0202 09:07:59.825369 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kwtmr" podUID="991faa9a-dd25-4f49-82bd-ce60cefd4af2" Feb 02 09:07:59 crc kubenswrapper[4764]: I0202 09:07:59.912468 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:07:59 crc kubenswrapper[4764]: I0202 09:07:59.912562 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:07:59 crc kubenswrapper[4764]: I0202 09:07:59.912580 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:07:59 crc kubenswrapper[4764]: I0202 09:07:59.912648 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:07:59 crc kubenswrapper[4764]: I0202 09:07:59.912667 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:07:59Z","lastTransitionTime":"2026-02-02T09:07:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:00 crc kubenswrapper[4764]: I0202 09:08:00.015950 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:00 crc kubenswrapper[4764]: I0202 09:08:00.015994 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:00 crc kubenswrapper[4764]: I0202 09:08:00.016004 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:00 crc kubenswrapper[4764]: I0202 09:08:00.016021 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:00 crc kubenswrapper[4764]: I0202 09:08:00.016033 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:00Z","lastTransitionTime":"2026-02-02T09:08:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:00 crc kubenswrapper[4764]: I0202 09:08:00.119489 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:00 crc kubenswrapper[4764]: I0202 09:08:00.119530 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:00 crc kubenswrapper[4764]: I0202 09:08:00.119545 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:00 crc kubenswrapper[4764]: I0202 09:08:00.119564 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:00 crc kubenswrapper[4764]: I0202 09:08:00.119578 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:00Z","lastTransitionTime":"2026-02-02T09:08:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:00 crc kubenswrapper[4764]: I0202 09:08:00.222921 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:00 crc kubenswrapper[4764]: I0202 09:08:00.223020 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:00 crc kubenswrapper[4764]: I0202 09:08:00.223037 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:00 crc kubenswrapper[4764]: I0202 09:08:00.223064 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:00 crc kubenswrapper[4764]: I0202 09:08:00.223082 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:00Z","lastTransitionTime":"2026-02-02T09:08:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:00 crc kubenswrapper[4764]: I0202 09:08:00.326012 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:00 crc kubenswrapper[4764]: I0202 09:08:00.326080 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:00 crc kubenswrapper[4764]: I0202 09:08:00.326103 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:00 crc kubenswrapper[4764]: I0202 09:08:00.326131 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:00 crc kubenswrapper[4764]: I0202 09:08:00.326152 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:00Z","lastTransitionTime":"2026-02-02T09:08:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:00 crc kubenswrapper[4764]: I0202 09:08:00.430098 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:00 crc kubenswrapper[4764]: I0202 09:08:00.430137 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:00 crc kubenswrapper[4764]: I0202 09:08:00.430156 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:00 crc kubenswrapper[4764]: I0202 09:08:00.430170 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:00 crc kubenswrapper[4764]: I0202 09:08:00.430181 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:00Z","lastTransitionTime":"2026-02-02T09:08:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:00 crc kubenswrapper[4764]: I0202 09:08:00.532414 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:00 crc kubenswrapper[4764]: I0202 09:08:00.532455 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:00 crc kubenswrapper[4764]: I0202 09:08:00.532472 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:00 crc kubenswrapper[4764]: I0202 09:08:00.532489 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:00 crc kubenswrapper[4764]: I0202 09:08:00.532502 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:00Z","lastTransitionTime":"2026-02-02T09:08:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:00 crc kubenswrapper[4764]: I0202 09:08:00.635871 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:00 crc kubenswrapper[4764]: I0202 09:08:00.635921 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:00 crc kubenswrapper[4764]: I0202 09:08:00.635958 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:00 crc kubenswrapper[4764]: I0202 09:08:00.635982 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:00 crc kubenswrapper[4764]: I0202 09:08:00.635996 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:00Z","lastTransitionTime":"2026-02-02T09:08:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:00 crc kubenswrapper[4764]: I0202 09:08:00.738820 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:00 crc kubenswrapper[4764]: I0202 09:08:00.738888 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:00 crc kubenswrapper[4764]: I0202 09:08:00.738910 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:00 crc kubenswrapper[4764]: I0202 09:08:00.738980 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:00 crc kubenswrapper[4764]: I0202 09:08:00.739005 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:00Z","lastTransitionTime":"2026-02-02T09:08:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:00 crc kubenswrapper[4764]: I0202 09:08:00.798003 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-17 16:49:20.024496324 +0000 UTC Feb 02 09:08:00 crc kubenswrapper[4764]: I0202 09:08:00.825025 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 09:08:00 crc kubenswrapper[4764]: E0202 09:08:00.825190 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 09:08:00 crc kubenswrapper[4764]: I0202 09:08:00.825282 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 09:08:00 crc kubenswrapper[4764]: E0202 09:08:00.825405 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 09:08:00 crc kubenswrapper[4764]: I0202 09:08:00.825428 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 09:08:00 crc kubenswrapper[4764]: E0202 09:08:00.825724 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 09:08:00 crc kubenswrapper[4764]: I0202 09:08:00.844027 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:00 crc kubenswrapper[4764]: I0202 09:08:00.844088 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:00 crc kubenswrapper[4764]: I0202 09:08:00.844107 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:00 crc kubenswrapper[4764]: I0202 09:08:00.844130 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:00 crc kubenswrapper[4764]: I0202 09:08:00.844149 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:00Z","lastTransitionTime":"2026-02-02T09:08:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:00 crc kubenswrapper[4764]: I0202 09:08:00.947232 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:00 crc kubenswrapper[4764]: I0202 09:08:00.947885 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:00 crc kubenswrapper[4764]: I0202 09:08:00.947947 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:00 crc kubenswrapper[4764]: I0202 09:08:00.947965 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:00 crc kubenswrapper[4764]: I0202 09:08:00.947976 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:00Z","lastTransitionTime":"2026-02-02T09:08:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:01 crc kubenswrapper[4764]: I0202 09:08:01.051013 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:01 crc kubenswrapper[4764]: I0202 09:08:01.051050 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:01 crc kubenswrapper[4764]: I0202 09:08:01.051060 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:01 crc kubenswrapper[4764]: I0202 09:08:01.051074 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:01 crc kubenswrapper[4764]: I0202 09:08:01.051083 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:01Z","lastTransitionTime":"2026-02-02T09:08:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:01 crc kubenswrapper[4764]: I0202 09:08:01.153510 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:01 crc kubenswrapper[4764]: I0202 09:08:01.153550 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:01 crc kubenswrapper[4764]: I0202 09:08:01.153560 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:01 crc kubenswrapper[4764]: I0202 09:08:01.153574 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:01 crc kubenswrapper[4764]: I0202 09:08:01.153583 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:01Z","lastTransitionTime":"2026-02-02T09:08:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:01 crc kubenswrapper[4764]: I0202 09:08:01.256256 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:01 crc kubenswrapper[4764]: I0202 09:08:01.256293 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:01 crc kubenswrapper[4764]: I0202 09:08:01.256304 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:01 crc kubenswrapper[4764]: I0202 09:08:01.256318 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:01 crc kubenswrapper[4764]: I0202 09:08:01.256328 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:01Z","lastTransitionTime":"2026-02-02T09:08:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:01 crc kubenswrapper[4764]: I0202 09:08:01.358208 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:01 crc kubenswrapper[4764]: I0202 09:08:01.358247 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:01 crc kubenswrapper[4764]: I0202 09:08:01.358255 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:01 crc kubenswrapper[4764]: I0202 09:08:01.358269 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:01 crc kubenswrapper[4764]: I0202 09:08:01.358281 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:01Z","lastTransitionTime":"2026-02-02T09:08:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:01 crc kubenswrapper[4764]: I0202 09:08:01.461570 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:01 crc kubenswrapper[4764]: I0202 09:08:01.461650 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:01 crc kubenswrapper[4764]: I0202 09:08:01.461673 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:01 crc kubenswrapper[4764]: I0202 09:08:01.461706 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:01 crc kubenswrapper[4764]: I0202 09:08:01.461735 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:01Z","lastTransitionTime":"2026-02-02T09:08:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:01 crc kubenswrapper[4764]: I0202 09:08:01.564275 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:01 crc kubenswrapper[4764]: I0202 09:08:01.564343 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:01 crc kubenswrapper[4764]: I0202 09:08:01.564369 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:01 crc kubenswrapper[4764]: I0202 09:08:01.564397 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:01 crc kubenswrapper[4764]: I0202 09:08:01.564419 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:01Z","lastTransitionTime":"2026-02-02T09:08:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:01 crc kubenswrapper[4764]: I0202 09:08:01.667413 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:01 crc kubenswrapper[4764]: I0202 09:08:01.667451 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:01 crc kubenswrapper[4764]: I0202 09:08:01.667479 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:01 crc kubenswrapper[4764]: I0202 09:08:01.667493 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:01 crc kubenswrapper[4764]: I0202 09:08:01.667504 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:01Z","lastTransitionTime":"2026-02-02T09:08:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:01 crc kubenswrapper[4764]: I0202 09:08:01.730172 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/991faa9a-dd25-4f49-82bd-ce60cefd4af2-metrics-certs\") pod \"network-metrics-daemon-kwtmr\" (UID: \"991faa9a-dd25-4f49-82bd-ce60cefd4af2\") " pod="openshift-multus/network-metrics-daemon-kwtmr" Feb 02 09:08:01 crc kubenswrapper[4764]: E0202 09:08:01.730409 4764 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 02 09:08:01 crc kubenswrapper[4764]: E0202 09:08:01.730513 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/991faa9a-dd25-4f49-82bd-ce60cefd4af2-metrics-certs podName:991faa9a-dd25-4f49-82bd-ce60cefd4af2 nodeName:}" failed. No retries permitted until 2026-02-02 09:08:09.7304896 +0000 UTC m=+52.664213758 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/991faa9a-dd25-4f49-82bd-ce60cefd4af2-metrics-certs") pod "network-metrics-daemon-kwtmr" (UID: "991faa9a-dd25-4f49-82bd-ce60cefd4af2") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 02 09:08:01 crc kubenswrapper[4764]: I0202 09:08:01.778798 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:01 crc kubenswrapper[4764]: I0202 09:08:01.778854 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:01 crc kubenswrapper[4764]: I0202 09:08:01.778864 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:01 crc kubenswrapper[4764]: I0202 09:08:01.778880 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:01 crc kubenswrapper[4764]: I0202 09:08:01.778907 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:01Z","lastTransitionTime":"2026-02-02T09:08:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:01 crc kubenswrapper[4764]: I0202 09:08:01.798435 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-01 13:32:47.974932454 +0000 UTC Feb 02 09:08:01 crc kubenswrapper[4764]: I0202 09:08:01.825069 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kwtmr" Feb 02 09:08:01 crc kubenswrapper[4764]: E0202 09:08:01.825317 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kwtmr" podUID="991faa9a-dd25-4f49-82bd-ce60cefd4af2" Feb 02 09:08:01 crc kubenswrapper[4764]: I0202 09:08:01.881744 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:01 crc kubenswrapper[4764]: I0202 09:08:01.882227 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:01 crc kubenswrapper[4764]: I0202 09:08:01.882312 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:01 crc kubenswrapper[4764]: I0202 09:08:01.882412 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:01 crc kubenswrapper[4764]: I0202 09:08:01.882495 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:01Z","lastTransitionTime":"2026-02-02T09:08:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:01 crc kubenswrapper[4764]: I0202 09:08:01.986177 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:01 crc kubenswrapper[4764]: I0202 09:08:01.986233 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:01 crc kubenswrapper[4764]: I0202 09:08:01.986243 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:01 crc kubenswrapper[4764]: I0202 09:08:01.986264 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:01 crc kubenswrapper[4764]: I0202 09:08:01.986276 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:01Z","lastTransitionTime":"2026-02-02T09:08:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:02 crc kubenswrapper[4764]: I0202 09:08:02.088700 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:02 crc kubenswrapper[4764]: I0202 09:08:02.088754 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:02 crc kubenswrapper[4764]: I0202 09:08:02.088766 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:02 crc kubenswrapper[4764]: I0202 09:08:02.088783 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:02 crc kubenswrapper[4764]: I0202 09:08:02.088795 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:02Z","lastTransitionTime":"2026-02-02T09:08:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:02 crc kubenswrapper[4764]: I0202 09:08:02.190478 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:02 crc kubenswrapper[4764]: I0202 09:08:02.190672 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:02 crc kubenswrapper[4764]: I0202 09:08:02.190760 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:02 crc kubenswrapper[4764]: I0202 09:08:02.190838 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:02 crc kubenswrapper[4764]: I0202 09:08:02.190913 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:02Z","lastTransitionTime":"2026-02-02T09:08:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:02 crc kubenswrapper[4764]: I0202 09:08:02.292522 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:02 crc kubenswrapper[4764]: I0202 09:08:02.292555 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:02 crc kubenswrapper[4764]: I0202 09:08:02.292566 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:02 crc kubenswrapper[4764]: I0202 09:08:02.292580 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:02 crc kubenswrapper[4764]: I0202 09:08:02.292591 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:02Z","lastTransitionTime":"2026-02-02T09:08:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:02 crc kubenswrapper[4764]: I0202 09:08:02.394988 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:02 crc kubenswrapper[4764]: I0202 09:08:02.395021 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:02 crc kubenswrapper[4764]: I0202 09:08:02.395029 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:02 crc kubenswrapper[4764]: I0202 09:08:02.395041 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:02 crc kubenswrapper[4764]: I0202 09:08:02.395050 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:02Z","lastTransitionTime":"2026-02-02T09:08:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:02 crc kubenswrapper[4764]: I0202 09:08:02.498202 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:02 crc kubenswrapper[4764]: I0202 09:08:02.498274 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:02 crc kubenswrapper[4764]: I0202 09:08:02.498308 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:02 crc kubenswrapper[4764]: I0202 09:08:02.498343 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:02 crc kubenswrapper[4764]: I0202 09:08:02.498364 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:02Z","lastTransitionTime":"2026-02-02T09:08:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:02 crc kubenswrapper[4764]: I0202 09:08:02.601783 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:02 crc kubenswrapper[4764]: I0202 09:08:02.601833 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:02 crc kubenswrapper[4764]: I0202 09:08:02.601845 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:02 crc kubenswrapper[4764]: I0202 09:08:02.601863 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:02 crc kubenswrapper[4764]: I0202 09:08:02.601879 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:02Z","lastTransitionTime":"2026-02-02T09:08:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:02 crc kubenswrapper[4764]: I0202 09:08:02.703911 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:02 crc kubenswrapper[4764]: I0202 09:08:02.703976 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:02 crc kubenswrapper[4764]: I0202 09:08:02.703987 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:02 crc kubenswrapper[4764]: I0202 09:08:02.704004 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:02 crc kubenswrapper[4764]: I0202 09:08:02.704015 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:02Z","lastTransitionTime":"2026-02-02T09:08:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:02 crc kubenswrapper[4764]: I0202 09:08:02.798550 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-14 07:43:21.135398838 +0000 UTC Feb 02 09:08:02 crc kubenswrapper[4764]: I0202 09:08:02.806764 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:02 crc kubenswrapper[4764]: I0202 09:08:02.807212 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:02 crc kubenswrapper[4764]: I0202 09:08:02.807417 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:02 crc kubenswrapper[4764]: I0202 09:08:02.807625 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:02 crc kubenswrapper[4764]: I0202 09:08:02.807854 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:02Z","lastTransitionTime":"2026-02-02T09:08:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:02 crc kubenswrapper[4764]: I0202 09:08:02.825226 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 09:08:02 crc kubenswrapper[4764]: E0202 09:08:02.825304 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 09:08:02 crc kubenswrapper[4764]: I0202 09:08:02.825473 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 09:08:02 crc kubenswrapper[4764]: I0202 09:08:02.825492 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 09:08:02 crc kubenswrapper[4764]: E0202 09:08:02.825766 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 09:08:02 crc kubenswrapper[4764]: E0202 09:08:02.825892 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 09:08:02 crc kubenswrapper[4764]: I0202 09:08:02.912337 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:02 crc kubenswrapper[4764]: I0202 09:08:02.912446 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:02 crc kubenswrapper[4764]: I0202 09:08:02.912474 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:02 crc kubenswrapper[4764]: I0202 09:08:02.912513 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:02 crc kubenswrapper[4764]: I0202 09:08:02.912548 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:02Z","lastTransitionTime":"2026-02-02T09:08:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:03 crc kubenswrapper[4764]: I0202 09:08:03.015735 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:03 crc kubenswrapper[4764]: I0202 09:08:03.015767 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:03 crc kubenswrapper[4764]: I0202 09:08:03.015777 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:03 crc kubenswrapper[4764]: I0202 09:08:03.015790 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:03 crc kubenswrapper[4764]: I0202 09:08:03.015800 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:03Z","lastTransitionTime":"2026-02-02T09:08:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:03 crc kubenswrapper[4764]: I0202 09:08:03.118921 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:03 crc kubenswrapper[4764]: I0202 09:08:03.119003 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:03 crc kubenswrapper[4764]: I0202 09:08:03.119015 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:03 crc kubenswrapper[4764]: I0202 09:08:03.119034 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:03 crc kubenswrapper[4764]: I0202 09:08:03.119047 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:03Z","lastTransitionTime":"2026-02-02T09:08:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:03 crc kubenswrapper[4764]: I0202 09:08:03.221490 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:03 crc kubenswrapper[4764]: I0202 09:08:03.221535 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:03 crc kubenswrapper[4764]: I0202 09:08:03.221548 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:03 crc kubenswrapper[4764]: I0202 09:08:03.221566 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:03 crc kubenswrapper[4764]: I0202 09:08:03.221579 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:03Z","lastTransitionTime":"2026-02-02T09:08:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:03 crc kubenswrapper[4764]: I0202 09:08:03.324877 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:03 crc kubenswrapper[4764]: I0202 09:08:03.324923 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:03 crc kubenswrapper[4764]: I0202 09:08:03.324953 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:03 crc kubenswrapper[4764]: I0202 09:08:03.324971 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:03 crc kubenswrapper[4764]: I0202 09:08:03.324982 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:03Z","lastTransitionTime":"2026-02-02T09:08:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:03 crc kubenswrapper[4764]: I0202 09:08:03.428429 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:03 crc kubenswrapper[4764]: I0202 09:08:03.428472 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:03 crc kubenswrapper[4764]: I0202 09:08:03.428485 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:03 crc kubenswrapper[4764]: I0202 09:08:03.428504 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:03 crc kubenswrapper[4764]: I0202 09:08:03.428518 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:03Z","lastTransitionTime":"2026-02-02T09:08:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:03 crc kubenswrapper[4764]: I0202 09:08:03.531276 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:03 crc kubenswrapper[4764]: I0202 09:08:03.531353 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:03 crc kubenswrapper[4764]: I0202 09:08:03.531376 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:03 crc kubenswrapper[4764]: I0202 09:08:03.531404 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:03 crc kubenswrapper[4764]: I0202 09:08:03.531425 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:03Z","lastTransitionTime":"2026-02-02T09:08:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:03 crc kubenswrapper[4764]: I0202 09:08:03.634655 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:03 crc kubenswrapper[4764]: I0202 09:08:03.634746 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:03 crc kubenswrapper[4764]: I0202 09:08:03.634772 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:03 crc kubenswrapper[4764]: I0202 09:08:03.634807 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:03 crc kubenswrapper[4764]: I0202 09:08:03.634832 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:03Z","lastTransitionTime":"2026-02-02T09:08:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:03 crc kubenswrapper[4764]: I0202 09:08:03.738605 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:03 crc kubenswrapper[4764]: I0202 09:08:03.738641 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:03 crc kubenswrapper[4764]: I0202 09:08:03.738650 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:03 crc kubenswrapper[4764]: I0202 09:08:03.738663 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:03 crc kubenswrapper[4764]: I0202 09:08:03.738674 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:03Z","lastTransitionTime":"2026-02-02T09:08:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:03 crc kubenswrapper[4764]: I0202 09:08:03.799088 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-18 01:00:47.568996803 +0000 UTC Feb 02 09:08:03 crc kubenswrapper[4764]: I0202 09:08:03.824651 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kwtmr" Feb 02 09:08:03 crc kubenswrapper[4764]: E0202 09:08:03.824977 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kwtmr" podUID="991faa9a-dd25-4f49-82bd-ce60cefd4af2" Feb 02 09:08:03 crc kubenswrapper[4764]: I0202 09:08:03.840843 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:03 crc kubenswrapper[4764]: I0202 09:08:03.840917 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:03 crc kubenswrapper[4764]: I0202 09:08:03.840955 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:03 crc kubenswrapper[4764]: I0202 09:08:03.840975 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:03 crc kubenswrapper[4764]: I0202 09:08:03.840988 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:03Z","lastTransitionTime":"2026-02-02T09:08:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:03 crc kubenswrapper[4764]: I0202 09:08:03.943700 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:03 crc kubenswrapper[4764]: I0202 09:08:03.943746 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:03 crc kubenswrapper[4764]: I0202 09:08:03.943755 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:03 crc kubenswrapper[4764]: I0202 09:08:03.943771 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:03 crc kubenswrapper[4764]: I0202 09:08:03.943782 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:03Z","lastTransitionTime":"2026-02-02T09:08:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:04 crc kubenswrapper[4764]: I0202 09:08:04.046373 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:04 crc kubenswrapper[4764]: I0202 09:08:04.046415 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:04 crc kubenswrapper[4764]: I0202 09:08:04.046427 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:04 crc kubenswrapper[4764]: I0202 09:08:04.046443 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:04 crc kubenswrapper[4764]: I0202 09:08:04.046454 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:04Z","lastTransitionTime":"2026-02-02T09:08:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:04 crc kubenswrapper[4764]: I0202 09:08:04.148205 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:04 crc kubenswrapper[4764]: I0202 09:08:04.148240 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:04 crc kubenswrapper[4764]: I0202 09:08:04.148250 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:04 crc kubenswrapper[4764]: I0202 09:08:04.148263 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:04 crc kubenswrapper[4764]: I0202 09:08:04.148271 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:04Z","lastTransitionTime":"2026-02-02T09:08:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:04 crc kubenswrapper[4764]: I0202 09:08:04.250808 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:04 crc kubenswrapper[4764]: I0202 09:08:04.250982 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:04 crc kubenswrapper[4764]: I0202 09:08:04.251027 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:04 crc kubenswrapper[4764]: I0202 09:08:04.251127 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:04 crc kubenswrapper[4764]: I0202 09:08:04.251144 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:04Z","lastTransitionTime":"2026-02-02T09:08:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:04 crc kubenswrapper[4764]: I0202 09:08:04.354138 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:04 crc kubenswrapper[4764]: I0202 09:08:04.354170 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:04 crc kubenswrapper[4764]: I0202 09:08:04.354182 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:04 crc kubenswrapper[4764]: I0202 09:08:04.354196 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:04 crc kubenswrapper[4764]: I0202 09:08:04.354205 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:04Z","lastTransitionTime":"2026-02-02T09:08:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:04 crc kubenswrapper[4764]: I0202 09:08:04.456798 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:04 crc kubenswrapper[4764]: I0202 09:08:04.456836 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:04 crc kubenswrapper[4764]: I0202 09:08:04.456848 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:04 crc kubenswrapper[4764]: I0202 09:08:04.456866 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:04 crc kubenswrapper[4764]: I0202 09:08:04.456887 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:04Z","lastTransitionTime":"2026-02-02T09:08:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:04 crc kubenswrapper[4764]: I0202 09:08:04.559685 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:04 crc kubenswrapper[4764]: I0202 09:08:04.559755 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:04 crc kubenswrapper[4764]: I0202 09:08:04.559777 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:04 crc kubenswrapper[4764]: I0202 09:08:04.559804 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:04 crc kubenswrapper[4764]: I0202 09:08:04.559820 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:04Z","lastTransitionTime":"2026-02-02T09:08:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:04 crc kubenswrapper[4764]: I0202 09:08:04.662547 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:04 crc kubenswrapper[4764]: I0202 09:08:04.662583 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:04 crc kubenswrapper[4764]: I0202 09:08:04.662604 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:04 crc kubenswrapper[4764]: I0202 09:08:04.662619 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:04 crc kubenswrapper[4764]: I0202 09:08:04.662631 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:04Z","lastTransitionTime":"2026-02-02T09:08:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:04 crc kubenswrapper[4764]: I0202 09:08:04.765517 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:04 crc kubenswrapper[4764]: I0202 09:08:04.765580 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:04 crc kubenswrapper[4764]: I0202 09:08:04.765593 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:04 crc kubenswrapper[4764]: I0202 09:08:04.765610 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:04 crc kubenswrapper[4764]: I0202 09:08:04.765640 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:04Z","lastTransitionTime":"2026-02-02T09:08:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:04 crc kubenswrapper[4764]: I0202 09:08:04.799972 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-17 20:22:22.102618421 +0000 UTC Feb 02 09:08:04 crc kubenswrapper[4764]: I0202 09:08:04.827249 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 09:08:04 crc kubenswrapper[4764]: I0202 09:08:04.827299 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 09:08:04 crc kubenswrapper[4764]: E0202 09:08:04.827438 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 09:08:04 crc kubenswrapper[4764]: I0202 09:08:04.827750 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 09:08:04 crc kubenswrapper[4764]: E0202 09:08:04.827868 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 09:08:04 crc kubenswrapper[4764]: E0202 09:08:04.828487 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 09:08:04 crc kubenswrapper[4764]: I0202 09:08:04.867842 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:04 crc kubenswrapper[4764]: I0202 09:08:04.867901 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:04 crc kubenswrapper[4764]: I0202 09:08:04.867913 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:04 crc kubenswrapper[4764]: I0202 09:08:04.867943 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:04 crc kubenswrapper[4764]: I0202 09:08:04.867952 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:04Z","lastTransitionTime":"2026-02-02T09:08:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:04 crc kubenswrapper[4764]: I0202 09:08:04.970890 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:04 crc kubenswrapper[4764]: I0202 09:08:04.970958 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:04 crc kubenswrapper[4764]: I0202 09:08:04.970972 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:04 crc kubenswrapper[4764]: I0202 09:08:04.970991 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:04 crc kubenswrapper[4764]: I0202 09:08:04.971005 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:04Z","lastTransitionTime":"2026-02-02T09:08:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:05 crc kubenswrapper[4764]: I0202 09:08:05.073414 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:05 crc kubenswrapper[4764]: I0202 09:08:05.073448 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:05 crc kubenswrapper[4764]: I0202 09:08:05.073458 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:05 crc kubenswrapper[4764]: I0202 09:08:05.073473 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:05 crc kubenswrapper[4764]: I0202 09:08:05.073483 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:05Z","lastTransitionTime":"2026-02-02T09:08:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:05 crc kubenswrapper[4764]: I0202 09:08:05.176077 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:05 crc kubenswrapper[4764]: I0202 09:08:05.176122 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:05 crc kubenswrapper[4764]: I0202 09:08:05.176134 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:05 crc kubenswrapper[4764]: I0202 09:08:05.176154 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:05 crc kubenswrapper[4764]: I0202 09:08:05.176167 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:05Z","lastTransitionTime":"2026-02-02T09:08:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:05 crc kubenswrapper[4764]: I0202 09:08:05.278621 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:05 crc kubenswrapper[4764]: I0202 09:08:05.278665 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:05 crc kubenswrapper[4764]: I0202 09:08:05.278674 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:05 crc kubenswrapper[4764]: I0202 09:08:05.278686 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:05 crc kubenswrapper[4764]: I0202 09:08:05.278695 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:05Z","lastTransitionTime":"2026-02-02T09:08:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:05 crc kubenswrapper[4764]: I0202 09:08:05.381626 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:05 crc kubenswrapper[4764]: I0202 09:08:05.381676 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:05 crc kubenswrapper[4764]: I0202 09:08:05.381690 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:05 crc kubenswrapper[4764]: I0202 09:08:05.381714 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:05 crc kubenswrapper[4764]: I0202 09:08:05.381730 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:05Z","lastTransitionTime":"2026-02-02T09:08:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:05 crc kubenswrapper[4764]: I0202 09:08:05.484882 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:05 crc kubenswrapper[4764]: I0202 09:08:05.484921 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:05 crc kubenswrapper[4764]: I0202 09:08:05.484942 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:05 crc kubenswrapper[4764]: I0202 09:08:05.484957 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:05 crc kubenswrapper[4764]: I0202 09:08:05.484966 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:05Z","lastTransitionTime":"2026-02-02T09:08:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:05 crc kubenswrapper[4764]: I0202 09:08:05.491971 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:05 crc kubenswrapper[4764]: I0202 09:08:05.492003 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:05 crc kubenswrapper[4764]: I0202 09:08:05.492012 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:05 crc kubenswrapper[4764]: I0202 09:08:05.492027 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:05 crc kubenswrapper[4764]: I0202 09:08:05.492037 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:05Z","lastTransitionTime":"2026-02-02T09:08:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:05 crc kubenswrapper[4764]: E0202 09:08:05.506966 4764 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:08:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:08:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:05Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:08:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:08:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:05Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"31973a70-ce17-4811-9a14-f5476c6979db\\\",\\\"systemUUID\\\":\\\"ca92679c-8580-43c3-be51-62441e93d22b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:05Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:05 crc kubenswrapper[4764]: I0202 09:08:05.511642 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:05 crc kubenswrapper[4764]: I0202 09:08:05.511684 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:05 crc kubenswrapper[4764]: I0202 09:08:05.511697 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:05 crc kubenswrapper[4764]: I0202 09:08:05.511715 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:05 crc kubenswrapper[4764]: I0202 09:08:05.511729 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:05Z","lastTransitionTime":"2026-02-02T09:08:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:05 crc kubenswrapper[4764]: E0202 09:08:05.525073 4764 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:08:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:08:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:05Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:08:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:08:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:05Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"31973a70-ce17-4811-9a14-f5476c6979db\\\",\\\"systemUUID\\\":\\\"ca92679c-8580-43c3-be51-62441e93d22b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:05Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:05 crc kubenswrapper[4764]: I0202 09:08:05.528711 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:05 crc kubenswrapper[4764]: I0202 09:08:05.528750 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:05 crc kubenswrapper[4764]: I0202 09:08:05.528758 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:05 crc kubenswrapper[4764]: I0202 09:08:05.528775 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:05 crc kubenswrapper[4764]: I0202 09:08:05.528788 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:05Z","lastTransitionTime":"2026-02-02T09:08:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:05 crc kubenswrapper[4764]: E0202 09:08:05.544448 4764 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:08:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:08:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:05Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:08:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:08:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:05Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"31973a70-ce17-4811-9a14-f5476c6979db\\\",\\\"systemUUID\\\":\\\"ca92679c-8580-43c3-be51-62441e93d22b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:05Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:05 crc kubenswrapper[4764]: I0202 09:08:05.548926 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:05 crc kubenswrapper[4764]: I0202 09:08:05.548982 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:05 crc kubenswrapper[4764]: I0202 09:08:05.548993 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:05 crc kubenswrapper[4764]: I0202 09:08:05.549006 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:05 crc kubenswrapper[4764]: I0202 09:08:05.549016 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:05Z","lastTransitionTime":"2026-02-02T09:08:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:05 crc kubenswrapper[4764]: E0202 09:08:05.560458 4764 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:08:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:08:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:05Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:08:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:08:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:05Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"31973a70-ce17-4811-9a14-f5476c6979db\\\",\\\"systemUUID\\\":\\\"ca92679c-8580-43c3-be51-62441e93d22b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:05Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:05 crc kubenswrapper[4764]: I0202 09:08:05.570691 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:05 crc kubenswrapper[4764]: I0202 09:08:05.570916 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:05 crc kubenswrapper[4764]: I0202 09:08:05.571058 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:05 crc kubenswrapper[4764]: I0202 09:08:05.571157 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:05 crc kubenswrapper[4764]: I0202 09:08:05.571237 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:05Z","lastTransitionTime":"2026-02-02T09:08:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:05 crc kubenswrapper[4764]: E0202 09:08:05.585106 4764 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:08:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:08:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:05Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:08:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:08:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:05Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"31973a70-ce17-4811-9a14-f5476c6979db\\\",\\\"systemUUID\\\":\\\"ca92679c-8580-43c3-be51-62441e93d22b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:05Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:05 crc kubenswrapper[4764]: E0202 09:08:05.585571 4764 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 02 09:08:05 crc kubenswrapper[4764]: I0202 09:08:05.587527 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:05 crc kubenswrapper[4764]: I0202 09:08:05.587696 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:05 crc kubenswrapper[4764]: I0202 09:08:05.587812 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:05 crc kubenswrapper[4764]: I0202 09:08:05.587962 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:05 crc kubenswrapper[4764]: I0202 09:08:05.588080 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:05Z","lastTransitionTime":"2026-02-02T09:08:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:05 crc kubenswrapper[4764]: I0202 09:08:05.690171 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:05 crc kubenswrapper[4764]: I0202 09:08:05.690418 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:05 crc kubenswrapper[4764]: I0202 09:08:05.690572 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:05 crc kubenswrapper[4764]: I0202 09:08:05.690664 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:05 crc kubenswrapper[4764]: I0202 09:08:05.690756 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:05Z","lastTransitionTime":"2026-02-02T09:08:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:05 crc kubenswrapper[4764]: I0202 09:08:05.793652 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:05 crc kubenswrapper[4764]: I0202 09:08:05.793765 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:05 crc kubenswrapper[4764]: I0202 09:08:05.793789 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:05 crc kubenswrapper[4764]: I0202 09:08:05.793827 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:05 crc kubenswrapper[4764]: I0202 09:08:05.793847 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:05Z","lastTransitionTime":"2026-02-02T09:08:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:05 crc kubenswrapper[4764]: I0202 09:08:05.800868 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-20 23:03:50.51392391 +0000 UTC Feb 02 09:08:05 crc kubenswrapper[4764]: I0202 09:08:05.824815 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kwtmr" Feb 02 09:08:05 crc kubenswrapper[4764]: E0202 09:08:05.825185 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kwtmr" podUID="991faa9a-dd25-4f49-82bd-ce60cefd4af2" Feb 02 09:08:05 crc kubenswrapper[4764]: I0202 09:08:05.895823 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:05 crc kubenswrapper[4764]: I0202 09:08:05.895869 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:05 crc kubenswrapper[4764]: I0202 09:08:05.895879 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:05 crc kubenswrapper[4764]: I0202 09:08:05.895895 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:05 crc kubenswrapper[4764]: I0202 09:08:05.895907 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:05Z","lastTransitionTime":"2026-02-02T09:08:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:05 crc kubenswrapper[4764]: I0202 09:08:05.998813 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:05 crc kubenswrapper[4764]: I0202 09:08:05.999099 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:05 crc kubenswrapper[4764]: I0202 09:08:05.999122 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:05 crc kubenswrapper[4764]: I0202 09:08:05.999138 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:05 crc kubenswrapper[4764]: I0202 09:08:05.999149 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:05Z","lastTransitionTime":"2026-02-02T09:08:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:06 crc kubenswrapper[4764]: I0202 09:08:06.102616 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:06 crc kubenswrapper[4764]: I0202 09:08:06.102680 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:06 crc kubenswrapper[4764]: I0202 09:08:06.102704 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:06 crc kubenswrapper[4764]: I0202 09:08:06.102734 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:06 crc kubenswrapper[4764]: I0202 09:08:06.102759 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:06Z","lastTransitionTime":"2026-02-02T09:08:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:06 crc kubenswrapper[4764]: I0202 09:08:06.205388 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:06 crc kubenswrapper[4764]: I0202 09:08:06.205440 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:06 crc kubenswrapper[4764]: I0202 09:08:06.205455 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:06 crc kubenswrapper[4764]: I0202 09:08:06.205477 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:06 crc kubenswrapper[4764]: I0202 09:08:06.205489 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:06Z","lastTransitionTime":"2026-02-02T09:08:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:06 crc kubenswrapper[4764]: I0202 09:08:06.308621 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:06 crc kubenswrapper[4764]: I0202 09:08:06.308688 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:06 crc kubenswrapper[4764]: I0202 09:08:06.308714 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:06 crc kubenswrapper[4764]: I0202 09:08:06.308745 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:06 crc kubenswrapper[4764]: I0202 09:08:06.308769 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:06Z","lastTransitionTime":"2026-02-02T09:08:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:06 crc kubenswrapper[4764]: I0202 09:08:06.413270 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:06 crc kubenswrapper[4764]: I0202 09:08:06.413487 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:06 crc kubenswrapper[4764]: I0202 09:08:06.413552 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:06 crc kubenswrapper[4764]: I0202 09:08:06.413617 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:06 crc kubenswrapper[4764]: I0202 09:08:06.413679 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:06Z","lastTransitionTime":"2026-02-02T09:08:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:06 crc kubenswrapper[4764]: I0202 09:08:06.517878 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:06 crc kubenswrapper[4764]: I0202 09:08:06.517922 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:06 crc kubenswrapper[4764]: I0202 09:08:06.517967 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:06 crc kubenswrapper[4764]: I0202 09:08:06.518001 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:06 crc kubenswrapper[4764]: I0202 09:08:06.518014 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:06Z","lastTransitionTime":"2026-02-02T09:08:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:06 crc kubenswrapper[4764]: I0202 09:08:06.621396 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:06 crc kubenswrapper[4764]: I0202 09:08:06.621477 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:06 crc kubenswrapper[4764]: I0202 09:08:06.621503 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:06 crc kubenswrapper[4764]: I0202 09:08:06.621534 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:06 crc kubenswrapper[4764]: I0202 09:08:06.621555 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:06Z","lastTransitionTime":"2026-02-02T09:08:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:06 crc kubenswrapper[4764]: I0202 09:08:06.724406 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:06 crc kubenswrapper[4764]: I0202 09:08:06.724451 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:06 crc kubenswrapper[4764]: I0202 09:08:06.724460 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:06 crc kubenswrapper[4764]: I0202 09:08:06.724473 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:06 crc kubenswrapper[4764]: I0202 09:08:06.724485 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:06Z","lastTransitionTime":"2026-02-02T09:08:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:06 crc kubenswrapper[4764]: I0202 09:08:06.801547 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-17 22:41:34.717833371 +0000 UTC Feb 02 09:08:06 crc kubenswrapper[4764]: I0202 09:08:06.825108 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 09:08:06 crc kubenswrapper[4764]: I0202 09:08:06.825107 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 09:08:06 crc kubenswrapper[4764]: E0202 09:08:06.825282 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 09:08:06 crc kubenswrapper[4764]: I0202 09:08:06.825381 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 09:08:06 crc kubenswrapper[4764]: E0202 09:08:06.825442 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 09:08:06 crc kubenswrapper[4764]: E0202 09:08:06.825588 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 09:08:06 crc kubenswrapper[4764]: I0202 09:08:06.827265 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:06 crc kubenswrapper[4764]: I0202 09:08:06.827316 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:06 crc kubenswrapper[4764]: I0202 09:08:06.827326 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:06 crc kubenswrapper[4764]: I0202 09:08:06.827338 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:06 crc kubenswrapper[4764]: I0202 09:08:06.827348 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:06Z","lastTransitionTime":"2026-02-02T09:08:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:06 crc kubenswrapper[4764]: I0202 09:08:06.931369 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:06 crc kubenswrapper[4764]: I0202 09:08:06.931438 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:06 crc kubenswrapper[4764]: I0202 09:08:06.931464 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:06 crc kubenswrapper[4764]: I0202 09:08:06.931501 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:06 crc kubenswrapper[4764]: I0202 09:08:06.931632 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:06Z","lastTransitionTime":"2026-02-02T09:08:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:07 crc kubenswrapper[4764]: I0202 09:08:07.036521 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:07 crc kubenswrapper[4764]: I0202 09:08:07.036587 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:07 crc kubenswrapper[4764]: I0202 09:08:07.036600 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:07 crc kubenswrapper[4764]: I0202 09:08:07.036620 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:07 crc kubenswrapper[4764]: I0202 09:08:07.036631 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:07Z","lastTransitionTime":"2026-02-02T09:08:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:07 crc kubenswrapper[4764]: I0202 09:08:07.140329 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:07 crc kubenswrapper[4764]: I0202 09:08:07.140775 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:07 crc kubenswrapper[4764]: I0202 09:08:07.140857 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:07 crc kubenswrapper[4764]: I0202 09:08:07.140927 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:07 crc kubenswrapper[4764]: I0202 09:08:07.141066 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:07Z","lastTransitionTime":"2026-02-02T09:08:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:07 crc kubenswrapper[4764]: I0202 09:08:07.243864 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:07 crc kubenswrapper[4764]: I0202 09:08:07.243922 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:07 crc kubenswrapper[4764]: I0202 09:08:07.243975 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:07 crc kubenswrapper[4764]: I0202 09:08:07.243996 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:07 crc kubenswrapper[4764]: I0202 09:08:07.244010 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:07Z","lastTransitionTime":"2026-02-02T09:08:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:07 crc kubenswrapper[4764]: I0202 09:08:07.347755 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:07 crc kubenswrapper[4764]: I0202 09:08:07.347800 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:07 crc kubenswrapper[4764]: I0202 09:08:07.347814 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:07 crc kubenswrapper[4764]: I0202 09:08:07.347833 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:07 crc kubenswrapper[4764]: I0202 09:08:07.347850 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:07Z","lastTransitionTime":"2026-02-02T09:08:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:07 crc kubenswrapper[4764]: I0202 09:08:07.450522 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:07 crc kubenswrapper[4764]: I0202 09:08:07.450595 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:07 crc kubenswrapper[4764]: I0202 09:08:07.450609 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:07 crc kubenswrapper[4764]: I0202 09:08:07.450632 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:07 crc kubenswrapper[4764]: I0202 09:08:07.450647 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:07Z","lastTransitionTime":"2026-02-02T09:08:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:07 crc kubenswrapper[4764]: I0202 09:08:07.553785 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:07 crc kubenswrapper[4764]: I0202 09:08:07.553844 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:07 crc kubenswrapper[4764]: I0202 09:08:07.553857 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:07 crc kubenswrapper[4764]: I0202 09:08:07.553874 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:07 crc kubenswrapper[4764]: I0202 09:08:07.553886 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:07Z","lastTransitionTime":"2026-02-02T09:08:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:07 crc kubenswrapper[4764]: I0202 09:08:07.656435 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:07 crc kubenswrapper[4764]: I0202 09:08:07.656474 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:07 crc kubenswrapper[4764]: I0202 09:08:07.656483 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:07 crc kubenswrapper[4764]: I0202 09:08:07.656496 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:07 crc kubenswrapper[4764]: I0202 09:08:07.656508 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:07Z","lastTransitionTime":"2026-02-02T09:08:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:07 crc kubenswrapper[4764]: I0202 09:08:07.759084 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:07 crc kubenswrapper[4764]: I0202 09:08:07.759141 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:07 crc kubenswrapper[4764]: I0202 09:08:07.759150 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:07 crc kubenswrapper[4764]: I0202 09:08:07.759182 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:07 crc kubenswrapper[4764]: I0202 09:08:07.759192 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:07Z","lastTransitionTime":"2026-02-02T09:08:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:07 crc kubenswrapper[4764]: I0202 09:08:07.802681 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-17 11:50:15.541248419 +0000 UTC Feb 02 09:08:07 crc kubenswrapper[4764]: I0202 09:08:07.825023 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kwtmr" Feb 02 09:08:07 crc kubenswrapper[4764]: E0202 09:08:07.825172 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kwtmr" podUID="991faa9a-dd25-4f49-82bd-ce60cefd4af2" Feb 02 09:08:07 crc kubenswrapper[4764]: I0202 09:08:07.838181 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kwtmr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"991faa9a-dd25-4f49-82bd-ce60cefd4af2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h865b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h865b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:53Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kwtmr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:07Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:07 crc kubenswrapper[4764]: I0202 09:08:07.855237 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87d61aea-46f6-4ee6-a7a4-767a6a67fea4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e78b6d08e4d616cefbb73e3fd43b3c6886dcd8aa997c7b4948bf1ff63401293\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d468715c2da33996561c40043216233fd12e21539d0c58dd1f3b892fd9f91a9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b184f7c936dc4046f74c0018269cbc2c4b7ceddb650b4e871d3f8a707e0df88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://269c5074371bd00f8363fb3888c11f1df383081a61497aea6578f08700881d50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b25bc9395831524c53f9e565b4efbe74f21792efafeb1dd507161bc91aa0ecdf\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T09:07:22Z\\\",\\\"message\\\":\\\"W0202 09:07:21.597488 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0202 09:07:21.598117 1 crypto.go:601] Generating new CA for check-endpoints-signer@1770023241 cert, and key in /tmp/serving-cert-3856183724/serving-signer.crt, /tmp/serving-cert-3856183724/serving-signer.key\\\\nI0202 09:07:22.012184 1 observer_polling.go:159] Starting file observer\\\\nW0202 09:07:22.015493 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0202 09:07:22.015797 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 09:07:22.016596 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3856183724/tls.crt::/tmp/serving-cert-3856183724/tls.key\\\\\\\"\\\\nF0202 09:07:22.270605 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b44e17b8ca01f549bc10081dd5f109cb1b9fff364bf5bd4fd90d9f0e3eeb04db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfb06a5ec9262cf39b108a66014d15e325cdbd5441a2d9e7a811dd57265c99a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfb06a5ec9262cf39b108a66014d15e325cdbd5441a2d9e7a811dd57265c99a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:07Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:07 crc kubenswrapper[4764]: I0202 09:08:07.861151 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:07 crc kubenswrapper[4764]: I0202 09:08:07.861189 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:07 crc kubenswrapper[4764]: I0202 09:08:07.861200 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:07 crc kubenswrapper[4764]: I0202 09:08:07.861215 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:07 crc kubenswrapper[4764]: I0202 09:08:07.861226 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:07Z","lastTransitionTime":"2026-02-02T09:08:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:07 crc kubenswrapper[4764]: I0202 09:08:07.869221 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5528225d91ac2778bef4554abb0bd036ac5d3b6cdac7357bab69e70c3894b6f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://071bbb80d0747cf3e109ba626d47f7d05f85f4cc0b3aba77276c795741f47ac6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:07Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:07 crc kubenswrapper[4764]: I0202 09:08:07.880504 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:07Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:07 crc kubenswrapper[4764]: I0202 09:08:07.893909 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nc7qt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1a2c7b22068b7165927fc158885d48bf9952b9eba521d8ae9a599eee53dc305\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2hz4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nc7qt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:07Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:07 crc kubenswrapper[4764]: I0202 09:08:07.910848 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9p5dc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"660ac006-2495-45c4-9fb3-e9c1dddcf7a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da812b124c97aa8ad1563f4bcfac8f3bfa025096fa4a2561e5847b1bf9cb30b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23e57e65a6f5c3059506b9db21e9ac2492c30bd9191a98db0281915e9ce39797\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23e57e65a6f5c3059506b9db21e9ac2492c30bd9191a98db0281915e9ce39797\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2417b66816da1a6cd1791a4b91eca4a3c4f3694462e066e7f0722f1b9c2105c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2417b66816da1a6cd1791a4b91eca4a3c4f3694462e066e7f0722f1b9c2105c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://883641f29e9120c2cea6b272bacf1dc63c752f55ab0408cdaf73db85c85ac369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://883641f29e9120c2cea6b272bacf1dc63c752f55ab0408cdaf73db85c85ac369\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2745baba1188ba5e8f820616e4117a12c50ecee5b68e71ad9cabbb621b6e131f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2745baba1188ba5e8f820616e4117a12c50ecee5b68e71ad9cabbb621b6e131f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8925e435be4d48d86824fc35f8b8819fb35d63e385164c4a20edb3bf3035c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a8925e435be4d48d86824fc35f8b8819fb35d63e385164c4a20edb3bf3035c3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d3fbf668ae7d04c8d9e574f43c2fe90c5604204ba38a3c60847afdfbceb7a57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d3fbf668ae7d04c8d9e574f43c2fe90c5604204ba38a3c60847afdfbceb7a57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9p5dc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:07Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:07 crc kubenswrapper[4764]: I0202 09:08:07.928527 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d66ab3c0-9b5b-41d5-b6c0-91ccddccd7e9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d92ac486f5dade46bc75de32099859a71fc2ed25a7abbe6ae8937df128e7df30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a4ac273a4f127d79e5fad7d4c30cd2044a346f5734dba5d33d68cd092c0c2d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d38c7cb5dcd22fb07524494ebe77aa1a7f4f808f566043563817e47e7ec12a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://408722c7f83d28919431532f8c797a1c795fd5d96736d14e7f06a05744899702\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f473b18cd8003f75bb1dafb82bcec45745a73ccd10434297356dac185a3d15c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e229f484d19cbdabedf53d90a822a4434db9e00b1301538e885e26acc19d25f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e229f484d19cbdabedf53d90a822a4434db9e00b1301538e885e26acc19d25f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a9cc673be8ff245dba81278e8cd34cfe45ed36ff0b8bd5b30a0ddf333ec6371\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a9cc673be8ff245dba81278e8cd34cfe45ed36ff0b8bd5b30a0ddf333ec6371\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bc7e49c7a5ae66ea939060192fa6a346dc0b0b9aa9750b29879d3c426cc0d276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc7e49c7a5ae66ea939060192fa6a346dc0b0b9aa9750b29879d3c426cc0d276\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:07Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:07 crc kubenswrapper[4764]: I0202 09:08:07.943459 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa9d037e66cb53b9dd698eff46284c9e718fd842cda6f7b525f0c79062d70bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:07Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:07 crc kubenswrapper[4764]: I0202 09:08:07.956068 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:07Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:07 crc kubenswrapper[4764]: I0202 09:08:07.963392 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:07 crc kubenswrapper[4764]: I0202 09:08:07.963425 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:07 crc kubenswrapper[4764]: I0202 09:08:07.963435 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:07 crc kubenswrapper[4764]: I0202 09:08:07.963450 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:07 crc kubenswrapper[4764]: I0202 09:08:07.963464 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:07Z","lastTransitionTime":"2026-02-02T09:08:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:07 crc kubenswrapper[4764]: I0202 09:08:07.974373 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24632dda-6100-4ab6-a28e-214ddae4360c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a6c67e9b7708cc0d85b0b6ffbbade0d34867e159a9c5487ca2606475c667a50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bffda48d7c881c1470251b501661324859373246cc1a21811bd36ca24887c7aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a989645cf485cd8ff80ffff1a86cd7808e1564eba5667535e161cc68db13b26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20b71d3ee7a381baa51b0b0d0470283022cfede30f491135737b5e82d631d763\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c380383b6a014d1b694a13ddc22f54b20b9e4f9d5c9b9000d76187621dce0a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5529f2bff87753428e9173826fc1880acf96a4c134fa9ecad195b15697cad73a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12244061eabaae7fa9057ac5cee894c3b308b466411bcb910682e7f65d240dd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://12244061eabaae7fa9057ac5cee894c3b308b466411bcb910682e7f65d240dd7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-02T09:07:52Z\\\",\\\"message\\\":\\\"l_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-kube-apiserver/apiserver]} name:Service_openshift-kube-apiserver/apiserver_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.93:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {d71b38eb-32af-4c0f-9490-7c317c111e3a}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0202 09:07:51.914338 6129 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-kube-apiserver/apiserver]} name:Service_openshift-kube-apiserver/apiserver_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.93:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {d71b38eb-32af-4c0f-9490-7c317c111e3a}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF0202 09:07:51.914404 6129 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:51Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-zhn7j_openshift-ovn-kubernetes(24632dda-6100-4ab6-a28e-214ddae4360c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b4cfbd4e682bcdcaff785013bdb72580ce36fced733e235538536827036b93d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://579903eadb183466a2564737e6b678e04308239c0cd75b6808892053b97afbc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://579903eadb183466a2564737e6b678e04308239c0cd75b6808892053b97afbc0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zhn7j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:07Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:07 crc kubenswrapper[4764]: I0202 09:08:07.987382 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed747fd8-e3ce-4319-a6da-bda0fd51b85a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6f064bdc66470aace3180d4988244026b54009a63cff199e41c5d02a1d18479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://60258f84a1d8dcfe48b7e9688ed50f077e201971f20f3f6663fbf1ba58a4611a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f84be2e08217a2ac3edde858d78b4410376ed9bae3bcabe2a7f9b2fbe2f69c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a69d25d352d7ceac5c735bc1fcf446a1fea6dc22c65a5296cfab828cea17377\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:07Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:08 crc kubenswrapper[4764]: I0202 09:08:08.002264 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f6904228e052fff69ad1fd66df97758b6b60ba4351e97a7fbd03eee4df1b653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:07Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:08 crc kubenswrapper[4764]: I0202 09:08:08.012323 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6zxff" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aaba93da-12a5-43df-b7c7-4c8b988163c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3db293678a907cd55b31b1d0c0bd7c30e1ba75ecd5e13cbc68ab98fbf0fbd9d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqwb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6zxff\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:08Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:08 crc kubenswrapper[4764]: I0202 09:08:08.023331 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d192f670-9f9d-4539-9641-e4bed73acdd4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://040c5e8a5659f85f0d52d1a1e064154d998ae104f711101d2204c0b1caebbb30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbksf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fdf06410349b7c99aa166e87fbe80bf68e2ae923635a6a978b7e2cd50383f7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbksf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4ndm4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:08Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:08 crc kubenswrapper[4764]: I0202 09:08:08.035826 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pcjmn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6e5ccd3b-b657-47a2-9c4d-bc9adb126533\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72863eef568d92a1abeb3cde2e89630a4f5fb7626c6949448ff5195774c98eab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2f7zl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pcjmn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:08Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:08 crc kubenswrapper[4764]: I0202 09:08:08.048018 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:08Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:08 crc kubenswrapper[4764]: I0202 09:08:08.056873 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lh6tm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2152e8a-f456-47b5-a547-b65ec837ecbc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38e176a00b07f2960fdd7ba2aadd1468a5e083b8d154e9d118d13f2585da81e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bb2sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://996e649011d5255783284421ddea8e0a5c383ef675b0920c3fdc0af6503d80e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bb2sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lh6tm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:08Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:08 crc kubenswrapper[4764]: I0202 09:08:08.065108 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:08 crc kubenswrapper[4764]: I0202 09:08:08.065141 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:08 crc kubenswrapper[4764]: I0202 09:08:08.065152 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:08 crc kubenswrapper[4764]: I0202 09:08:08.065168 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:08 crc kubenswrapper[4764]: I0202 09:08:08.065182 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:08Z","lastTransitionTime":"2026-02-02T09:08:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:08 crc kubenswrapper[4764]: I0202 09:08:08.167618 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:08 crc kubenswrapper[4764]: I0202 09:08:08.167670 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:08 crc kubenswrapper[4764]: I0202 09:08:08.167679 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:08 crc kubenswrapper[4764]: I0202 09:08:08.167693 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:08 crc kubenswrapper[4764]: I0202 09:08:08.167701 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:08Z","lastTransitionTime":"2026-02-02T09:08:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:08 crc kubenswrapper[4764]: I0202 09:08:08.270608 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:08 crc kubenswrapper[4764]: I0202 09:08:08.270648 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:08 crc kubenswrapper[4764]: I0202 09:08:08.270660 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:08 crc kubenswrapper[4764]: I0202 09:08:08.270676 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:08 crc kubenswrapper[4764]: I0202 09:08:08.270688 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:08Z","lastTransitionTime":"2026-02-02T09:08:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:08 crc kubenswrapper[4764]: I0202 09:08:08.373653 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:08 crc kubenswrapper[4764]: I0202 09:08:08.373721 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:08 crc kubenswrapper[4764]: I0202 09:08:08.373746 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:08 crc kubenswrapper[4764]: I0202 09:08:08.373778 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:08 crc kubenswrapper[4764]: I0202 09:08:08.373800 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:08Z","lastTransitionTime":"2026-02-02T09:08:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:08 crc kubenswrapper[4764]: I0202 09:08:08.477731 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:08 crc kubenswrapper[4764]: I0202 09:08:08.477775 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:08 crc kubenswrapper[4764]: I0202 09:08:08.477786 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:08 crc kubenswrapper[4764]: I0202 09:08:08.477802 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:08 crc kubenswrapper[4764]: I0202 09:08:08.477815 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:08Z","lastTransitionTime":"2026-02-02T09:08:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:08 crc kubenswrapper[4764]: I0202 09:08:08.580711 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:08 crc kubenswrapper[4764]: I0202 09:08:08.580754 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:08 crc kubenswrapper[4764]: I0202 09:08:08.580770 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:08 crc kubenswrapper[4764]: I0202 09:08:08.580791 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:08 crc kubenswrapper[4764]: I0202 09:08:08.580808 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:08Z","lastTransitionTime":"2026-02-02T09:08:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:08 crc kubenswrapper[4764]: I0202 09:08:08.684474 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:08 crc kubenswrapper[4764]: I0202 09:08:08.684545 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:08 crc kubenswrapper[4764]: I0202 09:08:08.684561 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:08 crc kubenswrapper[4764]: I0202 09:08:08.684588 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:08 crc kubenswrapper[4764]: I0202 09:08:08.684607 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:08Z","lastTransitionTime":"2026-02-02T09:08:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:08 crc kubenswrapper[4764]: I0202 09:08:08.717375 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 02 09:08:08 crc kubenswrapper[4764]: I0202 09:08:08.726463 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Feb 02 09:08:08 crc kubenswrapper[4764]: I0202 09:08:08.735305 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa9d037e66cb53b9dd698eff46284c9e718fd842cda6f7b525f0c79062d70bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:08Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:08 crc kubenswrapper[4764]: I0202 09:08:08.752033 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:08Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:08 crc kubenswrapper[4764]: I0202 09:08:08.781040 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24632dda-6100-4ab6-a28e-214ddae4360c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a6c67e9b7708cc0d85b0b6ffbbade0d34867e159a9c5487ca2606475c667a50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bffda48d7c881c1470251b501661324859373246cc1a21811bd36ca24887c7aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a989645cf485cd8ff80ffff1a86cd7808e1564eba5667535e161cc68db13b26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20b71d3ee7a381baa51b0b0d0470283022cfede30f491135737b5e82d631d763\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c380383b6a014d1b694a13ddc22f54b20b9e4f9d5c9b9000d76187621dce0a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5529f2bff87753428e9173826fc1880acf96a4c134fa9ecad195b15697cad73a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12244061eabaae7fa9057ac5cee894c3b308b466411bcb910682e7f65d240dd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://12244061eabaae7fa9057ac5cee894c3b308b466411bcb910682e7f65d240dd7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-02T09:07:52Z\\\",\\\"message\\\":\\\"l_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-kube-apiserver/apiserver]} name:Service_openshift-kube-apiserver/apiserver_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.93:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {d71b38eb-32af-4c0f-9490-7c317c111e3a}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0202 09:07:51.914338 6129 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-kube-apiserver/apiserver]} name:Service_openshift-kube-apiserver/apiserver_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.93:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {d71b38eb-32af-4c0f-9490-7c317c111e3a}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF0202 09:07:51.914404 6129 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:51Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-zhn7j_openshift-ovn-kubernetes(24632dda-6100-4ab6-a28e-214ddae4360c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b4cfbd4e682bcdcaff785013bdb72580ce36fced733e235538536827036b93d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://579903eadb183466a2564737e6b678e04308239c0cd75b6808892053b97afbc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://579903eadb183466a2564737e6b678e04308239c0cd75b6808892053b97afbc0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zhn7j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:08Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:08 crc kubenswrapper[4764]: I0202 09:08:08.786766 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:08 crc kubenswrapper[4764]: I0202 09:08:08.786852 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:08 crc kubenswrapper[4764]: I0202 09:08:08.786867 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:08 crc kubenswrapper[4764]: I0202 09:08:08.786891 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:08 crc kubenswrapper[4764]: I0202 09:08:08.786905 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:08Z","lastTransitionTime":"2026-02-02T09:08:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:08 crc kubenswrapper[4764]: I0202 09:08:08.803099 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-07 12:26:46.861641555 +0000 UTC Feb 02 09:08:08 crc kubenswrapper[4764]: I0202 09:08:08.809174 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d66ab3c0-9b5b-41d5-b6c0-91ccddccd7e9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d92ac486f5dade46bc75de32099859a71fc2ed25a7abbe6ae8937df128e7df30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a4ac273a4f127d79e5fad7d4c30cd2044a346f5734dba5d33d68cd092c0c2d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d38c7cb5dcd22fb07524494ebe77aa1a7f4f808f566043563817e47e7ec12a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://408722c7f83d28919431532f8c797a1c795fd5d96736d14e7f06a05744899702\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f473b18cd8003f75bb1dafb82bcec45745a73ccd10434297356dac185a3d15c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e229f484d19cbdabedf53d90a822a4434db9e00b1301538e885e26acc19d25f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e229f484d19cbdabedf53d90a822a4434db9e00b1301538e885e26acc19d25f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a9cc673be8ff245dba81278e8cd34cfe45ed36ff0b8bd5b30a0ddf333ec6371\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a9cc673be8ff245dba81278e8cd34cfe45ed36ff0b8bd5b30a0ddf333ec6371\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bc7e49c7a5ae66ea939060192fa6a346dc0b0b9aa9750b29879d3c426cc0d276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc7e49c7a5ae66ea939060192fa6a346dc0b0b9aa9750b29879d3c426cc0d276\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:08Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:08 crc kubenswrapper[4764]: I0202 09:08:08.824681 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 09:08:08 crc kubenswrapper[4764]: I0202 09:08:08.824722 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 09:08:08 crc kubenswrapper[4764]: I0202 09:08:08.824840 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 09:08:08 crc kubenswrapper[4764]: E0202 09:08:08.824875 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 09:08:08 crc kubenswrapper[4764]: E0202 09:08:08.825142 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 09:08:08 crc kubenswrapper[4764]: E0202 09:08:08.825543 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 09:08:08 crc kubenswrapper[4764]: I0202 09:08:08.825861 4764 scope.go:117] "RemoveContainer" containerID="12244061eabaae7fa9057ac5cee894c3b308b466411bcb910682e7f65d240dd7" Feb 02 09:08:08 crc kubenswrapper[4764]: I0202 09:08:08.829092 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f6904228e052fff69ad1fd66df97758b6b60ba4351e97a7fbd03eee4df1b653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:08Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:08 crc kubenswrapper[4764]: I0202 09:08:08.847264 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6zxff" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aaba93da-12a5-43df-b7c7-4c8b988163c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3db293678a907cd55b31b1d0c0bd7c30e1ba75ecd5e13cbc68ab98fbf0fbd9d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqwb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6zxff\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:08Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:08 crc kubenswrapper[4764]: I0202 09:08:08.860090 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d192f670-9f9d-4539-9641-e4bed73acdd4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://040c5e8a5659f85f0d52d1a1e064154d998ae104f711101d2204c0b1caebbb30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbksf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fdf06410349b7c99aa166e87fbe80bf68e2ae923635a6a978b7e2cd50383f7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbksf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4ndm4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:08Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:08 crc kubenswrapper[4764]: I0202 09:08:08.872499 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pcjmn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6e5ccd3b-b657-47a2-9c4d-bc9adb126533\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72863eef568d92a1abeb3cde2e89630a4f5fb7626c6949448ff5195774c98eab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2f7zl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pcjmn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:08Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:08 crc kubenswrapper[4764]: I0202 09:08:08.886972 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed747fd8-e3ce-4319-a6da-bda0fd51b85a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6f064bdc66470aace3180d4988244026b54009a63cff199e41c5d02a1d18479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://60258f84a1d8dcfe48b7e9688ed50f077e201971f20f3f6663fbf1ba58a4611a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f84be2e08217a2ac3edde858d78b4410376ed9bae3bcabe2a7f9b2fbe2f69c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a69d25d352d7ceac5c735bc1fcf446a1fea6dc22c65a5296cfab828cea17377\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:08Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:08 crc kubenswrapper[4764]: I0202 09:08:08.888721 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:08 crc kubenswrapper[4764]: I0202 09:08:08.888752 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:08 crc kubenswrapper[4764]: I0202 09:08:08.888763 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:08 crc kubenswrapper[4764]: I0202 09:08:08.888779 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:08 crc kubenswrapper[4764]: I0202 09:08:08.888788 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:08Z","lastTransitionTime":"2026-02-02T09:08:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:08 crc kubenswrapper[4764]: I0202 09:08:08.903701 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:08Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:08 crc kubenswrapper[4764]: I0202 09:08:08.918193 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lh6tm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2152e8a-f456-47b5-a547-b65ec837ecbc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38e176a00b07f2960fdd7ba2aadd1468a5e083b8d154e9d118d13f2585da81e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bb2sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://996e649011d5255783284421ddea8e0a5c383ef675b0920c3fdc0af6503d80e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bb2sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lh6tm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:08Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:08 crc kubenswrapper[4764]: I0202 09:08:08.938754 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5528225d91ac2778bef4554abb0bd036ac5d3b6cdac7357bab69e70c3894b6f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://071bbb80d0747cf3e109ba626d47f7d05f85f4cc0b3aba77276c795741f47ac6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:08Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:08 crc kubenswrapper[4764]: I0202 09:08:08.953235 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:08Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:08 crc kubenswrapper[4764]: I0202 09:08:08.965698 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nc7qt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1a2c7b22068b7165927fc158885d48bf9952b9eba521d8ae9a599eee53dc305\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2hz4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nc7qt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:08Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:08 crc kubenswrapper[4764]: I0202 09:08:08.981580 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9p5dc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"660ac006-2495-45c4-9fb3-e9c1dddcf7a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da812b124c97aa8ad1563f4bcfac8f3bfa025096fa4a2561e5847b1bf9cb30b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23e57e65a6f5c3059506b9db21e9ac2492c30bd9191a98db0281915e9ce39797\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23e57e65a6f5c3059506b9db21e9ac2492c30bd9191a98db0281915e9ce39797\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2417b66816da1a6cd1791a4b91eca4a3c4f3694462e066e7f0722f1b9c2105c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2417b66816da1a6cd1791a4b91eca4a3c4f3694462e066e7f0722f1b9c2105c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://883641f29e9120c2cea6b272bacf1dc63c752f55ab0408cdaf73db85c85ac369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://883641f29e9120c2cea6b272bacf1dc63c752f55ab0408cdaf73db85c85ac369\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2745baba1188ba5e8f820616e4117a12c50ecee5b68e71ad9cabbb621b6e131f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2745baba1188ba5e8f820616e4117a12c50ecee5b68e71ad9cabbb621b6e131f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8925e435be4d48d86824fc35f8b8819fb35d63e385164c4a20edb3bf3035c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a8925e435be4d48d86824fc35f8b8819fb35d63e385164c4a20edb3bf3035c3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d3fbf668ae7d04c8d9e574f43c2fe90c5604204ba38a3c60847afdfbceb7a57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d3fbf668ae7d04c8d9e574f43c2fe90c5604204ba38a3c60847afdfbceb7a57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9p5dc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:08Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:08 crc kubenswrapper[4764]: I0202 09:08:08.991243 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:08 crc kubenswrapper[4764]: I0202 09:08:08.991318 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:08 crc kubenswrapper[4764]: I0202 09:08:08.991331 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:08 crc kubenswrapper[4764]: I0202 09:08:08.991347 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:08 crc kubenswrapper[4764]: I0202 09:08:08.991357 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:08Z","lastTransitionTime":"2026-02-02T09:08:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:08 crc kubenswrapper[4764]: I0202 09:08:08.994371 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kwtmr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"991faa9a-dd25-4f49-82bd-ce60cefd4af2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h865b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h865b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:53Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kwtmr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:08Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:09 crc kubenswrapper[4764]: I0202 09:08:09.008911 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87d61aea-46f6-4ee6-a7a4-767a6a67fea4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e78b6d08e4d616cefbb73e3fd43b3c6886dcd8aa997c7b4948bf1ff63401293\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d468715c2da33996561c40043216233fd12e21539d0c58dd1f3b892fd9f91a9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b184f7c936dc4046f74c0018269cbc2c4b7ceddb650b4e871d3f8a707e0df88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://269c5074371bd00f8363fb3888c11f1df383081a61497aea6578f08700881d50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b25bc9395831524c53f9e565b4efbe74f21792efafeb1dd507161bc91aa0ecdf\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T09:07:22Z\\\",\\\"message\\\":\\\"W0202 09:07:21.597488 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0202 09:07:21.598117 1 crypto.go:601] Generating new CA for check-endpoints-signer@1770023241 cert, and key in /tmp/serving-cert-3856183724/serving-signer.crt, /tmp/serving-cert-3856183724/serving-signer.key\\\\nI0202 09:07:22.012184 1 observer_polling.go:159] Starting file observer\\\\nW0202 09:07:22.015493 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0202 09:07:22.015797 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 09:07:22.016596 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3856183724/tls.crt::/tmp/serving-cert-3856183724/tls.key\\\\\\\"\\\\nF0202 09:07:22.270605 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b44e17b8ca01f549bc10081dd5f109cb1b9fff364bf5bd4fd90d9f0e3eeb04db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfb06a5ec9262cf39b108a66014d15e325cdbd5441a2d9e7a811dd57265c99a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfb06a5ec9262cf39b108a66014d15e325cdbd5441a2d9e7a811dd57265c99a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:09Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:09 crc kubenswrapper[4764]: I0202 09:08:09.093169 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:09 crc kubenswrapper[4764]: I0202 09:08:09.093201 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:09 crc kubenswrapper[4764]: I0202 09:08:09.093209 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:09 crc kubenswrapper[4764]: I0202 09:08:09.093222 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:09 crc kubenswrapper[4764]: I0202 09:08:09.093230 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:09Z","lastTransitionTime":"2026-02-02T09:08:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:09 crc kubenswrapper[4764]: I0202 09:08:09.155918 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zhn7j_24632dda-6100-4ab6-a28e-214ddae4360c/ovnkube-controller/1.log" Feb 02 09:08:09 crc kubenswrapper[4764]: I0202 09:08:09.159640 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" event={"ID":"24632dda-6100-4ab6-a28e-214ddae4360c","Type":"ContainerStarted","Data":"334e828f8ef72570f2511ee38659941202530471f5943aa07b8b21483bd70550"} Feb 02 09:08:09 crc kubenswrapper[4764]: I0202 09:08:09.173244 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d192f670-9f9d-4539-9641-e4bed73acdd4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://040c5e8a5659f85f0d52d1a1e064154d998ae104f711101d2204c0b1caebbb30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbksf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fdf06410349b7c99aa166e87fbe80bf68e2ae923635a6a978b7e2cd50383f7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbksf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4ndm4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:09Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:09 crc kubenswrapper[4764]: I0202 09:08:09.183199 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pcjmn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6e5ccd3b-b657-47a2-9c4d-bc9adb126533\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72863eef568d92a1abeb3cde2e89630a4f5fb7626c6949448ff5195774c98eab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2f7zl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pcjmn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:09Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:09 crc kubenswrapper[4764]: I0202 09:08:09.196361 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed747fd8-e3ce-4319-a6da-bda0fd51b85a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6f064bdc66470aace3180d4988244026b54009a63cff199e41c5d02a1d18479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://60258f84a1d8dcfe48b7e9688ed50f077e201971f20f3f6663fbf1ba58a4611a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f84be2e08217a2ac3edde858d78b4410376ed9bae3bcabe2a7f9b2fbe2f69c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a69d25d352d7ceac5c735bc1fcf446a1fea6dc22c65a5296cfab828cea17377\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:09Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:09 crc kubenswrapper[4764]: I0202 09:08:09.196712 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:09 crc kubenswrapper[4764]: I0202 09:08:09.196734 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:09 crc kubenswrapper[4764]: I0202 09:08:09.196743 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:09 crc kubenswrapper[4764]: I0202 09:08:09.196757 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:09 crc kubenswrapper[4764]: I0202 09:08:09.196766 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:09Z","lastTransitionTime":"2026-02-02T09:08:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:09 crc kubenswrapper[4764]: I0202 09:08:09.208433 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f6904228e052fff69ad1fd66df97758b6b60ba4351e97a7fbd03eee4df1b653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:09Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:09 crc kubenswrapper[4764]: I0202 09:08:09.217875 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6zxff" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aaba93da-12a5-43df-b7c7-4c8b988163c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3db293678a907cd55b31b1d0c0bd7c30e1ba75ecd5e13cbc68ab98fbf0fbd9d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqwb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6zxff\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:09Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:09 crc kubenswrapper[4764]: I0202 09:08:09.230280 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fae745b-ea49-42f3-a486-dbc3dc2fb9ad\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://081d7f76d937ff71e35945f9b1d893888bb649d15bd84312fd4b61c9520ea71f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://429ead0993f5e648f6256d249ee12aea03514d3108618840552b672fb419cd74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fad17b01545eb20785f03e3c61bd5c982f6217e38176d7a3b23ac194298644a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85e2f75c56d63cacc93b9195c1e8bf46eb7a7234c45a024bd95a460db90edb6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85e2f75c56d63cacc93b9195c1e8bf46eb7a7234c45a024bd95a460db90edb6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:09Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:09 crc kubenswrapper[4764]: I0202 09:08:09.243313 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:09Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:09 crc kubenswrapper[4764]: I0202 09:08:09.255007 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lh6tm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2152e8a-f456-47b5-a547-b65ec837ecbc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38e176a00b07f2960fdd7ba2aadd1468a5e083b8d154e9d118d13f2585da81e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bb2sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://996e649011d5255783284421ddea8e0a5c383ef675b0920c3fdc0af6503d80e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bb2sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lh6tm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:09Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:09 crc kubenswrapper[4764]: I0202 09:08:09.266749 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:09Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:09 crc kubenswrapper[4764]: I0202 09:08:09.278380 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nc7qt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1a2c7b22068b7165927fc158885d48bf9952b9eba521d8ae9a599eee53dc305\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2hz4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nc7qt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:09Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:09 crc kubenswrapper[4764]: I0202 09:08:09.293524 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9p5dc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"660ac006-2495-45c4-9fb3-e9c1dddcf7a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da812b124c97aa8ad1563f4bcfac8f3bfa025096fa4a2561e5847b1bf9cb30b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23e57e65a6f5c3059506b9db21e9ac2492c30bd9191a98db0281915e9ce39797\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23e57e65a6f5c3059506b9db21e9ac2492c30bd9191a98db0281915e9ce39797\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2417b66816da1a6cd1791a4b91eca4a3c4f3694462e066e7f0722f1b9c2105c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2417b66816da1a6cd1791a4b91eca4a3c4f3694462e066e7f0722f1b9c2105c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://883641f29e9120c2cea6b272bacf1dc63c752f55ab0408cdaf73db85c85ac369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://883641f29e9120c2cea6b272bacf1dc63c752f55ab0408cdaf73db85c85ac369\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2745baba1188ba5e8f820616e4117a12c50ecee5b68e71ad9cabbb621b6e131f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2745baba1188ba5e8f820616e4117a12c50ecee5b68e71ad9cabbb621b6e131f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8925e435be4d48d86824fc35f8b8819fb35d63e385164c4a20edb3bf3035c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a8925e435be4d48d86824fc35f8b8819fb35d63e385164c4a20edb3bf3035c3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d3fbf668ae7d04c8d9e574f43c2fe90c5604204ba38a3c60847afdfbceb7a57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d3fbf668ae7d04c8d9e574f43c2fe90c5604204ba38a3c60847afdfbceb7a57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9p5dc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:09Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:09 crc kubenswrapper[4764]: I0202 09:08:09.299451 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:09 crc kubenswrapper[4764]: I0202 09:08:09.299486 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:09 crc kubenswrapper[4764]: I0202 09:08:09.299496 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:09 crc kubenswrapper[4764]: I0202 09:08:09.299513 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:09 crc kubenswrapper[4764]: I0202 09:08:09.299523 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:09Z","lastTransitionTime":"2026-02-02T09:08:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:09 crc kubenswrapper[4764]: I0202 09:08:09.303796 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kwtmr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"991faa9a-dd25-4f49-82bd-ce60cefd4af2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h865b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h865b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:53Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kwtmr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:09Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:09 crc kubenswrapper[4764]: I0202 09:08:09.317905 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87d61aea-46f6-4ee6-a7a4-767a6a67fea4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e78b6d08e4d616cefbb73e3fd43b3c6886dcd8aa997c7b4948bf1ff63401293\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d468715c2da33996561c40043216233fd12e21539d0c58dd1f3b892fd9f91a9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b184f7c936dc4046f74c0018269cbc2c4b7ceddb650b4e871d3f8a707e0df88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://269c5074371bd00f8363fb3888c11f1df383081a61497aea6578f08700881d50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b25bc9395831524c53f9e565b4efbe74f21792efafeb1dd507161bc91aa0ecdf\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T09:07:22Z\\\",\\\"message\\\":\\\"W0202 09:07:21.597488 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0202 09:07:21.598117 1 crypto.go:601] Generating new CA for check-endpoints-signer@1770023241 cert, and key in /tmp/serving-cert-3856183724/serving-signer.crt, /tmp/serving-cert-3856183724/serving-signer.key\\\\nI0202 09:07:22.012184 1 observer_polling.go:159] Starting file observer\\\\nW0202 09:07:22.015493 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0202 09:07:22.015797 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 09:07:22.016596 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3856183724/tls.crt::/tmp/serving-cert-3856183724/tls.key\\\\\\\"\\\\nF0202 09:07:22.270605 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b44e17b8ca01f549bc10081dd5f109cb1b9fff364bf5bd4fd90d9f0e3eeb04db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfb06a5ec9262cf39b108a66014d15e325cdbd5441a2d9e7a811dd57265c99a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfb06a5ec9262cf39b108a66014d15e325cdbd5441a2d9e7a811dd57265c99a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:09Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:09 crc kubenswrapper[4764]: I0202 09:08:09.332461 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5528225d91ac2778bef4554abb0bd036ac5d3b6cdac7357bab69e70c3894b6f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://071bbb80d0747cf3e109ba626d47f7d05f85f4cc0b3aba77276c795741f47ac6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:09Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:09 crc kubenswrapper[4764]: I0202 09:08:09.355175 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24632dda-6100-4ab6-a28e-214ddae4360c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a6c67e9b7708cc0d85b0b6ffbbade0d34867e159a9c5487ca2606475c667a50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bffda48d7c881c1470251b501661324859373246cc1a21811bd36ca24887c7aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a989645cf485cd8ff80ffff1a86cd7808e1564eba5667535e161cc68db13b26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20b71d3ee7a381baa51b0b0d0470283022cfede30f491135737b5e82d631d763\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c380383b6a014d1b694a13ddc22f54b20b9e4f9d5c9b9000d76187621dce0a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5529f2bff87753428e9173826fc1880acf96a4c134fa9ecad195b15697cad73a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://334e828f8ef72570f2511ee38659941202530471f5943aa07b8b21483bd70550\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://12244061eabaae7fa9057ac5cee894c3b308b466411bcb910682e7f65d240dd7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-02T09:07:52Z\\\",\\\"message\\\":\\\"l_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-kube-apiserver/apiserver]} name:Service_openshift-kube-apiserver/apiserver_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.93:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {d71b38eb-32af-4c0f-9490-7c317c111e3a}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0202 09:07:51.914338 6129 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-kube-apiserver/apiserver]} name:Service_openshift-kube-apiserver/apiserver_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.93:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {d71b38eb-32af-4c0f-9490-7c317c111e3a}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF0202 09:07:51.914404 6129 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:51Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b4cfbd4e682bcdcaff785013bdb72580ce36fced733e235538536827036b93d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://579903eadb183466a2564737e6b678e04308239c0cd75b6808892053b97afbc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://579903eadb183466a2564737e6b678e04308239c0cd75b6808892053b97afbc0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zhn7j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:09Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:09 crc kubenswrapper[4764]: I0202 09:08:09.383988 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d66ab3c0-9b5b-41d5-b6c0-91ccddccd7e9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d92ac486f5dade46bc75de32099859a71fc2ed25a7abbe6ae8937df128e7df30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a4ac273a4f127d79e5fad7d4c30cd2044a346f5734dba5d33d68cd092c0c2d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d38c7cb5dcd22fb07524494ebe77aa1a7f4f808f566043563817e47e7ec12a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://408722c7f83d28919431532f8c797a1c795fd5d96736d14e7f06a05744899702\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f473b18cd8003f75bb1dafb82bcec45745a73ccd10434297356dac185a3d15c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e229f484d19cbdabedf53d90a822a4434db9e00b1301538e885e26acc19d25f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e229f484d19cbdabedf53d90a822a4434db9e00b1301538e885e26acc19d25f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a9cc673be8ff245dba81278e8cd34cfe45ed36ff0b8bd5b30a0ddf333ec6371\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a9cc673be8ff245dba81278e8cd34cfe45ed36ff0b8bd5b30a0ddf333ec6371\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bc7e49c7a5ae66ea939060192fa6a346dc0b0b9aa9750b29879d3c426cc0d276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc7e49c7a5ae66ea939060192fa6a346dc0b0b9aa9750b29879d3c426cc0d276\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:09Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:09 crc kubenswrapper[4764]: I0202 09:08:09.402348 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:09 crc kubenswrapper[4764]: I0202 09:08:09.402381 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:09 crc kubenswrapper[4764]: I0202 09:08:09.402391 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:09 crc kubenswrapper[4764]: I0202 09:08:09.402404 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:09 crc kubenswrapper[4764]: I0202 09:08:09.402412 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:09Z","lastTransitionTime":"2026-02-02T09:08:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:09 crc kubenswrapper[4764]: I0202 09:08:09.409136 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa9d037e66cb53b9dd698eff46284c9e718fd842cda6f7b525f0c79062d70bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:09Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:09 crc kubenswrapper[4764]: I0202 09:08:09.422884 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:09Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:09 crc kubenswrapper[4764]: I0202 09:08:09.504367 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:09 crc kubenswrapper[4764]: I0202 09:08:09.504401 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:09 crc kubenswrapper[4764]: I0202 09:08:09.504409 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:09 crc kubenswrapper[4764]: I0202 09:08:09.504423 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:09 crc kubenswrapper[4764]: I0202 09:08:09.504432 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:09Z","lastTransitionTime":"2026-02-02T09:08:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:09 crc kubenswrapper[4764]: I0202 09:08:09.606741 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:09 crc kubenswrapper[4764]: I0202 09:08:09.606784 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:09 crc kubenswrapper[4764]: I0202 09:08:09.606792 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:09 crc kubenswrapper[4764]: I0202 09:08:09.606806 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:09 crc kubenswrapper[4764]: I0202 09:08:09.606815 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:09Z","lastTransitionTime":"2026-02-02T09:08:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:09 crc kubenswrapper[4764]: I0202 09:08:09.709356 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:09 crc kubenswrapper[4764]: I0202 09:08:09.709383 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:09 crc kubenswrapper[4764]: I0202 09:08:09.709393 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:09 crc kubenswrapper[4764]: I0202 09:08:09.709408 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:09 crc kubenswrapper[4764]: I0202 09:08:09.709419 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:09Z","lastTransitionTime":"2026-02-02T09:08:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:09 crc kubenswrapper[4764]: I0202 09:08:09.803533 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-27 10:49:46.624886608 +0000 UTC Feb 02 09:08:09 crc kubenswrapper[4764]: I0202 09:08:09.811309 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:09 crc kubenswrapper[4764]: I0202 09:08:09.811357 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:09 crc kubenswrapper[4764]: I0202 09:08:09.811370 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:09 crc kubenswrapper[4764]: I0202 09:08:09.811387 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:09 crc kubenswrapper[4764]: I0202 09:08:09.811400 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:09Z","lastTransitionTime":"2026-02-02T09:08:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:09 crc kubenswrapper[4764]: I0202 09:08:09.812699 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/991faa9a-dd25-4f49-82bd-ce60cefd4af2-metrics-certs\") pod \"network-metrics-daemon-kwtmr\" (UID: \"991faa9a-dd25-4f49-82bd-ce60cefd4af2\") " pod="openshift-multus/network-metrics-daemon-kwtmr" Feb 02 09:08:09 crc kubenswrapper[4764]: E0202 09:08:09.812877 4764 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 02 09:08:09 crc kubenswrapper[4764]: E0202 09:08:09.812969 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/991faa9a-dd25-4f49-82bd-ce60cefd4af2-metrics-certs podName:991faa9a-dd25-4f49-82bd-ce60cefd4af2 nodeName:}" failed. No retries permitted until 2026-02-02 09:08:25.81294704 +0000 UTC m=+68.746671138 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/991faa9a-dd25-4f49-82bd-ce60cefd4af2-metrics-certs") pod "network-metrics-daemon-kwtmr" (UID: "991faa9a-dd25-4f49-82bd-ce60cefd4af2") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 02 09:08:09 crc kubenswrapper[4764]: I0202 09:08:09.825556 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kwtmr" Feb 02 09:08:09 crc kubenswrapper[4764]: E0202 09:08:09.825751 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kwtmr" podUID="991faa9a-dd25-4f49-82bd-ce60cefd4af2" Feb 02 09:08:09 crc kubenswrapper[4764]: I0202 09:08:09.914292 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:09 crc kubenswrapper[4764]: I0202 09:08:09.914331 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:09 crc kubenswrapper[4764]: I0202 09:08:09.914343 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:09 crc kubenswrapper[4764]: I0202 09:08:09.914360 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:09 crc kubenswrapper[4764]: I0202 09:08:09.914372 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:09Z","lastTransitionTime":"2026-02-02T09:08:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:10 crc kubenswrapper[4764]: I0202 09:08:10.103568 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:10 crc kubenswrapper[4764]: I0202 09:08:10.103612 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:10 crc kubenswrapper[4764]: I0202 09:08:10.103623 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:10 crc kubenswrapper[4764]: I0202 09:08:10.103642 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:10 crc kubenswrapper[4764]: I0202 09:08:10.103653 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:10Z","lastTransitionTime":"2026-02-02T09:08:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:10 crc kubenswrapper[4764]: I0202 09:08:10.164783 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zhn7j_24632dda-6100-4ab6-a28e-214ddae4360c/ovnkube-controller/2.log" Feb 02 09:08:10 crc kubenswrapper[4764]: I0202 09:08:10.165543 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zhn7j_24632dda-6100-4ab6-a28e-214ddae4360c/ovnkube-controller/1.log" Feb 02 09:08:10 crc kubenswrapper[4764]: I0202 09:08:10.168303 4764 generic.go:334] "Generic (PLEG): container finished" podID="24632dda-6100-4ab6-a28e-214ddae4360c" containerID="334e828f8ef72570f2511ee38659941202530471f5943aa07b8b21483bd70550" exitCode=1 Feb 02 09:08:10 crc kubenswrapper[4764]: I0202 09:08:10.168381 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" event={"ID":"24632dda-6100-4ab6-a28e-214ddae4360c","Type":"ContainerDied","Data":"334e828f8ef72570f2511ee38659941202530471f5943aa07b8b21483bd70550"} Feb 02 09:08:10 crc kubenswrapper[4764]: I0202 09:08:10.168460 4764 scope.go:117] "RemoveContainer" containerID="12244061eabaae7fa9057ac5cee894c3b308b466411bcb910682e7f65d240dd7" Feb 02 09:08:10 crc kubenswrapper[4764]: I0202 09:08:10.169430 4764 scope.go:117] "RemoveContainer" containerID="334e828f8ef72570f2511ee38659941202530471f5943aa07b8b21483bd70550" Feb 02 09:08:10 crc kubenswrapper[4764]: E0202 09:08:10.169692 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-zhn7j_openshift-ovn-kubernetes(24632dda-6100-4ab6-a28e-214ddae4360c)\"" pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" podUID="24632dda-6100-4ab6-a28e-214ddae4360c" Feb 02 09:08:10 crc kubenswrapper[4764]: I0202 09:08:10.189560 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nc7qt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1a2c7b22068b7165927fc158885d48bf9952b9eba521d8ae9a599eee53dc305\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2hz4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nc7qt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:10Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:10 crc kubenswrapper[4764]: I0202 09:08:10.205337 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:10 crc kubenswrapper[4764]: I0202 09:08:10.205363 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:10 crc kubenswrapper[4764]: I0202 09:08:10.205371 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:10 crc kubenswrapper[4764]: I0202 09:08:10.205383 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:10 crc kubenswrapper[4764]: I0202 09:08:10.205392 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:10Z","lastTransitionTime":"2026-02-02T09:08:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:10 crc kubenswrapper[4764]: I0202 09:08:10.206344 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9p5dc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"660ac006-2495-45c4-9fb3-e9c1dddcf7a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da812b124c97aa8ad1563f4bcfac8f3bfa025096fa4a2561e5847b1bf9cb30b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23e57e65a6f5c3059506b9db21e9ac2492c30bd9191a98db0281915e9ce39797\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23e57e65a6f5c3059506b9db21e9ac2492c30bd9191a98db0281915e9ce39797\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2417b66816da1a6cd1791a4b91eca4a3c4f3694462e066e7f0722f1b9c2105c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2417b66816da1a6cd1791a4b91eca4a3c4f3694462e066e7f0722f1b9c2105c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://883641f29e9120c2cea6b272bacf1dc63c752f55ab0408cdaf73db85c85ac369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://883641f29e9120c2cea6b272bacf1dc63c752f55ab0408cdaf73db85c85ac369\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2745baba1188ba5e8f820616e4117a12c50ecee5b68e71ad9cabbb621b6e131f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2745baba1188ba5e8f820616e4117a12c50ecee5b68e71ad9cabbb621b6e131f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8925e435be4d48d86824fc35f8b8819fb35d63e385164c4a20edb3bf3035c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a8925e435be4d48d86824fc35f8b8819fb35d63e385164c4a20edb3bf3035c3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d3fbf668ae7d04c8d9e574f43c2fe90c5604204ba38a3c60847afdfbceb7a57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d3fbf668ae7d04c8d9e574f43c2fe90c5604204ba38a3c60847afdfbceb7a57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9p5dc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:10Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:10 crc kubenswrapper[4764]: I0202 09:08:10.216464 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kwtmr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"991faa9a-dd25-4f49-82bd-ce60cefd4af2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h865b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h865b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:53Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kwtmr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:10Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:10 crc kubenswrapper[4764]: I0202 09:08:10.230974 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87d61aea-46f6-4ee6-a7a4-767a6a67fea4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e78b6d08e4d616cefbb73e3fd43b3c6886dcd8aa997c7b4948bf1ff63401293\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d468715c2da33996561c40043216233fd12e21539d0c58dd1f3b892fd9f91a9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b184f7c936dc4046f74c0018269cbc2c4b7ceddb650b4e871d3f8a707e0df88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://269c5074371bd00f8363fb3888c11f1df383081a61497aea6578f08700881d50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b25bc9395831524c53f9e565b4efbe74f21792efafeb1dd507161bc91aa0ecdf\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T09:07:22Z\\\",\\\"message\\\":\\\"W0202 09:07:21.597488 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0202 09:07:21.598117 1 crypto.go:601] Generating new CA for check-endpoints-signer@1770023241 cert, and key in /tmp/serving-cert-3856183724/serving-signer.crt, /tmp/serving-cert-3856183724/serving-signer.key\\\\nI0202 09:07:22.012184 1 observer_polling.go:159] Starting file observer\\\\nW0202 09:07:22.015493 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0202 09:07:22.015797 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 09:07:22.016596 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3856183724/tls.crt::/tmp/serving-cert-3856183724/tls.key\\\\\\\"\\\\nF0202 09:07:22.270605 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b44e17b8ca01f549bc10081dd5f109cb1b9fff364bf5bd4fd90d9f0e3eeb04db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfb06a5ec9262cf39b108a66014d15e325cdbd5441a2d9e7a811dd57265c99a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfb06a5ec9262cf39b108a66014d15e325cdbd5441a2d9e7a811dd57265c99a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:10Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:10 crc kubenswrapper[4764]: I0202 09:08:10.243609 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5528225d91ac2778bef4554abb0bd036ac5d3b6cdac7357bab69e70c3894b6f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://071bbb80d0747cf3e109ba626d47f7d05f85f4cc0b3aba77276c795741f47ac6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:10Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:10 crc kubenswrapper[4764]: I0202 09:08:10.259248 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:10Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:10 crc kubenswrapper[4764]: I0202 09:08:10.277393 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d66ab3c0-9b5b-41d5-b6c0-91ccddccd7e9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d92ac486f5dade46bc75de32099859a71fc2ed25a7abbe6ae8937df128e7df30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a4ac273a4f127d79e5fad7d4c30cd2044a346f5734dba5d33d68cd092c0c2d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d38c7cb5dcd22fb07524494ebe77aa1a7f4f808f566043563817e47e7ec12a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://408722c7f83d28919431532f8c797a1c795fd5d96736d14e7f06a05744899702\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f473b18cd8003f75bb1dafb82bcec45745a73ccd10434297356dac185a3d15c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e229f484d19cbdabedf53d90a822a4434db9e00b1301538e885e26acc19d25f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e229f484d19cbdabedf53d90a822a4434db9e00b1301538e885e26acc19d25f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a9cc673be8ff245dba81278e8cd34cfe45ed36ff0b8bd5b30a0ddf333ec6371\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a9cc673be8ff245dba81278e8cd34cfe45ed36ff0b8bd5b30a0ddf333ec6371\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bc7e49c7a5ae66ea939060192fa6a346dc0b0b9aa9750b29879d3c426cc0d276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc7e49c7a5ae66ea939060192fa6a346dc0b0b9aa9750b29879d3c426cc0d276\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:10Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:10 crc kubenswrapper[4764]: I0202 09:08:10.290044 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa9d037e66cb53b9dd698eff46284c9e718fd842cda6f7b525f0c79062d70bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:10Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:10 crc kubenswrapper[4764]: I0202 09:08:10.301718 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:10Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:10 crc kubenswrapper[4764]: I0202 09:08:10.307722 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:10 crc kubenswrapper[4764]: I0202 09:08:10.307759 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:10 crc kubenswrapper[4764]: I0202 09:08:10.307769 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:10 crc kubenswrapper[4764]: I0202 09:08:10.307787 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:10 crc kubenswrapper[4764]: I0202 09:08:10.307800 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:10Z","lastTransitionTime":"2026-02-02T09:08:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:10 crc kubenswrapper[4764]: I0202 09:08:10.319922 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24632dda-6100-4ab6-a28e-214ddae4360c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a6c67e9b7708cc0d85b0b6ffbbade0d34867e159a9c5487ca2606475c667a50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bffda48d7c881c1470251b501661324859373246cc1a21811bd36ca24887c7aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a989645cf485cd8ff80ffff1a86cd7808e1564eba5667535e161cc68db13b26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20b71d3ee7a381baa51b0b0d0470283022cfede30f491135737b5e82d631d763\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c380383b6a014d1b694a13ddc22f54b20b9e4f9d5c9b9000d76187621dce0a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5529f2bff87753428e9173826fc1880acf96a4c134fa9ecad195b15697cad73a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://334e828f8ef72570f2511ee38659941202530471f5943aa07b8b21483bd70550\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://12244061eabaae7fa9057ac5cee894c3b308b466411bcb910682e7f65d240dd7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-02T09:07:52Z\\\",\\\"message\\\":\\\"l_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-kube-apiserver/apiserver]} name:Service_openshift-kube-apiserver/apiserver_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.93:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {d71b38eb-32af-4c0f-9490-7c317c111e3a}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0202 09:07:51.914338 6129 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-kube-apiserver/apiserver]} name:Service_openshift-kube-apiserver/apiserver_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.93:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {d71b38eb-32af-4c0f-9490-7c317c111e3a}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF0202 09:07:51.914404 6129 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:51Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://334e828f8ef72570f2511ee38659941202530471f5943aa07b8b21483bd70550\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-02T09:08:10Z\\\",\\\"message\\\":\\\"ing OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-dns/dns-default]} name:Service_openshift-dns/dns-default_TCP_node_router+switch_crc options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.10:53: 10.217.4.10:9154:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {be9dcc9e-c16a-4962-a6d2-4adeb0b929c4}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-dns/dns-default]} name:Service_openshift-dns/dns-default_UDP_node_router+switch_crc options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[udp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.10:53:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {4c1be812-05d3-4f45-91b5-a853a5c8de71}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF0202 09:08:09.911531 6327 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T09:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b4cfbd4e682bcdcaff785013bdb72580ce36fced733e235538536827036b93d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://579903eadb183466a2564737e6b678e04308239c0cd75b6808892053b97afbc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://579903eadb183466a2564737e6b678e04308239c0cd75b6808892053b97afbc0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zhn7j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:10Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:10 crc kubenswrapper[4764]: I0202 09:08:10.330671 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pcjmn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6e5ccd3b-b657-47a2-9c4d-bc9adb126533\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72863eef568d92a1abeb3cde2e89630a4f5fb7626c6949448ff5195774c98eab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2f7zl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pcjmn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:10Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:10 crc kubenswrapper[4764]: I0202 09:08:10.345451 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed747fd8-e3ce-4319-a6da-bda0fd51b85a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6f064bdc66470aace3180d4988244026b54009a63cff199e41c5d02a1d18479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://60258f84a1d8dcfe48b7e9688ed50f077e201971f20f3f6663fbf1ba58a4611a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f84be2e08217a2ac3edde858d78b4410376ed9bae3bcabe2a7f9b2fbe2f69c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a69d25d352d7ceac5c735bc1fcf446a1fea6dc22c65a5296cfab828cea17377\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:10Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:10 crc kubenswrapper[4764]: I0202 09:08:10.361977 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f6904228e052fff69ad1fd66df97758b6b60ba4351e97a7fbd03eee4df1b653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:10Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:10 crc kubenswrapper[4764]: I0202 09:08:10.374880 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6zxff" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aaba93da-12a5-43df-b7c7-4c8b988163c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3db293678a907cd55b31b1d0c0bd7c30e1ba75ecd5e13cbc68ab98fbf0fbd9d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqwb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6zxff\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:10Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:10 crc kubenswrapper[4764]: I0202 09:08:10.387297 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d192f670-9f9d-4539-9641-e4bed73acdd4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://040c5e8a5659f85f0d52d1a1e064154d998ae104f711101d2204c0b1caebbb30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbksf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fdf06410349b7c99aa166e87fbe80bf68e2ae923635a6a978b7e2cd50383f7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbksf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4ndm4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:10Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:10 crc kubenswrapper[4764]: I0202 09:08:10.400057 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fae745b-ea49-42f3-a486-dbc3dc2fb9ad\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://081d7f76d937ff71e35945f9b1d893888bb649d15bd84312fd4b61c9520ea71f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://429ead0993f5e648f6256d249ee12aea03514d3108618840552b672fb419cd74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fad17b01545eb20785f03e3c61bd5c982f6217e38176d7a3b23ac194298644a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85e2f75c56d63cacc93b9195c1e8bf46eb7a7234c45a024bd95a460db90edb6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85e2f75c56d63cacc93b9195c1e8bf46eb7a7234c45a024bd95a460db90edb6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:10Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:10 crc kubenswrapper[4764]: I0202 09:08:10.410813 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:10 crc kubenswrapper[4764]: I0202 09:08:10.410853 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:10 crc kubenswrapper[4764]: I0202 09:08:10.410865 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:10 crc kubenswrapper[4764]: I0202 09:08:10.410882 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:10 crc kubenswrapper[4764]: I0202 09:08:10.410896 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:10Z","lastTransitionTime":"2026-02-02T09:08:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:10 crc kubenswrapper[4764]: I0202 09:08:10.413479 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:10Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:10 crc kubenswrapper[4764]: I0202 09:08:10.423513 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lh6tm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2152e8a-f456-47b5-a547-b65ec837ecbc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38e176a00b07f2960fdd7ba2aadd1468a5e083b8d154e9d118d13f2585da81e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bb2sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://996e649011d5255783284421ddea8e0a5c383ef675b0920c3fdc0af6503d80e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bb2sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lh6tm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:10Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:10 crc kubenswrapper[4764]: I0202 09:08:10.513272 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:10 crc kubenswrapper[4764]: I0202 09:08:10.513352 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:10 crc kubenswrapper[4764]: I0202 09:08:10.513376 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:10 crc kubenswrapper[4764]: I0202 09:08:10.513408 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:10 crc kubenswrapper[4764]: I0202 09:08:10.513431 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:10Z","lastTransitionTime":"2026-02-02T09:08:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:10 crc kubenswrapper[4764]: I0202 09:08:10.520097 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 09:08:10 crc kubenswrapper[4764]: E0202 09:08:10.520231 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 09:08:42.520204111 +0000 UTC m=+85.453928239 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:08:10 crc kubenswrapper[4764]: I0202 09:08:10.616630 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:10 crc kubenswrapper[4764]: I0202 09:08:10.616690 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:10 crc kubenswrapper[4764]: I0202 09:08:10.616715 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:10 crc kubenswrapper[4764]: I0202 09:08:10.616744 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:10 crc kubenswrapper[4764]: I0202 09:08:10.616765 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:10Z","lastTransitionTime":"2026-02-02T09:08:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:10 crc kubenswrapper[4764]: I0202 09:08:10.621678 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 09:08:10 crc kubenswrapper[4764]: I0202 09:08:10.621776 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 09:08:10 crc kubenswrapper[4764]: I0202 09:08:10.621819 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 09:08:10 crc kubenswrapper[4764]: I0202 09:08:10.621855 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 09:08:10 crc kubenswrapper[4764]: E0202 09:08:10.621908 4764 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 02 09:08:10 crc kubenswrapper[4764]: E0202 09:08:10.622001 4764 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 02 09:08:10 crc kubenswrapper[4764]: E0202 09:08:10.622026 4764 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 02 09:08:10 crc kubenswrapper[4764]: E0202 09:08:10.622036 4764 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 02 09:08:10 crc kubenswrapper[4764]: E0202 09:08:10.622047 4764 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 02 09:08:10 crc kubenswrapper[4764]: E0202 09:08:10.622066 4764 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 02 09:08:10 crc kubenswrapper[4764]: E0202 09:08:10.622036 4764 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 02 09:08:10 crc kubenswrapper[4764]: E0202 09:08:10.622088 4764 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 02 09:08:10 crc kubenswrapper[4764]: E0202 09:08:10.622120 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-02 09:08:42.622101135 +0000 UTC m=+85.555825233 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 02 09:08:10 crc kubenswrapper[4764]: E0202 09:08:10.622140 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-02 09:08:42.622132286 +0000 UTC m=+85.555856384 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 02 09:08:10 crc kubenswrapper[4764]: E0202 09:08:10.622170 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-02 09:08:42.622149307 +0000 UTC m=+85.555873475 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 02 09:08:10 crc kubenswrapper[4764]: E0202 09:08:10.622196 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-02 09:08:42.622185257 +0000 UTC m=+85.555909385 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 02 09:08:10 crc kubenswrapper[4764]: I0202 09:08:10.719296 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:10 crc kubenswrapper[4764]: I0202 09:08:10.719342 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:10 crc kubenswrapper[4764]: I0202 09:08:10.719358 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:10 crc kubenswrapper[4764]: I0202 09:08:10.719378 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:10 crc kubenswrapper[4764]: I0202 09:08:10.719393 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:10Z","lastTransitionTime":"2026-02-02T09:08:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:10 crc kubenswrapper[4764]: I0202 09:08:10.804399 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-07 10:37:53.933269457 +0000 UTC Feb 02 09:08:10 crc kubenswrapper[4764]: I0202 09:08:10.822632 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:10 crc kubenswrapper[4764]: I0202 09:08:10.822666 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:10 crc kubenswrapper[4764]: I0202 09:08:10.822682 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:10 crc kubenswrapper[4764]: I0202 09:08:10.822701 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:10 crc kubenswrapper[4764]: I0202 09:08:10.822715 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:10Z","lastTransitionTime":"2026-02-02T09:08:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:10 crc kubenswrapper[4764]: I0202 09:08:10.825026 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 09:08:10 crc kubenswrapper[4764]: I0202 09:08:10.825134 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 09:08:10 crc kubenswrapper[4764]: E0202 09:08:10.825299 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 09:08:10 crc kubenswrapper[4764]: I0202 09:08:10.825396 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 09:08:10 crc kubenswrapper[4764]: E0202 09:08:10.825467 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 09:08:10 crc kubenswrapper[4764]: E0202 09:08:10.825870 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 09:08:10 crc kubenswrapper[4764]: I0202 09:08:10.926761 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:10 crc kubenswrapper[4764]: I0202 09:08:10.926831 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:10 crc kubenswrapper[4764]: I0202 09:08:10.926850 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:10 crc kubenswrapper[4764]: I0202 09:08:10.926875 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:10 crc kubenswrapper[4764]: I0202 09:08:10.926893 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:10Z","lastTransitionTime":"2026-02-02T09:08:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:11 crc kubenswrapper[4764]: I0202 09:08:11.029924 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:11 crc kubenswrapper[4764]: I0202 09:08:11.030031 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:11 crc kubenswrapper[4764]: I0202 09:08:11.030057 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:11 crc kubenswrapper[4764]: I0202 09:08:11.030086 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:11 crc kubenswrapper[4764]: I0202 09:08:11.030111 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:11Z","lastTransitionTime":"2026-02-02T09:08:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:11 crc kubenswrapper[4764]: I0202 09:08:11.133049 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:11 crc kubenswrapper[4764]: I0202 09:08:11.133117 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:11 crc kubenswrapper[4764]: I0202 09:08:11.133133 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:11 crc kubenswrapper[4764]: I0202 09:08:11.133157 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:11 crc kubenswrapper[4764]: I0202 09:08:11.133178 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:11Z","lastTransitionTime":"2026-02-02T09:08:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:11 crc kubenswrapper[4764]: I0202 09:08:11.173835 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zhn7j_24632dda-6100-4ab6-a28e-214ddae4360c/ovnkube-controller/2.log" Feb 02 09:08:11 crc kubenswrapper[4764]: I0202 09:08:11.236249 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:11 crc kubenswrapper[4764]: I0202 09:08:11.236325 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:11 crc kubenswrapper[4764]: I0202 09:08:11.236349 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:11 crc kubenswrapper[4764]: I0202 09:08:11.236381 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:11 crc kubenswrapper[4764]: I0202 09:08:11.236407 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:11Z","lastTransitionTime":"2026-02-02T09:08:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:11 crc kubenswrapper[4764]: I0202 09:08:11.339895 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:11 crc kubenswrapper[4764]: I0202 09:08:11.340209 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:11 crc kubenswrapper[4764]: I0202 09:08:11.340313 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:11 crc kubenswrapper[4764]: I0202 09:08:11.340385 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:11 crc kubenswrapper[4764]: I0202 09:08:11.340451 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:11Z","lastTransitionTime":"2026-02-02T09:08:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:11 crc kubenswrapper[4764]: I0202 09:08:11.443307 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:11 crc kubenswrapper[4764]: I0202 09:08:11.443342 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:11 crc kubenswrapper[4764]: I0202 09:08:11.443352 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:11 crc kubenswrapper[4764]: I0202 09:08:11.443366 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:11 crc kubenswrapper[4764]: I0202 09:08:11.443374 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:11Z","lastTransitionTime":"2026-02-02T09:08:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:11 crc kubenswrapper[4764]: I0202 09:08:11.547097 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:11 crc kubenswrapper[4764]: I0202 09:08:11.547183 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:11 crc kubenswrapper[4764]: I0202 09:08:11.547203 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:11 crc kubenswrapper[4764]: I0202 09:08:11.547223 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:11 crc kubenswrapper[4764]: I0202 09:08:11.547234 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:11Z","lastTransitionTime":"2026-02-02T09:08:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:11 crc kubenswrapper[4764]: I0202 09:08:11.650116 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:11 crc kubenswrapper[4764]: I0202 09:08:11.650190 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:11 crc kubenswrapper[4764]: I0202 09:08:11.650203 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:11 crc kubenswrapper[4764]: I0202 09:08:11.650222 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:11 crc kubenswrapper[4764]: I0202 09:08:11.650236 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:11Z","lastTransitionTime":"2026-02-02T09:08:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:11 crc kubenswrapper[4764]: I0202 09:08:11.753740 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:11 crc kubenswrapper[4764]: I0202 09:08:11.753818 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:11 crc kubenswrapper[4764]: I0202 09:08:11.753838 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:11 crc kubenswrapper[4764]: I0202 09:08:11.753877 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:11 crc kubenswrapper[4764]: I0202 09:08:11.753896 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:11Z","lastTransitionTime":"2026-02-02T09:08:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:11 crc kubenswrapper[4764]: I0202 09:08:11.804560 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-11 23:11:04.071369507 +0000 UTC Feb 02 09:08:11 crc kubenswrapper[4764]: I0202 09:08:11.825186 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kwtmr" Feb 02 09:08:11 crc kubenswrapper[4764]: E0202 09:08:11.825540 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kwtmr" podUID="991faa9a-dd25-4f49-82bd-ce60cefd4af2" Feb 02 09:08:11 crc kubenswrapper[4764]: I0202 09:08:11.857034 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:11 crc kubenswrapper[4764]: I0202 09:08:11.857312 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:11 crc kubenswrapper[4764]: I0202 09:08:11.857416 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:11 crc kubenswrapper[4764]: I0202 09:08:11.857536 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:11 crc kubenswrapper[4764]: I0202 09:08:11.857631 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:11Z","lastTransitionTime":"2026-02-02T09:08:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:11 crc kubenswrapper[4764]: I0202 09:08:11.959533 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:11 crc kubenswrapper[4764]: I0202 09:08:11.959998 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:11 crc kubenswrapper[4764]: I0202 09:08:11.960021 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:11 crc kubenswrapper[4764]: I0202 09:08:11.960045 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:11 crc kubenswrapper[4764]: I0202 09:08:11.960062 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:11Z","lastTransitionTime":"2026-02-02T09:08:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:12 crc kubenswrapper[4764]: I0202 09:08:12.062821 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:12 crc kubenswrapper[4764]: I0202 09:08:12.063270 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:12 crc kubenswrapper[4764]: I0202 09:08:12.063423 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:12 crc kubenswrapper[4764]: I0202 09:08:12.063604 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:12 crc kubenswrapper[4764]: I0202 09:08:12.063842 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:12Z","lastTransitionTime":"2026-02-02T09:08:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:12 crc kubenswrapper[4764]: I0202 09:08:12.166161 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:12 crc kubenswrapper[4764]: I0202 09:08:12.166857 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:12 crc kubenswrapper[4764]: I0202 09:08:12.166973 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:12 crc kubenswrapper[4764]: I0202 09:08:12.167072 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:12 crc kubenswrapper[4764]: I0202 09:08:12.167142 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:12Z","lastTransitionTime":"2026-02-02T09:08:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:12 crc kubenswrapper[4764]: I0202 09:08:12.270016 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:12 crc kubenswrapper[4764]: I0202 09:08:12.270052 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:12 crc kubenswrapper[4764]: I0202 09:08:12.270062 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:12 crc kubenswrapper[4764]: I0202 09:08:12.270077 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:12 crc kubenswrapper[4764]: I0202 09:08:12.270086 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:12Z","lastTransitionTime":"2026-02-02T09:08:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:12 crc kubenswrapper[4764]: I0202 09:08:12.372017 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:12 crc kubenswrapper[4764]: I0202 09:08:12.372293 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:12 crc kubenswrapper[4764]: I0202 09:08:12.372369 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:12 crc kubenswrapper[4764]: I0202 09:08:12.372436 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:12 crc kubenswrapper[4764]: I0202 09:08:12.372503 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:12Z","lastTransitionTime":"2026-02-02T09:08:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:12 crc kubenswrapper[4764]: I0202 09:08:12.475206 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:12 crc kubenswrapper[4764]: I0202 09:08:12.475275 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:12 crc kubenswrapper[4764]: I0202 09:08:12.475298 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:12 crc kubenswrapper[4764]: I0202 09:08:12.475320 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:12 crc kubenswrapper[4764]: I0202 09:08:12.475338 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:12Z","lastTransitionTime":"2026-02-02T09:08:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:12 crc kubenswrapper[4764]: I0202 09:08:12.578091 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:12 crc kubenswrapper[4764]: I0202 09:08:12.578155 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:12 crc kubenswrapper[4764]: I0202 09:08:12.578174 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:12 crc kubenswrapper[4764]: I0202 09:08:12.578197 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:12 crc kubenswrapper[4764]: I0202 09:08:12.578214 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:12Z","lastTransitionTime":"2026-02-02T09:08:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:12 crc kubenswrapper[4764]: I0202 09:08:12.679640 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:12 crc kubenswrapper[4764]: I0202 09:08:12.679669 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:12 crc kubenswrapper[4764]: I0202 09:08:12.679676 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:12 crc kubenswrapper[4764]: I0202 09:08:12.679688 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:12 crc kubenswrapper[4764]: I0202 09:08:12.679698 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:12Z","lastTransitionTime":"2026-02-02T09:08:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:12 crc kubenswrapper[4764]: I0202 09:08:12.782256 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:12 crc kubenswrapper[4764]: I0202 09:08:12.782348 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:12 crc kubenswrapper[4764]: I0202 09:08:12.782365 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:12 crc kubenswrapper[4764]: I0202 09:08:12.782385 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:12 crc kubenswrapper[4764]: I0202 09:08:12.782425 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:12Z","lastTransitionTime":"2026-02-02T09:08:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:12 crc kubenswrapper[4764]: I0202 09:08:12.804852 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-21 13:56:56.783639507 +0000 UTC Feb 02 09:08:12 crc kubenswrapper[4764]: I0202 09:08:12.824518 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 09:08:12 crc kubenswrapper[4764]: I0202 09:08:12.824524 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 09:08:12 crc kubenswrapper[4764]: E0202 09:08:12.824745 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 09:08:12 crc kubenswrapper[4764]: E0202 09:08:12.824649 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 09:08:12 crc kubenswrapper[4764]: I0202 09:08:12.824531 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 09:08:12 crc kubenswrapper[4764]: E0202 09:08:12.824823 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 09:08:12 crc kubenswrapper[4764]: I0202 09:08:12.885224 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:12 crc kubenswrapper[4764]: I0202 09:08:12.885351 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:12 crc kubenswrapper[4764]: I0202 09:08:12.885431 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:12 crc kubenswrapper[4764]: I0202 09:08:12.885461 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:12 crc kubenswrapper[4764]: I0202 09:08:12.885522 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:12Z","lastTransitionTime":"2026-02-02T09:08:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:12 crc kubenswrapper[4764]: I0202 09:08:12.989130 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:12 crc kubenswrapper[4764]: I0202 09:08:12.989226 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:12 crc kubenswrapper[4764]: I0202 09:08:12.989234 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:12 crc kubenswrapper[4764]: I0202 09:08:12.989260 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:12 crc kubenswrapper[4764]: I0202 09:08:12.989269 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:12Z","lastTransitionTime":"2026-02-02T09:08:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:13 crc kubenswrapper[4764]: I0202 09:08:13.091636 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:13 crc kubenswrapper[4764]: I0202 09:08:13.091682 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:13 crc kubenswrapper[4764]: I0202 09:08:13.091697 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:13 crc kubenswrapper[4764]: I0202 09:08:13.091719 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:13 crc kubenswrapper[4764]: I0202 09:08:13.091735 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:13Z","lastTransitionTime":"2026-02-02T09:08:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:13 crc kubenswrapper[4764]: I0202 09:08:13.194356 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:13 crc kubenswrapper[4764]: I0202 09:08:13.194421 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:13 crc kubenswrapper[4764]: I0202 09:08:13.194440 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:13 crc kubenswrapper[4764]: I0202 09:08:13.194468 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:13 crc kubenswrapper[4764]: I0202 09:08:13.194485 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:13Z","lastTransitionTime":"2026-02-02T09:08:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:13 crc kubenswrapper[4764]: I0202 09:08:13.296959 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:13 crc kubenswrapper[4764]: I0202 09:08:13.297039 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:13 crc kubenswrapper[4764]: I0202 09:08:13.297052 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:13 crc kubenswrapper[4764]: I0202 09:08:13.297068 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:13 crc kubenswrapper[4764]: I0202 09:08:13.297080 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:13Z","lastTransitionTime":"2026-02-02T09:08:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:13 crc kubenswrapper[4764]: I0202 09:08:13.400212 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:13 crc kubenswrapper[4764]: I0202 09:08:13.400284 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:13 crc kubenswrapper[4764]: I0202 09:08:13.400302 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:13 crc kubenswrapper[4764]: I0202 09:08:13.400326 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:13 crc kubenswrapper[4764]: I0202 09:08:13.400342 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:13Z","lastTransitionTime":"2026-02-02T09:08:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:13 crc kubenswrapper[4764]: I0202 09:08:13.502858 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:13 crc kubenswrapper[4764]: I0202 09:08:13.502921 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:13 crc kubenswrapper[4764]: I0202 09:08:13.502987 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:13 crc kubenswrapper[4764]: I0202 09:08:13.503012 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:13 crc kubenswrapper[4764]: I0202 09:08:13.503029 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:13Z","lastTransitionTime":"2026-02-02T09:08:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:13 crc kubenswrapper[4764]: I0202 09:08:13.606415 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:13 crc kubenswrapper[4764]: I0202 09:08:13.606492 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:13 crc kubenswrapper[4764]: I0202 09:08:13.606516 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:13 crc kubenswrapper[4764]: I0202 09:08:13.606549 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:13 crc kubenswrapper[4764]: I0202 09:08:13.606572 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:13Z","lastTransitionTime":"2026-02-02T09:08:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:13 crc kubenswrapper[4764]: I0202 09:08:13.708199 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:13 crc kubenswrapper[4764]: I0202 09:08:13.708236 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:13 crc kubenswrapper[4764]: I0202 09:08:13.708244 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:13 crc kubenswrapper[4764]: I0202 09:08:13.708257 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:13 crc kubenswrapper[4764]: I0202 09:08:13.708267 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:13Z","lastTransitionTime":"2026-02-02T09:08:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:13 crc kubenswrapper[4764]: I0202 09:08:13.805206 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-03 09:07:58.658460866 +0000 UTC Feb 02 09:08:13 crc kubenswrapper[4764]: I0202 09:08:13.811031 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:13 crc kubenswrapper[4764]: I0202 09:08:13.811071 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:13 crc kubenswrapper[4764]: I0202 09:08:13.811083 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:13 crc kubenswrapper[4764]: I0202 09:08:13.811107 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:13 crc kubenswrapper[4764]: I0202 09:08:13.811120 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:13Z","lastTransitionTime":"2026-02-02T09:08:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:13 crc kubenswrapper[4764]: I0202 09:08:13.825074 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kwtmr" Feb 02 09:08:13 crc kubenswrapper[4764]: E0202 09:08:13.825238 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kwtmr" podUID="991faa9a-dd25-4f49-82bd-ce60cefd4af2" Feb 02 09:08:13 crc kubenswrapper[4764]: I0202 09:08:13.913531 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:13 crc kubenswrapper[4764]: I0202 09:08:13.913585 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:13 crc kubenswrapper[4764]: I0202 09:08:13.913598 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:13 crc kubenswrapper[4764]: I0202 09:08:13.913619 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:13 crc kubenswrapper[4764]: I0202 09:08:13.913631 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:13Z","lastTransitionTime":"2026-02-02T09:08:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:14 crc kubenswrapper[4764]: I0202 09:08:14.016604 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:14 crc kubenswrapper[4764]: I0202 09:08:14.016660 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:14 crc kubenswrapper[4764]: I0202 09:08:14.016678 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:14 crc kubenswrapper[4764]: I0202 09:08:14.016701 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:14 crc kubenswrapper[4764]: I0202 09:08:14.016726 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:14Z","lastTransitionTime":"2026-02-02T09:08:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:14 crc kubenswrapper[4764]: I0202 09:08:14.118719 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:14 crc kubenswrapper[4764]: I0202 09:08:14.118759 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:14 crc kubenswrapper[4764]: I0202 09:08:14.118771 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:14 crc kubenswrapper[4764]: I0202 09:08:14.118787 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:14 crc kubenswrapper[4764]: I0202 09:08:14.118799 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:14Z","lastTransitionTime":"2026-02-02T09:08:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:14 crc kubenswrapper[4764]: I0202 09:08:14.221335 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:14 crc kubenswrapper[4764]: I0202 09:08:14.221632 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:14 crc kubenswrapper[4764]: I0202 09:08:14.221652 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:14 crc kubenswrapper[4764]: I0202 09:08:14.221676 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:14 crc kubenswrapper[4764]: I0202 09:08:14.221693 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:14Z","lastTransitionTime":"2026-02-02T09:08:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:14 crc kubenswrapper[4764]: I0202 09:08:14.325093 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:14 crc kubenswrapper[4764]: I0202 09:08:14.325154 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:14 crc kubenswrapper[4764]: I0202 09:08:14.325166 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:14 crc kubenswrapper[4764]: I0202 09:08:14.325180 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:14 crc kubenswrapper[4764]: I0202 09:08:14.325191 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:14Z","lastTransitionTime":"2026-02-02T09:08:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:14 crc kubenswrapper[4764]: I0202 09:08:14.427695 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:14 crc kubenswrapper[4764]: I0202 09:08:14.427762 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:14 crc kubenswrapper[4764]: I0202 09:08:14.427779 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:14 crc kubenswrapper[4764]: I0202 09:08:14.427800 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:14 crc kubenswrapper[4764]: I0202 09:08:14.427817 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:14Z","lastTransitionTime":"2026-02-02T09:08:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:14 crc kubenswrapper[4764]: I0202 09:08:14.530019 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:14 crc kubenswrapper[4764]: I0202 09:08:14.530058 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:14 crc kubenswrapper[4764]: I0202 09:08:14.530068 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:14 crc kubenswrapper[4764]: I0202 09:08:14.530084 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:14 crc kubenswrapper[4764]: I0202 09:08:14.530095 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:14Z","lastTransitionTime":"2026-02-02T09:08:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:14 crc kubenswrapper[4764]: I0202 09:08:14.632840 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:14 crc kubenswrapper[4764]: I0202 09:08:14.632997 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:14 crc kubenswrapper[4764]: I0202 09:08:14.633046 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:14 crc kubenswrapper[4764]: I0202 09:08:14.633095 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:14 crc kubenswrapper[4764]: I0202 09:08:14.633121 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:14Z","lastTransitionTime":"2026-02-02T09:08:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:14 crc kubenswrapper[4764]: I0202 09:08:14.735253 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:14 crc kubenswrapper[4764]: I0202 09:08:14.735285 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:14 crc kubenswrapper[4764]: I0202 09:08:14.735294 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:14 crc kubenswrapper[4764]: I0202 09:08:14.735308 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:14 crc kubenswrapper[4764]: I0202 09:08:14.735319 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:14Z","lastTransitionTime":"2026-02-02T09:08:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:14 crc kubenswrapper[4764]: I0202 09:08:14.806381 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-24 21:16:28.063556044 +0000 UTC Feb 02 09:08:14 crc kubenswrapper[4764]: I0202 09:08:14.824551 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 09:08:14 crc kubenswrapper[4764]: I0202 09:08:14.824652 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 09:08:14 crc kubenswrapper[4764]: E0202 09:08:14.824750 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 09:08:14 crc kubenswrapper[4764]: I0202 09:08:14.824793 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 09:08:14 crc kubenswrapper[4764]: E0202 09:08:14.824921 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 09:08:14 crc kubenswrapper[4764]: E0202 09:08:14.825027 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 09:08:14 crc kubenswrapper[4764]: I0202 09:08:14.837462 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:14 crc kubenswrapper[4764]: I0202 09:08:14.837512 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:14 crc kubenswrapper[4764]: I0202 09:08:14.837523 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:14 crc kubenswrapper[4764]: I0202 09:08:14.837538 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:14 crc kubenswrapper[4764]: I0202 09:08:14.837550 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:14Z","lastTransitionTime":"2026-02-02T09:08:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:14 crc kubenswrapper[4764]: I0202 09:08:14.940064 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:14 crc kubenswrapper[4764]: I0202 09:08:14.940109 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:14 crc kubenswrapper[4764]: I0202 09:08:14.940125 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:14 crc kubenswrapper[4764]: I0202 09:08:14.940144 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:14 crc kubenswrapper[4764]: I0202 09:08:14.940160 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:14Z","lastTransitionTime":"2026-02-02T09:08:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:15 crc kubenswrapper[4764]: I0202 09:08:15.042086 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:15 crc kubenswrapper[4764]: I0202 09:08:15.042134 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:15 crc kubenswrapper[4764]: I0202 09:08:15.042148 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:15 crc kubenswrapper[4764]: I0202 09:08:15.042167 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:15 crc kubenswrapper[4764]: I0202 09:08:15.042178 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:15Z","lastTransitionTime":"2026-02-02T09:08:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:15 crc kubenswrapper[4764]: I0202 09:08:15.145843 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:15 crc kubenswrapper[4764]: I0202 09:08:15.145906 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:15 crc kubenswrapper[4764]: I0202 09:08:15.145924 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:15 crc kubenswrapper[4764]: I0202 09:08:15.145983 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:15 crc kubenswrapper[4764]: I0202 09:08:15.146002 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:15Z","lastTransitionTime":"2026-02-02T09:08:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:15 crc kubenswrapper[4764]: I0202 09:08:15.248632 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:15 crc kubenswrapper[4764]: I0202 09:08:15.248697 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:15 crc kubenswrapper[4764]: I0202 09:08:15.248715 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:15 crc kubenswrapper[4764]: I0202 09:08:15.248739 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:15 crc kubenswrapper[4764]: I0202 09:08:15.248756 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:15Z","lastTransitionTime":"2026-02-02T09:08:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:15 crc kubenswrapper[4764]: I0202 09:08:15.351834 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:15 crc kubenswrapper[4764]: I0202 09:08:15.351894 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:15 crc kubenswrapper[4764]: I0202 09:08:15.351911 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:15 crc kubenswrapper[4764]: I0202 09:08:15.351966 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:15 crc kubenswrapper[4764]: I0202 09:08:15.351984 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:15Z","lastTransitionTime":"2026-02-02T09:08:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:15 crc kubenswrapper[4764]: I0202 09:08:15.455051 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:15 crc kubenswrapper[4764]: I0202 09:08:15.455131 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:15 crc kubenswrapper[4764]: I0202 09:08:15.455154 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:15 crc kubenswrapper[4764]: I0202 09:08:15.455183 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:15 crc kubenswrapper[4764]: I0202 09:08:15.455208 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:15Z","lastTransitionTime":"2026-02-02T09:08:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:15 crc kubenswrapper[4764]: I0202 09:08:15.558257 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:15 crc kubenswrapper[4764]: I0202 09:08:15.558366 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:15 crc kubenswrapper[4764]: I0202 09:08:15.558391 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:15 crc kubenswrapper[4764]: I0202 09:08:15.558422 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:15 crc kubenswrapper[4764]: I0202 09:08:15.558444 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:15Z","lastTransitionTime":"2026-02-02T09:08:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:15 crc kubenswrapper[4764]: I0202 09:08:15.661573 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:15 crc kubenswrapper[4764]: I0202 09:08:15.661618 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:15 crc kubenswrapper[4764]: I0202 09:08:15.661627 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:15 crc kubenswrapper[4764]: I0202 09:08:15.661640 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:15 crc kubenswrapper[4764]: I0202 09:08:15.661650 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:15Z","lastTransitionTime":"2026-02-02T09:08:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:15 crc kubenswrapper[4764]: I0202 09:08:15.765110 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:15 crc kubenswrapper[4764]: I0202 09:08:15.765176 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:15 crc kubenswrapper[4764]: I0202 09:08:15.765241 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:15 crc kubenswrapper[4764]: I0202 09:08:15.765280 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:15 crc kubenswrapper[4764]: I0202 09:08:15.765305 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:15Z","lastTransitionTime":"2026-02-02T09:08:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:15 crc kubenswrapper[4764]: I0202 09:08:15.806604 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-15 13:37:26.706035525 +0000 UTC Feb 02 09:08:15 crc kubenswrapper[4764]: I0202 09:08:15.825093 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kwtmr" Feb 02 09:08:15 crc kubenswrapper[4764]: E0202 09:08:15.825305 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kwtmr" podUID="991faa9a-dd25-4f49-82bd-ce60cefd4af2" Feb 02 09:08:15 crc kubenswrapper[4764]: I0202 09:08:15.861229 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:15 crc kubenswrapper[4764]: I0202 09:08:15.861271 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:15 crc kubenswrapper[4764]: I0202 09:08:15.861287 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:15 crc kubenswrapper[4764]: I0202 09:08:15.861309 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:15 crc kubenswrapper[4764]: I0202 09:08:15.861336 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:15Z","lastTransitionTime":"2026-02-02T09:08:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:15 crc kubenswrapper[4764]: E0202 09:08:15.882173 4764 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:08:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:08:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:08:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:08:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"31973a70-ce17-4811-9a14-f5476c6979db\\\",\\\"systemUUID\\\":\\\"ca92679c-8580-43c3-be51-62441e93d22b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:15Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:15 crc kubenswrapper[4764]: I0202 09:08:15.887401 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:15 crc kubenswrapper[4764]: I0202 09:08:15.887448 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:15 crc kubenswrapper[4764]: I0202 09:08:15.887465 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:15 crc kubenswrapper[4764]: I0202 09:08:15.887488 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:15 crc kubenswrapper[4764]: I0202 09:08:15.887506 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:15Z","lastTransitionTime":"2026-02-02T09:08:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:15 crc kubenswrapper[4764]: E0202 09:08:15.908238 4764 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:08:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:08:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:08:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:08:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"31973a70-ce17-4811-9a14-f5476c6979db\\\",\\\"systemUUID\\\":\\\"ca92679c-8580-43c3-be51-62441e93d22b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:15Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:15 crc kubenswrapper[4764]: I0202 09:08:15.918884 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:15 crc kubenswrapper[4764]: I0202 09:08:15.918955 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:15 crc kubenswrapper[4764]: I0202 09:08:15.918966 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:15 crc kubenswrapper[4764]: I0202 09:08:15.918983 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:15 crc kubenswrapper[4764]: I0202 09:08:15.919011 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:15Z","lastTransitionTime":"2026-02-02T09:08:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:15 crc kubenswrapper[4764]: E0202 09:08:15.939391 4764 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:08:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:08:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:08:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:08:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"31973a70-ce17-4811-9a14-f5476c6979db\\\",\\\"systemUUID\\\":\\\"ca92679c-8580-43c3-be51-62441e93d22b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:15Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:15 crc kubenswrapper[4764]: I0202 09:08:15.943295 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:15 crc kubenswrapper[4764]: I0202 09:08:15.943345 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:15 crc kubenswrapper[4764]: I0202 09:08:15.943358 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:15 crc kubenswrapper[4764]: I0202 09:08:15.943376 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:15 crc kubenswrapper[4764]: I0202 09:08:15.943389 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:15Z","lastTransitionTime":"2026-02-02T09:08:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:15 crc kubenswrapper[4764]: E0202 09:08:15.957537 4764 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:08:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:08:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:08:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:08:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"31973a70-ce17-4811-9a14-f5476c6979db\\\",\\\"systemUUID\\\":\\\"ca92679c-8580-43c3-be51-62441e93d22b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:15Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:15 crc kubenswrapper[4764]: I0202 09:08:15.962199 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:15 crc kubenswrapper[4764]: I0202 09:08:15.962300 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:15 crc kubenswrapper[4764]: I0202 09:08:15.962402 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:15 crc kubenswrapper[4764]: I0202 09:08:15.962481 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:15 crc kubenswrapper[4764]: I0202 09:08:15.962522 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:15Z","lastTransitionTime":"2026-02-02T09:08:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:15 crc kubenswrapper[4764]: E0202 09:08:15.982772 4764 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:08:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:08:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:08:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:08:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"31973a70-ce17-4811-9a14-f5476c6979db\\\",\\\"systemUUID\\\":\\\"ca92679c-8580-43c3-be51-62441e93d22b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:15Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:15 crc kubenswrapper[4764]: E0202 09:08:15.983046 4764 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 02 09:08:15 crc kubenswrapper[4764]: I0202 09:08:15.984612 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:15 crc kubenswrapper[4764]: I0202 09:08:15.984685 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:15 crc kubenswrapper[4764]: I0202 09:08:15.984704 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:15 crc kubenswrapper[4764]: I0202 09:08:15.984723 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:15 crc kubenswrapper[4764]: I0202 09:08:15.984738 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:15Z","lastTransitionTime":"2026-02-02T09:08:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:16 crc kubenswrapper[4764]: I0202 09:08:16.087205 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:16 crc kubenswrapper[4764]: I0202 09:08:16.087269 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:16 crc kubenswrapper[4764]: I0202 09:08:16.087288 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:16 crc kubenswrapper[4764]: I0202 09:08:16.087314 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:16 crc kubenswrapper[4764]: I0202 09:08:16.087333 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:16Z","lastTransitionTime":"2026-02-02T09:08:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:16 crc kubenswrapper[4764]: I0202 09:08:16.190128 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:16 crc kubenswrapper[4764]: I0202 09:08:16.190173 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:16 crc kubenswrapper[4764]: I0202 09:08:16.190189 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:16 crc kubenswrapper[4764]: I0202 09:08:16.190208 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:16 crc kubenswrapper[4764]: I0202 09:08:16.190224 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:16Z","lastTransitionTime":"2026-02-02T09:08:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:16 crc kubenswrapper[4764]: I0202 09:08:16.292667 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:16 crc kubenswrapper[4764]: I0202 09:08:16.292710 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:16 crc kubenswrapper[4764]: I0202 09:08:16.292725 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:16 crc kubenswrapper[4764]: I0202 09:08:16.292745 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:16 crc kubenswrapper[4764]: I0202 09:08:16.292758 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:16Z","lastTransitionTime":"2026-02-02T09:08:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:16 crc kubenswrapper[4764]: I0202 09:08:16.395051 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:16 crc kubenswrapper[4764]: I0202 09:08:16.395096 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:16 crc kubenswrapper[4764]: I0202 09:08:16.395105 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:16 crc kubenswrapper[4764]: I0202 09:08:16.395120 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:16 crc kubenswrapper[4764]: I0202 09:08:16.395130 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:16Z","lastTransitionTime":"2026-02-02T09:08:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:16 crc kubenswrapper[4764]: I0202 09:08:16.497110 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:16 crc kubenswrapper[4764]: I0202 09:08:16.497154 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:16 crc kubenswrapper[4764]: I0202 09:08:16.497165 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:16 crc kubenswrapper[4764]: I0202 09:08:16.497178 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:16 crc kubenswrapper[4764]: I0202 09:08:16.497189 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:16Z","lastTransitionTime":"2026-02-02T09:08:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:16 crc kubenswrapper[4764]: I0202 09:08:16.599267 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:16 crc kubenswrapper[4764]: I0202 09:08:16.599305 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:16 crc kubenswrapper[4764]: I0202 09:08:16.599314 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:16 crc kubenswrapper[4764]: I0202 09:08:16.599330 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:16 crc kubenswrapper[4764]: I0202 09:08:16.599339 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:16Z","lastTransitionTime":"2026-02-02T09:08:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:16 crc kubenswrapper[4764]: I0202 09:08:16.701788 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:16 crc kubenswrapper[4764]: I0202 09:08:16.701871 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:16 crc kubenswrapper[4764]: I0202 09:08:16.701896 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:16 crc kubenswrapper[4764]: I0202 09:08:16.701927 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:16 crc kubenswrapper[4764]: I0202 09:08:16.702002 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:16Z","lastTransitionTime":"2026-02-02T09:08:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:16 crc kubenswrapper[4764]: I0202 09:08:16.804152 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:16 crc kubenswrapper[4764]: I0202 09:08:16.804208 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:16 crc kubenswrapper[4764]: I0202 09:08:16.804217 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:16 crc kubenswrapper[4764]: I0202 09:08:16.804232 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:16 crc kubenswrapper[4764]: I0202 09:08:16.804241 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:16Z","lastTransitionTime":"2026-02-02T09:08:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:16 crc kubenswrapper[4764]: I0202 09:08:16.807435 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-19 20:26:11.345369943 +0000 UTC Feb 02 09:08:16 crc kubenswrapper[4764]: I0202 09:08:16.824734 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 09:08:16 crc kubenswrapper[4764]: I0202 09:08:16.824743 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 09:08:16 crc kubenswrapper[4764]: I0202 09:08:16.824794 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 09:08:16 crc kubenswrapper[4764]: E0202 09:08:16.824875 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 09:08:16 crc kubenswrapper[4764]: E0202 09:08:16.825042 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 09:08:16 crc kubenswrapper[4764]: E0202 09:08:16.825140 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 09:08:16 crc kubenswrapper[4764]: I0202 09:08:16.907540 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:16 crc kubenswrapper[4764]: I0202 09:08:16.907593 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:16 crc kubenswrapper[4764]: I0202 09:08:16.907614 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:16 crc kubenswrapper[4764]: I0202 09:08:16.907638 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:16 crc kubenswrapper[4764]: I0202 09:08:16.907655 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:16Z","lastTransitionTime":"2026-02-02T09:08:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:17 crc kubenswrapper[4764]: I0202 09:08:17.010264 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:17 crc kubenswrapper[4764]: I0202 09:08:17.010327 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:17 crc kubenswrapper[4764]: I0202 09:08:17.010337 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:17 crc kubenswrapper[4764]: I0202 09:08:17.010349 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:17 crc kubenswrapper[4764]: I0202 09:08:17.010359 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:17Z","lastTransitionTime":"2026-02-02T09:08:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:17 crc kubenswrapper[4764]: I0202 09:08:17.112609 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:17 crc kubenswrapper[4764]: I0202 09:08:17.112647 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:17 crc kubenswrapper[4764]: I0202 09:08:17.112656 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:17 crc kubenswrapper[4764]: I0202 09:08:17.112669 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:17 crc kubenswrapper[4764]: I0202 09:08:17.112695 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:17Z","lastTransitionTime":"2026-02-02T09:08:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:17 crc kubenswrapper[4764]: I0202 09:08:17.215188 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:17 crc kubenswrapper[4764]: I0202 09:08:17.215240 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:17 crc kubenswrapper[4764]: I0202 09:08:17.215257 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:17 crc kubenswrapper[4764]: I0202 09:08:17.215274 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:17 crc kubenswrapper[4764]: I0202 09:08:17.215287 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:17Z","lastTransitionTime":"2026-02-02T09:08:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:17 crc kubenswrapper[4764]: I0202 09:08:17.318274 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:17 crc kubenswrapper[4764]: I0202 09:08:17.318328 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:17 crc kubenswrapper[4764]: I0202 09:08:17.318340 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:17 crc kubenswrapper[4764]: I0202 09:08:17.318358 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:17 crc kubenswrapper[4764]: I0202 09:08:17.318369 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:17Z","lastTransitionTime":"2026-02-02T09:08:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:17 crc kubenswrapper[4764]: I0202 09:08:17.420670 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:17 crc kubenswrapper[4764]: I0202 09:08:17.420715 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:17 crc kubenswrapper[4764]: I0202 09:08:17.420733 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:17 crc kubenswrapper[4764]: I0202 09:08:17.420755 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:17 crc kubenswrapper[4764]: I0202 09:08:17.420773 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:17Z","lastTransitionTime":"2026-02-02T09:08:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:17 crc kubenswrapper[4764]: I0202 09:08:17.523676 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:17 crc kubenswrapper[4764]: I0202 09:08:17.523728 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:17 crc kubenswrapper[4764]: I0202 09:08:17.523742 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:17 crc kubenswrapper[4764]: I0202 09:08:17.523760 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:17 crc kubenswrapper[4764]: I0202 09:08:17.523771 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:17Z","lastTransitionTime":"2026-02-02T09:08:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:17 crc kubenswrapper[4764]: I0202 09:08:17.626653 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:17 crc kubenswrapper[4764]: I0202 09:08:17.626719 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:17 crc kubenswrapper[4764]: I0202 09:08:17.626735 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:17 crc kubenswrapper[4764]: I0202 09:08:17.626777 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:17 crc kubenswrapper[4764]: I0202 09:08:17.626792 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:17Z","lastTransitionTime":"2026-02-02T09:08:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:17 crc kubenswrapper[4764]: I0202 09:08:17.728807 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:17 crc kubenswrapper[4764]: I0202 09:08:17.728855 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:17 crc kubenswrapper[4764]: I0202 09:08:17.728868 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:17 crc kubenswrapper[4764]: I0202 09:08:17.728886 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:17 crc kubenswrapper[4764]: I0202 09:08:17.728900 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:17Z","lastTransitionTime":"2026-02-02T09:08:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:17 crc kubenswrapper[4764]: I0202 09:08:17.808057 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-07 21:19:53.32195012 +0000 UTC Feb 02 09:08:17 crc kubenswrapper[4764]: I0202 09:08:17.824769 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kwtmr" Feb 02 09:08:17 crc kubenswrapper[4764]: E0202 09:08:17.824879 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kwtmr" podUID="991faa9a-dd25-4f49-82bd-ce60cefd4af2" Feb 02 09:08:17 crc kubenswrapper[4764]: I0202 09:08:17.831142 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:17 crc kubenswrapper[4764]: I0202 09:08:17.831174 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:17 crc kubenswrapper[4764]: I0202 09:08:17.831182 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:17 crc kubenswrapper[4764]: I0202 09:08:17.831196 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:17 crc kubenswrapper[4764]: I0202 09:08:17.831204 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:17Z","lastTransitionTime":"2026-02-02T09:08:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:17 crc kubenswrapper[4764]: I0202 09:08:17.840888 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed747fd8-e3ce-4319-a6da-bda0fd51b85a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6f064bdc66470aace3180d4988244026b54009a63cff199e41c5d02a1d18479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://60258f84a1d8dcfe48b7e9688ed50f077e201971f20f3f6663fbf1ba58a4611a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f84be2e08217a2ac3edde858d78b4410376ed9bae3bcabe2a7f9b2fbe2f69c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a69d25d352d7ceac5c735bc1fcf446a1fea6dc22c65a5296cfab828cea17377\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:17Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:17 crc kubenswrapper[4764]: I0202 09:08:17.857524 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f6904228e052fff69ad1fd66df97758b6b60ba4351e97a7fbd03eee4df1b653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:17Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:17 crc kubenswrapper[4764]: I0202 09:08:17.871791 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6zxff" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aaba93da-12a5-43df-b7c7-4c8b988163c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3db293678a907cd55b31b1d0c0bd7c30e1ba75ecd5e13cbc68ab98fbf0fbd9d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqwb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6zxff\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:17Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:17 crc kubenswrapper[4764]: I0202 09:08:17.884643 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d192f670-9f9d-4539-9641-e4bed73acdd4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://040c5e8a5659f85f0d52d1a1e064154d998ae104f711101d2204c0b1caebbb30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbksf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fdf06410349b7c99aa166e87fbe80bf68e2ae923635a6a978b7e2cd50383f7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbksf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4ndm4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:17Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:17 crc kubenswrapper[4764]: I0202 09:08:17.899354 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pcjmn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6e5ccd3b-b657-47a2-9c4d-bc9adb126533\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72863eef568d92a1abeb3cde2e89630a4f5fb7626c6949448ff5195774c98eab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2f7zl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pcjmn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:17Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:17 crc kubenswrapper[4764]: I0202 09:08:17.917780 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fae745b-ea49-42f3-a486-dbc3dc2fb9ad\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://081d7f76d937ff71e35945f9b1d893888bb649d15bd84312fd4b61c9520ea71f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://429ead0993f5e648f6256d249ee12aea03514d3108618840552b672fb419cd74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fad17b01545eb20785f03e3c61bd5c982f6217e38176d7a3b23ac194298644a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85e2f75c56d63cacc93b9195c1e8bf46eb7a7234c45a024bd95a460db90edb6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85e2f75c56d63cacc93b9195c1e8bf46eb7a7234c45a024bd95a460db90edb6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:17Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:17 crc kubenswrapper[4764]: I0202 09:08:17.930180 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:17Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:17 crc kubenswrapper[4764]: I0202 09:08:17.933571 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:17 crc kubenswrapper[4764]: I0202 09:08:17.933598 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:17 crc kubenswrapper[4764]: I0202 09:08:17.933609 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:17 crc kubenswrapper[4764]: I0202 09:08:17.933626 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:17 crc kubenswrapper[4764]: I0202 09:08:17.933644 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:17Z","lastTransitionTime":"2026-02-02T09:08:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:17 crc kubenswrapper[4764]: I0202 09:08:17.945246 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lh6tm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2152e8a-f456-47b5-a547-b65ec837ecbc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38e176a00b07f2960fdd7ba2aadd1468a5e083b8d154e9d118d13f2585da81e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bb2sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://996e649011d5255783284421ddea8e0a5c383ef675b0920c3fdc0af6503d80e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bb2sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lh6tm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:17Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:17 crc kubenswrapper[4764]: I0202 09:08:17.956890 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kwtmr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"991faa9a-dd25-4f49-82bd-ce60cefd4af2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h865b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h865b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:53Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kwtmr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:17Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:17 crc kubenswrapper[4764]: I0202 09:08:17.976654 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87d61aea-46f6-4ee6-a7a4-767a6a67fea4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e78b6d08e4d616cefbb73e3fd43b3c6886dcd8aa997c7b4948bf1ff63401293\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d468715c2da33996561c40043216233fd12e21539d0c58dd1f3b892fd9f91a9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b184f7c936dc4046f74c0018269cbc2c4b7ceddb650b4e871d3f8a707e0df88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://269c5074371bd00f8363fb3888c11f1df383081a61497aea6578f08700881d50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b25bc9395831524c53f9e565b4efbe74f21792efafeb1dd507161bc91aa0ecdf\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T09:07:22Z\\\",\\\"message\\\":\\\"W0202 09:07:21.597488 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0202 09:07:21.598117 1 crypto.go:601] Generating new CA for check-endpoints-signer@1770023241 cert, and key in /tmp/serving-cert-3856183724/serving-signer.crt, /tmp/serving-cert-3856183724/serving-signer.key\\\\nI0202 09:07:22.012184 1 observer_polling.go:159] Starting file observer\\\\nW0202 09:07:22.015493 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0202 09:07:22.015797 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 09:07:22.016596 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3856183724/tls.crt::/tmp/serving-cert-3856183724/tls.key\\\\\\\"\\\\nF0202 09:07:22.270605 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b44e17b8ca01f549bc10081dd5f109cb1b9fff364bf5bd4fd90d9f0e3eeb04db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfb06a5ec9262cf39b108a66014d15e325cdbd5441a2d9e7a811dd57265c99a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfb06a5ec9262cf39b108a66014d15e325cdbd5441a2d9e7a811dd57265c99a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:17Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:17 crc kubenswrapper[4764]: I0202 09:08:17.989588 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5528225d91ac2778bef4554abb0bd036ac5d3b6cdac7357bab69e70c3894b6f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://071bbb80d0747cf3e109ba626d47f7d05f85f4cc0b3aba77276c795741f47ac6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:17Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:18 crc kubenswrapper[4764]: I0202 09:08:18.003417 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:18Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:18 crc kubenswrapper[4764]: I0202 09:08:18.016628 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nc7qt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1a2c7b22068b7165927fc158885d48bf9952b9eba521d8ae9a599eee53dc305\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2hz4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nc7qt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:18Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:18 crc kubenswrapper[4764]: I0202 09:08:18.031550 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9p5dc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"660ac006-2495-45c4-9fb3-e9c1dddcf7a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da812b124c97aa8ad1563f4bcfac8f3bfa025096fa4a2561e5847b1bf9cb30b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23e57e65a6f5c3059506b9db21e9ac2492c30bd9191a98db0281915e9ce39797\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23e57e65a6f5c3059506b9db21e9ac2492c30bd9191a98db0281915e9ce39797\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2417b66816da1a6cd1791a4b91eca4a3c4f3694462e066e7f0722f1b9c2105c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2417b66816da1a6cd1791a4b91eca4a3c4f3694462e066e7f0722f1b9c2105c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://883641f29e9120c2cea6b272bacf1dc63c752f55ab0408cdaf73db85c85ac369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://883641f29e9120c2cea6b272bacf1dc63c752f55ab0408cdaf73db85c85ac369\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2745baba1188ba5e8f820616e4117a12c50ecee5b68e71ad9cabbb621b6e131f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2745baba1188ba5e8f820616e4117a12c50ecee5b68e71ad9cabbb621b6e131f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8925e435be4d48d86824fc35f8b8819fb35d63e385164c4a20edb3bf3035c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a8925e435be4d48d86824fc35f8b8819fb35d63e385164c4a20edb3bf3035c3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d3fbf668ae7d04c8d9e574f43c2fe90c5604204ba38a3c60847afdfbceb7a57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d3fbf668ae7d04c8d9e574f43c2fe90c5604204ba38a3c60847afdfbceb7a57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9p5dc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:18Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:18 crc kubenswrapper[4764]: I0202 09:08:18.036101 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:18 crc kubenswrapper[4764]: I0202 09:08:18.036128 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:18 crc kubenswrapper[4764]: I0202 09:08:18.036137 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:18 crc kubenswrapper[4764]: I0202 09:08:18.036155 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:18 crc kubenswrapper[4764]: I0202 09:08:18.036167 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:18Z","lastTransitionTime":"2026-02-02T09:08:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:18 crc kubenswrapper[4764]: I0202 09:08:18.054449 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d66ab3c0-9b5b-41d5-b6c0-91ccddccd7e9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d92ac486f5dade46bc75de32099859a71fc2ed25a7abbe6ae8937df128e7df30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a4ac273a4f127d79e5fad7d4c30cd2044a346f5734dba5d33d68cd092c0c2d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d38c7cb5dcd22fb07524494ebe77aa1a7f4f808f566043563817e47e7ec12a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://408722c7f83d28919431532f8c797a1c795fd5d96736d14e7f06a05744899702\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f473b18cd8003f75bb1dafb82bcec45745a73ccd10434297356dac185a3d15c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e229f484d19cbdabedf53d90a822a4434db9e00b1301538e885e26acc19d25f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e229f484d19cbdabedf53d90a822a4434db9e00b1301538e885e26acc19d25f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a9cc673be8ff245dba81278e8cd34cfe45ed36ff0b8bd5b30a0ddf333ec6371\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a9cc673be8ff245dba81278e8cd34cfe45ed36ff0b8bd5b30a0ddf333ec6371\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bc7e49c7a5ae66ea939060192fa6a346dc0b0b9aa9750b29879d3c426cc0d276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc7e49c7a5ae66ea939060192fa6a346dc0b0b9aa9750b29879d3c426cc0d276\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:18Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:18 crc kubenswrapper[4764]: I0202 09:08:18.072825 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa9d037e66cb53b9dd698eff46284c9e718fd842cda6f7b525f0c79062d70bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:18Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:18 crc kubenswrapper[4764]: I0202 09:08:18.088254 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:18Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:18 crc kubenswrapper[4764]: I0202 09:08:18.106956 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24632dda-6100-4ab6-a28e-214ddae4360c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a6c67e9b7708cc0d85b0b6ffbbade0d34867e159a9c5487ca2606475c667a50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bffda48d7c881c1470251b501661324859373246cc1a21811bd36ca24887c7aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a989645cf485cd8ff80ffff1a86cd7808e1564eba5667535e161cc68db13b26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20b71d3ee7a381baa51b0b0d0470283022cfede30f491135737b5e82d631d763\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c380383b6a014d1b694a13ddc22f54b20b9e4f9d5c9b9000d76187621dce0a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5529f2bff87753428e9173826fc1880acf96a4c134fa9ecad195b15697cad73a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://334e828f8ef72570f2511ee38659941202530471f5943aa07b8b21483bd70550\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://12244061eabaae7fa9057ac5cee894c3b308b466411bcb910682e7f65d240dd7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-02T09:07:52Z\\\",\\\"message\\\":\\\"l_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-kube-apiserver/apiserver]} name:Service_openshift-kube-apiserver/apiserver_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.93:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {d71b38eb-32af-4c0f-9490-7c317c111e3a}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0202 09:07:51.914338 6129 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-kube-apiserver/apiserver]} name:Service_openshift-kube-apiserver/apiserver_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.93:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {d71b38eb-32af-4c0f-9490-7c317c111e3a}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF0202 09:07:51.914404 6129 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:51Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://334e828f8ef72570f2511ee38659941202530471f5943aa07b8b21483bd70550\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-02T09:08:10Z\\\",\\\"message\\\":\\\"ing OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-dns/dns-default]} name:Service_openshift-dns/dns-default_TCP_node_router+switch_crc options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.10:53: 10.217.4.10:9154:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {be9dcc9e-c16a-4962-a6d2-4adeb0b929c4}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-dns/dns-default]} name:Service_openshift-dns/dns-default_UDP_node_router+switch_crc options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[udp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.10:53:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {4c1be812-05d3-4f45-91b5-a853a5c8de71}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF0202 09:08:09.911531 6327 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T09:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b4cfbd4e682bcdcaff785013bdb72580ce36fced733e235538536827036b93d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://579903eadb183466a2564737e6b678e04308239c0cd75b6808892053b97afbc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://579903eadb183466a2564737e6b678e04308239c0cd75b6808892053b97afbc0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zhn7j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:18Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:18 crc kubenswrapper[4764]: I0202 09:08:18.139331 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:18 crc kubenswrapper[4764]: I0202 09:08:18.139367 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:18 crc kubenswrapper[4764]: I0202 09:08:18.139376 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:18 crc kubenswrapper[4764]: I0202 09:08:18.139391 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:18 crc kubenswrapper[4764]: I0202 09:08:18.139404 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:18Z","lastTransitionTime":"2026-02-02T09:08:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:18 crc kubenswrapper[4764]: I0202 09:08:18.242528 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:18 crc kubenswrapper[4764]: I0202 09:08:18.242595 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:18 crc kubenswrapper[4764]: I0202 09:08:18.242611 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:18 crc kubenswrapper[4764]: I0202 09:08:18.242637 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:18 crc kubenswrapper[4764]: I0202 09:08:18.242654 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:18Z","lastTransitionTime":"2026-02-02T09:08:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:18 crc kubenswrapper[4764]: I0202 09:08:18.345275 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:18 crc kubenswrapper[4764]: I0202 09:08:18.346019 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:18 crc kubenswrapper[4764]: I0202 09:08:18.346074 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:18 crc kubenswrapper[4764]: I0202 09:08:18.346103 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:18 crc kubenswrapper[4764]: I0202 09:08:18.346121 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:18Z","lastTransitionTime":"2026-02-02T09:08:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:18 crc kubenswrapper[4764]: I0202 09:08:18.449726 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:18 crc kubenswrapper[4764]: I0202 09:08:18.449800 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:18 crc kubenswrapper[4764]: I0202 09:08:18.449814 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:18 crc kubenswrapper[4764]: I0202 09:08:18.449839 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:18 crc kubenswrapper[4764]: I0202 09:08:18.449855 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:18Z","lastTransitionTime":"2026-02-02T09:08:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:18 crc kubenswrapper[4764]: I0202 09:08:18.553095 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:18 crc kubenswrapper[4764]: I0202 09:08:18.553141 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:18 crc kubenswrapper[4764]: I0202 09:08:18.553150 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:18 crc kubenswrapper[4764]: I0202 09:08:18.553165 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:18 crc kubenswrapper[4764]: I0202 09:08:18.553175 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:18Z","lastTransitionTime":"2026-02-02T09:08:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:18 crc kubenswrapper[4764]: I0202 09:08:18.656874 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:18 crc kubenswrapper[4764]: I0202 09:08:18.656971 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:18 crc kubenswrapper[4764]: I0202 09:08:18.656993 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:18 crc kubenswrapper[4764]: I0202 09:08:18.657020 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:18 crc kubenswrapper[4764]: I0202 09:08:18.657042 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:18Z","lastTransitionTime":"2026-02-02T09:08:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:18 crc kubenswrapper[4764]: I0202 09:08:18.760698 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:18 crc kubenswrapper[4764]: I0202 09:08:18.761052 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:18 crc kubenswrapper[4764]: I0202 09:08:18.761173 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:18 crc kubenswrapper[4764]: I0202 09:08:18.761274 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:18 crc kubenswrapper[4764]: I0202 09:08:18.761363 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:18Z","lastTransitionTime":"2026-02-02T09:08:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:18 crc kubenswrapper[4764]: I0202 09:08:18.809021 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-19 21:15:40.583400847 +0000 UTC Feb 02 09:08:18 crc kubenswrapper[4764]: I0202 09:08:18.824864 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 09:08:18 crc kubenswrapper[4764]: I0202 09:08:18.825007 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 09:08:18 crc kubenswrapper[4764]: E0202 09:08:18.825359 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 09:08:18 crc kubenswrapper[4764]: E0202 09:08:18.825538 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 09:08:18 crc kubenswrapper[4764]: I0202 09:08:18.825103 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 09:08:18 crc kubenswrapper[4764]: E0202 09:08:18.825900 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 09:08:18 crc kubenswrapper[4764]: I0202 09:08:18.864865 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:18 crc kubenswrapper[4764]: I0202 09:08:18.864914 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:18 crc kubenswrapper[4764]: I0202 09:08:18.864971 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:18 crc kubenswrapper[4764]: I0202 09:08:18.865005 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:18 crc kubenswrapper[4764]: I0202 09:08:18.865022 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:18Z","lastTransitionTime":"2026-02-02T09:08:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:18 crc kubenswrapper[4764]: I0202 09:08:18.968540 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:18 crc kubenswrapper[4764]: I0202 09:08:18.968614 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:18 crc kubenswrapper[4764]: I0202 09:08:18.968631 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:18 crc kubenswrapper[4764]: I0202 09:08:18.968657 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:18 crc kubenswrapper[4764]: I0202 09:08:18.968676 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:18Z","lastTransitionTime":"2026-02-02T09:08:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:19 crc kubenswrapper[4764]: I0202 09:08:19.071262 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:19 crc kubenswrapper[4764]: I0202 09:08:19.071540 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:19 crc kubenswrapper[4764]: I0202 09:08:19.071616 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:19 crc kubenswrapper[4764]: I0202 09:08:19.071679 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:19 crc kubenswrapper[4764]: I0202 09:08:19.071735 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:19Z","lastTransitionTime":"2026-02-02T09:08:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:19 crc kubenswrapper[4764]: I0202 09:08:19.174327 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:19 crc kubenswrapper[4764]: I0202 09:08:19.174362 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:19 crc kubenswrapper[4764]: I0202 09:08:19.174371 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:19 crc kubenswrapper[4764]: I0202 09:08:19.174387 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:19 crc kubenswrapper[4764]: I0202 09:08:19.174399 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:19Z","lastTransitionTime":"2026-02-02T09:08:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:19 crc kubenswrapper[4764]: I0202 09:08:19.277005 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:19 crc kubenswrapper[4764]: I0202 09:08:19.277047 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:19 crc kubenswrapper[4764]: I0202 09:08:19.277061 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:19 crc kubenswrapper[4764]: I0202 09:08:19.277079 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:19 crc kubenswrapper[4764]: I0202 09:08:19.277092 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:19Z","lastTransitionTime":"2026-02-02T09:08:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:19 crc kubenswrapper[4764]: I0202 09:08:19.379736 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:19 crc kubenswrapper[4764]: I0202 09:08:19.379763 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:19 crc kubenswrapper[4764]: I0202 09:08:19.379771 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:19 crc kubenswrapper[4764]: I0202 09:08:19.379784 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:19 crc kubenswrapper[4764]: I0202 09:08:19.379793 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:19Z","lastTransitionTime":"2026-02-02T09:08:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:19 crc kubenswrapper[4764]: I0202 09:08:19.481612 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:19 crc kubenswrapper[4764]: I0202 09:08:19.481746 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:19 crc kubenswrapper[4764]: I0202 09:08:19.481777 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:19 crc kubenswrapper[4764]: I0202 09:08:19.481807 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:19 crc kubenswrapper[4764]: I0202 09:08:19.481830 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:19Z","lastTransitionTime":"2026-02-02T09:08:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:19 crc kubenswrapper[4764]: I0202 09:08:19.584136 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:19 crc kubenswrapper[4764]: I0202 09:08:19.584205 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:19 crc kubenswrapper[4764]: I0202 09:08:19.584221 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:19 crc kubenswrapper[4764]: I0202 09:08:19.584243 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:19 crc kubenswrapper[4764]: I0202 09:08:19.584255 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:19Z","lastTransitionTime":"2026-02-02T09:08:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:19 crc kubenswrapper[4764]: I0202 09:08:19.686401 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:19 crc kubenswrapper[4764]: I0202 09:08:19.686439 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:19 crc kubenswrapper[4764]: I0202 09:08:19.686448 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:19 crc kubenswrapper[4764]: I0202 09:08:19.686463 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:19 crc kubenswrapper[4764]: I0202 09:08:19.686473 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:19Z","lastTransitionTime":"2026-02-02T09:08:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:19 crc kubenswrapper[4764]: I0202 09:08:19.788624 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:19 crc kubenswrapper[4764]: I0202 09:08:19.788665 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:19 crc kubenswrapper[4764]: I0202 09:08:19.788675 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:19 crc kubenswrapper[4764]: I0202 09:08:19.788692 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:19 crc kubenswrapper[4764]: I0202 09:08:19.788703 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:19Z","lastTransitionTime":"2026-02-02T09:08:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:19 crc kubenswrapper[4764]: I0202 09:08:19.809928 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-07 01:26:55.413707083 +0000 UTC Feb 02 09:08:19 crc kubenswrapper[4764]: I0202 09:08:19.825279 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kwtmr" Feb 02 09:08:19 crc kubenswrapper[4764]: E0202 09:08:19.825965 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kwtmr" podUID="991faa9a-dd25-4f49-82bd-ce60cefd4af2" Feb 02 09:08:19 crc kubenswrapper[4764]: I0202 09:08:19.891360 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:19 crc kubenswrapper[4764]: I0202 09:08:19.891399 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:19 crc kubenswrapper[4764]: I0202 09:08:19.891410 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:19 crc kubenswrapper[4764]: I0202 09:08:19.891423 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:19 crc kubenswrapper[4764]: I0202 09:08:19.891433 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:19Z","lastTransitionTime":"2026-02-02T09:08:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:19 crc kubenswrapper[4764]: I0202 09:08:19.994357 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:19 crc kubenswrapper[4764]: I0202 09:08:19.994404 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:19 crc kubenswrapper[4764]: I0202 09:08:19.994413 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:19 crc kubenswrapper[4764]: I0202 09:08:19.994427 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:19 crc kubenswrapper[4764]: I0202 09:08:19.994435 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:19Z","lastTransitionTime":"2026-02-02T09:08:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:20 crc kubenswrapper[4764]: I0202 09:08:20.096887 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:20 crc kubenswrapper[4764]: I0202 09:08:20.096948 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:20 crc kubenswrapper[4764]: I0202 09:08:20.097004 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:20 crc kubenswrapper[4764]: I0202 09:08:20.097022 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:20 crc kubenswrapper[4764]: I0202 09:08:20.097035 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:20Z","lastTransitionTime":"2026-02-02T09:08:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:20 crc kubenswrapper[4764]: I0202 09:08:20.199700 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:20 crc kubenswrapper[4764]: I0202 09:08:20.199751 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:20 crc kubenswrapper[4764]: I0202 09:08:20.199767 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:20 crc kubenswrapper[4764]: I0202 09:08:20.199788 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:20 crc kubenswrapper[4764]: I0202 09:08:20.199802 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:20Z","lastTransitionTime":"2026-02-02T09:08:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:20 crc kubenswrapper[4764]: I0202 09:08:20.302509 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:20 crc kubenswrapper[4764]: I0202 09:08:20.302880 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:20 crc kubenswrapper[4764]: I0202 09:08:20.303079 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:20 crc kubenswrapper[4764]: I0202 09:08:20.303224 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:20 crc kubenswrapper[4764]: I0202 09:08:20.303355 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:20Z","lastTransitionTime":"2026-02-02T09:08:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:20 crc kubenswrapper[4764]: I0202 09:08:20.405870 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:20 crc kubenswrapper[4764]: I0202 09:08:20.405965 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:20 crc kubenswrapper[4764]: I0202 09:08:20.405979 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:20 crc kubenswrapper[4764]: I0202 09:08:20.406003 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:20 crc kubenswrapper[4764]: I0202 09:08:20.406019 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:20Z","lastTransitionTime":"2026-02-02T09:08:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:20 crc kubenswrapper[4764]: I0202 09:08:20.508413 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:20 crc kubenswrapper[4764]: I0202 09:08:20.508711 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:20 crc kubenswrapper[4764]: I0202 09:08:20.508785 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:20 crc kubenswrapper[4764]: I0202 09:08:20.508886 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:20 crc kubenswrapper[4764]: I0202 09:08:20.508993 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:20Z","lastTransitionTime":"2026-02-02T09:08:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:20 crc kubenswrapper[4764]: I0202 09:08:20.611831 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:20 crc kubenswrapper[4764]: I0202 09:08:20.611871 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:20 crc kubenswrapper[4764]: I0202 09:08:20.611880 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:20 crc kubenswrapper[4764]: I0202 09:08:20.611895 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:20 crc kubenswrapper[4764]: I0202 09:08:20.611906 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:20Z","lastTransitionTime":"2026-02-02T09:08:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:20 crc kubenswrapper[4764]: I0202 09:08:20.714405 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:20 crc kubenswrapper[4764]: I0202 09:08:20.714458 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:20 crc kubenswrapper[4764]: I0202 09:08:20.714471 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:20 crc kubenswrapper[4764]: I0202 09:08:20.714488 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:20 crc kubenswrapper[4764]: I0202 09:08:20.714501 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:20Z","lastTransitionTime":"2026-02-02T09:08:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:20 crc kubenswrapper[4764]: I0202 09:08:20.810475 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-14 22:25:03.23140997 +0000 UTC Feb 02 09:08:20 crc kubenswrapper[4764]: I0202 09:08:20.818017 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:20 crc kubenswrapper[4764]: I0202 09:08:20.818048 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:20 crc kubenswrapper[4764]: I0202 09:08:20.818059 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:20 crc kubenswrapper[4764]: I0202 09:08:20.818074 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:20 crc kubenswrapper[4764]: I0202 09:08:20.818085 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:20Z","lastTransitionTime":"2026-02-02T09:08:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:20 crc kubenswrapper[4764]: I0202 09:08:20.825108 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 09:08:20 crc kubenswrapper[4764]: I0202 09:08:20.825133 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 09:08:20 crc kubenswrapper[4764]: E0202 09:08:20.825232 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 09:08:20 crc kubenswrapper[4764]: I0202 09:08:20.825416 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 09:08:20 crc kubenswrapper[4764]: E0202 09:08:20.825478 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 09:08:20 crc kubenswrapper[4764]: E0202 09:08:20.825632 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 09:08:20 crc kubenswrapper[4764]: I0202 09:08:20.826325 4764 scope.go:117] "RemoveContainer" containerID="334e828f8ef72570f2511ee38659941202530471f5943aa07b8b21483bd70550" Feb 02 09:08:20 crc kubenswrapper[4764]: E0202 09:08:20.826538 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-zhn7j_openshift-ovn-kubernetes(24632dda-6100-4ab6-a28e-214ddae4360c)\"" pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" podUID="24632dda-6100-4ab6-a28e-214ddae4360c" Feb 02 09:08:20 crc kubenswrapper[4764]: I0202 09:08:20.858342 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24632dda-6100-4ab6-a28e-214ddae4360c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a6c67e9b7708cc0d85b0b6ffbbade0d34867e159a9c5487ca2606475c667a50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bffda48d7c881c1470251b501661324859373246cc1a21811bd36ca24887c7aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a989645cf485cd8ff80ffff1a86cd7808e1564eba5667535e161cc68db13b26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20b71d3ee7a381baa51b0b0d0470283022cfede30f491135737b5e82d631d763\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c380383b6a014d1b694a13ddc22f54b20b9e4f9d5c9b9000d76187621dce0a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5529f2bff87753428e9173826fc1880acf96a4c134fa9ecad195b15697cad73a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://334e828f8ef72570f2511ee38659941202530471f5943aa07b8b21483bd70550\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://334e828f8ef72570f2511ee38659941202530471f5943aa07b8b21483bd70550\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-02T09:08:10Z\\\",\\\"message\\\":\\\"ing OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-dns/dns-default]} name:Service_openshift-dns/dns-default_TCP_node_router+switch_crc options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.10:53: 10.217.4.10:9154:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {be9dcc9e-c16a-4962-a6d2-4adeb0b929c4}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-dns/dns-default]} name:Service_openshift-dns/dns-default_UDP_node_router+switch_crc options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[udp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.10:53:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {4c1be812-05d3-4f45-91b5-a853a5c8de71}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF0202 09:08:09.911531 6327 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T09:08:08Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-zhn7j_openshift-ovn-kubernetes(24632dda-6100-4ab6-a28e-214ddae4360c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b4cfbd4e682bcdcaff785013bdb72580ce36fced733e235538536827036b93d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://579903eadb183466a2564737e6b678e04308239c0cd75b6808892053b97afbc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://579903eadb183466a2564737e6b678e04308239c0cd75b6808892053b97afbc0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zhn7j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:20Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:20 crc kubenswrapper[4764]: I0202 09:08:20.879569 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d66ab3c0-9b5b-41d5-b6c0-91ccddccd7e9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d92ac486f5dade46bc75de32099859a71fc2ed25a7abbe6ae8937df128e7df30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a4ac273a4f127d79e5fad7d4c30cd2044a346f5734dba5d33d68cd092c0c2d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d38c7cb5dcd22fb07524494ebe77aa1a7f4f808f566043563817e47e7ec12a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://408722c7f83d28919431532f8c797a1c795fd5d96736d14e7f06a05744899702\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f473b18cd8003f75bb1dafb82bcec45745a73ccd10434297356dac185a3d15c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e229f484d19cbdabedf53d90a822a4434db9e00b1301538e885e26acc19d25f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e229f484d19cbdabedf53d90a822a4434db9e00b1301538e885e26acc19d25f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a9cc673be8ff245dba81278e8cd34cfe45ed36ff0b8bd5b30a0ddf333ec6371\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a9cc673be8ff245dba81278e8cd34cfe45ed36ff0b8bd5b30a0ddf333ec6371\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bc7e49c7a5ae66ea939060192fa6a346dc0b0b9aa9750b29879d3c426cc0d276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc7e49c7a5ae66ea939060192fa6a346dc0b0b9aa9750b29879d3c426cc0d276\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:20Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:20 crc kubenswrapper[4764]: I0202 09:08:20.891765 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa9d037e66cb53b9dd698eff46284c9e718fd842cda6f7b525f0c79062d70bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:20Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:20 crc kubenswrapper[4764]: I0202 09:08:20.909061 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:20Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:20 crc kubenswrapper[4764]: I0202 09:08:20.920196 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:20 crc kubenswrapper[4764]: I0202 09:08:20.920239 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:20 crc kubenswrapper[4764]: I0202 09:08:20.920269 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:20 crc kubenswrapper[4764]: I0202 09:08:20.920289 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:20 crc kubenswrapper[4764]: I0202 09:08:20.920302 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:20Z","lastTransitionTime":"2026-02-02T09:08:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:20 crc kubenswrapper[4764]: I0202 09:08:20.922462 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d192f670-9f9d-4539-9641-e4bed73acdd4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://040c5e8a5659f85f0d52d1a1e064154d998ae104f711101d2204c0b1caebbb30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbksf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fdf06410349b7c99aa166e87fbe80bf68e2ae923635a6a978b7e2cd50383f7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbksf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4ndm4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:20Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:20 crc kubenswrapper[4764]: I0202 09:08:20.939202 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pcjmn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6e5ccd3b-b657-47a2-9c4d-bc9adb126533\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72863eef568d92a1abeb3cde2e89630a4f5fb7626c6949448ff5195774c98eab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2f7zl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pcjmn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:20Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:20 crc kubenswrapper[4764]: I0202 09:08:20.954396 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed747fd8-e3ce-4319-a6da-bda0fd51b85a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6f064bdc66470aace3180d4988244026b54009a63cff199e41c5d02a1d18479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://60258f84a1d8dcfe48b7e9688ed50f077e201971f20f3f6663fbf1ba58a4611a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f84be2e08217a2ac3edde858d78b4410376ed9bae3bcabe2a7f9b2fbe2f69c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a69d25d352d7ceac5c735bc1fcf446a1fea6dc22c65a5296cfab828cea17377\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:20Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:20 crc kubenswrapper[4764]: I0202 09:08:20.967793 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f6904228e052fff69ad1fd66df97758b6b60ba4351e97a7fbd03eee4df1b653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:20Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:20 crc kubenswrapper[4764]: I0202 09:08:20.979222 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6zxff" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aaba93da-12a5-43df-b7c7-4c8b988163c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3db293678a907cd55b31b1d0c0bd7c30e1ba75ecd5e13cbc68ab98fbf0fbd9d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqwb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6zxff\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:20Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:20 crc kubenswrapper[4764]: I0202 09:08:20.992663 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fae745b-ea49-42f3-a486-dbc3dc2fb9ad\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://081d7f76d937ff71e35945f9b1d893888bb649d15bd84312fd4b61c9520ea71f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://429ead0993f5e648f6256d249ee12aea03514d3108618840552b672fb419cd74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fad17b01545eb20785f03e3c61bd5c982f6217e38176d7a3b23ac194298644a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85e2f75c56d63cacc93b9195c1e8bf46eb7a7234c45a024bd95a460db90edb6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85e2f75c56d63cacc93b9195c1e8bf46eb7a7234c45a024bd95a460db90edb6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:20Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:21 crc kubenswrapper[4764]: I0202 09:08:21.006532 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:21Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:21 crc kubenswrapper[4764]: I0202 09:08:21.020551 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lh6tm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2152e8a-f456-47b5-a547-b65ec837ecbc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38e176a00b07f2960fdd7ba2aadd1468a5e083b8d154e9d118d13f2585da81e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bb2sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://996e649011d5255783284421ddea8e0a5c383ef675b0920c3fdc0af6503d80e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bb2sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lh6tm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:21Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:21 crc kubenswrapper[4764]: I0202 09:08:21.022368 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:21 crc kubenswrapper[4764]: I0202 09:08:21.022405 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:21 crc kubenswrapper[4764]: I0202 09:08:21.022414 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:21 crc kubenswrapper[4764]: I0202 09:08:21.022429 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:21 crc kubenswrapper[4764]: I0202 09:08:21.022441 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:21Z","lastTransitionTime":"2026-02-02T09:08:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:21 crc kubenswrapper[4764]: I0202 09:08:21.034278 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:21Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:21 crc kubenswrapper[4764]: I0202 09:08:21.052749 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nc7qt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1a2c7b22068b7165927fc158885d48bf9952b9eba521d8ae9a599eee53dc305\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2hz4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nc7qt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:21Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:21 crc kubenswrapper[4764]: I0202 09:08:21.069885 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9p5dc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"660ac006-2495-45c4-9fb3-e9c1dddcf7a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da812b124c97aa8ad1563f4bcfac8f3bfa025096fa4a2561e5847b1bf9cb30b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23e57e65a6f5c3059506b9db21e9ac2492c30bd9191a98db0281915e9ce39797\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23e57e65a6f5c3059506b9db21e9ac2492c30bd9191a98db0281915e9ce39797\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2417b66816da1a6cd1791a4b91eca4a3c4f3694462e066e7f0722f1b9c2105c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2417b66816da1a6cd1791a4b91eca4a3c4f3694462e066e7f0722f1b9c2105c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://883641f29e9120c2cea6b272bacf1dc63c752f55ab0408cdaf73db85c85ac369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://883641f29e9120c2cea6b272bacf1dc63c752f55ab0408cdaf73db85c85ac369\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2745baba1188ba5e8f820616e4117a12c50ecee5b68e71ad9cabbb621b6e131f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2745baba1188ba5e8f820616e4117a12c50ecee5b68e71ad9cabbb621b6e131f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8925e435be4d48d86824fc35f8b8819fb35d63e385164c4a20edb3bf3035c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a8925e435be4d48d86824fc35f8b8819fb35d63e385164c4a20edb3bf3035c3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d3fbf668ae7d04c8d9e574f43c2fe90c5604204ba38a3c60847afdfbceb7a57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d3fbf668ae7d04c8d9e574f43c2fe90c5604204ba38a3c60847afdfbceb7a57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9p5dc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:21Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:21 crc kubenswrapper[4764]: I0202 09:08:21.080304 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kwtmr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"991faa9a-dd25-4f49-82bd-ce60cefd4af2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h865b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h865b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:53Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kwtmr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:21Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:21 crc kubenswrapper[4764]: I0202 09:08:21.097907 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87d61aea-46f6-4ee6-a7a4-767a6a67fea4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e78b6d08e4d616cefbb73e3fd43b3c6886dcd8aa997c7b4948bf1ff63401293\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d468715c2da33996561c40043216233fd12e21539d0c58dd1f3b892fd9f91a9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b184f7c936dc4046f74c0018269cbc2c4b7ceddb650b4e871d3f8a707e0df88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://269c5074371bd00f8363fb3888c11f1df383081a61497aea6578f08700881d50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b25bc9395831524c53f9e565b4efbe74f21792efafeb1dd507161bc91aa0ecdf\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T09:07:22Z\\\",\\\"message\\\":\\\"W0202 09:07:21.597488 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0202 09:07:21.598117 1 crypto.go:601] Generating new CA for check-endpoints-signer@1770023241 cert, and key in /tmp/serving-cert-3856183724/serving-signer.crt, /tmp/serving-cert-3856183724/serving-signer.key\\\\nI0202 09:07:22.012184 1 observer_polling.go:159] Starting file observer\\\\nW0202 09:07:22.015493 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0202 09:07:22.015797 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 09:07:22.016596 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3856183724/tls.crt::/tmp/serving-cert-3856183724/tls.key\\\\\\\"\\\\nF0202 09:07:22.270605 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b44e17b8ca01f549bc10081dd5f109cb1b9fff364bf5bd4fd90d9f0e3eeb04db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfb06a5ec9262cf39b108a66014d15e325cdbd5441a2d9e7a811dd57265c99a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfb06a5ec9262cf39b108a66014d15e325cdbd5441a2d9e7a811dd57265c99a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:21Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:21 crc kubenswrapper[4764]: I0202 09:08:21.111417 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5528225d91ac2778bef4554abb0bd036ac5d3b6cdac7357bab69e70c3894b6f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://071bbb80d0747cf3e109ba626d47f7d05f85f4cc0b3aba77276c795741f47ac6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:21Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:21 crc kubenswrapper[4764]: I0202 09:08:21.125347 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:21 crc kubenswrapper[4764]: I0202 09:08:21.125468 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:21 crc kubenswrapper[4764]: I0202 09:08:21.125485 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:21 crc kubenswrapper[4764]: I0202 09:08:21.125504 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:21 crc kubenswrapper[4764]: I0202 09:08:21.125526 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:21Z","lastTransitionTime":"2026-02-02T09:08:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:21 crc kubenswrapper[4764]: I0202 09:08:21.228193 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:21 crc kubenswrapper[4764]: I0202 09:08:21.228224 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:21 crc kubenswrapper[4764]: I0202 09:08:21.228234 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:21 crc kubenswrapper[4764]: I0202 09:08:21.228249 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:21 crc kubenswrapper[4764]: I0202 09:08:21.228258 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:21Z","lastTransitionTime":"2026-02-02T09:08:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:21 crc kubenswrapper[4764]: I0202 09:08:21.330851 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:21 crc kubenswrapper[4764]: I0202 09:08:21.330888 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:21 crc kubenswrapper[4764]: I0202 09:08:21.330895 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:21 crc kubenswrapper[4764]: I0202 09:08:21.330909 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:21 crc kubenswrapper[4764]: I0202 09:08:21.330918 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:21Z","lastTransitionTime":"2026-02-02T09:08:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:21 crc kubenswrapper[4764]: I0202 09:08:21.432996 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:21 crc kubenswrapper[4764]: I0202 09:08:21.433027 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:21 crc kubenswrapper[4764]: I0202 09:08:21.433038 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:21 crc kubenswrapper[4764]: I0202 09:08:21.433057 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:21 crc kubenswrapper[4764]: I0202 09:08:21.433071 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:21Z","lastTransitionTime":"2026-02-02T09:08:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:21 crc kubenswrapper[4764]: I0202 09:08:21.535322 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:21 crc kubenswrapper[4764]: I0202 09:08:21.535391 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:21 crc kubenswrapper[4764]: I0202 09:08:21.535402 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:21 crc kubenswrapper[4764]: I0202 09:08:21.535420 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:21 crc kubenswrapper[4764]: I0202 09:08:21.535432 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:21Z","lastTransitionTime":"2026-02-02T09:08:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:21 crc kubenswrapper[4764]: I0202 09:08:21.637087 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:21 crc kubenswrapper[4764]: I0202 09:08:21.637117 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:21 crc kubenswrapper[4764]: I0202 09:08:21.637125 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:21 crc kubenswrapper[4764]: I0202 09:08:21.637138 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:21 crc kubenswrapper[4764]: I0202 09:08:21.637149 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:21Z","lastTransitionTime":"2026-02-02T09:08:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:21 crc kubenswrapper[4764]: I0202 09:08:21.740668 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:21 crc kubenswrapper[4764]: I0202 09:08:21.740700 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:21 crc kubenswrapper[4764]: I0202 09:08:21.740708 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:21 crc kubenswrapper[4764]: I0202 09:08:21.740723 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:21 crc kubenswrapper[4764]: I0202 09:08:21.740732 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:21Z","lastTransitionTime":"2026-02-02T09:08:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:21 crc kubenswrapper[4764]: I0202 09:08:21.811523 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-15 18:00:57.987150598 +0000 UTC Feb 02 09:08:21 crc kubenswrapper[4764]: I0202 09:08:21.824955 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kwtmr" Feb 02 09:08:21 crc kubenswrapper[4764]: E0202 09:08:21.825080 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kwtmr" podUID="991faa9a-dd25-4f49-82bd-ce60cefd4af2" Feb 02 09:08:21 crc kubenswrapper[4764]: I0202 09:08:21.842483 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:21 crc kubenswrapper[4764]: I0202 09:08:21.842523 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:21 crc kubenswrapper[4764]: I0202 09:08:21.842531 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:21 crc kubenswrapper[4764]: I0202 09:08:21.842546 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:21 crc kubenswrapper[4764]: I0202 09:08:21.842555 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:21Z","lastTransitionTime":"2026-02-02T09:08:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:21 crc kubenswrapper[4764]: I0202 09:08:21.944793 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:21 crc kubenswrapper[4764]: I0202 09:08:21.944827 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:21 crc kubenswrapper[4764]: I0202 09:08:21.944835 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:21 crc kubenswrapper[4764]: I0202 09:08:21.944848 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:21 crc kubenswrapper[4764]: I0202 09:08:21.944857 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:21Z","lastTransitionTime":"2026-02-02T09:08:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:22 crc kubenswrapper[4764]: I0202 09:08:22.047261 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:22 crc kubenswrapper[4764]: I0202 09:08:22.047307 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:22 crc kubenswrapper[4764]: I0202 09:08:22.047321 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:22 crc kubenswrapper[4764]: I0202 09:08:22.047338 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:22 crc kubenswrapper[4764]: I0202 09:08:22.047350 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:22Z","lastTransitionTime":"2026-02-02T09:08:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:22 crc kubenswrapper[4764]: I0202 09:08:22.150124 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:22 crc kubenswrapper[4764]: I0202 09:08:22.150156 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:22 crc kubenswrapper[4764]: I0202 09:08:22.150164 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:22 crc kubenswrapper[4764]: I0202 09:08:22.150178 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:22 crc kubenswrapper[4764]: I0202 09:08:22.150188 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:22Z","lastTransitionTime":"2026-02-02T09:08:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:22 crc kubenswrapper[4764]: I0202 09:08:22.251988 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:22 crc kubenswrapper[4764]: I0202 09:08:22.252025 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:22 crc kubenswrapper[4764]: I0202 09:08:22.252036 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:22 crc kubenswrapper[4764]: I0202 09:08:22.252050 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:22 crc kubenswrapper[4764]: I0202 09:08:22.252062 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:22Z","lastTransitionTime":"2026-02-02T09:08:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:22 crc kubenswrapper[4764]: I0202 09:08:22.354634 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:22 crc kubenswrapper[4764]: I0202 09:08:22.354674 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:22 crc kubenswrapper[4764]: I0202 09:08:22.354685 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:22 crc kubenswrapper[4764]: I0202 09:08:22.354702 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:22 crc kubenswrapper[4764]: I0202 09:08:22.354716 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:22Z","lastTransitionTime":"2026-02-02T09:08:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:22 crc kubenswrapper[4764]: I0202 09:08:22.457207 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:22 crc kubenswrapper[4764]: I0202 09:08:22.457241 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:22 crc kubenswrapper[4764]: I0202 09:08:22.457251 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:22 crc kubenswrapper[4764]: I0202 09:08:22.457266 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:22 crc kubenswrapper[4764]: I0202 09:08:22.457277 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:22Z","lastTransitionTime":"2026-02-02T09:08:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:22 crc kubenswrapper[4764]: I0202 09:08:22.559369 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:22 crc kubenswrapper[4764]: I0202 09:08:22.559398 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:22 crc kubenswrapper[4764]: I0202 09:08:22.559406 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:22 crc kubenswrapper[4764]: I0202 09:08:22.559434 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:22 crc kubenswrapper[4764]: I0202 09:08:22.559443 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:22Z","lastTransitionTime":"2026-02-02T09:08:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:22 crc kubenswrapper[4764]: I0202 09:08:22.662376 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:22 crc kubenswrapper[4764]: I0202 09:08:22.662413 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:22 crc kubenswrapper[4764]: I0202 09:08:22.662423 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:22 crc kubenswrapper[4764]: I0202 09:08:22.662438 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:22 crc kubenswrapper[4764]: I0202 09:08:22.662449 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:22Z","lastTransitionTime":"2026-02-02T09:08:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:22 crc kubenswrapper[4764]: I0202 09:08:22.764405 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:22 crc kubenswrapper[4764]: I0202 09:08:22.764449 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:22 crc kubenswrapper[4764]: I0202 09:08:22.764459 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:22 crc kubenswrapper[4764]: I0202 09:08:22.764475 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:22 crc kubenswrapper[4764]: I0202 09:08:22.764484 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:22Z","lastTransitionTime":"2026-02-02T09:08:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:22 crc kubenswrapper[4764]: I0202 09:08:22.811970 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-09 15:27:52.565620697 +0000 UTC Feb 02 09:08:22 crc kubenswrapper[4764]: I0202 09:08:22.825246 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 09:08:22 crc kubenswrapper[4764]: I0202 09:08:22.825332 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 09:08:22 crc kubenswrapper[4764]: E0202 09:08:22.825374 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 09:08:22 crc kubenswrapper[4764]: E0202 09:08:22.825446 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 09:08:22 crc kubenswrapper[4764]: I0202 09:08:22.825267 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 09:08:22 crc kubenswrapper[4764]: E0202 09:08:22.825520 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 09:08:22 crc kubenswrapper[4764]: I0202 09:08:22.867423 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:22 crc kubenswrapper[4764]: I0202 09:08:22.867457 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:22 crc kubenswrapper[4764]: I0202 09:08:22.867467 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:22 crc kubenswrapper[4764]: I0202 09:08:22.867485 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:22 crc kubenswrapper[4764]: I0202 09:08:22.867497 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:22Z","lastTransitionTime":"2026-02-02T09:08:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:22 crc kubenswrapper[4764]: I0202 09:08:22.969560 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:22 crc kubenswrapper[4764]: I0202 09:08:22.969594 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:22 crc kubenswrapper[4764]: I0202 09:08:22.969602 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:22 crc kubenswrapper[4764]: I0202 09:08:22.969615 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:22 crc kubenswrapper[4764]: I0202 09:08:22.969626 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:22Z","lastTransitionTime":"2026-02-02T09:08:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:23 crc kubenswrapper[4764]: I0202 09:08:23.071613 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:23 crc kubenswrapper[4764]: I0202 09:08:23.071875 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:23 crc kubenswrapper[4764]: I0202 09:08:23.071972 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:23 crc kubenswrapper[4764]: I0202 09:08:23.072083 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:23 crc kubenswrapper[4764]: I0202 09:08:23.072158 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:23Z","lastTransitionTime":"2026-02-02T09:08:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:23 crc kubenswrapper[4764]: I0202 09:08:23.173766 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:23 crc kubenswrapper[4764]: I0202 09:08:23.174032 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:23 crc kubenswrapper[4764]: I0202 09:08:23.174140 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:23 crc kubenswrapper[4764]: I0202 09:08:23.174225 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:23 crc kubenswrapper[4764]: I0202 09:08:23.174321 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:23Z","lastTransitionTime":"2026-02-02T09:08:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:23 crc kubenswrapper[4764]: I0202 09:08:23.279008 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:23 crc kubenswrapper[4764]: I0202 09:08:23.279043 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:23 crc kubenswrapper[4764]: I0202 09:08:23.279054 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:23 crc kubenswrapper[4764]: I0202 09:08:23.279087 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:23 crc kubenswrapper[4764]: I0202 09:08:23.279097 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:23Z","lastTransitionTime":"2026-02-02T09:08:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:23 crc kubenswrapper[4764]: I0202 09:08:23.387716 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:23 crc kubenswrapper[4764]: I0202 09:08:23.387762 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:23 crc kubenswrapper[4764]: I0202 09:08:23.387780 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:23 crc kubenswrapper[4764]: I0202 09:08:23.387801 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:23 crc kubenswrapper[4764]: I0202 09:08:23.387815 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:23Z","lastTransitionTime":"2026-02-02T09:08:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:23 crc kubenswrapper[4764]: I0202 09:08:23.489781 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:23 crc kubenswrapper[4764]: I0202 09:08:23.489831 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:23 crc kubenswrapper[4764]: I0202 09:08:23.489841 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:23 crc kubenswrapper[4764]: I0202 09:08:23.489855 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:23 crc kubenswrapper[4764]: I0202 09:08:23.489864 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:23Z","lastTransitionTime":"2026-02-02T09:08:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:23 crc kubenswrapper[4764]: I0202 09:08:23.591583 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:23 crc kubenswrapper[4764]: I0202 09:08:23.591618 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:23 crc kubenswrapper[4764]: I0202 09:08:23.591626 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:23 crc kubenswrapper[4764]: I0202 09:08:23.591640 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:23 crc kubenswrapper[4764]: I0202 09:08:23.591650 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:23Z","lastTransitionTime":"2026-02-02T09:08:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:23 crc kubenswrapper[4764]: I0202 09:08:23.694259 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:23 crc kubenswrapper[4764]: I0202 09:08:23.694316 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:23 crc kubenswrapper[4764]: I0202 09:08:23.694336 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:23 crc kubenswrapper[4764]: I0202 09:08:23.694361 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:23 crc kubenswrapper[4764]: I0202 09:08:23.694378 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:23Z","lastTransitionTime":"2026-02-02T09:08:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:23 crc kubenswrapper[4764]: I0202 09:08:23.796973 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:23 crc kubenswrapper[4764]: I0202 09:08:23.797171 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:23 crc kubenswrapper[4764]: I0202 09:08:23.797229 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:23 crc kubenswrapper[4764]: I0202 09:08:23.797318 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:23 crc kubenswrapper[4764]: I0202 09:08:23.797373 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:23Z","lastTransitionTime":"2026-02-02T09:08:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:23 crc kubenswrapper[4764]: I0202 09:08:23.812084 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-12 11:35:46.699490559 +0000 UTC Feb 02 09:08:23 crc kubenswrapper[4764]: I0202 09:08:23.825167 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kwtmr" Feb 02 09:08:23 crc kubenswrapper[4764]: E0202 09:08:23.825299 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kwtmr" podUID="991faa9a-dd25-4f49-82bd-ce60cefd4af2" Feb 02 09:08:23 crc kubenswrapper[4764]: I0202 09:08:23.900158 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:23 crc kubenswrapper[4764]: I0202 09:08:23.900217 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:23 crc kubenswrapper[4764]: I0202 09:08:23.900228 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:23 crc kubenswrapper[4764]: I0202 09:08:23.900262 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:23 crc kubenswrapper[4764]: I0202 09:08:23.900275 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:23Z","lastTransitionTime":"2026-02-02T09:08:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:24 crc kubenswrapper[4764]: I0202 09:08:24.003026 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:24 crc kubenswrapper[4764]: I0202 09:08:24.003078 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:24 crc kubenswrapper[4764]: I0202 09:08:24.003096 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:24 crc kubenswrapper[4764]: I0202 09:08:24.003116 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:24 crc kubenswrapper[4764]: I0202 09:08:24.003128 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:24Z","lastTransitionTime":"2026-02-02T09:08:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:24 crc kubenswrapper[4764]: I0202 09:08:24.105382 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:24 crc kubenswrapper[4764]: I0202 09:08:24.105578 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:24 crc kubenswrapper[4764]: I0202 09:08:24.105654 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:24 crc kubenswrapper[4764]: I0202 09:08:24.105716 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:24 crc kubenswrapper[4764]: I0202 09:08:24.105807 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:24Z","lastTransitionTime":"2026-02-02T09:08:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:24 crc kubenswrapper[4764]: I0202 09:08:24.208034 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:24 crc kubenswrapper[4764]: I0202 09:08:24.208101 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:24 crc kubenswrapper[4764]: I0202 09:08:24.208115 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:24 crc kubenswrapper[4764]: I0202 09:08:24.208130 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:24 crc kubenswrapper[4764]: I0202 09:08:24.208141 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:24Z","lastTransitionTime":"2026-02-02T09:08:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:24 crc kubenswrapper[4764]: I0202 09:08:24.310699 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:24 crc kubenswrapper[4764]: I0202 09:08:24.310735 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:24 crc kubenswrapper[4764]: I0202 09:08:24.310744 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:24 crc kubenswrapper[4764]: I0202 09:08:24.310761 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:24 crc kubenswrapper[4764]: I0202 09:08:24.310774 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:24Z","lastTransitionTime":"2026-02-02T09:08:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:24 crc kubenswrapper[4764]: I0202 09:08:24.412840 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:24 crc kubenswrapper[4764]: I0202 09:08:24.412882 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:24 crc kubenswrapper[4764]: I0202 09:08:24.412892 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:24 crc kubenswrapper[4764]: I0202 09:08:24.412908 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:24 crc kubenswrapper[4764]: I0202 09:08:24.412919 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:24Z","lastTransitionTime":"2026-02-02T09:08:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:24 crc kubenswrapper[4764]: I0202 09:08:24.515122 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:24 crc kubenswrapper[4764]: I0202 09:08:24.515157 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:24 crc kubenswrapper[4764]: I0202 09:08:24.515170 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:24 crc kubenswrapper[4764]: I0202 09:08:24.515186 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:24 crc kubenswrapper[4764]: I0202 09:08:24.515196 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:24Z","lastTransitionTime":"2026-02-02T09:08:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:24 crc kubenswrapper[4764]: I0202 09:08:24.618318 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:24 crc kubenswrapper[4764]: I0202 09:08:24.618345 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:24 crc kubenswrapper[4764]: I0202 09:08:24.618357 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:24 crc kubenswrapper[4764]: I0202 09:08:24.618375 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:24 crc kubenswrapper[4764]: I0202 09:08:24.618386 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:24Z","lastTransitionTime":"2026-02-02T09:08:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:24 crc kubenswrapper[4764]: I0202 09:08:24.720797 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:24 crc kubenswrapper[4764]: I0202 09:08:24.720837 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:24 crc kubenswrapper[4764]: I0202 09:08:24.720848 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:24 crc kubenswrapper[4764]: I0202 09:08:24.720864 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:24 crc kubenswrapper[4764]: I0202 09:08:24.720876 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:24Z","lastTransitionTime":"2026-02-02T09:08:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:24 crc kubenswrapper[4764]: I0202 09:08:24.813142 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-17 14:11:43.372814303 +0000 UTC Feb 02 09:08:24 crc kubenswrapper[4764]: I0202 09:08:24.818397 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" Feb 02 09:08:24 crc kubenswrapper[4764]: I0202 09:08:24.819317 4764 scope.go:117] "RemoveContainer" containerID="334e828f8ef72570f2511ee38659941202530471f5943aa07b8b21483bd70550" Feb 02 09:08:24 crc kubenswrapper[4764]: E0202 09:08:24.819488 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-zhn7j_openshift-ovn-kubernetes(24632dda-6100-4ab6-a28e-214ddae4360c)\"" pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" podUID="24632dda-6100-4ab6-a28e-214ddae4360c" Feb 02 09:08:24 crc kubenswrapper[4764]: I0202 09:08:24.822608 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:24 crc kubenswrapper[4764]: I0202 09:08:24.822637 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:24 crc kubenswrapper[4764]: I0202 09:08:24.822650 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:24 crc kubenswrapper[4764]: I0202 09:08:24.822665 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:24 crc kubenswrapper[4764]: I0202 09:08:24.822677 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:24Z","lastTransitionTime":"2026-02-02T09:08:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:24 crc kubenswrapper[4764]: I0202 09:08:24.824872 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 09:08:24 crc kubenswrapper[4764]: E0202 09:08:24.825153 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 09:08:24 crc kubenswrapper[4764]: I0202 09:08:24.825343 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 09:08:24 crc kubenswrapper[4764]: I0202 09:08:24.825417 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 09:08:24 crc kubenswrapper[4764]: E0202 09:08:24.825529 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 09:08:24 crc kubenswrapper[4764]: E0202 09:08:24.825624 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 09:08:24 crc kubenswrapper[4764]: I0202 09:08:24.925015 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:24 crc kubenswrapper[4764]: I0202 09:08:24.925249 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:24 crc kubenswrapper[4764]: I0202 09:08:24.925341 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:24 crc kubenswrapper[4764]: I0202 09:08:24.925405 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:24 crc kubenswrapper[4764]: I0202 09:08:24.925471 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:24Z","lastTransitionTime":"2026-02-02T09:08:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:25 crc kubenswrapper[4764]: I0202 09:08:25.028059 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:25 crc kubenswrapper[4764]: I0202 09:08:25.028091 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:25 crc kubenswrapper[4764]: I0202 09:08:25.028103 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:25 crc kubenswrapper[4764]: I0202 09:08:25.028120 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:25 crc kubenswrapper[4764]: I0202 09:08:25.028131 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:25Z","lastTransitionTime":"2026-02-02T09:08:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:25 crc kubenswrapper[4764]: I0202 09:08:25.130796 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:25 crc kubenswrapper[4764]: I0202 09:08:25.130832 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:25 crc kubenswrapper[4764]: I0202 09:08:25.130843 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:25 crc kubenswrapper[4764]: I0202 09:08:25.130860 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:25 crc kubenswrapper[4764]: I0202 09:08:25.130875 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:25Z","lastTransitionTime":"2026-02-02T09:08:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:25 crc kubenswrapper[4764]: I0202 09:08:25.232672 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:25 crc kubenswrapper[4764]: I0202 09:08:25.232717 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:25 crc kubenswrapper[4764]: I0202 09:08:25.232729 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:25 crc kubenswrapper[4764]: I0202 09:08:25.232745 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:25 crc kubenswrapper[4764]: I0202 09:08:25.232758 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:25Z","lastTransitionTime":"2026-02-02T09:08:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:25 crc kubenswrapper[4764]: I0202 09:08:25.335086 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:25 crc kubenswrapper[4764]: I0202 09:08:25.335132 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:25 crc kubenswrapper[4764]: I0202 09:08:25.335141 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:25 crc kubenswrapper[4764]: I0202 09:08:25.335156 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:25 crc kubenswrapper[4764]: I0202 09:08:25.335166 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:25Z","lastTransitionTime":"2026-02-02T09:08:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:25 crc kubenswrapper[4764]: I0202 09:08:25.437063 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:25 crc kubenswrapper[4764]: I0202 09:08:25.437094 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:25 crc kubenswrapper[4764]: I0202 09:08:25.437102 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:25 crc kubenswrapper[4764]: I0202 09:08:25.437116 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:25 crc kubenswrapper[4764]: I0202 09:08:25.437125 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:25Z","lastTransitionTime":"2026-02-02T09:08:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:25 crc kubenswrapper[4764]: I0202 09:08:25.539791 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:25 crc kubenswrapper[4764]: I0202 09:08:25.539819 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:25 crc kubenswrapper[4764]: I0202 09:08:25.539827 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:25 crc kubenswrapper[4764]: I0202 09:08:25.539839 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:25 crc kubenswrapper[4764]: I0202 09:08:25.539848 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:25Z","lastTransitionTime":"2026-02-02T09:08:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:25 crc kubenswrapper[4764]: I0202 09:08:25.641714 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:25 crc kubenswrapper[4764]: I0202 09:08:25.641742 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:25 crc kubenswrapper[4764]: I0202 09:08:25.641769 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:25 crc kubenswrapper[4764]: I0202 09:08:25.641785 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:25 crc kubenswrapper[4764]: I0202 09:08:25.641795 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:25Z","lastTransitionTime":"2026-02-02T09:08:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:25 crc kubenswrapper[4764]: I0202 09:08:25.743573 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:25 crc kubenswrapper[4764]: I0202 09:08:25.743597 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:25 crc kubenswrapper[4764]: I0202 09:08:25.743605 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:25 crc kubenswrapper[4764]: I0202 09:08:25.743618 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:25 crc kubenswrapper[4764]: I0202 09:08:25.743626 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:25Z","lastTransitionTime":"2026-02-02T09:08:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:25 crc kubenswrapper[4764]: I0202 09:08:25.814008 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-19 04:20:48.795956623 +0000 UTC Feb 02 09:08:25 crc kubenswrapper[4764]: I0202 09:08:25.825371 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kwtmr" Feb 02 09:08:25 crc kubenswrapper[4764]: E0202 09:08:25.825492 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kwtmr" podUID="991faa9a-dd25-4f49-82bd-ce60cefd4af2" Feb 02 09:08:25 crc kubenswrapper[4764]: I0202 09:08:25.845205 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:25 crc kubenswrapper[4764]: I0202 09:08:25.845236 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:25 crc kubenswrapper[4764]: I0202 09:08:25.845245 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:25 crc kubenswrapper[4764]: I0202 09:08:25.845259 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:25 crc kubenswrapper[4764]: I0202 09:08:25.845271 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:25Z","lastTransitionTime":"2026-02-02T09:08:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:25 crc kubenswrapper[4764]: I0202 09:08:25.919462 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/991faa9a-dd25-4f49-82bd-ce60cefd4af2-metrics-certs\") pod \"network-metrics-daemon-kwtmr\" (UID: \"991faa9a-dd25-4f49-82bd-ce60cefd4af2\") " pod="openshift-multus/network-metrics-daemon-kwtmr" Feb 02 09:08:25 crc kubenswrapper[4764]: E0202 09:08:25.919628 4764 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 02 09:08:25 crc kubenswrapper[4764]: E0202 09:08:25.919683 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/991faa9a-dd25-4f49-82bd-ce60cefd4af2-metrics-certs podName:991faa9a-dd25-4f49-82bd-ce60cefd4af2 nodeName:}" failed. No retries permitted until 2026-02-02 09:08:57.919665214 +0000 UTC m=+100.853389302 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/991faa9a-dd25-4f49-82bd-ce60cefd4af2-metrics-certs") pod "network-metrics-daemon-kwtmr" (UID: "991faa9a-dd25-4f49-82bd-ce60cefd4af2") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 02 09:08:25 crc kubenswrapper[4764]: I0202 09:08:25.946980 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:25 crc kubenswrapper[4764]: I0202 09:08:25.947018 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:25 crc kubenswrapper[4764]: I0202 09:08:25.947027 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:25 crc kubenswrapper[4764]: I0202 09:08:25.947042 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:25 crc kubenswrapper[4764]: I0202 09:08:25.947052 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:25Z","lastTransitionTime":"2026-02-02T09:08:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:26 crc kubenswrapper[4764]: I0202 09:08:26.049292 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:26 crc kubenswrapper[4764]: I0202 09:08:26.049336 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:26 crc kubenswrapper[4764]: I0202 09:08:26.049344 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:26 crc kubenswrapper[4764]: I0202 09:08:26.049360 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:26 crc kubenswrapper[4764]: I0202 09:08:26.049369 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:26Z","lastTransitionTime":"2026-02-02T09:08:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:26 crc kubenswrapper[4764]: I0202 09:08:26.151822 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:26 crc kubenswrapper[4764]: I0202 09:08:26.151861 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:26 crc kubenswrapper[4764]: I0202 09:08:26.151872 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:26 crc kubenswrapper[4764]: I0202 09:08:26.151889 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:26 crc kubenswrapper[4764]: I0202 09:08:26.151900 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:26Z","lastTransitionTime":"2026-02-02T09:08:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:26 crc kubenswrapper[4764]: I0202 09:08:26.239244 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:26 crc kubenswrapper[4764]: I0202 09:08:26.239287 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:26 crc kubenswrapper[4764]: I0202 09:08:26.239301 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:26 crc kubenswrapper[4764]: I0202 09:08:26.239318 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:26 crc kubenswrapper[4764]: I0202 09:08:26.239331 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:26Z","lastTransitionTime":"2026-02-02T09:08:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:26 crc kubenswrapper[4764]: E0202 09:08:26.253536 4764 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:08:26Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:08:26Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:08:26Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:08:26Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"31973a70-ce17-4811-9a14-f5476c6979db\\\",\\\"systemUUID\\\":\\\"ca92679c-8580-43c3-be51-62441e93d22b\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:26Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:26 crc kubenswrapper[4764]: I0202 09:08:26.256892 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:26 crc kubenswrapper[4764]: I0202 09:08:26.256960 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:26 crc kubenswrapper[4764]: I0202 09:08:26.256977 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:26 crc kubenswrapper[4764]: I0202 09:08:26.256996 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:26 crc kubenswrapper[4764]: I0202 09:08:26.257009 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:26Z","lastTransitionTime":"2026-02-02T09:08:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:26 crc kubenswrapper[4764]: E0202 09:08:26.270427 4764 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:08:26Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:08:26Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:08:26Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:08:26Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"31973a70-ce17-4811-9a14-f5476c6979db\\\",\\\"systemUUID\\\":\\\"ca92679c-8580-43c3-be51-62441e93d22b\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:26Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:26 crc kubenswrapper[4764]: I0202 09:08:26.274788 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:26 crc kubenswrapper[4764]: I0202 09:08:26.274830 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:26 crc kubenswrapper[4764]: I0202 09:08:26.274841 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:26 crc kubenswrapper[4764]: I0202 09:08:26.274861 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:26 crc kubenswrapper[4764]: I0202 09:08:26.274871 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:26Z","lastTransitionTime":"2026-02-02T09:08:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:26 crc kubenswrapper[4764]: E0202 09:08:26.288241 4764 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:08:26Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:08:26Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:08:26Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:08:26Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"31973a70-ce17-4811-9a14-f5476c6979db\\\",\\\"systemUUID\\\":\\\"ca92679c-8580-43c3-be51-62441e93d22b\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:26Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:26 crc kubenswrapper[4764]: I0202 09:08:26.292978 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:26 crc kubenswrapper[4764]: I0202 09:08:26.293195 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:26 crc kubenswrapper[4764]: I0202 09:08:26.293322 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:26 crc kubenswrapper[4764]: I0202 09:08:26.293418 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:26 crc kubenswrapper[4764]: I0202 09:08:26.293514 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:26Z","lastTransitionTime":"2026-02-02T09:08:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:26 crc kubenswrapper[4764]: E0202 09:08:26.307435 4764 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:08:26Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:08:26Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:08:26Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:08:26Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"31973a70-ce17-4811-9a14-f5476c6979db\\\",\\\"systemUUID\\\":\\\"ca92679c-8580-43c3-be51-62441e93d22b\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:26Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:26 crc kubenswrapper[4764]: I0202 09:08:26.311501 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:26 crc kubenswrapper[4764]: I0202 09:08:26.311565 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:26 crc kubenswrapper[4764]: I0202 09:08:26.311578 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:26 crc kubenswrapper[4764]: I0202 09:08:26.311596 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:26 crc kubenswrapper[4764]: I0202 09:08:26.311608 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:26Z","lastTransitionTime":"2026-02-02T09:08:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:26 crc kubenswrapper[4764]: E0202 09:08:26.327399 4764 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:08:26Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:08:26Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:08:26Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:08:26Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"31973a70-ce17-4811-9a14-f5476c6979db\\\",\\\"systemUUID\\\":\\\"ca92679c-8580-43c3-be51-62441e93d22b\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:26Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:26 crc kubenswrapper[4764]: E0202 09:08:26.327614 4764 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 02 09:08:26 crc kubenswrapper[4764]: I0202 09:08:26.329260 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:26 crc kubenswrapper[4764]: I0202 09:08:26.329290 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:26 crc kubenswrapper[4764]: I0202 09:08:26.329302 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:26 crc kubenswrapper[4764]: I0202 09:08:26.329319 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:26 crc kubenswrapper[4764]: I0202 09:08:26.329332 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:26Z","lastTransitionTime":"2026-02-02T09:08:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:26 crc kubenswrapper[4764]: I0202 09:08:26.431829 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:26 crc kubenswrapper[4764]: I0202 09:08:26.431862 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:26 crc kubenswrapper[4764]: I0202 09:08:26.431874 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:26 crc kubenswrapper[4764]: I0202 09:08:26.431890 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:26 crc kubenswrapper[4764]: I0202 09:08:26.431900 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:26Z","lastTransitionTime":"2026-02-02T09:08:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:26 crc kubenswrapper[4764]: I0202 09:08:26.534271 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:26 crc kubenswrapper[4764]: I0202 09:08:26.534310 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:26 crc kubenswrapper[4764]: I0202 09:08:26.534321 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:26 crc kubenswrapper[4764]: I0202 09:08:26.534337 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:26 crc kubenswrapper[4764]: I0202 09:08:26.534350 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:26Z","lastTransitionTime":"2026-02-02T09:08:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:26 crc kubenswrapper[4764]: I0202 09:08:26.636222 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:26 crc kubenswrapper[4764]: I0202 09:08:26.636762 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:26 crc kubenswrapper[4764]: I0202 09:08:26.636838 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:26 crc kubenswrapper[4764]: I0202 09:08:26.636920 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:26 crc kubenswrapper[4764]: I0202 09:08:26.637030 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:26Z","lastTransitionTime":"2026-02-02T09:08:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:26 crc kubenswrapper[4764]: I0202 09:08:26.739788 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:26 crc kubenswrapper[4764]: I0202 09:08:26.739835 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:26 crc kubenswrapper[4764]: I0202 09:08:26.739847 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:26 crc kubenswrapper[4764]: I0202 09:08:26.739864 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:26 crc kubenswrapper[4764]: I0202 09:08:26.739875 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:26Z","lastTransitionTime":"2026-02-02T09:08:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:26 crc kubenswrapper[4764]: I0202 09:08:26.815073 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-13 18:06:44.034290884 +0000 UTC Feb 02 09:08:26 crc kubenswrapper[4764]: I0202 09:08:26.825346 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 09:08:26 crc kubenswrapper[4764]: I0202 09:08:26.825394 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 09:08:26 crc kubenswrapper[4764]: E0202 09:08:26.825458 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 09:08:26 crc kubenswrapper[4764]: E0202 09:08:26.825570 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 09:08:26 crc kubenswrapper[4764]: I0202 09:08:26.825840 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 09:08:26 crc kubenswrapper[4764]: E0202 09:08:26.826139 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 09:08:26 crc kubenswrapper[4764]: I0202 09:08:26.842267 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:26 crc kubenswrapper[4764]: I0202 09:08:26.842297 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:26 crc kubenswrapper[4764]: I0202 09:08:26.842308 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:26 crc kubenswrapper[4764]: I0202 09:08:26.842322 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:26 crc kubenswrapper[4764]: I0202 09:08:26.842333 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:26Z","lastTransitionTime":"2026-02-02T09:08:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:26 crc kubenswrapper[4764]: I0202 09:08:26.945233 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:26 crc kubenswrapper[4764]: I0202 09:08:26.945288 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:26 crc kubenswrapper[4764]: I0202 09:08:26.945302 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:26 crc kubenswrapper[4764]: I0202 09:08:26.945320 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:26 crc kubenswrapper[4764]: I0202 09:08:26.945337 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:26Z","lastTransitionTime":"2026-02-02T09:08:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:27 crc kubenswrapper[4764]: I0202 09:08:27.047689 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:27 crc kubenswrapper[4764]: I0202 09:08:27.047721 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:27 crc kubenswrapper[4764]: I0202 09:08:27.047729 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:27 crc kubenswrapper[4764]: I0202 09:08:27.047742 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:27 crc kubenswrapper[4764]: I0202 09:08:27.047750 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:27Z","lastTransitionTime":"2026-02-02T09:08:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:27 crc kubenswrapper[4764]: I0202 09:08:27.150285 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:27 crc kubenswrapper[4764]: I0202 09:08:27.150319 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:27 crc kubenswrapper[4764]: I0202 09:08:27.150328 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:27 crc kubenswrapper[4764]: I0202 09:08:27.150340 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:27 crc kubenswrapper[4764]: I0202 09:08:27.150349 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:27Z","lastTransitionTime":"2026-02-02T09:08:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:27 crc kubenswrapper[4764]: I0202 09:08:27.252507 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:27 crc kubenswrapper[4764]: I0202 09:08:27.252549 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:27 crc kubenswrapper[4764]: I0202 09:08:27.252558 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:27 crc kubenswrapper[4764]: I0202 09:08:27.252571 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:27 crc kubenswrapper[4764]: I0202 09:08:27.252580 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:27Z","lastTransitionTime":"2026-02-02T09:08:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:27 crc kubenswrapper[4764]: I0202 09:08:27.355436 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:27 crc kubenswrapper[4764]: I0202 09:08:27.355483 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:27 crc kubenswrapper[4764]: I0202 09:08:27.355496 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:27 crc kubenswrapper[4764]: I0202 09:08:27.355520 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:27 crc kubenswrapper[4764]: I0202 09:08:27.355533 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:27Z","lastTransitionTime":"2026-02-02T09:08:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:27 crc kubenswrapper[4764]: I0202 09:08:27.457957 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:27 crc kubenswrapper[4764]: I0202 09:08:27.458005 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:27 crc kubenswrapper[4764]: I0202 09:08:27.458023 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:27 crc kubenswrapper[4764]: I0202 09:08:27.458041 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:27 crc kubenswrapper[4764]: I0202 09:08:27.458052 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:27Z","lastTransitionTime":"2026-02-02T09:08:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:27 crc kubenswrapper[4764]: I0202 09:08:27.559816 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:27 crc kubenswrapper[4764]: I0202 09:08:27.559845 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:27 crc kubenswrapper[4764]: I0202 09:08:27.559855 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:27 crc kubenswrapper[4764]: I0202 09:08:27.559871 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:27 crc kubenswrapper[4764]: I0202 09:08:27.559882 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:27Z","lastTransitionTime":"2026-02-02T09:08:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:27 crc kubenswrapper[4764]: I0202 09:08:27.662121 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:27 crc kubenswrapper[4764]: I0202 09:08:27.662170 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:27 crc kubenswrapper[4764]: I0202 09:08:27.662178 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:27 crc kubenswrapper[4764]: I0202 09:08:27.662193 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:27 crc kubenswrapper[4764]: I0202 09:08:27.662201 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:27Z","lastTransitionTime":"2026-02-02T09:08:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:27 crc kubenswrapper[4764]: I0202 09:08:27.764391 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:27 crc kubenswrapper[4764]: I0202 09:08:27.764429 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:27 crc kubenswrapper[4764]: I0202 09:08:27.764437 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:27 crc kubenswrapper[4764]: I0202 09:08:27.764452 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:27 crc kubenswrapper[4764]: I0202 09:08:27.764464 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:27Z","lastTransitionTime":"2026-02-02T09:08:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:27 crc kubenswrapper[4764]: I0202 09:08:27.815647 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-26 02:34:51.704461425 +0000 UTC Feb 02 09:08:27 crc kubenswrapper[4764]: I0202 09:08:27.825293 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kwtmr" Feb 02 09:08:27 crc kubenswrapper[4764]: E0202 09:08:27.825452 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kwtmr" podUID="991faa9a-dd25-4f49-82bd-ce60cefd4af2" Feb 02 09:08:27 crc kubenswrapper[4764]: I0202 09:08:27.836752 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f6904228e052fff69ad1fd66df97758b6b60ba4351e97a7fbd03eee4df1b653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:27Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:27 crc kubenswrapper[4764]: I0202 09:08:27.846764 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6zxff" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aaba93da-12a5-43df-b7c7-4c8b988163c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3db293678a907cd55b31b1d0c0bd7c30e1ba75ecd5e13cbc68ab98fbf0fbd9d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqwb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6zxff\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:27Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:27 crc kubenswrapper[4764]: I0202 09:08:27.859874 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d192f670-9f9d-4539-9641-e4bed73acdd4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://040c5e8a5659f85f0d52d1a1e064154d998ae104f711101d2204c0b1caebbb30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbksf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fdf06410349b7c99aa166e87fbe80bf68e2ae923635a6a978b7e2cd50383f7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbksf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4ndm4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:27Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:27 crc kubenswrapper[4764]: I0202 09:08:27.866186 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:27 crc kubenswrapper[4764]: I0202 09:08:27.866251 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:27 crc kubenswrapper[4764]: I0202 09:08:27.866262 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:27 crc kubenswrapper[4764]: I0202 09:08:27.866276 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:27 crc kubenswrapper[4764]: I0202 09:08:27.866288 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:27Z","lastTransitionTime":"2026-02-02T09:08:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:27 crc kubenswrapper[4764]: I0202 09:08:27.870348 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pcjmn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6e5ccd3b-b657-47a2-9c4d-bc9adb126533\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72863eef568d92a1abeb3cde2e89630a4f5fb7626c6949448ff5195774c98eab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2f7zl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pcjmn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:27Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:27 crc kubenswrapper[4764]: I0202 09:08:27.882358 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed747fd8-e3ce-4319-a6da-bda0fd51b85a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6f064bdc66470aace3180d4988244026b54009a63cff199e41c5d02a1d18479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://60258f84a1d8dcfe48b7e9688ed50f077e201971f20f3f6663fbf1ba58a4611a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f84be2e08217a2ac3edde858d78b4410376ed9bae3bcabe2a7f9b2fbe2f69c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a69d25d352d7ceac5c735bc1fcf446a1fea6dc22c65a5296cfab828cea17377\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:27Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:27 crc kubenswrapper[4764]: I0202 09:08:27.899320 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:27Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:27 crc kubenswrapper[4764]: I0202 09:08:27.919412 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lh6tm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2152e8a-f456-47b5-a547-b65ec837ecbc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38e176a00b07f2960fdd7ba2aadd1468a5e083b8d154e9d118d13f2585da81e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bb2sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://996e649011d5255783284421ddea8e0a5c383ef675b0920c3fdc0af6503d80e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bb2sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lh6tm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:27Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:27 crc kubenswrapper[4764]: I0202 09:08:27.931730 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fae745b-ea49-42f3-a486-dbc3dc2fb9ad\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://081d7f76d937ff71e35945f9b1d893888bb649d15bd84312fd4b61c9520ea71f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://429ead0993f5e648f6256d249ee12aea03514d3108618840552b672fb419cd74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fad17b01545eb20785f03e3c61bd5c982f6217e38176d7a3b23ac194298644a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85e2f75c56d63cacc93b9195c1e8bf46eb7a7234c45a024bd95a460db90edb6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85e2f75c56d63cacc93b9195c1e8bf46eb7a7234c45a024bd95a460db90edb6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:27Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:27 crc kubenswrapper[4764]: I0202 09:08:27.944705 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5528225d91ac2778bef4554abb0bd036ac5d3b6cdac7357bab69e70c3894b6f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://071bbb80d0747cf3e109ba626d47f7d05f85f4cc0b3aba77276c795741f47ac6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:27Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:27 crc kubenswrapper[4764]: I0202 09:08:27.958236 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:27Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:27 crc kubenswrapper[4764]: I0202 09:08:27.968607 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:27 crc kubenswrapper[4764]: I0202 09:08:27.968645 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:27 crc kubenswrapper[4764]: I0202 09:08:27.968656 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:27 crc kubenswrapper[4764]: I0202 09:08:27.968670 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:27 crc kubenswrapper[4764]: I0202 09:08:27.968681 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:27Z","lastTransitionTime":"2026-02-02T09:08:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:27 crc kubenswrapper[4764]: I0202 09:08:27.972313 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nc7qt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1a2c7b22068b7165927fc158885d48bf9952b9eba521d8ae9a599eee53dc305\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2hz4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nc7qt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:27Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:27 crc kubenswrapper[4764]: I0202 09:08:27.985913 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9p5dc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"660ac006-2495-45c4-9fb3-e9c1dddcf7a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da812b124c97aa8ad1563f4bcfac8f3bfa025096fa4a2561e5847b1bf9cb30b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23e57e65a6f5c3059506b9db21e9ac2492c30bd9191a98db0281915e9ce39797\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23e57e65a6f5c3059506b9db21e9ac2492c30bd9191a98db0281915e9ce39797\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2417b66816da1a6cd1791a4b91eca4a3c4f3694462e066e7f0722f1b9c2105c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2417b66816da1a6cd1791a4b91eca4a3c4f3694462e066e7f0722f1b9c2105c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://883641f29e9120c2cea6b272bacf1dc63c752f55ab0408cdaf73db85c85ac369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://883641f29e9120c2cea6b272bacf1dc63c752f55ab0408cdaf73db85c85ac369\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2745baba1188ba5e8f820616e4117a12c50ecee5b68e71ad9cabbb621b6e131f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2745baba1188ba5e8f820616e4117a12c50ecee5b68e71ad9cabbb621b6e131f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8925e435be4d48d86824fc35f8b8819fb35d63e385164c4a20edb3bf3035c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a8925e435be4d48d86824fc35f8b8819fb35d63e385164c4a20edb3bf3035c3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d3fbf668ae7d04c8d9e574f43c2fe90c5604204ba38a3c60847afdfbceb7a57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d3fbf668ae7d04c8d9e574f43c2fe90c5604204ba38a3c60847afdfbceb7a57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9p5dc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:27Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:27 crc kubenswrapper[4764]: I0202 09:08:27.998065 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kwtmr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"991faa9a-dd25-4f49-82bd-ce60cefd4af2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h865b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h865b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:53Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kwtmr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:27Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:28 crc kubenswrapper[4764]: I0202 09:08:28.010442 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87d61aea-46f6-4ee6-a7a4-767a6a67fea4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e78b6d08e4d616cefbb73e3fd43b3c6886dcd8aa997c7b4948bf1ff63401293\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d468715c2da33996561c40043216233fd12e21539d0c58dd1f3b892fd9f91a9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b184f7c936dc4046f74c0018269cbc2c4b7ceddb650b4e871d3f8a707e0df88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://269c5074371bd00f8363fb3888c11f1df383081a61497aea6578f08700881d50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b25bc9395831524c53f9e565b4efbe74f21792efafeb1dd507161bc91aa0ecdf\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T09:07:22Z\\\",\\\"message\\\":\\\"W0202 09:07:21.597488 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0202 09:07:21.598117 1 crypto.go:601] Generating new CA for check-endpoints-signer@1770023241 cert, and key in /tmp/serving-cert-3856183724/serving-signer.crt, /tmp/serving-cert-3856183724/serving-signer.key\\\\nI0202 09:07:22.012184 1 observer_polling.go:159] Starting file observer\\\\nW0202 09:07:22.015493 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0202 09:07:22.015797 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 09:07:22.016596 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3856183724/tls.crt::/tmp/serving-cert-3856183724/tls.key\\\\\\\"\\\\nF0202 09:07:22.270605 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b44e17b8ca01f549bc10081dd5f109cb1b9fff364bf5bd4fd90d9f0e3eeb04db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfb06a5ec9262cf39b108a66014d15e325cdbd5441a2d9e7a811dd57265c99a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfb06a5ec9262cf39b108a66014d15e325cdbd5441a2d9e7a811dd57265c99a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:28Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:28 crc kubenswrapper[4764]: I0202 09:08:28.028892 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa9d037e66cb53b9dd698eff46284c9e718fd842cda6f7b525f0c79062d70bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:28Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:28 crc kubenswrapper[4764]: I0202 09:08:28.042490 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:28Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:28 crc kubenswrapper[4764]: I0202 09:08:28.059644 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24632dda-6100-4ab6-a28e-214ddae4360c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a6c67e9b7708cc0d85b0b6ffbbade0d34867e159a9c5487ca2606475c667a50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bffda48d7c881c1470251b501661324859373246cc1a21811bd36ca24887c7aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a989645cf485cd8ff80ffff1a86cd7808e1564eba5667535e161cc68db13b26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20b71d3ee7a381baa51b0b0d0470283022cfede30f491135737b5e82d631d763\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c380383b6a014d1b694a13ddc22f54b20b9e4f9d5c9b9000d76187621dce0a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5529f2bff87753428e9173826fc1880acf96a4c134fa9ecad195b15697cad73a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://334e828f8ef72570f2511ee38659941202530471f5943aa07b8b21483bd70550\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://334e828f8ef72570f2511ee38659941202530471f5943aa07b8b21483bd70550\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-02T09:08:10Z\\\",\\\"message\\\":\\\"ing OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-dns/dns-default]} name:Service_openshift-dns/dns-default_TCP_node_router+switch_crc options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.10:53: 10.217.4.10:9154:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {be9dcc9e-c16a-4962-a6d2-4adeb0b929c4}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-dns/dns-default]} name:Service_openshift-dns/dns-default_UDP_node_router+switch_crc options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[udp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.10:53:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {4c1be812-05d3-4f45-91b5-a853a5c8de71}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF0202 09:08:09.911531 6327 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T09:08:08Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-zhn7j_openshift-ovn-kubernetes(24632dda-6100-4ab6-a28e-214ddae4360c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b4cfbd4e682bcdcaff785013bdb72580ce36fced733e235538536827036b93d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://579903eadb183466a2564737e6b678e04308239c0cd75b6808892053b97afbc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://579903eadb183466a2564737e6b678e04308239c0cd75b6808892053b97afbc0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zhn7j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:28Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:28 crc kubenswrapper[4764]: I0202 09:08:28.071265 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:28 crc kubenswrapper[4764]: I0202 09:08:28.071303 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:28 crc kubenswrapper[4764]: I0202 09:08:28.071312 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:28 crc kubenswrapper[4764]: I0202 09:08:28.071329 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:28 crc kubenswrapper[4764]: I0202 09:08:28.071338 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:28Z","lastTransitionTime":"2026-02-02T09:08:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:28 crc kubenswrapper[4764]: I0202 09:08:28.076541 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d66ab3c0-9b5b-41d5-b6c0-91ccddccd7e9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d92ac486f5dade46bc75de32099859a71fc2ed25a7abbe6ae8937df128e7df30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a4ac273a4f127d79e5fad7d4c30cd2044a346f5734dba5d33d68cd092c0c2d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d38c7cb5dcd22fb07524494ebe77aa1a7f4f808f566043563817e47e7ec12a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://408722c7f83d28919431532f8c797a1c795fd5d96736d14e7f06a05744899702\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f473b18cd8003f75bb1dafb82bcec45745a73ccd10434297356dac185a3d15c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e229f484d19cbdabedf53d90a822a4434db9e00b1301538e885e26acc19d25f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e229f484d19cbdabedf53d90a822a4434db9e00b1301538e885e26acc19d25f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a9cc673be8ff245dba81278e8cd34cfe45ed36ff0b8bd5b30a0ddf333ec6371\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a9cc673be8ff245dba81278e8cd34cfe45ed36ff0b8bd5b30a0ddf333ec6371\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bc7e49c7a5ae66ea939060192fa6a346dc0b0b9aa9750b29879d3c426cc0d276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc7e49c7a5ae66ea939060192fa6a346dc0b0b9aa9750b29879d3c426cc0d276\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:28Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:28 crc kubenswrapper[4764]: I0202 09:08:28.173713 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:28 crc kubenswrapper[4764]: I0202 09:08:28.173747 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:28 crc kubenswrapper[4764]: I0202 09:08:28.173755 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:28 crc kubenswrapper[4764]: I0202 09:08:28.173767 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:28 crc kubenswrapper[4764]: I0202 09:08:28.173775 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:28Z","lastTransitionTime":"2026-02-02T09:08:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:28 crc kubenswrapper[4764]: I0202 09:08:28.229273 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-nc7qt_6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e/kube-multus/0.log" Feb 02 09:08:28 crc kubenswrapper[4764]: I0202 09:08:28.229335 4764 generic.go:334] "Generic (PLEG): container finished" podID="6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e" containerID="b1a2c7b22068b7165927fc158885d48bf9952b9eba521d8ae9a599eee53dc305" exitCode=1 Feb 02 09:08:28 crc kubenswrapper[4764]: I0202 09:08:28.229370 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-nc7qt" event={"ID":"6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e","Type":"ContainerDied","Data":"b1a2c7b22068b7165927fc158885d48bf9952b9eba521d8ae9a599eee53dc305"} Feb 02 09:08:28 crc kubenswrapper[4764]: I0202 09:08:28.229811 4764 scope.go:117] "RemoveContainer" containerID="b1a2c7b22068b7165927fc158885d48bf9952b9eba521d8ae9a599eee53dc305" Feb 02 09:08:28 crc kubenswrapper[4764]: I0202 09:08:28.249903 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6zxff" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aaba93da-12a5-43df-b7c7-4c8b988163c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3db293678a907cd55b31b1d0c0bd7c30e1ba75ecd5e13cbc68ab98fbf0fbd9d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqwb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6zxff\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:28Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:28 crc kubenswrapper[4764]: I0202 09:08:28.262689 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d192f670-9f9d-4539-9641-e4bed73acdd4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://040c5e8a5659f85f0d52d1a1e064154d998ae104f711101d2204c0b1caebbb30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbksf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fdf06410349b7c99aa166e87fbe80bf68e2ae923635a6a978b7e2cd50383f7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbksf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4ndm4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:28Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:28 crc kubenswrapper[4764]: I0202 09:08:28.272587 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pcjmn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6e5ccd3b-b657-47a2-9c4d-bc9adb126533\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72863eef568d92a1abeb3cde2e89630a4f5fb7626c6949448ff5195774c98eab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2f7zl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pcjmn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:28Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:28 crc kubenswrapper[4764]: I0202 09:08:28.275899 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:28 crc kubenswrapper[4764]: I0202 09:08:28.275927 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:28 crc kubenswrapper[4764]: I0202 09:08:28.275952 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:28 crc kubenswrapper[4764]: I0202 09:08:28.275965 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:28 crc kubenswrapper[4764]: I0202 09:08:28.275973 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:28Z","lastTransitionTime":"2026-02-02T09:08:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:28 crc kubenswrapper[4764]: I0202 09:08:28.285902 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed747fd8-e3ce-4319-a6da-bda0fd51b85a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6f064bdc66470aace3180d4988244026b54009a63cff199e41c5d02a1d18479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://60258f84a1d8dcfe48b7e9688ed50f077e201971f20f3f6663fbf1ba58a4611a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f84be2e08217a2ac3edde858d78b4410376ed9bae3bcabe2a7f9b2fbe2f69c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a69d25d352d7ceac5c735bc1fcf446a1fea6dc22c65a5296cfab828cea17377\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:28Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:28 crc kubenswrapper[4764]: I0202 09:08:28.296701 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f6904228e052fff69ad1fd66df97758b6b60ba4351e97a7fbd03eee4df1b653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:28Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:28 crc kubenswrapper[4764]: I0202 09:08:28.307416 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lh6tm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2152e8a-f456-47b5-a547-b65ec837ecbc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38e176a00b07f2960fdd7ba2aadd1468a5e083b8d154e9d118d13f2585da81e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bb2sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://996e649011d5255783284421ddea8e0a5c383ef675b0920c3fdc0af6503d80e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bb2sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lh6tm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:28Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:28 crc kubenswrapper[4764]: I0202 09:08:28.319142 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fae745b-ea49-42f3-a486-dbc3dc2fb9ad\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://081d7f76d937ff71e35945f9b1d893888bb649d15bd84312fd4b61c9520ea71f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://429ead0993f5e648f6256d249ee12aea03514d3108618840552b672fb419cd74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fad17b01545eb20785f03e3c61bd5c982f6217e38176d7a3b23ac194298644a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85e2f75c56d63cacc93b9195c1e8bf46eb7a7234c45a024bd95a460db90edb6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85e2f75c56d63cacc93b9195c1e8bf46eb7a7234c45a024bd95a460db90edb6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:28Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:28 crc kubenswrapper[4764]: I0202 09:08:28.334243 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:28Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:28 crc kubenswrapper[4764]: I0202 09:08:28.346347 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5528225d91ac2778bef4554abb0bd036ac5d3b6cdac7357bab69e70c3894b6f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://071bbb80d0747cf3e109ba626d47f7d05f85f4cc0b3aba77276c795741f47ac6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:28Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:28 crc kubenswrapper[4764]: I0202 09:08:28.358094 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:28Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:28 crc kubenswrapper[4764]: I0202 09:08:28.369606 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nc7qt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1a2c7b22068b7165927fc158885d48bf9952b9eba521d8ae9a599eee53dc305\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1a2c7b22068b7165927fc158885d48bf9952b9eba521d8ae9a599eee53dc305\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-02T09:08:28Z\\\",\\\"message\\\":\\\"2026-02-02T09:07:42+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_32cbaf6a-865c-4787-8f89-a40249873b40\\\\n2026-02-02T09:07:42+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_32cbaf6a-865c-4787-8f89-a40249873b40 to /host/opt/cni/bin/\\\\n2026-02-02T09:07:43Z [verbose] multus-daemon started\\\\n2026-02-02T09:07:43Z [verbose] Readiness Indicator file check\\\\n2026-02-02T09:08:28Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2hz4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nc7qt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:28Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:28 crc kubenswrapper[4764]: I0202 09:08:28.378861 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:28 crc kubenswrapper[4764]: I0202 09:08:28.378900 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:28 crc kubenswrapper[4764]: I0202 09:08:28.378910 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:28 crc kubenswrapper[4764]: I0202 09:08:28.378927 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:28 crc kubenswrapper[4764]: I0202 09:08:28.378951 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:28Z","lastTransitionTime":"2026-02-02T09:08:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:28 crc kubenswrapper[4764]: I0202 09:08:28.383826 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9p5dc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"660ac006-2495-45c4-9fb3-e9c1dddcf7a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da812b124c97aa8ad1563f4bcfac8f3bfa025096fa4a2561e5847b1bf9cb30b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23e57e65a6f5c3059506b9db21e9ac2492c30bd9191a98db0281915e9ce39797\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23e57e65a6f5c3059506b9db21e9ac2492c30bd9191a98db0281915e9ce39797\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2417b66816da1a6cd1791a4b91eca4a3c4f3694462e066e7f0722f1b9c2105c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2417b66816da1a6cd1791a4b91eca4a3c4f3694462e066e7f0722f1b9c2105c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://883641f29e9120c2cea6b272bacf1dc63c752f55ab0408cdaf73db85c85ac369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://883641f29e9120c2cea6b272bacf1dc63c752f55ab0408cdaf73db85c85ac369\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2745baba1188ba5e8f820616e4117a12c50ecee5b68e71ad9cabbb621b6e131f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2745baba1188ba5e8f820616e4117a12c50ecee5b68e71ad9cabbb621b6e131f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8925e435be4d48d86824fc35f8b8819fb35d63e385164c4a20edb3bf3035c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a8925e435be4d48d86824fc35f8b8819fb35d63e385164c4a20edb3bf3035c3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d3fbf668ae7d04c8d9e574f43c2fe90c5604204ba38a3c60847afdfbceb7a57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d3fbf668ae7d04c8d9e574f43c2fe90c5604204ba38a3c60847afdfbceb7a57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9p5dc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:28Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:28 crc kubenswrapper[4764]: I0202 09:08:28.393854 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kwtmr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"991faa9a-dd25-4f49-82bd-ce60cefd4af2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h865b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h865b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:53Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kwtmr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:28Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:28 crc kubenswrapper[4764]: I0202 09:08:28.406214 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87d61aea-46f6-4ee6-a7a4-767a6a67fea4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e78b6d08e4d616cefbb73e3fd43b3c6886dcd8aa997c7b4948bf1ff63401293\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d468715c2da33996561c40043216233fd12e21539d0c58dd1f3b892fd9f91a9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b184f7c936dc4046f74c0018269cbc2c4b7ceddb650b4e871d3f8a707e0df88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://269c5074371bd00f8363fb3888c11f1df383081a61497aea6578f08700881d50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b25bc9395831524c53f9e565b4efbe74f21792efafeb1dd507161bc91aa0ecdf\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T09:07:22Z\\\",\\\"message\\\":\\\"W0202 09:07:21.597488 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0202 09:07:21.598117 1 crypto.go:601] Generating new CA for check-endpoints-signer@1770023241 cert, and key in /tmp/serving-cert-3856183724/serving-signer.crt, /tmp/serving-cert-3856183724/serving-signer.key\\\\nI0202 09:07:22.012184 1 observer_polling.go:159] Starting file observer\\\\nW0202 09:07:22.015493 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0202 09:07:22.015797 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 09:07:22.016596 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3856183724/tls.crt::/tmp/serving-cert-3856183724/tls.key\\\\\\\"\\\\nF0202 09:07:22.270605 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b44e17b8ca01f549bc10081dd5f109cb1b9fff364bf5bd4fd90d9f0e3eeb04db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfb06a5ec9262cf39b108a66014d15e325cdbd5441a2d9e7a811dd57265c99a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfb06a5ec9262cf39b108a66014d15e325cdbd5441a2d9e7a811dd57265c99a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:28Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:28 crc kubenswrapper[4764]: I0202 09:08:28.417778 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:28Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:28 crc kubenswrapper[4764]: I0202 09:08:28.434761 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24632dda-6100-4ab6-a28e-214ddae4360c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a6c67e9b7708cc0d85b0b6ffbbade0d34867e159a9c5487ca2606475c667a50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bffda48d7c881c1470251b501661324859373246cc1a21811bd36ca24887c7aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a989645cf485cd8ff80ffff1a86cd7808e1564eba5667535e161cc68db13b26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20b71d3ee7a381baa51b0b0d0470283022cfede30f491135737b5e82d631d763\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c380383b6a014d1b694a13ddc22f54b20b9e4f9d5c9b9000d76187621dce0a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5529f2bff87753428e9173826fc1880acf96a4c134fa9ecad195b15697cad73a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://334e828f8ef72570f2511ee38659941202530471f5943aa07b8b21483bd70550\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://334e828f8ef72570f2511ee38659941202530471f5943aa07b8b21483bd70550\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-02T09:08:10Z\\\",\\\"message\\\":\\\"ing OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-dns/dns-default]} name:Service_openshift-dns/dns-default_TCP_node_router+switch_crc options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.10:53: 10.217.4.10:9154:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {be9dcc9e-c16a-4962-a6d2-4adeb0b929c4}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-dns/dns-default]} name:Service_openshift-dns/dns-default_UDP_node_router+switch_crc options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[udp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.10:53:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {4c1be812-05d3-4f45-91b5-a853a5c8de71}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF0202 09:08:09.911531 6327 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T09:08:08Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-zhn7j_openshift-ovn-kubernetes(24632dda-6100-4ab6-a28e-214ddae4360c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b4cfbd4e682bcdcaff785013bdb72580ce36fced733e235538536827036b93d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://579903eadb183466a2564737e6b678e04308239c0cd75b6808892053b97afbc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://579903eadb183466a2564737e6b678e04308239c0cd75b6808892053b97afbc0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zhn7j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:28Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:28 crc kubenswrapper[4764]: I0202 09:08:28.452560 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d66ab3c0-9b5b-41d5-b6c0-91ccddccd7e9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d92ac486f5dade46bc75de32099859a71fc2ed25a7abbe6ae8937df128e7df30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a4ac273a4f127d79e5fad7d4c30cd2044a346f5734dba5d33d68cd092c0c2d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d38c7cb5dcd22fb07524494ebe77aa1a7f4f808f566043563817e47e7ec12a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://408722c7f83d28919431532f8c797a1c795fd5d96736d14e7f06a05744899702\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f473b18cd8003f75bb1dafb82bcec45745a73ccd10434297356dac185a3d15c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e229f484d19cbdabedf53d90a822a4434db9e00b1301538e885e26acc19d25f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e229f484d19cbdabedf53d90a822a4434db9e00b1301538e885e26acc19d25f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a9cc673be8ff245dba81278e8cd34cfe45ed36ff0b8bd5b30a0ddf333ec6371\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a9cc673be8ff245dba81278e8cd34cfe45ed36ff0b8bd5b30a0ddf333ec6371\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bc7e49c7a5ae66ea939060192fa6a346dc0b0b9aa9750b29879d3c426cc0d276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc7e49c7a5ae66ea939060192fa6a346dc0b0b9aa9750b29879d3c426cc0d276\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:28Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:28 crc kubenswrapper[4764]: I0202 09:08:28.465545 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa9d037e66cb53b9dd698eff46284c9e718fd842cda6f7b525f0c79062d70bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:28Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:28 crc kubenswrapper[4764]: I0202 09:08:28.481269 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:28 crc kubenswrapper[4764]: I0202 09:08:28.481309 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:28 crc kubenswrapper[4764]: I0202 09:08:28.481321 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:28 crc kubenswrapper[4764]: I0202 09:08:28.481338 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:28 crc kubenswrapper[4764]: I0202 09:08:28.481349 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:28Z","lastTransitionTime":"2026-02-02T09:08:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:28 crc kubenswrapper[4764]: I0202 09:08:28.583487 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:28 crc kubenswrapper[4764]: I0202 09:08:28.583519 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:28 crc kubenswrapper[4764]: I0202 09:08:28.583529 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:28 crc kubenswrapper[4764]: I0202 09:08:28.583543 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:28 crc kubenswrapper[4764]: I0202 09:08:28.583554 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:28Z","lastTransitionTime":"2026-02-02T09:08:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:28 crc kubenswrapper[4764]: I0202 09:08:28.685560 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:28 crc kubenswrapper[4764]: I0202 09:08:28.685600 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:28 crc kubenswrapper[4764]: I0202 09:08:28.685610 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:28 crc kubenswrapper[4764]: I0202 09:08:28.685625 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:28 crc kubenswrapper[4764]: I0202 09:08:28.685636 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:28Z","lastTransitionTime":"2026-02-02T09:08:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:28 crc kubenswrapper[4764]: I0202 09:08:28.787555 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:28 crc kubenswrapper[4764]: I0202 09:08:28.787598 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:28 crc kubenswrapper[4764]: I0202 09:08:28.787610 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:28 crc kubenswrapper[4764]: I0202 09:08:28.787626 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:28 crc kubenswrapper[4764]: I0202 09:08:28.787638 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:28Z","lastTransitionTime":"2026-02-02T09:08:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:28 crc kubenswrapper[4764]: I0202 09:08:28.815788 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-11 07:49:31.427568456 +0000 UTC Feb 02 09:08:28 crc kubenswrapper[4764]: I0202 09:08:28.825208 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 09:08:28 crc kubenswrapper[4764]: I0202 09:08:28.825216 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 09:08:28 crc kubenswrapper[4764]: E0202 09:08:28.825314 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 09:08:28 crc kubenswrapper[4764]: I0202 09:08:28.825208 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 09:08:28 crc kubenswrapper[4764]: E0202 09:08:28.825392 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 09:08:28 crc kubenswrapper[4764]: E0202 09:08:28.825450 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 09:08:28 crc kubenswrapper[4764]: I0202 09:08:28.889407 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:28 crc kubenswrapper[4764]: I0202 09:08:28.889445 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:28 crc kubenswrapper[4764]: I0202 09:08:28.889456 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:28 crc kubenswrapper[4764]: I0202 09:08:28.889471 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:28 crc kubenswrapper[4764]: I0202 09:08:28.889484 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:28Z","lastTransitionTime":"2026-02-02T09:08:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:28 crc kubenswrapper[4764]: I0202 09:08:28.991818 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:28 crc kubenswrapper[4764]: I0202 09:08:28.991841 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:28 crc kubenswrapper[4764]: I0202 09:08:28.991849 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:28 crc kubenswrapper[4764]: I0202 09:08:28.991861 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:28 crc kubenswrapper[4764]: I0202 09:08:28.991870 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:28Z","lastTransitionTime":"2026-02-02T09:08:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:29 crc kubenswrapper[4764]: I0202 09:08:29.094139 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:29 crc kubenswrapper[4764]: I0202 09:08:29.094170 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:29 crc kubenswrapper[4764]: I0202 09:08:29.094182 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:29 crc kubenswrapper[4764]: I0202 09:08:29.094197 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:29 crc kubenswrapper[4764]: I0202 09:08:29.094208 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:29Z","lastTransitionTime":"2026-02-02T09:08:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:29 crc kubenswrapper[4764]: I0202 09:08:29.198159 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:29 crc kubenswrapper[4764]: I0202 09:08:29.198196 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:29 crc kubenswrapper[4764]: I0202 09:08:29.198208 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:29 crc kubenswrapper[4764]: I0202 09:08:29.198229 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:29 crc kubenswrapper[4764]: I0202 09:08:29.198241 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:29Z","lastTransitionTime":"2026-02-02T09:08:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:29 crc kubenswrapper[4764]: I0202 09:08:29.232768 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-nc7qt_6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e/kube-multus/0.log" Feb 02 09:08:29 crc kubenswrapper[4764]: I0202 09:08:29.232853 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-nc7qt" event={"ID":"6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e","Type":"ContainerStarted","Data":"b52b499d17b4e2e359475bd13730c72ef5cedb052e13def6dfe5bc5fd7487d2d"} Feb 02 09:08:29 crc kubenswrapper[4764]: I0202 09:08:29.243872 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6zxff" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aaba93da-12a5-43df-b7c7-4c8b988163c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3db293678a907cd55b31b1d0c0bd7c30e1ba75ecd5e13cbc68ab98fbf0fbd9d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqwb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6zxff\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:29Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:29 crc kubenswrapper[4764]: I0202 09:08:29.255336 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d192f670-9f9d-4539-9641-e4bed73acdd4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://040c5e8a5659f85f0d52d1a1e064154d998ae104f711101d2204c0b1caebbb30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbksf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fdf06410349b7c99aa166e87fbe80bf68e2ae923635a6a978b7e2cd50383f7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbksf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4ndm4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:29Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:29 crc kubenswrapper[4764]: I0202 09:08:29.263973 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pcjmn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6e5ccd3b-b657-47a2-9c4d-bc9adb126533\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72863eef568d92a1abeb3cde2e89630a4f5fb7626c6949448ff5195774c98eab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2f7zl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pcjmn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:29Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:29 crc kubenswrapper[4764]: I0202 09:08:29.274902 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed747fd8-e3ce-4319-a6da-bda0fd51b85a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6f064bdc66470aace3180d4988244026b54009a63cff199e41c5d02a1d18479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://60258f84a1d8dcfe48b7e9688ed50f077e201971f20f3f6663fbf1ba58a4611a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f84be2e08217a2ac3edde858d78b4410376ed9bae3bcabe2a7f9b2fbe2f69c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a69d25d352d7ceac5c735bc1fcf446a1fea6dc22c65a5296cfab828cea17377\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:29Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:29 crc kubenswrapper[4764]: I0202 09:08:29.285849 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f6904228e052fff69ad1fd66df97758b6b60ba4351e97a7fbd03eee4df1b653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:29Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:29 crc kubenswrapper[4764]: I0202 09:08:29.295483 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lh6tm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2152e8a-f456-47b5-a547-b65ec837ecbc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38e176a00b07f2960fdd7ba2aadd1468a5e083b8d154e9d118d13f2585da81e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bb2sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://996e649011d5255783284421ddea8e0a5c383ef675b0920c3fdc0af6503d80e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bb2sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lh6tm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:29Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:29 crc kubenswrapper[4764]: I0202 09:08:29.300537 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:29 crc kubenswrapper[4764]: I0202 09:08:29.300575 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:29 crc kubenswrapper[4764]: I0202 09:08:29.300584 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:29 crc kubenswrapper[4764]: I0202 09:08:29.300602 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:29 crc kubenswrapper[4764]: I0202 09:08:29.300612 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:29Z","lastTransitionTime":"2026-02-02T09:08:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:29 crc kubenswrapper[4764]: I0202 09:08:29.308252 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fae745b-ea49-42f3-a486-dbc3dc2fb9ad\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://081d7f76d937ff71e35945f9b1d893888bb649d15bd84312fd4b61c9520ea71f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://429ead0993f5e648f6256d249ee12aea03514d3108618840552b672fb419cd74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fad17b01545eb20785f03e3c61bd5c982f6217e38176d7a3b23ac194298644a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85e2f75c56d63cacc93b9195c1e8bf46eb7a7234c45a024bd95a460db90edb6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85e2f75c56d63cacc93b9195c1e8bf46eb7a7234c45a024bd95a460db90edb6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:29Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:29 crc kubenswrapper[4764]: I0202 09:08:29.318971 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:29Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:29 crc kubenswrapper[4764]: I0202 09:08:29.329620 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5528225d91ac2778bef4554abb0bd036ac5d3b6cdac7357bab69e70c3894b6f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://071bbb80d0747cf3e109ba626d47f7d05f85f4cc0b3aba77276c795741f47ac6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:29Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:29 crc kubenswrapper[4764]: I0202 09:08:29.344250 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:29Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:29 crc kubenswrapper[4764]: I0202 09:08:29.356777 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nc7qt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b52b499d17b4e2e359475bd13730c72ef5cedb052e13def6dfe5bc5fd7487d2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1a2c7b22068b7165927fc158885d48bf9952b9eba521d8ae9a599eee53dc305\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-02T09:08:28Z\\\",\\\"message\\\":\\\"2026-02-02T09:07:42+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_32cbaf6a-865c-4787-8f89-a40249873b40\\\\n2026-02-02T09:07:42+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_32cbaf6a-865c-4787-8f89-a40249873b40 to /host/opt/cni/bin/\\\\n2026-02-02T09:07:43Z [verbose] multus-daemon started\\\\n2026-02-02T09:07:43Z [verbose] Readiness Indicator file check\\\\n2026-02-02T09:08:28Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:08:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2hz4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nc7qt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:29Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:29 crc kubenswrapper[4764]: I0202 09:08:29.370280 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9p5dc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"660ac006-2495-45c4-9fb3-e9c1dddcf7a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da812b124c97aa8ad1563f4bcfac8f3bfa025096fa4a2561e5847b1bf9cb30b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23e57e65a6f5c3059506b9db21e9ac2492c30bd9191a98db0281915e9ce39797\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23e57e65a6f5c3059506b9db21e9ac2492c30bd9191a98db0281915e9ce39797\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2417b66816da1a6cd1791a4b91eca4a3c4f3694462e066e7f0722f1b9c2105c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2417b66816da1a6cd1791a4b91eca4a3c4f3694462e066e7f0722f1b9c2105c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://883641f29e9120c2cea6b272bacf1dc63c752f55ab0408cdaf73db85c85ac369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://883641f29e9120c2cea6b272bacf1dc63c752f55ab0408cdaf73db85c85ac369\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2745baba1188ba5e8f820616e4117a12c50ecee5b68e71ad9cabbb621b6e131f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2745baba1188ba5e8f820616e4117a12c50ecee5b68e71ad9cabbb621b6e131f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8925e435be4d48d86824fc35f8b8819fb35d63e385164c4a20edb3bf3035c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a8925e435be4d48d86824fc35f8b8819fb35d63e385164c4a20edb3bf3035c3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d3fbf668ae7d04c8d9e574f43c2fe90c5604204ba38a3c60847afdfbceb7a57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d3fbf668ae7d04c8d9e574f43c2fe90c5604204ba38a3c60847afdfbceb7a57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9p5dc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:29Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:29 crc kubenswrapper[4764]: I0202 09:08:29.381087 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kwtmr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"991faa9a-dd25-4f49-82bd-ce60cefd4af2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h865b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h865b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:53Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kwtmr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:29Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:29 crc kubenswrapper[4764]: I0202 09:08:29.397316 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87d61aea-46f6-4ee6-a7a4-767a6a67fea4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e78b6d08e4d616cefbb73e3fd43b3c6886dcd8aa997c7b4948bf1ff63401293\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d468715c2da33996561c40043216233fd12e21539d0c58dd1f3b892fd9f91a9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b184f7c936dc4046f74c0018269cbc2c4b7ceddb650b4e871d3f8a707e0df88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://269c5074371bd00f8363fb3888c11f1df383081a61497aea6578f08700881d50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b25bc9395831524c53f9e565b4efbe74f21792efafeb1dd507161bc91aa0ecdf\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T09:07:22Z\\\",\\\"message\\\":\\\"W0202 09:07:21.597488 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0202 09:07:21.598117 1 crypto.go:601] Generating new CA for check-endpoints-signer@1770023241 cert, and key in /tmp/serving-cert-3856183724/serving-signer.crt, /tmp/serving-cert-3856183724/serving-signer.key\\\\nI0202 09:07:22.012184 1 observer_polling.go:159] Starting file observer\\\\nW0202 09:07:22.015493 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0202 09:07:22.015797 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 09:07:22.016596 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3856183724/tls.crt::/tmp/serving-cert-3856183724/tls.key\\\\\\\"\\\\nF0202 09:07:22.270605 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b44e17b8ca01f549bc10081dd5f109cb1b9fff364bf5bd4fd90d9f0e3eeb04db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfb06a5ec9262cf39b108a66014d15e325cdbd5441a2d9e7a811dd57265c99a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfb06a5ec9262cf39b108a66014d15e325cdbd5441a2d9e7a811dd57265c99a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:29Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:29 crc kubenswrapper[4764]: I0202 09:08:29.403599 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:29 crc kubenswrapper[4764]: I0202 09:08:29.403635 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:29 crc kubenswrapper[4764]: I0202 09:08:29.403644 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:29 crc kubenswrapper[4764]: I0202 09:08:29.403657 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:29 crc kubenswrapper[4764]: I0202 09:08:29.403667 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:29Z","lastTransitionTime":"2026-02-02T09:08:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:29 crc kubenswrapper[4764]: I0202 09:08:29.411474 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:29Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:29 crc kubenswrapper[4764]: I0202 09:08:29.433343 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24632dda-6100-4ab6-a28e-214ddae4360c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a6c67e9b7708cc0d85b0b6ffbbade0d34867e159a9c5487ca2606475c667a50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bffda48d7c881c1470251b501661324859373246cc1a21811bd36ca24887c7aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a989645cf485cd8ff80ffff1a86cd7808e1564eba5667535e161cc68db13b26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20b71d3ee7a381baa51b0b0d0470283022cfede30f491135737b5e82d631d763\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c380383b6a014d1b694a13ddc22f54b20b9e4f9d5c9b9000d76187621dce0a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5529f2bff87753428e9173826fc1880acf96a4c134fa9ecad195b15697cad73a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://334e828f8ef72570f2511ee38659941202530471f5943aa07b8b21483bd70550\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://334e828f8ef72570f2511ee38659941202530471f5943aa07b8b21483bd70550\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-02T09:08:10Z\\\",\\\"message\\\":\\\"ing OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-dns/dns-default]} name:Service_openshift-dns/dns-default_TCP_node_router+switch_crc options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.10:53: 10.217.4.10:9154:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {be9dcc9e-c16a-4962-a6d2-4adeb0b929c4}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-dns/dns-default]} name:Service_openshift-dns/dns-default_UDP_node_router+switch_crc options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[udp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.10:53:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {4c1be812-05d3-4f45-91b5-a853a5c8de71}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF0202 09:08:09.911531 6327 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T09:08:08Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-zhn7j_openshift-ovn-kubernetes(24632dda-6100-4ab6-a28e-214ddae4360c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b4cfbd4e682bcdcaff785013bdb72580ce36fced733e235538536827036b93d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://579903eadb183466a2564737e6b678e04308239c0cd75b6808892053b97afbc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://579903eadb183466a2564737e6b678e04308239c0cd75b6808892053b97afbc0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zhn7j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:29Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:29 crc kubenswrapper[4764]: I0202 09:08:29.452447 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d66ab3c0-9b5b-41d5-b6c0-91ccddccd7e9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d92ac486f5dade46bc75de32099859a71fc2ed25a7abbe6ae8937df128e7df30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a4ac273a4f127d79e5fad7d4c30cd2044a346f5734dba5d33d68cd092c0c2d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d38c7cb5dcd22fb07524494ebe77aa1a7f4f808f566043563817e47e7ec12a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://408722c7f83d28919431532f8c797a1c795fd5d96736d14e7f06a05744899702\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f473b18cd8003f75bb1dafb82bcec45745a73ccd10434297356dac185a3d15c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e229f484d19cbdabedf53d90a822a4434db9e00b1301538e885e26acc19d25f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e229f484d19cbdabedf53d90a822a4434db9e00b1301538e885e26acc19d25f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a9cc673be8ff245dba81278e8cd34cfe45ed36ff0b8bd5b30a0ddf333ec6371\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a9cc673be8ff245dba81278e8cd34cfe45ed36ff0b8bd5b30a0ddf333ec6371\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bc7e49c7a5ae66ea939060192fa6a346dc0b0b9aa9750b29879d3c426cc0d276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc7e49c7a5ae66ea939060192fa6a346dc0b0b9aa9750b29879d3c426cc0d276\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:29Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:29 crc kubenswrapper[4764]: I0202 09:08:29.465209 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa9d037e66cb53b9dd698eff46284c9e718fd842cda6f7b525f0c79062d70bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:29Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:29 crc kubenswrapper[4764]: I0202 09:08:29.506347 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:29 crc kubenswrapper[4764]: I0202 09:08:29.506380 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:29 crc kubenswrapper[4764]: I0202 09:08:29.506388 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:29 crc kubenswrapper[4764]: I0202 09:08:29.506402 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:29 crc kubenswrapper[4764]: I0202 09:08:29.506414 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:29Z","lastTransitionTime":"2026-02-02T09:08:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:29 crc kubenswrapper[4764]: I0202 09:08:29.608883 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:29 crc kubenswrapper[4764]: I0202 09:08:29.608990 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:29 crc kubenswrapper[4764]: I0202 09:08:29.609011 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:29 crc kubenswrapper[4764]: I0202 09:08:29.609035 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:29 crc kubenswrapper[4764]: I0202 09:08:29.609051 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:29Z","lastTransitionTime":"2026-02-02T09:08:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:29 crc kubenswrapper[4764]: I0202 09:08:29.710904 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:29 crc kubenswrapper[4764]: I0202 09:08:29.710967 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:29 crc kubenswrapper[4764]: I0202 09:08:29.710977 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:29 crc kubenswrapper[4764]: I0202 09:08:29.711019 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:29 crc kubenswrapper[4764]: I0202 09:08:29.711030 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:29Z","lastTransitionTime":"2026-02-02T09:08:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:29 crc kubenswrapper[4764]: I0202 09:08:29.813607 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:29 crc kubenswrapper[4764]: I0202 09:08:29.813638 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:29 crc kubenswrapper[4764]: I0202 09:08:29.813646 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:29 crc kubenswrapper[4764]: I0202 09:08:29.813660 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:29 crc kubenswrapper[4764]: I0202 09:08:29.813669 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:29Z","lastTransitionTime":"2026-02-02T09:08:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:29 crc kubenswrapper[4764]: I0202 09:08:29.816163 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-03 00:40:35.083180563 +0000 UTC Feb 02 09:08:29 crc kubenswrapper[4764]: I0202 09:08:29.825560 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kwtmr" Feb 02 09:08:29 crc kubenswrapper[4764]: E0202 09:08:29.825689 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kwtmr" podUID="991faa9a-dd25-4f49-82bd-ce60cefd4af2" Feb 02 09:08:29 crc kubenswrapper[4764]: I0202 09:08:29.916049 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:29 crc kubenswrapper[4764]: I0202 09:08:29.916084 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:29 crc kubenswrapper[4764]: I0202 09:08:29.916093 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:29 crc kubenswrapper[4764]: I0202 09:08:29.916106 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:29 crc kubenswrapper[4764]: I0202 09:08:29.916115 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:29Z","lastTransitionTime":"2026-02-02T09:08:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:30 crc kubenswrapper[4764]: I0202 09:08:30.018535 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:30 crc kubenswrapper[4764]: I0202 09:08:30.018572 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:30 crc kubenswrapper[4764]: I0202 09:08:30.018580 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:30 crc kubenswrapper[4764]: I0202 09:08:30.018596 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:30 crc kubenswrapper[4764]: I0202 09:08:30.018609 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:30Z","lastTransitionTime":"2026-02-02T09:08:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:30 crc kubenswrapper[4764]: I0202 09:08:30.121427 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:30 crc kubenswrapper[4764]: I0202 09:08:30.121476 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:30 crc kubenswrapper[4764]: I0202 09:08:30.121487 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:30 crc kubenswrapper[4764]: I0202 09:08:30.121501 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:30 crc kubenswrapper[4764]: I0202 09:08:30.121514 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:30Z","lastTransitionTime":"2026-02-02T09:08:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:30 crc kubenswrapper[4764]: I0202 09:08:30.224296 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:30 crc kubenswrapper[4764]: I0202 09:08:30.224352 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:30 crc kubenswrapper[4764]: I0202 09:08:30.224370 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:30 crc kubenswrapper[4764]: I0202 09:08:30.224388 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:30 crc kubenswrapper[4764]: I0202 09:08:30.224400 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:30Z","lastTransitionTime":"2026-02-02T09:08:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:30 crc kubenswrapper[4764]: I0202 09:08:30.327501 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:30 crc kubenswrapper[4764]: I0202 09:08:30.327561 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:30 crc kubenswrapper[4764]: I0202 09:08:30.327573 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:30 crc kubenswrapper[4764]: I0202 09:08:30.327597 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:30 crc kubenswrapper[4764]: I0202 09:08:30.327609 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:30Z","lastTransitionTime":"2026-02-02T09:08:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:30 crc kubenswrapper[4764]: I0202 09:08:30.429925 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:30 crc kubenswrapper[4764]: I0202 09:08:30.430019 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:30 crc kubenswrapper[4764]: I0202 09:08:30.430036 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:30 crc kubenswrapper[4764]: I0202 09:08:30.430058 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:30 crc kubenswrapper[4764]: I0202 09:08:30.430075 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:30Z","lastTransitionTime":"2026-02-02T09:08:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:30 crc kubenswrapper[4764]: I0202 09:08:30.531966 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:30 crc kubenswrapper[4764]: I0202 09:08:30.532004 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:30 crc kubenswrapper[4764]: I0202 09:08:30.532014 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:30 crc kubenswrapper[4764]: I0202 09:08:30.532028 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:30 crc kubenswrapper[4764]: I0202 09:08:30.532039 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:30Z","lastTransitionTime":"2026-02-02T09:08:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:30 crc kubenswrapper[4764]: I0202 09:08:30.635499 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:30 crc kubenswrapper[4764]: I0202 09:08:30.635810 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:30 crc kubenswrapper[4764]: I0202 09:08:30.635907 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:30 crc kubenswrapper[4764]: I0202 09:08:30.635960 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:30 crc kubenswrapper[4764]: I0202 09:08:30.635979 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:30Z","lastTransitionTime":"2026-02-02T09:08:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:30 crc kubenswrapper[4764]: I0202 09:08:30.739262 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:30 crc kubenswrapper[4764]: I0202 09:08:30.739305 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:30 crc kubenswrapper[4764]: I0202 09:08:30.739313 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:30 crc kubenswrapper[4764]: I0202 09:08:30.739329 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:30 crc kubenswrapper[4764]: I0202 09:08:30.739338 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:30Z","lastTransitionTime":"2026-02-02T09:08:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:30 crc kubenswrapper[4764]: I0202 09:08:30.817170 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-31 20:20:23.146363371 +0000 UTC Feb 02 09:08:30 crc kubenswrapper[4764]: I0202 09:08:30.824688 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 09:08:30 crc kubenswrapper[4764]: E0202 09:08:30.824798 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 09:08:30 crc kubenswrapper[4764]: I0202 09:08:30.824824 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 09:08:30 crc kubenswrapper[4764]: E0202 09:08:30.824924 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 09:08:30 crc kubenswrapper[4764]: I0202 09:08:30.824993 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 09:08:30 crc kubenswrapper[4764]: E0202 09:08:30.825038 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 09:08:30 crc kubenswrapper[4764]: I0202 09:08:30.843088 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Feb 02 09:08:30 crc kubenswrapper[4764]: I0202 09:08:30.844514 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:30 crc kubenswrapper[4764]: I0202 09:08:30.844570 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:30 crc kubenswrapper[4764]: I0202 09:08:30.844582 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:30 crc kubenswrapper[4764]: I0202 09:08:30.844599 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:30 crc kubenswrapper[4764]: I0202 09:08:30.844612 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:30Z","lastTransitionTime":"2026-02-02T09:08:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:30 crc kubenswrapper[4764]: I0202 09:08:30.946270 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:30 crc kubenswrapper[4764]: I0202 09:08:30.946304 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:30 crc kubenswrapper[4764]: I0202 09:08:30.946312 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:30 crc kubenswrapper[4764]: I0202 09:08:30.946326 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:30 crc kubenswrapper[4764]: I0202 09:08:30.946337 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:30Z","lastTransitionTime":"2026-02-02T09:08:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:31 crc kubenswrapper[4764]: I0202 09:08:31.048411 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:31 crc kubenswrapper[4764]: I0202 09:08:31.048440 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:31 crc kubenswrapper[4764]: I0202 09:08:31.048449 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:31 crc kubenswrapper[4764]: I0202 09:08:31.048462 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:31 crc kubenswrapper[4764]: I0202 09:08:31.048470 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:31Z","lastTransitionTime":"2026-02-02T09:08:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:31 crc kubenswrapper[4764]: I0202 09:08:31.150612 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:31 crc kubenswrapper[4764]: I0202 09:08:31.150649 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:31 crc kubenswrapper[4764]: I0202 09:08:31.150663 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:31 crc kubenswrapper[4764]: I0202 09:08:31.150711 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:31 crc kubenswrapper[4764]: I0202 09:08:31.150732 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:31Z","lastTransitionTime":"2026-02-02T09:08:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:31 crc kubenswrapper[4764]: I0202 09:08:31.253268 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:31 crc kubenswrapper[4764]: I0202 09:08:31.253319 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:31 crc kubenswrapper[4764]: I0202 09:08:31.253329 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:31 crc kubenswrapper[4764]: I0202 09:08:31.253346 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:31 crc kubenswrapper[4764]: I0202 09:08:31.253358 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:31Z","lastTransitionTime":"2026-02-02T09:08:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:31 crc kubenswrapper[4764]: I0202 09:08:31.355702 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:31 crc kubenswrapper[4764]: I0202 09:08:31.355762 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:31 crc kubenswrapper[4764]: I0202 09:08:31.355779 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:31 crc kubenswrapper[4764]: I0202 09:08:31.355803 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:31 crc kubenswrapper[4764]: I0202 09:08:31.355820 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:31Z","lastTransitionTime":"2026-02-02T09:08:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:31 crc kubenswrapper[4764]: I0202 09:08:31.458191 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:31 crc kubenswrapper[4764]: I0202 09:08:31.458228 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:31 crc kubenswrapper[4764]: I0202 09:08:31.458241 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:31 crc kubenswrapper[4764]: I0202 09:08:31.458258 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:31 crc kubenswrapper[4764]: I0202 09:08:31.458270 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:31Z","lastTransitionTime":"2026-02-02T09:08:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:31 crc kubenswrapper[4764]: I0202 09:08:31.560835 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:31 crc kubenswrapper[4764]: I0202 09:08:31.561101 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:31 crc kubenswrapper[4764]: I0202 09:08:31.561228 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:31 crc kubenswrapper[4764]: I0202 09:08:31.561377 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:31 crc kubenswrapper[4764]: I0202 09:08:31.561509 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:31Z","lastTransitionTime":"2026-02-02T09:08:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:31 crc kubenswrapper[4764]: I0202 09:08:31.663846 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:31 crc kubenswrapper[4764]: I0202 09:08:31.663876 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:31 crc kubenswrapper[4764]: I0202 09:08:31.663884 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:31 crc kubenswrapper[4764]: I0202 09:08:31.663897 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:31 crc kubenswrapper[4764]: I0202 09:08:31.663908 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:31Z","lastTransitionTime":"2026-02-02T09:08:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:31 crc kubenswrapper[4764]: I0202 09:08:31.766244 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:31 crc kubenswrapper[4764]: I0202 09:08:31.766277 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:31 crc kubenswrapper[4764]: I0202 09:08:31.766286 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:31 crc kubenswrapper[4764]: I0202 09:08:31.766301 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:31 crc kubenswrapper[4764]: I0202 09:08:31.766311 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:31Z","lastTransitionTime":"2026-02-02T09:08:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:31 crc kubenswrapper[4764]: I0202 09:08:31.817948 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-17 04:19:29.953459986 +0000 UTC Feb 02 09:08:31 crc kubenswrapper[4764]: I0202 09:08:31.827158 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kwtmr" Feb 02 09:08:31 crc kubenswrapper[4764]: E0202 09:08:31.827296 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kwtmr" podUID="991faa9a-dd25-4f49-82bd-ce60cefd4af2" Feb 02 09:08:31 crc kubenswrapper[4764]: I0202 09:08:31.868209 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:31 crc kubenswrapper[4764]: I0202 09:08:31.868247 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:31 crc kubenswrapper[4764]: I0202 09:08:31.868256 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:31 crc kubenswrapper[4764]: I0202 09:08:31.868271 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:31 crc kubenswrapper[4764]: I0202 09:08:31.868282 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:31Z","lastTransitionTime":"2026-02-02T09:08:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:31 crc kubenswrapper[4764]: I0202 09:08:31.970336 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:31 crc kubenswrapper[4764]: I0202 09:08:31.970378 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:31 crc kubenswrapper[4764]: I0202 09:08:31.970390 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:31 crc kubenswrapper[4764]: I0202 09:08:31.970408 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:31 crc kubenswrapper[4764]: I0202 09:08:31.970420 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:31Z","lastTransitionTime":"2026-02-02T09:08:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:32 crc kubenswrapper[4764]: I0202 09:08:32.072288 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:32 crc kubenswrapper[4764]: I0202 09:08:32.072329 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:32 crc kubenswrapper[4764]: I0202 09:08:32.072337 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:32 crc kubenswrapper[4764]: I0202 09:08:32.072350 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:32 crc kubenswrapper[4764]: I0202 09:08:32.072359 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:32Z","lastTransitionTime":"2026-02-02T09:08:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:32 crc kubenswrapper[4764]: I0202 09:08:32.174567 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:32 crc kubenswrapper[4764]: I0202 09:08:32.174599 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:32 crc kubenswrapper[4764]: I0202 09:08:32.174610 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:32 crc kubenswrapper[4764]: I0202 09:08:32.174626 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:32 crc kubenswrapper[4764]: I0202 09:08:32.174635 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:32Z","lastTransitionTime":"2026-02-02T09:08:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:32 crc kubenswrapper[4764]: I0202 09:08:32.276507 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:32 crc kubenswrapper[4764]: I0202 09:08:32.276545 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:32 crc kubenswrapper[4764]: I0202 09:08:32.276571 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:32 crc kubenswrapper[4764]: I0202 09:08:32.276588 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:32 crc kubenswrapper[4764]: I0202 09:08:32.276600 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:32Z","lastTransitionTime":"2026-02-02T09:08:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:32 crc kubenswrapper[4764]: I0202 09:08:32.378541 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:32 crc kubenswrapper[4764]: I0202 09:08:32.378634 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:32 crc kubenswrapper[4764]: I0202 09:08:32.378658 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:32 crc kubenswrapper[4764]: I0202 09:08:32.378685 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:32 crc kubenswrapper[4764]: I0202 09:08:32.378705 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:32Z","lastTransitionTime":"2026-02-02T09:08:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:32 crc kubenswrapper[4764]: I0202 09:08:32.480752 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:32 crc kubenswrapper[4764]: I0202 09:08:32.480806 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:32 crc kubenswrapper[4764]: I0202 09:08:32.480823 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:32 crc kubenswrapper[4764]: I0202 09:08:32.480844 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:32 crc kubenswrapper[4764]: I0202 09:08:32.480861 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:32Z","lastTransitionTime":"2026-02-02T09:08:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:32 crc kubenswrapper[4764]: I0202 09:08:32.583516 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:32 crc kubenswrapper[4764]: I0202 09:08:32.583569 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:32 crc kubenswrapper[4764]: I0202 09:08:32.583580 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:32 crc kubenswrapper[4764]: I0202 09:08:32.583596 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:32 crc kubenswrapper[4764]: I0202 09:08:32.583607 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:32Z","lastTransitionTime":"2026-02-02T09:08:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:32 crc kubenswrapper[4764]: I0202 09:08:32.685638 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:32 crc kubenswrapper[4764]: I0202 09:08:32.685672 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:32 crc kubenswrapper[4764]: I0202 09:08:32.685684 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:32 crc kubenswrapper[4764]: I0202 09:08:32.685701 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:32 crc kubenswrapper[4764]: I0202 09:08:32.685714 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:32Z","lastTransitionTime":"2026-02-02T09:08:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:32 crc kubenswrapper[4764]: I0202 09:08:32.789027 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:32 crc kubenswrapper[4764]: I0202 09:08:32.789082 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:32 crc kubenswrapper[4764]: I0202 09:08:32.789097 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:32 crc kubenswrapper[4764]: I0202 09:08:32.789119 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:32 crc kubenswrapper[4764]: I0202 09:08:32.789137 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:32Z","lastTransitionTime":"2026-02-02T09:08:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:32 crc kubenswrapper[4764]: I0202 09:08:32.818311 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-21 21:42:11.998248532 +0000 UTC Feb 02 09:08:32 crc kubenswrapper[4764]: I0202 09:08:32.825298 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 09:08:32 crc kubenswrapper[4764]: I0202 09:08:32.825359 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 09:08:32 crc kubenswrapper[4764]: I0202 09:08:32.825368 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 09:08:32 crc kubenswrapper[4764]: E0202 09:08:32.825437 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 09:08:32 crc kubenswrapper[4764]: E0202 09:08:32.825578 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 09:08:32 crc kubenswrapper[4764]: E0202 09:08:32.825717 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 09:08:32 crc kubenswrapper[4764]: I0202 09:08:32.891428 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:32 crc kubenswrapper[4764]: I0202 09:08:32.891471 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:32 crc kubenswrapper[4764]: I0202 09:08:32.891481 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:32 crc kubenswrapper[4764]: I0202 09:08:32.891494 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:32 crc kubenswrapper[4764]: I0202 09:08:32.891503 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:32Z","lastTransitionTime":"2026-02-02T09:08:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:32 crc kubenswrapper[4764]: I0202 09:08:32.993838 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:32 crc kubenswrapper[4764]: I0202 09:08:32.993879 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:32 crc kubenswrapper[4764]: I0202 09:08:32.993888 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:32 crc kubenswrapper[4764]: I0202 09:08:32.993903 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:32 crc kubenswrapper[4764]: I0202 09:08:32.993913 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:32Z","lastTransitionTime":"2026-02-02T09:08:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:33 crc kubenswrapper[4764]: I0202 09:08:33.097730 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:33 crc kubenswrapper[4764]: I0202 09:08:33.097779 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:33 crc kubenswrapper[4764]: I0202 09:08:33.097793 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:33 crc kubenswrapper[4764]: I0202 09:08:33.097823 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:33 crc kubenswrapper[4764]: I0202 09:08:33.097834 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:33Z","lastTransitionTime":"2026-02-02T09:08:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:33 crc kubenswrapper[4764]: I0202 09:08:33.200599 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:33 crc kubenswrapper[4764]: I0202 09:08:33.200643 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:33 crc kubenswrapper[4764]: I0202 09:08:33.200653 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:33 crc kubenswrapper[4764]: I0202 09:08:33.200668 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:33 crc kubenswrapper[4764]: I0202 09:08:33.200677 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:33Z","lastTransitionTime":"2026-02-02T09:08:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:33 crc kubenswrapper[4764]: I0202 09:08:33.304335 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:33 crc kubenswrapper[4764]: I0202 09:08:33.304390 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:33 crc kubenswrapper[4764]: I0202 09:08:33.304399 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:33 crc kubenswrapper[4764]: I0202 09:08:33.304413 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:33 crc kubenswrapper[4764]: I0202 09:08:33.304424 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:33Z","lastTransitionTime":"2026-02-02T09:08:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:33 crc kubenswrapper[4764]: I0202 09:08:33.408018 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:33 crc kubenswrapper[4764]: I0202 09:08:33.408078 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:33 crc kubenswrapper[4764]: I0202 09:08:33.408095 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:33 crc kubenswrapper[4764]: I0202 09:08:33.408118 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:33 crc kubenswrapper[4764]: I0202 09:08:33.408135 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:33Z","lastTransitionTime":"2026-02-02T09:08:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:33 crc kubenswrapper[4764]: I0202 09:08:33.511805 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:33 crc kubenswrapper[4764]: I0202 09:08:33.511882 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:33 crc kubenswrapper[4764]: I0202 09:08:33.512144 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:33 crc kubenswrapper[4764]: I0202 09:08:33.512181 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:33 crc kubenswrapper[4764]: I0202 09:08:33.512199 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:33Z","lastTransitionTime":"2026-02-02T09:08:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:33 crc kubenswrapper[4764]: I0202 09:08:33.615568 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:33 crc kubenswrapper[4764]: I0202 09:08:33.615603 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:33 crc kubenswrapper[4764]: I0202 09:08:33.615615 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:33 crc kubenswrapper[4764]: I0202 09:08:33.615630 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:33 crc kubenswrapper[4764]: I0202 09:08:33.615640 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:33Z","lastTransitionTime":"2026-02-02T09:08:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:33 crc kubenswrapper[4764]: I0202 09:08:33.718266 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:33 crc kubenswrapper[4764]: I0202 09:08:33.718498 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:33 crc kubenswrapper[4764]: I0202 09:08:33.718584 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:33 crc kubenswrapper[4764]: I0202 09:08:33.718664 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:33 crc kubenswrapper[4764]: I0202 09:08:33.718750 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:33Z","lastTransitionTime":"2026-02-02T09:08:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:33 crc kubenswrapper[4764]: I0202 09:08:33.819322 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-26 00:30:57.232376681 +0000 UTC Feb 02 09:08:33 crc kubenswrapper[4764]: I0202 09:08:33.821184 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:33 crc kubenswrapper[4764]: I0202 09:08:33.821220 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:33 crc kubenswrapper[4764]: I0202 09:08:33.821232 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:33 crc kubenswrapper[4764]: I0202 09:08:33.821252 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:33 crc kubenswrapper[4764]: I0202 09:08:33.821263 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:33Z","lastTransitionTime":"2026-02-02T09:08:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:33 crc kubenswrapper[4764]: I0202 09:08:33.825564 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kwtmr" Feb 02 09:08:33 crc kubenswrapper[4764]: E0202 09:08:33.825703 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kwtmr" podUID="991faa9a-dd25-4f49-82bd-ce60cefd4af2" Feb 02 09:08:33 crc kubenswrapper[4764]: I0202 09:08:33.923728 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:33 crc kubenswrapper[4764]: I0202 09:08:33.923768 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:33 crc kubenswrapper[4764]: I0202 09:08:33.923783 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:33 crc kubenswrapper[4764]: I0202 09:08:33.923805 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:33 crc kubenswrapper[4764]: I0202 09:08:33.923816 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:33Z","lastTransitionTime":"2026-02-02T09:08:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:34 crc kubenswrapper[4764]: I0202 09:08:34.026124 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:34 crc kubenswrapper[4764]: I0202 09:08:34.026174 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:34 crc kubenswrapper[4764]: I0202 09:08:34.026187 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:34 crc kubenswrapper[4764]: I0202 09:08:34.026206 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:34 crc kubenswrapper[4764]: I0202 09:08:34.026218 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:34Z","lastTransitionTime":"2026-02-02T09:08:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:34 crc kubenswrapper[4764]: I0202 09:08:34.128738 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:34 crc kubenswrapper[4764]: I0202 09:08:34.128782 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:34 crc kubenswrapper[4764]: I0202 09:08:34.128798 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:34 crc kubenswrapper[4764]: I0202 09:08:34.128823 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:34 crc kubenswrapper[4764]: I0202 09:08:34.128840 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:34Z","lastTransitionTime":"2026-02-02T09:08:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:34 crc kubenswrapper[4764]: I0202 09:08:34.231782 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:34 crc kubenswrapper[4764]: I0202 09:08:34.231847 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:34 crc kubenswrapper[4764]: I0202 09:08:34.231869 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:34 crc kubenswrapper[4764]: I0202 09:08:34.231896 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:34 crc kubenswrapper[4764]: I0202 09:08:34.231917 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:34Z","lastTransitionTime":"2026-02-02T09:08:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:34 crc kubenswrapper[4764]: I0202 09:08:34.335205 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:34 crc kubenswrapper[4764]: I0202 09:08:34.335261 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:34 crc kubenswrapper[4764]: I0202 09:08:34.335276 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:34 crc kubenswrapper[4764]: I0202 09:08:34.335295 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:34 crc kubenswrapper[4764]: I0202 09:08:34.335310 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:34Z","lastTransitionTime":"2026-02-02T09:08:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:34 crc kubenswrapper[4764]: I0202 09:08:34.437664 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:34 crc kubenswrapper[4764]: I0202 09:08:34.437695 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:34 crc kubenswrapper[4764]: I0202 09:08:34.437705 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:34 crc kubenswrapper[4764]: I0202 09:08:34.437721 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:34 crc kubenswrapper[4764]: I0202 09:08:34.437732 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:34Z","lastTransitionTime":"2026-02-02T09:08:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:34 crc kubenswrapper[4764]: I0202 09:08:34.542065 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:34 crc kubenswrapper[4764]: I0202 09:08:34.542146 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:34 crc kubenswrapper[4764]: I0202 09:08:34.542168 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:34 crc kubenswrapper[4764]: I0202 09:08:34.542199 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:34 crc kubenswrapper[4764]: I0202 09:08:34.542220 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:34Z","lastTransitionTime":"2026-02-02T09:08:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:34 crc kubenswrapper[4764]: I0202 09:08:34.646409 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:34 crc kubenswrapper[4764]: I0202 09:08:34.646472 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:34 crc kubenswrapper[4764]: I0202 09:08:34.646487 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:34 crc kubenswrapper[4764]: I0202 09:08:34.646520 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:34 crc kubenswrapper[4764]: I0202 09:08:34.646545 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:34Z","lastTransitionTime":"2026-02-02T09:08:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:34 crc kubenswrapper[4764]: I0202 09:08:34.749494 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:34 crc kubenswrapper[4764]: I0202 09:08:34.749552 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:34 crc kubenswrapper[4764]: I0202 09:08:34.749568 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:34 crc kubenswrapper[4764]: I0202 09:08:34.749593 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:34 crc kubenswrapper[4764]: I0202 09:08:34.749610 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:34Z","lastTransitionTime":"2026-02-02T09:08:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:34 crc kubenswrapper[4764]: I0202 09:08:34.819923 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-13 18:25:44.530846245 +0000 UTC Feb 02 09:08:34 crc kubenswrapper[4764]: I0202 09:08:34.825475 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 09:08:34 crc kubenswrapper[4764]: I0202 09:08:34.825541 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 09:08:34 crc kubenswrapper[4764]: E0202 09:08:34.826274 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 09:08:34 crc kubenswrapper[4764]: E0202 09:08:34.826320 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 09:08:34 crc kubenswrapper[4764]: I0202 09:08:34.825541 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 09:08:34 crc kubenswrapper[4764]: E0202 09:08:34.826625 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 09:08:34 crc kubenswrapper[4764]: I0202 09:08:34.852892 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:34 crc kubenswrapper[4764]: I0202 09:08:34.852997 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:34 crc kubenswrapper[4764]: I0202 09:08:34.853018 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:34 crc kubenswrapper[4764]: I0202 09:08:34.853051 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:34 crc kubenswrapper[4764]: I0202 09:08:34.853072 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:34Z","lastTransitionTime":"2026-02-02T09:08:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:34 crc kubenswrapper[4764]: I0202 09:08:34.955900 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:34 crc kubenswrapper[4764]: I0202 09:08:34.955976 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:34 crc kubenswrapper[4764]: I0202 09:08:34.955990 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:34 crc kubenswrapper[4764]: I0202 09:08:34.956015 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:34 crc kubenswrapper[4764]: I0202 09:08:34.956031 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:34Z","lastTransitionTime":"2026-02-02T09:08:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:35 crc kubenswrapper[4764]: I0202 09:08:35.059347 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:35 crc kubenswrapper[4764]: I0202 09:08:35.059995 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:35 crc kubenswrapper[4764]: I0202 09:08:35.060357 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:35 crc kubenswrapper[4764]: I0202 09:08:35.060534 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:35 crc kubenswrapper[4764]: I0202 09:08:35.060689 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:35Z","lastTransitionTime":"2026-02-02T09:08:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:35 crc kubenswrapper[4764]: I0202 09:08:35.164375 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:35 crc kubenswrapper[4764]: I0202 09:08:35.164420 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:35 crc kubenswrapper[4764]: I0202 09:08:35.164429 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:35 crc kubenswrapper[4764]: I0202 09:08:35.164447 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:35 crc kubenswrapper[4764]: I0202 09:08:35.164459 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:35Z","lastTransitionTime":"2026-02-02T09:08:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:35 crc kubenswrapper[4764]: I0202 09:08:35.268470 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:35 crc kubenswrapper[4764]: I0202 09:08:35.268526 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:35 crc kubenswrapper[4764]: I0202 09:08:35.268546 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:35 crc kubenswrapper[4764]: I0202 09:08:35.268572 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:35 crc kubenswrapper[4764]: I0202 09:08:35.268590 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:35Z","lastTransitionTime":"2026-02-02T09:08:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:35 crc kubenswrapper[4764]: I0202 09:08:35.372296 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:35 crc kubenswrapper[4764]: I0202 09:08:35.372350 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:35 crc kubenswrapper[4764]: I0202 09:08:35.372365 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:35 crc kubenswrapper[4764]: I0202 09:08:35.372388 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:35 crc kubenswrapper[4764]: I0202 09:08:35.372403 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:35Z","lastTransitionTime":"2026-02-02T09:08:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:35 crc kubenswrapper[4764]: I0202 09:08:35.476159 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:35 crc kubenswrapper[4764]: I0202 09:08:35.476240 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:35 crc kubenswrapper[4764]: I0202 09:08:35.476267 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:35 crc kubenswrapper[4764]: I0202 09:08:35.476311 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:35 crc kubenswrapper[4764]: I0202 09:08:35.476336 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:35Z","lastTransitionTime":"2026-02-02T09:08:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:35 crc kubenswrapper[4764]: I0202 09:08:35.579226 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:35 crc kubenswrapper[4764]: I0202 09:08:35.579316 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:35 crc kubenswrapper[4764]: I0202 09:08:35.579337 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:35 crc kubenswrapper[4764]: I0202 09:08:35.579370 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:35 crc kubenswrapper[4764]: I0202 09:08:35.579392 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:35Z","lastTransitionTime":"2026-02-02T09:08:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:35 crc kubenswrapper[4764]: I0202 09:08:35.684644 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:35 crc kubenswrapper[4764]: I0202 09:08:35.684743 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:35 crc kubenswrapper[4764]: I0202 09:08:35.684876 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:35 crc kubenswrapper[4764]: I0202 09:08:35.684918 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:35 crc kubenswrapper[4764]: I0202 09:08:35.686336 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:35Z","lastTransitionTime":"2026-02-02T09:08:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:35 crc kubenswrapper[4764]: I0202 09:08:35.791005 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:35 crc kubenswrapper[4764]: I0202 09:08:35.791072 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:35 crc kubenswrapper[4764]: I0202 09:08:35.791085 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:35 crc kubenswrapper[4764]: I0202 09:08:35.791105 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:35 crc kubenswrapper[4764]: I0202 09:08:35.791119 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:35Z","lastTransitionTime":"2026-02-02T09:08:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:35 crc kubenswrapper[4764]: I0202 09:08:35.820465 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-30 05:31:27.852480172 +0000 UTC Feb 02 09:08:35 crc kubenswrapper[4764]: I0202 09:08:35.825397 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kwtmr" Feb 02 09:08:35 crc kubenswrapper[4764]: E0202 09:08:35.825630 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kwtmr" podUID="991faa9a-dd25-4f49-82bd-ce60cefd4af2" Feb 02 09:08:35 crc kubenswrapper[4764]: I0202 09:08:35.895484 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:35 crc kubenswrapper[4764]: I0202 09:08:35.895551 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:35 crc kubenswrapper[4764]: I0202 09:08:35.895566 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:35 crc kubenswrapper[4764]: I0202 09:08:35.895593 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:35 crc kubenswrapper[4764]: I0202 09:08:35.895617 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:35Z","lastTransitionTime":"2026-02-02T09:08:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:35 crc kubenswrapper[4764]: I0202 09:08:35.999095 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:35 crc kubenswrapper[4764]: I0202 09:08:35.999807 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:36 crc kubenswrapper[4764]: I0202 09:08:36.000137 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:36 crc kubenswrapper[4764]: I0202 09:08:36.000288 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:36 crc kubenswrapper[4764]: I0202 09:08:36.000421 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:36Z","lastTransitionTime":"2026-02-02T09:08:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:36 crc kubenswrapper[4764]: I0202 09:08:36.104481 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:36 crc kubenswrapper[4764]: I0202 09:08:36.104543 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:36 crc kubenswrapper[4764]: I0202 09:08:36.104566 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:36 crc kubenswrapper[4764]: I0202 09:08:36.104595 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:36 crc kubenswrapper[4764]: I0202 09:08:36.104615 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:36Z","lastTransitionTime":"2026-02-02T09:08:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:36 crc kubenswrapper[4764]: I0202 09:08:36.208065 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:36 crc kubenswrapper[4764]: I0202 09:08:36.208439 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:36 crc kubenswrapper[4764]: I0202 09:08:36.208527 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:36 crc kubenswrapper[4764]: I0202 09:08:36.208671 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:36 crc kubenswrapper[4764]: I0202 09:08:36.208755 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:36Z","lastTransitionTime":"2026-02-02T09:08:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:36 crc kubenswrapper[4764]: I0202 09:08:36.311891 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:36 crc kubenswrapper[4764]: I0202 09:08:36.311953 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:36 crc kubenswrapper[4764]: I0202 09:08:36.311969 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:36 crc kubenswrapper[4764]: I0202 09:08:36.311986 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:36 crc kubenswrapper[4764]: I0202 09:08:36.311997 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:36Z","lastTransitionTime":"2026-02-02T09:08:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:36 crc kubenswrapper[4764]: I0202 09:08:36.403495 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:36 crc kubenswrapper[4764]: I0202 09:08:36.403587 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:36 crc kubenswrapper[4764]: I0202 09:08:36.403608 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:36 crc kubenswrapper[4764]: I0202 09:08:36.403636 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:36 crc kubenswrapper[4764]: I0202 09:08:36.403655 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:36Z","lastTransitionTime":"2026-02-02T09:08:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:36 crc kubenswrapper[4764]: E0202 09:08:36.421131 4764 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:08:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:08:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:08:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:08:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"31973a70-ce17-4811-9a14-f5476c6979db\\\",\\\"systemUUID\\\":\\\"ca92679c-8580-43c3-be51-62441e93d22b\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:36Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:36 crc kubenswrapper[4764]: I0202 09:08:36.425755 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:36 crc kubenswrapper[4764]: I0202 09:08:36.425806 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:36 crc kubenswrapper[4764]: I0202 09:08:36.425824 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:36 crc kubenswrapper[4764]: I0202 09:08:36.425852 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:36 crc kubenswrapper[4764]: I0202 09:08:36.425870 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:36Z","lastTransitionTime":"2026-02-02T09:08:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:36 crc kubenswrapper[4764]: E0202 09:08:36.448103 4764 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:08:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:08:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:08:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:08:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"31973a70-ce17-4811-9a14-f5476c6979db\\\",\\\"systemUUID\\\":\\\"ca92679c-8580-43c3-be51-62441e93d22b\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:36Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:36 crc kubenswrapper[4764]: I0202 09:08:36.454406 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:36 crc kubenswrapper[4764]: I0202 09:08:36.454584 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:36 crc kubenswrapper[4764]: I0202 09:08:36.454654 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:36 crc kubenswrapper[4764]: I0202 09:08:36.454721 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:36 crc kubenswrapper[4764]: I0202 09:08:36.454806 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:36Z","lastTransitionTime":"2026-02-02T09:08:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:36 crc kubenswrapper[4764]: E0202 09:08:36.475668 4764 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:08:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:08:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:08:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:08:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"31973a70-ce17-4811-9a14-f5476c6979db\\\",\\\"systemUUID\\\":\\\"ca92679c-8580-43c3-be51-62441e93d22b\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:36Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:36 crc kubenswrapper[4764]: I0202 09:08:36.480741 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:36 crc kubenswrapper[4764]: I0202 09:08:36.480776 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:36 crc kubenswrapper[4764]: I0202 09:08:36.480784 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:36 crc kubenswrapper[4764]: I0202 09:08:36.480798 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:36 crc kubenswrapper[4764]: I0202 09:08:36.480808 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:36Z","lastTransitionTime":"2026-02-02T09:08:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:36 crc kubenswrapper[4764]: E0202 09:08:36.501184 4764 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:08:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:08:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:08:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:08:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"31973a70-ce17-4811-9a14-f5476c6979db\\\",\\\"systemUUID\\\":\\\"ca92679c-8580-43c3-be51-62441e93d22b\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:36Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:36 crc kubenswrapper[4764]: I0202 09:08:36.506107 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:36 crc kubenswrapper[4764]: I0202 09:08:36.506430 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:36 crc kubenswrapper[4764]: I0202 09:08:36.506676 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:36 crc kubenswrapper[4764]: I0202 09:08:36.506887 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:36 crc kubenswrapper[4764]: I0202 09:08:36.507152 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:36Z","lastTransitionTime":"2026-02-02T09:08:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:36 crc kubenswrapper[4764]: E0202 09:08:36.529274 4764 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:08:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:08:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:08:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:08:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"31973a70-ce17-4811-9a14-f5476c6979db\\\",\\\"systemUUID\\\":\\\"ca92679c-8580-43c3-be51-62441e93d22b\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:36Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:36 crc kubenswrapper[4764]: E0202 09:08:36.529843 4764 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 02 09:08:36 crc kubenswrapper[4764]: I0202 09:08:36.531700 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:36 crc kubenswrapper[4764]: I0202 09:08:36.531742 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:36 crc kubenswrapper[4764]: I0202 09:08:36.531756 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:36 crc kubenswrapper[4764]: I0202 09:08:36.531774 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:36 crc kubenswrapper[4764]: I0202 09:08:36.531787 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:36Z","lastTransitionTime":"2026-02-02T09:08:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:36 crc kubenswrapper[4764]: I0202 09:08:36.636094 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:36 crc kubenswrapper[4764]: I0202 09:08:36.636576 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:36 crc kubenswrapper[4764]: I0202 09:08:36.636673 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:36 crc kubenswrapper[4764]: I0202 09:08:36.636763 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:36 crc kubenswrapper[4764]: I0202 09:08:36.636875 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:36Z","lastTransitionTime":"2026-02-02T09:08:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:36 crc kubenswrapper[4764]: I0202 09:08:36.739447 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:36 crc kubenswrapper[4764]: I0202 09:08:36.739739 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:36 crc kubenswrapper[4764]: I0202 09:08:36.739837 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:36 crc kubenswrapper[4764]: I0202 09:08:36.739976 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:36 crc kubenswrapper[4764]: I0202 09:08:36.740083 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:36Z","lastTransitionTime":"2026-02-02T09:08:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:36 crc kubenswrapper[4764]: I0202 09:08:36.822645 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-09 06:00:46.53741418 +0000 UTC Feb 02 09:08:36 crc kubenswrapper[4764]: I0202 09:08:36.825028 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 09:08:36 crc kubenswrapper[4764]: I0202 09:08:36.825051 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 09:08:36 crc kubenswrapper[4764]: E0202 09:08:36.825192 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 09:08:36 crc kubenswrapper[4764]: E0202 09:08:36.825402 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 09:08:36 crc kubenswrapper[4764]: I0202 09:08:36.825374 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 09:08:36 crc kubenswrapper[4764]: E0202 09:08:36.825622 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 09:08:36 crc kubenswrapper[4764]: I0202 09:08:36.844137 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:36 crc kubenswrapper[4764]: I0202 09:08:36.844199 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:36 crc kubenswrapper[4764]: I0202 09:08:36.844213 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:36 crc kubenswrapper[4764]: I0202 09:08:36.844236 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:36 crc kubenswrapper[4764]: I0202 09:08:36.844253 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:36Z","lastTransitionTime":"2026-02-02T09:08:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:36 crc kubenswrapper[4764]: I0202 09:08:36.947774 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:36 crc kubenswrapper[4764]: I0202 09:08:36.948163 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:36 crc kubenswrapper[4764]: I0202 09:08:36.948375 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:36 crc kubenswrapper[4764]: I0202 09:08:36.948584 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:36 crc kubenswrapper[4764]: I0202 09:08:36.948737 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:36Z","lastTransitionTime":"2026-02-02T09:08:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:37 crc kubenswrapper[4764]: I0202 09:08:37.052095 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:37 crc kubenswrapper[4764]: I0202 09:08:37.052144 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:37 crc kubenswrapper[4764]: I0202 09:08:37.052162 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:37 crc kubenswrapper[4764]: I0202 09:08:37.052181 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:37 crc kubenswrapper[4764]: I0202 09:08:37.052193 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:37Z","lastTransitionTime":"2026-02-02T09:08:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:37 crc kubenswrapper[4764]: I0202 09:08:37.155270 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:37 crc kubenswrapper[4764]: I0202 09:08:37.155510 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:37 crc kubenswrapper[4764]: I0202 09:08:37.155596 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:37 crc kubenswrapper[4764]: I0202 09:08:37.155685 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:37 crc kubenswrapper[4764]: I0202 09:08:37.155768 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:37Z","lastTransitionTime":"2026-02-02T09:08:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:37 crc kubenswrapper[4764]: I0202 09:08:37.259914 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:37 crc kubenswrapper[4764]: I0202 09:08:37.260450 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:37 crc kubenswrapper[4764]: I0202 09:08:37.260555 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:37 crc kubenswrapper[4764]: I0202 09:08:37.260632 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:37 crc kubenswrapper[4764]: I0202 09:08:37.260697 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:37Z","lastTransitionTime":"2026-02-02T09:08:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:37 crc kubenswrapper[4764]: I0202 09:08:37.364172 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:37 crc kubenswrapper[4764]: I0202 09:08:37.364455 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:37 crc kubenswrapper[4764]: I0202 09:08:37.364570 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:37 crc kubenswrapper[4764]: I0202 09:08:37.364720 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:37 crc kubenswrapper[4764]: I0202 09:08:37.364837 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:37Z","lastTransitionTime":"2026-02-02T09:08:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:37 crc kubenswrapper[4764]: I0202 09:08:37.468582 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:37 crc kubenswrapper[4764]: I0202 09:08:37.468649 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:37 crc kubenswrapper[4764]: I0202 09:08:37.468665 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:37 crc kubenswrapper[4764]: I0202 09:08:37.468688 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:37 crc kubenswrapper[4764]: I0202 09:08:37.468702 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:37Z","lastTransitionTime":"2026-02-02T09:08:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:37 crc kubenswrapper[4764]: I0202 09:08:37.571826 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:37 crc kubenswrapper[4764]: I0202 09:08:37.571924 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:37 crc kubenswrapper[4764]: I0202 09:08:37.571977 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:37 crc kubenswrapper[4764]: I0202 09:08:37.572002 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:37 crc kubenswrapper[4764]: I0202 09:08:37.572020 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:37Z","lastTransitionTime":"2026-02-02T09:08:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:37 crc kubenswrapper[4764]: I0202 09:08:37.674477 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:37 crc kubenswrapper[4764]: I0202 09:08:37.674521 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:37 crc kubenswrapper[4764]: I0202 09:08:37.674534 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:37 crc kubenswrapper[4764]: I0202 09:08:37.674551 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:37 crc kubenswrapper[4764]: I0202 09:08:37.674564 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:37Z","lastTransitionTime":"2026-02-02T09:08:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:37 crc kubenswrapper[4764]: I0202 09:08:37.777273 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:37 crc kubenswrapper[4764]: I0202 09:08:37.777331 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:37 crc kubenswrapper[4764]: I0202 09:08:37.777344 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:37 crc kubenswrapper[4764]: I0202 09:08:37.777362 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:37 crc kubenswrapper[4764]: I0202 09:08:37.777375 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:37Z","lastTransitionTime":"2026-02-02T09:08:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:37 crc kubenswrapper[4764]: I0202 09:08:37.823852 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-06 02:09:57.045543531 +0000 UTC Feb 02 09:08:37 crc kubenswrapper[4764]: I0202 09:08:37.825299 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kwtmr" Feb 02 09:08:37 crc kubenswrapper[4764]: E0202 09:08:37.825846 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kwtmr" podUID="991faa9a-dd25-4f49-82bd-ce60cefd4af2" Feb 02 09:08:37 crc kubenswrapper[4764]: I0202 09:08:37.826516 4764 scope.go:117] "RemoveContainer" containerID="334e828f8ef72570f2511ee38659941202530471f5943aa07b8b21483bd70550" Feb 02 09:08:37 crc kubenswrapper[4764]: I0202 09:08:37.845633 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed747fd8-e3ce-4319-a6da-bda0fd51b85a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6f064bdc66470aace3180d4988244026b54009a63cff199e41c5d02a1d18479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://60258f84a1d8dcfe48b7e9688ed50f077e201971f20f3f6663fbf1ba58a4611a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f84be2e08217a2ac3edde858d78b4410376ed9bae3bcabe2a7f9b2fbe2f69c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a69d25d352d7ceac5c735bc1fcf446a1fea6dc22c65a5296cfab828cea17377\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:37Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:37 crc kubenswrapper[4764]: I0202 09:08:37.861737 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f6904228e052fff69ad1fd66df97758b6b60ba4351e97a7fbd03eee4df1b653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:37Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:37 crc kubenswrapper[4764]: I0202 09:08:37.878229 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6zxff" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aaba93da-12a5-43df-b7c7-4c8b988163c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3db293678a907cd55b31b1d0c0bd7c30e1ba75ecd5e13cbc68ab98fbf0fbd9d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqwb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6zxff\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:37Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:37 crc kubenswrapper[4764]: I0202 09:08:37.880000 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:37 crc kubenswrapper[4764]: I0202 09:08:37.880668 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:37 crc kubenswrapper[4764]: I0202 09:08:37.880822 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:37 crc kubenswrapper[4764]: I0202 09:08:37.880901 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:37 crc kubenswrapper[4764]: I0202 09:08:37.881001 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:37Z","lastTransitionTime":"2026-02-02T09:08:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:37 crc kubenswrapper[4764]: I0202 09:08:37.893985 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d192f670-9f9d-4539-9641-e4bed73acdd4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://040c5e8a5659f85f0d52d1a1e064154d998ae104f711101d2204c0b1caebbb30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbksf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fdf06410349b7c99aa166e87fbe80bf68e2ae923635a6a978b7e2cd50383f7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbksf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4ndm4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:37Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:37 crc kubenswrapper[4764]: I0202 09:08:37.906783 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pcjmn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6e5ccd3b-b657-47a2-9c4d-bc9adb126533\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72863eef568d92a1abeb3cde2e89630a4f5fb7626c6949448ff5195774c98eab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2f7zl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pcjmn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:37Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:37 crc kubenswrapper[4764]: I0202 09:08:37.919642 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fae745b-ea49-42f3-a486-dbc3dc2fb9ad\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://081d7f76d937ff71e35945f9b1d893888bb649d15bd84312fd4b61c9520ea71f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://429ead0993f5e648f6256d249ee12aea03514d3108618840552b672fb419cd74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fad17b01545eb20785f03e3c61bd5c982f6217e38176d7a3b23ac194298644a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85e2f75c56d63cacc93b9195c1e8bf46eb7a7234c45a024bd95a460db90edb6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85e2f75c56d63cacc93b9195c1e8bf46eb7a7234c45a024bd95a460db90edb6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:37Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:37 crc kubenswrapper[4764]: I0202 09:08:37.937322 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:37Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:37 crc kubenswrapper[4764]: I0202 09:08:37.951881 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lh6tm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2152e8a-f456-47b5-a547-b65ec837ecbc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38e176a00b07f2960fdd7ba2aadd1468a5e083b8d154e9d118d13f2585da81e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bb2sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://996e649011d5255783284421ddea8e0a5c383ef675b0920c3fdc0af6503d80e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bb2sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lh6tm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:37Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:37 crc kubenswrapper[4764]: I0202 09:08:37.965720 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9p5dc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"660ac006-2495-45c4-9fb3-e9c1dddcf7a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da812b124c97aa8ad1563f4bcfac8f3bfa025096fa4a2561e5847b1bf9cb30b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23e57e65a6f5c3059506b9db21e9ac2492c30bd9191a98db0281915e9ce39797\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23e57e65a6f5c3059506b9db21e9ac2492c30bd9191a98db0281915e9ce39797\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2417b66816da1a6cd1791a4b91eca4a3c4f3694462e066e7f0722f1b9c2105c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2417b66816da1a6cd1791a4b91eca4a3c4f3694462e066e7f0722f1b9c2105c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://883641f29e9120c2cea6b272bacf1dc63c752f55ab0408cdaf73db85c85ac369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://883641f29e9120c2cea6b272bacf1dc63c752f55ab0408cdaf73db85c85ac369\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2745baba1188ba5e8f820616e4117a12c50ecee5b68e71ad9cabbb621b6e131f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2745baba1188ba5e8f820616e4117a12c50ecee5b68e71ad9cabbb621b6e131f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8925e435be4d48d86824fc35f8b8819fb35d63e385164c4a20edb3bf3035c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a8925e435be4d48d86824fc35f8b8819fb35d63e385164c4a20edb3bf3035c3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d3fbf668ae7d04c8d9e574f43c2fe90c5604204ba38a3c60847afdfbceb7a57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d3fbf668ae7d04c8d9e574f43c2fe90c5604204ba38a3c60847afdfbceb7a57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9p5dc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:37Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:37 crc kubenswrapper[4764]: I0202 09:08:37.980277 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kwtmr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"991faa9a-dd25-4f49-82bd-ce60cefd4af2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h865b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h865b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:53Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kwtmr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:37Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:37 crc kubenswrapper[4764]: I0202 09:08:37.991769 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:37 crc kubenswrapper[4764]: I0202 09:08:37.991807 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:37 crc kubenswrapper[4764]: I0202 09:08:37.991815 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:37 crc kubenswrapper[4764]: I0202 09:08:37.991832 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:37 crc kubenswrapper[4764]: I0202 09:08:37.991849 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:37Z","lastTransitionTime":"2026-02-02T09:08:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:37 crc kubenswrapper[4764]: I0202 09:08:37.994434 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87d61aea-46f6-4ee6-a7a4-767a6a67fea4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e78b6d08e4d616cefbb73e3fd43b3c6886dcd8aa997c7b4948bf1ff63401293\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d468715c2da33996561c40043216233fd12e21539d0c58dd1f3b892fd9f91a9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b184f7c936dc4046f74c0018269cbc2c4b7ceddb650b4e871d3f8a707e0df88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://269c5074371bd00f8363fb3888c11f1df383081a61497aea6578f08700881d50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b25bc9395831524c53f9e565b4efbe74f21792efafeb1dd507161bc91aa0ecdf\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T09:07:22Z\\\",\\\"message\\\":\\\"W0202 09:07:21.597488 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0202 09:07:21.598117 1 crypto.go:601] Generating new CA for check-endpoints-signer@1770023241 cert, and key in /tmp/serving-cert-3856183724/serving-signer.crt, /tmp/serving-cert-3856183724/serving-signer.key\\\\nI0202 09:07:22.012184 1 observer_polling.go:159] Starting file observer\\\\nW0202 09:07:22.015493 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0202 09:07:22.015797 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 09:07:22.016596 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3856183724/tls.crt::/tmp/serving-cert-3856183724/tls.key\\\\\\\"\\\\nF0202 09:07:22.270605 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b44e17b8ca01f549bc10081dd5f109cb1b9fff364bf5bd4fd90d9f0e3eeb04db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfb06a5ec9262cf39b108a66014d15e325cdbd5441a2d9e7a811dd57265c99a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfb06a5ec9262cf39b108a66014d15e325cdbd5441a2d9e7a811dd57265c99a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:37Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:38 crc kubenswrapper[4764]: I0202 09:08:38.005129 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fea07bdc-9134-4993-934c-316c19d0dc76\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://664e0519c2a2c60904dfcc8deebaf5d638dac0bedb13d870c0aaea5b4f5428f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db5a54d3a2117252a1ff50884683f99185874d2904c0960161642280f614679d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db5a54d3a2117252a1ff50884683f99185874d2904c0960161642280f614679d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:38Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:38 crc kubenswrapper[4764]: I0202 09:08:38.019374 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5528225d91ac2778bef4554abb0bd036ac5d3b6cdac7357bab69e70c3894b6f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://071bbb80d0747cf3e109ba626d47f7d05f85f4cc0b3aba77276c795741f47ac6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:38Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:38 crc kubenswrapper[4764]: I0202 09:08:38.030611 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:38Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:38 crc kubenswrapper[4764]: I0202 09:08:38.042125 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nc7qt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b52b499d17b4e2e359475bd13730c72ef5cedb052e13def6dfe5bc5fd7487d2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1a2c7b22068b7165927fc158885d48bf9952b9eba521d8ae9a599eee53dc305\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-02T09:08:28Z\\\",\\\"message\\\":\\\"2026-02-02T09:07:42+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_32cbaf6a-865c-4787-8f89-a40249873b40\\\\n2026-02-02T09:07:42+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_32cbaf6a-865c-4787-8f89-a40249873b40 to /host/opt/cni/bin/\\\\n2026-02-02T09:07:43Z [verbose] multus-daemon started\\\\n2026-02-02T09:07:43Z [verbose] Readiness Indicator file check\\\\n2026-02-02T09:08:28Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:08:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2hz4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nc7qt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:38Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:38 crc kubenswrapper[4764]: I0202 09:08:38.059703 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d66ab3c0-9b5b-41d5-b6c0-91ccddccd7e9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d92ac486f5dade46bc75de32099859a71fc2ed25a7abbe6ae8937df128e7df30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a4ac273a4f127d79e5fad7d4c30cd2044a346f5734dba5d33d68cd092c0c2d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d38c7cb5dcd22fb07524494ebe77aa1a7f4f808f566043563817e47e7ec12a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://408722c7f83d28919431532f8c797a1c795fd5d96736d14e7f06a05744899702\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f473b18cd8003f75bb1dafb82bcec45745a73ccd10434297356dac185a3d15c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e229f484d19cbdabedf53d90a822a4434db9e00b1301538e885e26acc19d25f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e229f484d19cbdabedf53d90a822a4434db9e00b1301538e885e26acc19d25f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a9cc673be8ff245dba81278e8cd34cfe45ed36ff0b8bd5b30a0ddf333ec6371\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a9cc673be8ff245dba81278e8cd34cfe45ed36ff0b8bd5b30a0ddf333ec6371\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bc7e49c7a5ae66ea939060192fa6a346dc0b0b9aa9750b29879d3c426cc0d276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc7e49c7a5ae66ea939060192fa6a346dc0b0b9aa9750b29879d3c426cc0d276\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:38Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:38 crc kubenswrapper[4764]: I0202 09:08:38.073468 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa9d037e66cb53b9dd698eff46284c9e718fd842cda6f7b525f0c79062d70bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:38Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:38 crc kubenswrapper[4764]: I0202 09:08:38.087752 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:38Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:38 crc kubenswrapper[4764]: I0202 09:08:38.096574 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:38 crc kubenswrapper[4764]: I0202 09:08:38.096621 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:38 crc kubenswrapper[4764]: I0202 09:08:38.096637 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:38 crc kubenswrapper[4764]: I0202 09:08:38.096658 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:38 crc kubenswrapper[4764]: I0202 09:08:38.096673 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:38Z","lastTransitionTime":"2026-02-02T09:08:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:38 crc kubenswrapper[4764]: I0202 09:08:38.109989 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24632dda-6100-4ab6-a28e-214ddae4360c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a6c67e9b7708cc0d85b0b6ffbbade0d34867e159a9c5487ca2606475c667a50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bffda48d7c881c1470251b501661324859373246cc1a21811bd36ca24887c7aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a989645cf485cd8ff80ffff1a86cd7808e1564eba5667535e161cc68db13b26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20b71d3ee7a381baa51b0b0d0470283022cfede30f491135737b5e82d631d763\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c380383b6a014d1b694a13ddc22f54b20b9e4f9d5c9b9000d76187621dce0a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5529f2bff87753428e9173826fc1880acf96a4c134fa9ecad195b15697cad73a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://334e828f8ef72570f2511ee38659941202530471f5943aa07b8b21483bd70550\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://334e828f8ef72570f2511ee38659941202530471f5943aa07b8b21483bd70550\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-02T09:08:10Z\\\",\\\"message\\\":\\\"ing OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-dns/dns-default]} name:Service_openshift-dns/dns-default_TCP_node_router+switch_crc options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.10:53: 10.217.4.10:9154:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {be9dcc9e-c16a-4962-a6d2-4adeb0b929c4}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-dns/dns-default]} name:Service_openshift-dns/dns-default_UDP_node_router+switch_crc options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[udp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.10:53:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {4c1be812-05d3-4f45-91b5-a853a5c8de71}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF0202 09:08:09.911531 6327 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T09:08:08Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-zhn7j_openshift-ovn-kubernetes(24632dda-6100-4ab6-a28e-214ddae4360c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b4cfbd4e682bcdcaff785013bdb72580ce36fced733e235538536827036b93d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://579903eadb183466a2564737e6b678e04308239c0cd75b6808892053b97afbc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://579903eadb183466a2564737e6b678e04308239c0cd75b6808892053b97afbc0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zhn7j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:38Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:38 crc kubenswrapper[4764]: I0202 09:08:38.200606 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:38 crc kubenswrapper[4764]: I0202 09:08:38.200640 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:38 crc kubenswrapper[4764]: I0202 09:08:38.200648 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:38 crc kubenswrapper[4764]: I0202 09:08:38.200661 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:38 crc kubenswrapper[4764]: I0202 09:08:38.200670 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:38Z","lastTransitionTime":"2026-02-02T09:08:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:38 crc kubenswrapper[4764]: I0202 09:08:38.267146 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zhn7j_24632dda-6100-4ab6-a28e-214ddae4360c/ovnkube-controller/2.log" Feb 02 09:08:38 crc kubenswrapper[4764]: I0202 09:08:38.270497 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" event={"ID":"24632dda-6100-4ab6-a28e-214ddae4360c","Type":"ContainerStarted","Data":"3cb255e14eb02f7c08364c41fb9906e58b2155eb507f694b77ec8eb1f4251343"} Feb 02 09:08:38 crc kubenswrapper[4764]: I0202 09:08:38.271002 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" Feb 02 09:08:38 crc kubenswrapper[4764]: I0202 09:08:38.288632 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kwtmr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"991faa9a-dd25-4f49-82bd-ce60cefd4af2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h865b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h865b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:53Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kwtmr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:38Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:38 crc kubenswrapper[4764]: I0202 09:08:38.303349 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:38 crc kubenswrapper[4764]: I0202 09:08:38.303391 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:38 crc kubenswrapper[4764]: I0202 09:08:38.303402 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:38 crc kubenswrapper[4764]: I0202 09:08:38.303417 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:38 crc kubenswrapper[4764]: I0202 09:08:38.303428 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:38Z","lastTransitionTime":"2026-02-02T09:08:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:38 crc kubenswrapper[4764]: I0202 09:08:38.307446 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87d61aea-46f6-4ee6-a7a4-767a6a67fea4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e78b6d08e4d616cefbb73e3fd43b3c6886dcd8aa997c7b4948bf1ff63401293\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d468715c2da33996561c40043216233fd12e21539d0c58dd1f3b892fd9f91a9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b184f7c936dc4046f74c0018269cbc2c4b7ceddb650b4e871d3f8a707e0df88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://269c5074371bd00f8363fb3888c11f1df383081a61497aea6578f08700881d50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b25bc9395831524c53f9e565b4efbe74f21792efafeb1dd507161bc91aa0ecdf\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T09:07:22Z\\\",\\\"message\\\":\\\"W0202 09:07:21.597488 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0202 09:07:21.598117 1 crypto.go:601] Generating new CA for check-endpoints-signer@1770023241 cert, and key in /tmp/serving-cert-3856183724/serving-signer.crt, /tmp/serving-cert-3856183724/serving-signer.key\\\\nI0202 09:07:22.012184 1 observer_polling.go:159] Starting file observer\\\\nW0202 09:07:22.015493 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0202 09:07:22.015797 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 09:07:22.016596 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3856183724/tls.crt::/tmp/serving-cert-3856183724/tls.key\\\\\\\"\\\\nF0202 09:07:22.270605 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b44e17b8ca01f549bc10081dd5f109cb1b9fff364bf5bd4fd90d9f0e3eeb04db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfb06a5ec9262cf39b108a66014d15e325cdbd5441a2d9e7a811dd57265c99a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfb06a5ec9262cf39b108a66014d15e325cdbd5441a2d9e7a811dd57265c99a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:38Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:38 crc kubenswrapper[4764]: I0202 09:08:38.337474 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fea07bdc-9134-4993-934c-316c19d0dc76\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://664e0519c2a2c60904dfcc8deebaf5d638dac0bedb13d870c0aaea5b4f5428f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db5a54d3a2117252a1ff50884683f99185874d2904c0960161642280f614679d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db5a54d3a2117252a1ff50884683f99185874d2904c0960161642280f614679d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:38Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:38 crc kubenswrapper[4764]: I0202 09:08:38.371238 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5528225d91ac2778bef4554abb0bd036ac5d3b6cdac7357bab69e70c3894b6f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://071bbb80d0747cf3e109ba626d47f7d05f85f4cc0b3aba77276c795741f47ac6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:38Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:38 crc kubenswrapper[4764]: I0202 09:08:38.390301 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:38Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:38 crc kubenswrapper[4764]: I0202 09:08:38.406128 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:38 crc kubenswrapper[4764]: I0202 09:08:38.406159 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:38 crc kubenswrapper[4764]: I0202 09:08:38.406167 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:38 crc kubenswrapper[4764]: I0202 09:08:38.406180 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:38 crc kubenswrapper[4764]: I0202 09:08:38.406188 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:38Z","lastTransitionTime":"2026-02-02T09:08:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:38 crc kubenswrapper[4764]: I0202 09:08:38.406385 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nc7qt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b52b499d17b4e2e359475bd13730c72ef5cedb052e13def6dfe5bc5fd7487d2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1a2c7b22068b7165927fc158885d48bf9952b9eba521d8ae9a599eee53dc305\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-02T09:08:28Z\\\",\\\"message\\\":\\\"2026-02-02T09:07:42+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_32cbaf6a-865c-4787-8f89-a40249873b40\\\\n2026-02-02T09:07:42+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_32cbaf6a-865c-4787-8f89-a40249873b40 to /host/opt/cni/bin/\\\\n2026-02-02T09:07:43Z [verbose] multus-daemon started\\\\n2026-02-02T09:07:43Z [verbose] Readiness Indicator file check\\\\n2026-02-02T09:08:28Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:08:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2hz4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nc7qt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:38Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:38 crc kubenswrapper[4764]: I0202 09:08:38.422206 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9p5dc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"660ac006-2495-45c4-9fb3-e9c1dddcf7a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da812b124c97aa8ad1563f4bcfac8f3bfa025096fa4a2561e5847b1bf9cb30b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23e57e65a6f5c3059506b9db21e9ac2492c30bd9191a98db0281915e9ce39797\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23e57e65a6f5c3059506b9db21e9ac2492c30bd9191a98db0281915e9ce39797\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2417b66816da1a6cd1791a4b91eca4a3c4f3694462e066e7f0722f1b9c2105c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2417b66816da1a6cd1791a4b91eca4a3c4f3694462e066e7f0722f1b9c2105c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://883641f29e9120c2cea6b272bacf1dc63c752f55ab0408cdaf73db85c85ac369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://883641f29e9120c2cea6b272bacf1dc63c752f55ab0408cdaf73db85c85ac369\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2745baba1188ba5e8f820616e4117a12c50ecee5b68e71ad9cabbb621b6e131f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2745baba1188ba5e8f820616e4117a12c50ecee5b68e71ad9cabbb621b6e131f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8925e435be4d48d86824fc35f8b8819fb35d63e385164c4a20edb3bf3035c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a8925e435be4d48d86824fc35f8b8819fb35d63e385164c4a20edb3bf3035c3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d3fbf668ae7d04c8d9e574f43c2fe90c5604204ba38a3c60847afdfbceb7a57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d3fbf668ae7d04c8d9e574f43c2fe90c5604204ba38a3c60847afdfbceb7a57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9p5dc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:38Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:38 crc kubenswrapper[4764]: I0202 09:08:38.440361 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d66ab3c0-9b5b-41d5-b6c0-91ccddccd7e9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d92ac486f5dade46bc75de32099859a71fc2ed25a7abbe6ae8937df128e7df30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a4ac273a4f127d79e5fad7d4c30cd2044a346f5734dba5d33d68cd092c0c2d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d38c7cb5dcd22fb07524494ebe77aa1a7f4f808f566043563817e47e7ec12a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://408722c7f83d28919431532f8c797a1c795fd5d96736d14e7f06a05744899702\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f473b18cd8003f75bb1dafb82bcec45745a73ccd10434297356dac185a3d15c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e229f484d19cbdabedf53d90a822a4434db9e00b1301538e885e26acc19d25f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e229f484d19cbdabedf53d90a822a4434db9e00b1301538e885e26acc19d25f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a9cc673be8ff245dba81278e8cd34cfe45ed36ff0b8bd5b30a0ddf333ec6371\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a9cc673be8ff245dba81278e8cd34cfe45ed36ff0b8bd5b30a0ddf333ec6371\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bc7e49c7a5ae66ea939060192fa6a346dc0b0b9aa9750b29879d3c426cc0d276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc7e49c7a5ae66ea939060192fa6a346dc0b0b9aa9750b29879d3c426cc0d276\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:38Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:38 crc kubenswrapper[4764]: I0202 09:08:38.456248 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa9d037e66cb53b9dd698eff46284c9e718fd842cda6f7b525f0c79062d70bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:38Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:38 crc kubenswrapper[4764]: I0202 09:08:38.469130 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:38Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:38 crc kubenswrapper[4764]: I0202 09:08:38.489684 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24632dda-6100-4ab6-a28e-214ddae4360c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a6c67e9b7708cc0d85b0b6ffbbade0d34867e159a9c5487ca2606475c667a50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bffda48d7c881c1470251b501661324859373246cc1a21811bd36ca24887c7aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a989645cf485cd8ff80ffff1a86cd7808e1564eba5667535e161cc68db13b26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20b71d3ee7a381baa51b0b0d0470283022cfede30f491135737b5e82d631d763\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c380383b6a014d1b694a13ddc22f54b20b9e4f9d5c9b9000d76187621dce0a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5529f2bff87753428e9173826fc1880acf96a4c134fa9ecad195b15697cad73a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cb255e14eb02f7c08364c41fb9906e58b2155eb507f694b77ec8eb1f4251343\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://334e828f8ef72570f2511ee38659941202530471f5943aa07b8b21483bd70550\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-02T09:08:10Z\\\",\\\"message\\\":\\\"ing OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-dns/dns-default]} name:Service_openshift-dns/dns-default_TCP_node_router+switch_crc options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.10:53: 10.217.4.10:9154:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {be9dcc9e-c16a-4962-a6d2-4adeb0b929c4}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-dns/dns-default]} name:Service_openshift-dns/dns-default_UDP_node_router+switch_crc options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[udp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.10:53:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {4c1be812-05d3-4f45-91b5-a853a5c8de71}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF0202 09:08:09.911531 6327 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T09:08:08Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b4cfbd4e682bcdcaff785013bdb72580ce36fced733e235538536827036b93d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://579903eadb183466a2564737e6b678e04308239c0cd75b6808892053b97afbc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://579903eadb183466a2564737e6b678e04308239c0cd75b6808892053b97afbc0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zhn7j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:38Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:38 crc kubenswrapper[4764]: I0202 09:08:38.504226 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed747fd8-e3ce-4319-a6da-bda0fd51b85a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6f064bdc66470aace3180d4988244026b54009a63cff199e41c5d02a1d18479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://60258f84a1d8dcfe48b7e9688ed50f077e201971f20f3f6663fbf1ba58a4611a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f84be2e08217a2ac3edde858d78b4410376ed9bae3bcabe2a7f9b2fbe2f69c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a69d25d352d7ceac5c735bc1fcf446a1fea6dc22c65a5296cfab828cea17377\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:38Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:38 crc kubenswrapper[4764]: I0202 09:08:38.507945 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:38 crc kubenswrapper[4764]: I0202 09:08:38.508193 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:38 crc kubenswrapper[4764]: I0202 09:08:38.508309 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:38 crc kubenswrapper[4764]: I0202 09:08:38.508425 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:38 crc kubenswrapper[4764]: I0202 09:08:38.508535 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:38Z","lastTransitionTime":"2026-02-02T09:08:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:38 crc kubenswrapper[4764]: I0202 09:08:38.517278 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f6904228e052fff69ad1fd66df97758b6b60ba4351e97a7fbd03eee4df1b653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:38Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:38 crc kubenswrapper[4764]: I0202 09:08:38.528292 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6zxff" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aaba93da-12a5-43df-b7c7-4c8b988163c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3db293678a907cd55b31b1d0c0bd7c30e1ba75ecd5e13cbc68ab98fbf0fbd9d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqwb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6zxff\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:38Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:38 crc kubenswrapper[4764]: I0202 09:08:38.541501 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d192f670-9f9d-4539-9641-e4bed73acdd4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://040c5e8a5659f85f0d52d1a1e064154d998ae104f711101d2204c0b1caebbb30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbksf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fdf06410349b7c99aa166e87fbe80bf68e2ae923635a6a978b7e2cd50383f7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbksf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4ndm4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:38Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:38 crc kubenswrapper[4764]: I0202 09:08:38.552876 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pcjmn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6e5ccd3b-b657-47a2-9c4d-bc9adb126533\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72863eef568d92a1abeb3cde2e89630a4f5fb7626c6949448ff5195774c98eab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2f7zl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pcjmn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:38Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:38 crc kubenswrapper[4764]: I0202 09:08:38.565590 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fae745b-ea49-42f3-a486-dbc3dc2fb9ad\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://081d7f76d937ff71e35945f9b1d893888bb649d15bd84312fd4b61c9520ea71f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://429ead0993f5e648f6256d249ee12aea03514d3108618840552b672fb419cd74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fad17b01545eb20785f03e3c61bd5c982f6217e38176d7a3b23ac194298644a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85e2f75c56d63cacc93b9195c1e8bf46eb7a7234c45a024bd95a460db90edb6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85e2f75c56d63cacc93b9195c1e8bf46eb7a7234c45a024bd95a460db90edb6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:38Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:38 crc kubenswrapper[4764]: I0202 09:08:38.578615 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:38Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:38 crc kubenswrapper[4764]: I0202 09:08:38.590308 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lh6tm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2152e8a-f456-47b5-a547-b65ec837ecbc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38e176a00b07f2960fdd7ba2aadd1468a5e083b8d154e9d118d13f2585da81e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bb2sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://996e649011d5255783284421ddea8e0a5c383ef675b0920c3fdc0af6503d80e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bb2sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lh6tm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:38Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:38 crc kubenswrapper[4764]: I0202 09:08:38.610914 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:38 crc kubenswrapper[4764]: I0202 09:08:38.610965 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:38 crc kubenswrapper[4764]: I0202 09:08:38.610975 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:38 crc kubenswrapper[4764]: I0202 09:08:38.610988 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:38 crc kubenswrapper[4764]: I0202 09:08:38.611000 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:38Z","lastTransitionTime":"2026-02-02T09:08:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:38 crc kubenswrapper[4764]: I0202 09:08:38.713487 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:38 crc kubenswrapper[4764]: I0202 09:08:38.713734 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:38 crc kubenswrapper[4764]: I0202 09:08:38.713821 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:38 crc kubenswrapper[4764]: I0202 09:08:38.713905 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:38 crc kubenswrapper[4764]: I0202 09:08:38.714003 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:38Z","lastTransitionTime":"2026-02-02T09:08:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:38 crc kubenswrapper[4764]: I0202 09:08:38.816848 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:38 crc kubenswrapper[4764]: I0202 09:08:38.816884 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:38 crc kubenswrapper[4764]: I0202 09:08:38.816895 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:38 crc kubenswrapper[4764]: I0202 09:08:38.816913 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:38 crc kubenswrapper[4764]: I0202 09:08:38.816924 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:38Z","lastTransitionTime":"2026-02-02T09:08:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:38 crc kubenswrapper[4764]: I0202 09:08:38.824154 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-29 14:36:13.913430306 +0000 UTC Feb 02 09:08:38 crc kubenswrapper[4764]: I0202 09:08:38.825348 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 09:08:38 crc kubenswrapper[4764]: I0202 09:08:38.825367 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 09:08:38 crc kubenswrapper[4764]: E0202 09:08:38.825450 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 09:08:38 crc kubenswrapper[4764]: I0202 09:08:38.825356 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 09:08:38 crc kubenswrapper[4764]: E0202 09:08:38.825544 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 09:08:38 crc kubenswrapper[4764]: E0202 09:08:38.825724 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 09:08:38 crc kubenswrapper[4764]: I0202 09:08:38.919556 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:38 crc kubenswrapper[4764]: I0202 09:08:38.919593 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:38 crc kubenswrapper[4764]: I0202 09:08:38.919606 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:38 crc kubenswrapper[4764]: I0202 09:08:38.919622 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:38 crc kubenswrapper[4764]: I0202 09:08:38.919632 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:38Z","lastTransitionTime":"2026-02-02T09:08:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:39 crc kubenswrapper[4764]: I0202 09:08:39.023227 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:39 crc kubenswrapper[4764]: I0202 09:08:39.023683 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:39 crc kubenswrapper[4764]: I0202 09:08:39.023906 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:39 crc kubenswrapper[4764]: I0202 09:08:39.024135 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:39 crc kubenswrapper[4764]: I0202 09:08:39.024346 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:39Z","lastTransitionTime":"2026-02-02T09:08:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:39 crc kubenswrapper[4764]: I0202 09:08:39.127598 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:39 crc kubenswrapper[4764]: I0202 09:08:39.127901 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:39 crc kubenswrapper[4764]: I0202 09:08:39.128213 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:39 crc kubenswrapper[4764]: I0202 09:08:39.128370 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:39 crc kubenswrapper[4764]: I0202 09:08:39.128496 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:39Z","lastTransitionTime":"2026-02-02T09:08:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:39 crc kubenswrapper[4764]: I0202 09:08:39.231233 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:39 crc kubenswrapper[4764]: I0202 09:08:39.231285 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:39 crc kubenswrapper[4764]: I0202 09:08:39.231305 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:39 crc kubenswrapper[4764]: I0202 09:08:39.231329 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:39 crc kubenswrapper[4764]: I0202 09:08:39.231346 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:39Z","lastTransitionTime":"2026-02-02T09:08:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:39 crc kubenswrapper[4764]: I0202 09:08:39.276448 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zhn7j_24632dda-6100-4ab6-a28e-214ddae4360c/ovnkube-controller/3.log" Feb 02 09:08:39 crc kubenswrapper[4764]: I0202 09:08:39.277253 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zhn7j_24632dda-6100-4ab6-a28e-214ddae4360c/ovnkube-controller/2.log" Feb 02 09:08:39 crc kubenswrapper[4764]: I0202 09:08:39.280530 4764 generic.go:334] "Generic (PLEG): container finished" podID="24632dda-6100-4ab6-a28e-214ddae4360c" containerID="3cb255e14eb02f7c08364c41fb9906e58b2155eb507f694b77ec8eb1f4251343" exitCode=1 Feb 02 09:08:39 crc kubenswrapper[4764]: I0202 09:08:39.280577 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" event={"ID":"24632dda-6100-4ab6-a28e-214ddae4360c","Type":"ContainerDied","Data":"3cb255e14eb02f7c08364c41fb9906e58b2155eb507f694b77ec8eb1f4251343"} Feb 02 09:08:39 crc kubenswrapper[4764]: I0202 09:08:39.280617 4764 scope.go:117] "RemoveContainer" containerID="334e828f8ef72570f2511ee38659941202530471f5943aa07b8b21483bd70550" Feb 02 09:08:39 crc kubenswrapper[4764]: I0202 09:08:39.282502 4764 scope.go:117] "RemoveContainer" containerID="3cb255e14eb02f7c08364c41fb9906e58b2155eb507f694b77ec8eb1f4251343" Feb 02 09:08:39 crc kubenswrapper[4764]: E0202 09:08:39.282997 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-zhn7j_openshift-ovn-kubernetes(24632dda-6100-4ab6-a28e-214ddae4360c)\"" pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" podUID="24632dda-6100-4ab6-a28e-214ddae4360c" Feb 02 09:08:39 crc kubenswrapper[4764]: I0202 09:08:39.317660 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d66ab3c0-9b5b-41d5-b6c0-91ccddccd7e9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d92ac486f5dade46bc75de32099859a71fc2ed25a7abbe6ae8937df128e7df30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a4ac273a4f127d79e5fad7d4c30cd2044a346f5734dba5d33d68cd092c0c2d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d38c7cb5dcd22fb07524494ebe77aa1a7f4f808f566043563817e47e7ec12a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://408722c7f83d28919431532f8c797a1c795fd5d96736d14e7f06a05744899702\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f473b18cd8003f75bb1dafb82bcec45745a73ccd10434297356dac185a3d15c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e229f484d19cbdabedf53d90a822a4434db9e00b1301538e885e26acc19d25f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e229f484d19cbdabedf53d90a822a4434db9e00b1301538e885e26acc19d25f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a9cc673be8ff245dba81278e8cd34cfe45ed36ff0b8bd5b30a0ddf333ec6371\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a9cc673be8ff245dba81278e8cd34cfe45ed36ff0b8bd5b30a0ddf333ec6371\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bc7e49c7a5ae66ea939060192fa6a346dc0b0b9aa9750b29879d3c426cc0d276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc7e49c7a5ae66ea939060192fa6a346dc0b0b9aa9750b29879d3c426cc0d276\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:39Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:39 crc kubenswrapper[4764]: I0202 09:08:39.332562 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa9d037e66cb53b9dd698eff46284c9e718fd842cda6f7b525f0c79062d70bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:39Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:39 crc kubenswrapper[4764]: I0202 09:08:39.335415 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:39 crc kubenswrapper[4764]: I0202 09:08:39.335557 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:39 crc kubenswrapper[4764]: I0202 09:08:39.335576 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:39 crc kubenswrapper[4764]: I0202 09:08:39.335632 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:39 crc kubenswrapper[4764]: I0202 09:08:39.335651 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:39Z","lastTransitionTime":"2026-02-02T09:08:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:39 crc kubenswrapper[4764]: I0202 09:08:39.350027 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:39Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:39 crc kubenswrapper[4764]: I0202 09:08:39.379770 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24632dda-6100-4ab6-a28e-214ddae4360c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a6c67e9b7708cc0d85b0b6ffbbade0d34867e159a9c5487ca2606475c667a50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bffda48d7c881c1470251b501661324859373246cc1a21811bd36ca24887c7aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a989645cf485cd8ff80ffff1a86cd7808e1564eba5667535e161cc68db13b26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20b71d3ee7a381baa51b0b0d0470283022cfede30f491135737b5e82d631d763\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c380383b6a014d1b694a13ddc22f54b20b9e4f9d5c9b9000d76187621dce0a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5529f2bff87753428e9173826fc1880acf96a4c134fa9ecad195b15697cad73a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cb255e14eb02f7c08364c41fb9906e58b2155eb507f694b77ec8eb1f4251343\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://334e828f8ef72570f2511ee38659941202530471f5943aa07b8b21483bd70550\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-02T09:08:10Z\\\",\\\"message\\\":\\\"ing OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-dns/dns-default]} name:Service_openshift-dns/dns-default_TCP_node_router+switch_crc options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.10:53: 10.217.4.10:9154:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {be9dcc9e-c16a-4962-a6d2-4adeb0b929c4}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-dns/dns-default]} name:Service_openshift-dns/dns-default_UDP_node_router+switch_crc options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[udp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.10:53:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {4c1be812-05d3-4f45-91b5-a853a5c8de71}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF0202 09:08:09.911531 6327 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T09:08:08Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3cb255e14eb02f7c08364c41fb9906e58b2155eb507f694b77ec8eb1f4251343\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-02T09:08:38Z\\\",\\\"message\\\":\\\":38.747848 6731 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-diagnostics/network-check-target-xd92c\\\\nI0202 09:08:38.747852 6731 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf in node crc\\\\nI0202 09:08:38.747854 6731 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-target-xd92c\\\\nI0202 09:08:38.747780 6731 obj_retry.go:386] Retry successful for *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc after 0 failed attempt(s)\\\\nI0202 09:08:38.747862 6731 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-target-xd92c in node crc\\\\nI0202 09:08:38.747864 6731 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf after 0 failed attempt(s)\\\\nI0202 09:08:38.747869 6731 default_network_controller.go:776] Recording success event on pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI0202 09:08:38.747837 6731 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-additional-cni-plugins-9p5dc in node crc\\\\nI0202 09:08:38.747884 6731 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-target-xd92c] creating logical port openshift-network-diagnostics_network-check-target-\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T09:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b4cfbd4e682bcdcaff785013bdb72580ce36fced733e235538536827036b93d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://579903eadb183466a2564737e6b678e04308239c0cd75b6808892053b97afbc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://579903eadb183466a2564737e6b678e04308239c0cd75b6808892053b97afbc0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zhn7j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:39Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:39 crc kubenswrapper[4764]: I0202 09:08:39.390897 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pcjmn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6e5ccd3b-b657-47a2-9c4d-bc9adb126533\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72863eef568d92a1abeb3cde2e89630a4f5fb7626c6949448ff5195774c98eab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2f7zl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pcjmn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:39Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:39 crc kubenswrapper[4764]: I0202 09:08:39.408081 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed747fd8-e3ce-4319-a6da-bda0fd51b85a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6f064bdc66470aace3180d4988244026b54009a63cff199e41c5d02a1d18479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://60258f84a1d8dcfe48b7e9688ed50f077e201971f20f3f6663fbf1ba58a4611a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f84be2e08217a2ac3edde858d78b4410376ed9bae3bcabe2a7f9b2fbe2f69c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a69d25d352d7ceac5c735bc1fcf446a1fea6dc22c65a5296cfab828cea17377\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:39Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:39 crc kubenswrapper[4764]: I0202 09:08:39.420671 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f6904228e052fff69ad1fd66df97758b6b60ba4351e97a7fbd03eee4df1b653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:39Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:39 crc kubenswrapper[4764]: I0202 09:08:39.431386 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6zxff" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aaba93da-12a5-43df-b7c7-4c8b988163c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3db293678a907cd55b31b1d0c0bd7c30e1ba75ecd5e13cbc68ab98fbf0fbd9d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqwb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6zxff\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:39Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:39 crc kubenswrapper[4764]: I0202 09:08:39.438364 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:39 crc kubenswrapper[4764]: I0202 09:08:39.438397 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:39 crc kubenswrapper[4764]: I0202 09:08:39.438426 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:39 crc kubenswrapper[4764]: I0202 09:08:39.438456 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:39 crc kubenswrapper[4764]: I0202 09:08:39.438465 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:39Z","lastTransitionTime":"2026-02-02T09:08:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:39 crc kubenswrapper[4764]: I0202 09:08:39.442960 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d192f670-9f9d-4539-9641-e4bed73acdd4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://040c5e8a5659f85f0d52d1a1e064154d998ae104f711101d2204c0b1caebbb30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbksf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fdf06410349b7c99aa166e87fbe80bf68e2ae923635a6a978b7e2cd50383f7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbksf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4ndm4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:39Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:39 crc kubenswrapper[4764]: I0202 09:08:39.454670 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fae745b-ea49-42f3-a486-dbc3dc2fb9ad\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://081d7f76d937ff71e35945f9b1d893888bb649d15bd84312fd4b61c9520ea71f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://429ead0993f5e648f6256d249ee12aea03514d3108618840552b672fb419cd74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fad17b01545eb20785f03e3c61bd5c982f6217e38176d7a3b23ac194298644a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85e2f75c56d63cacc93b9195c1e8bf46eb7a7234c45a024bd95a460db90edb6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85e2f75c56d63cacc93b9195c1e8bf46eb7a7234c45a024bd95a460db90edb6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:39Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:39 crc kubenswrapper[4764]: I0202 09:08:39.468318 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:39Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:39 crc kubenswrapper[4764]: I0202 09:08:39.478771 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lh6tm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2152e8a-f456-47b5-a547-b65ec837ecbc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38e176a00b07f2960fdd7ba2aadd1468a5e083b8d154e9d118d13f2585da81e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bb2sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://996e649011d5255783284421ddea8e0a5c383ef675b0920c3fdc0af6503d80e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bb2sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lh6tm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:39Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:39 crc kubenswrapper[4764]: I0202 09:08:39.492335 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nc7qt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b52b499d17b4e2e359475bd13730c72ef5cedb052e13def6dfe5bc5fd7487d2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1a2c7b22068b7165927fc158885d48bf9952b9eba521d8ae9a599eee53dc305\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-02T09:08:28Z\\\",\\\"message\\\":\\\"2026-02-02T09:07:42+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_32cbaf6a-865c-4787-8f89-a40249873b40\\\\n2026-02-02T09:07:42+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_32cbaf6a-865c-4787-8f89-a40249873b40 to /host/opt/cni/bin/\\\\n2026-02-02T09:07:43Z [verbose] multus-daemon started\\\\n2026-02-02T09:07:43Z [verbose] Readiness Indicator file check\\\\n2026-02-02T09:08:28Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:08:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2hz4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nc7qt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:39Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:39 crc kubenswrapper[4764]: I0202 09:08:39.507209 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9p5dc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"660ac006-2495-45c4-9fb3-e9c1dddcf7a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da812b124c97aa8ad1563f4bcfac8f3bfa025096fa4a2561e5847b1bf9cb30b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23e57e65a6f5c3059506b9db21e9ac2492c30bd9191a98db0281915e9ce39797\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23e57e65a6f5c3059506b9db21e9ac2492c30bd9191a98db0281915e9ce39797\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2417b66816da1a6cd1791a4b91eca4a3c4f3694462e066e7f0722f1b9c2105c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2417b66816da1a6cd1791a4b91eca4a3c4f3694462e066e7f0722f1b9c2105c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://883641f29e9120c2cea6b272bacf1dc63c752f55ab0408cdaf73db85c85ac369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://883641f29e9120c2cea6b272bacf1dc63c752f55ab0408cdaf73db85c85ac369\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2745baba1188ba5e8f820616e4117a12c50ecee5b68e71ad9cabbb621b6e131f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2745baba1188ba5e8f820616e4117a12c50ecee5b68e71ad9cabbb621b6e131f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8925e435be4d48d86824fc35f8b8819fb35d63e385164c4a20edb3bf3035c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a8925e435be4d48d86824fc35f8b8819fb35d63e385164c4a20edb3bf3035c3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d3fbf668ae7d04c8d9e574f43c2fe90c5604204ba38a3c60847afdfbceb7a57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d3fbf668ae7d04c8d9e574f43c2fe90c5604204ba38a3c60847afdfbceb7a57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9p5dc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:39Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:39 crc kubenswrapper[4764]: I0202 09:08:39.516836 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kwtmr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"991faa9a-dd25-4f49-82bd-ce60cefd4af2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h865b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h865b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:53Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kwtmr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:39Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:39 crc kubenswrapper[4764]: I0202 09:08:39.530119 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87d61aea-46f6-4ee6-a7a4-767a6a67fea4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e78b6d08e4d616cefbb73e3fd43b3c6886dcd8aa997c7b4948bf1ff63401293\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d468715c2da33996561c40043216233fd12e21539d0c58dd1f3b892fd9f91a9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b184f7c936dc4046f74c0018269cbc2c4b7ceddb650b4e871d3f8a707e0df88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://269c5074371bd00f8363fb3888c11f1df383081a61497aea6578f08700881d50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b25bc9395831524c53f9e565b4efbe74f21792efafeb1dd507161bc91aa0ecdf\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T09:07:22Z\\\",\\\"message\\\":\\\"W0202 09:07:21.597488 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0202 09:07:21.598117 1 crypto.go:601] Generating new CA for check-endpoints-signer@1770023241 cert, and key in /tmp/serving-cert-3856183724/serving-signer.crt, /tmp/serving-cert-3856183724/serving-signer.key\\\\nI0202 09:07:22.012184 1 observer_polling.go:159] Starting file observer\\\\nW0202 09:07:22.015493 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0202 09:07:22.015797 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 09:07:22.016596 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3856183724/tls.crt::/tmp/serving-cert-3856183724/tls.key\\\\\\\"\\\\nF0202 09:07:22.270605 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b44e17b8ca01f549bc10081dd5f109cb1b9fff364bf5bd4fd90d9f0e3eeb04db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfb06a5ec9262cf39b108a66014d15e325cdbd5441a2d9e7a811dd57265c99a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfb06a5ec9262cf39b108a66014d15e325cdbd5441a2d9e7a811dd57265c99a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:39Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:39 crc kubenswrapper[4764]: I0202 09:08:39.540527 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fea07bdc-9134-4993-934c-316c19d0dc76\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://664e0519c2a2c60904dfcc8deebaf5d638dac0bedb13d870c0aaea5b4f5428f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db5a54d3a2117252a1ff50884683f99185874d2904c0960161642280f614679d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db5a54d3a2117252a1ff50884683f99185874d2904c0960161642280f614679d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:39Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:39 crc kubenswrapper[4764]: I0202 09:08:39.541045 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:39 crc kubenswrapper[4764]: I0202 09:08:39.541069 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:39 crc kubenswrapper[4764]: I0202 09:08:39.541077 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:39 crc kubenswrapper[4764]: I0202 09:08:39.541091 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:39 crc kubenswrapper[4764]: I0202 09:08:39.541101 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:39Z","lastTransitionTime":"2026-02-02T09:08:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:39 crc kubenswrapper[4764]: I0202 09:08:39.554214 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5528225d91ac2778bef4554abb0bd036ac5d3b6cdac7357bab69e70c3894b6f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://071bbb80d0747cf3e109ba626d47f7d05f85f4cc0b3aba77276c795741f47ac6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:39Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:39 crc kubenswrapper[4764]: I0202 09:08:39.566741 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:39Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:39 crc kubenswrapper[4764]: I0202 09:08:39.643973 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:39 crc kubenswrapper[4764]: I0202 09:08:39.644275 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:39 crc kubenswrapper[4764]: I0202 09:08:39.644405 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:39 crc kubenswrapper[4764]: I0202 09:08:39.644531 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:39 crc kubenswrapper[4764]: I0202 09:08:39.644647 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:39Z","lastTransitionTime":"2026-02-02T09:08:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:39 crc kubenswrapper[4764]: I0202 09:08:39.747349 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:39 crc kubenswrapper[4764]: I0202 09:08:39.747384 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:39 crc kubenswrapper[4764]: I0202 09:08:39.747395 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:39 crc kubenswrapper[4764]: I0202 09:08:39.747407 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:39 crc kubenswrapper[4764]: I0202 09:08:39.747416 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:39Z","lastTransitionTime":"2026-02-02T09:08:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:39 crc kubenswrapper[4764]: I0202 09:08:39.824362 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-11 18:48:23.519341851 +0000 UTC Feb 02 09:08:39 crc kubenswrapper[4764]: I0202 09:08:39.824724 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kwtmr" Feb 02 09:08:39 crc kubenswrapper[4764]: E0202 09:08:39.824854 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kwtmr" podUID="991faa9a-dd25-4f49-82bd-ce60cefd4af2" Feb 02 09:08:39 crc kubenswrapper[4764]: I0202 09:08:39.849082 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:39 crc kubenswrapper[4764]: I0202 09:08:39.849114 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:39 crc kubenswrapper[4764]: I0202 09:08:39.849123 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:39 crc kubenswrapper[4764]: I0202 09:08:39.849135 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:39 crc kubenswrapper[4764]: I0202 09:08:39.849144 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:39Z","lastTransitionTime":"2026-02-02T09:08:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:39 crc kubenswrapper[4764]: I0202 09:08:39.951334 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:39 crc kubenswrapper[4764]: I0202 09:08:39.951363 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:39 crc kubenswrapper[4764]: I0202 09:08:39.951372 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:39 crc kubenswrapper[4764]: I0202 09:08:39.951384 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:39 crc kubenswrapper[4764]: I0202 09:08:39.951393 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:39Z","lastTransitionTime":"2026-02-02T09:08:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:40 crc kubenswrapper[4764]: I0202 09:08:40.054781 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:40 crc kubenswrapper[4764]: I0202 09:08:40.054832 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:40 crc kubenswrapper[4764]: I0202 09:08:40.054848 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:40 crc kubenswrapper[4764]: I0202 09:08:40.054874 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:40 crc kubenswrapper[4764]: I0202 09:08:40.054891 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:40Z","lastTransitionTime":"2026-02-02T09:08:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:40 crc kubenswrapper[4764]: I0202 09:08:40.158194 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:40 crc kubenswrapper[4764]: I0202 09:08:40.158479 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:40 crc kubenswrapper[4764]: I0202 09:08:40.158666 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:40 crc kubenswrapper[4764]: I0202 09:08:40.158824 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:40 crc kubenswrapper[4764]: I0202 09:08:40.158986 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:40Z","lastTransitionTime":"2026-02-02T09:08:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:40 crc kubenswrapper[4764]: I0202 09:08:40.262083 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:40 crc kubenswrapper[4764]: I0202 09:08:40.262318 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:40 crc kubenswrapper[4764]: I0202 09:08:40.262437 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:40 crc kubenswrapper[4764]: I0202 09:08:40.262533 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:40 crc kubenswrapper[4764]: I0202 09:08:40.262616 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:40Z","lastTransitionTime":"2026-02-02T09:08:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:40 crc kubenswrapper[4764]: I0202 09:08:40.287597 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zhn7j_24632dda-6100-4ab6-a28e-214ddae4360c/ovnkube-controller/3.log" Feb 02 09:08:40 crc kubenswrapper[4764]: I0202 09:08:40.293117 4764 scope.go:117] "RemoveContainer" containerID="3cb255e14eb02f7c08364c41fb9906e58b2155eb507f694b77ec8eb1f4251343" Feb 02 09:08:40 crc kubenswrapper[4764]: E0202 09:08:40.293375 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-zhn7j_openshift-ovn-kubernetes(24632dda-6100-4ab6-a28e-214ddae4360c)\"" pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" podUID="24632dda-6100-4ab6-a28e-214ddae4360c" Feb 02 09:08:40 crc kubenswrapper[4764]: I0202 09:08:40.313049 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fae745b-ea49-42f3-a486-dbc3dc2fb9ad\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://081d7f76d937ff71e35945f9b1d893888bb649d15bd84312fd4b61c9520ea71f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://429ead0993f5e648f6256d249ee12aea03514d3108618840552b672fb419cd74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fad17b01545eb20785f03e3c61bd5c982f6217e38176d7a3b23ac194298644a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85e2f75c56d63cacc93b9195c1e8bf46eb7a7234c45a024bd95a460db90edb6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85e2f75c56d63cacc93b9195c1e8bf46eb7a7234c45a024bd95a460db90edb6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:40Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:40 crc kubenswrapper[4764]: I0202 09:08:40.335719 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:40Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:40 crc kubenswrapper[4764]: I0202 09:08:40.356686 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lh6tm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2152e8a-f456-47b5-a547-b65ec837ecbc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38e176a00b07f2960fdd7ba2aadd1468a5e083b8d154e9d118d13f2585da81e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bb2sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://996e649011d5255783284421ddea8e0a5c383ef675b0920c3fdc0af6503d80e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bb2sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lh6tm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:40Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:40 crc kubenswrapper[4764]: I0202 09:08:40.365997 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:40 crc kubenswrapper[4764]: I0202 09:08:40.366062 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:40 crc kubenswrapper[4764]: I0202 09:08:40.366083 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:40 crc kubenswrapper[4764]: I0202 09:08:40.366110 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:40 crc kubenswrapper[4764]: I0202 09:08:40.366128 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:40Z","lastTransitionTime":"2026-02-02T09:08:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:40 crc kubenswrapper[4764]: I0202 09:08:40.376970 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nc7qt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b52b499d17b4e2e359475bd13730c72ef5cedb052e13def6dfe5bc5fd7487d2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1a2c7b22068b7165927fc158885d48bf9952b9eba521d8ae9a599eee53dc305\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-02T09:08:28Z\\\",\\\"message\\\":\\\"2026-02-02T09:07:42+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_32cbaf6a-865c-4787-8f89-a40249873b40\\\\n2026-02-02T09:07:42+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_32cbaf6a-865c-4787-8f89-a40249873b40 to /host/opt/cni/bin/\\\\n2026-02-02T09:07:43Z [verbose] multus-daemon started\\\\n2026-02-02T09:07:43Z [verbose] Readiness Indicator file check\\\\n2026-02-02T09:08:28Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:08:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2hz4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nc7qt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:40Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:40 crc kubenswrapper[4764]: I0202 09:08:40.392638 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9p5dc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"660ac006-2495-45c4-9fb3-e9c1dddcf7a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da812b124c97aa8ad1563f4bcfac8f3bfa025096fa4a2561e5847b1bf9cb30b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23e57e65a6f5c3059506b9db21e9ac2492c30bd9191a98db0281915e9ce39797\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23e57e65a6f5c3059506b9db21e9ac2492c30bd9191a98db0281915e9ce39797\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2417b66816da1a6cd1791a4b91eca4a3c4f3694462e066e7f0722f1b9c2105c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2417b66816da1a6cd1791a4b91eca4a3c4f3694462e066e7f0722f1b9c2105c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://883641f29e9120c2cea6b272bacf1dc63c752f55ab0408cdaf73db85c85ac369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://883641f29e9120c2cea6b272bacf1dc63c752f55ab0408cdaf73db85c85ac369\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2745baba1188ba5e8f820616e4117a12c50ecee5b68e71ad9cabbb621b6e131f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2745baba1188ba5e8f820616e4117a12c50ecee5b68e71ad9cabbb621b6e131f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8925e435be4d48d86824fc35f8b8819fb35d63e385164c4a20edb3bf3035c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a8925e435be4d48d86824fc35f8b8819fb35d63e385164c4a20edb3bf3035c3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d3fbf668ae7d04c8d9e574f43c2fe90c5604204ba38a3c60847afdfbceb7a57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d3fbf668ae7d04c8d9e574f43c2fe90c5604204ba38a3c60847afdfbceb7a57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9p5dc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:40Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:40 crc kubenswrapper[4764]: I0202 09:08:40.406178 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kwtmr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"991faa9a-dd25-4f49-82bd-ce60cefd4af2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h865b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h865b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:53Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kwtmr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:40Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:40 crc kubenswrapper[4764]: I0202 09:08:40.420249 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87d61aea-46f6-4ee6-a7a4-767a6a67fea4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e78b6d08e4d616cefbb73e3fd43b3c6886dcd8aa997c7b4948bf1ff63401293\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d468715c2da33996561c40043216233fd12e21539d0c58dd1f3b892fd9f91a9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b184f7c936dc4046f74c0018269cbc2c4b7ceddb650b4e871d3f8a707e0df88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://269c5074371bd00f8363fb3888c11f1df383081a61497aea6578f08700881d50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b25bc9395831524c53f9e565b4efbe74f21792efafeb1dd507161bc91aa0ecdf\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T09:07:22Z\\\",\\\"message\\\":\\\"W0202 09:07:21.597488 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0202 09:07:21.598117 1 crypto.go:601] Generating new CA for check-endpoints-signer@1770023241 cert, and key in /tmp/serving-cert-3856183724/serving-signer.crt, /tmp/serving-cert-3856183724/serving-signer.key\\\\nI0202 09:07:22.012184 1 observer_polling.go:159] Starting file observer\\\\nW0202 09:07:22.015493 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0202 09:07:22.015797 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 09:07:22.016596 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3856183724/tls.crt::/tmp/serving-cert-3856183724/tls.key\\\\\\\"\\\\nF0202 09:07:22.270605 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b44e17b8ca01f549bc10081dd5f109cb1b9fff364bf5bd4fd90d9f0e3eeb04db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfb06a5ec9262cf39b108a66014d15e325cdbd5441a2d9e7a811dd57265c99a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfb06a5ec9262cf39b108a66014d15e325cdbd5441a2d9e7a811dd57265c99a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:40Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:40 crc kubenswrapper[4764]: I0202 09:08:40.432587 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fea07bdc-9134-4993-934c-316c19d0dc76\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://664e0519c2a2c60904dfcc8deebaf5d638dac0bedb13d870c0aaea5b4f5428f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db5a54d3a2117252a1ff50884683f99185874d2904c0960161642280f614679d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db5a54d3a2117252a1ff50884683f99185874d2904c0960161642280f614679d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:40Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:40 crc kubenswrapper[4764]: I0202 09:08:40.446809 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5528225d91ac2778bef4554abb0bd036ac5d3b6cdac7357bab69e70c3894b6f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://071bbb80d0747cf3e109ba626d47f7d05f85f4cc0b3aba77276c795741f47ac6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:40Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:40 crc kubenswrapper[4764]: I0202 09:08:40.459895 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:40Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:40 crc kubenswrapper[4764]: I0202 09:08:40.468590 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:40 crc kubenswrapper[4764]: I0202 09:08:40.468617 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:40 crc kubenswrapper[4764]: I0202 09:08:40.468626 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:40 crc kubenswrapper[4764]: I0202 09:08:40.468639 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:40 crc kubenswrapper[4764]: I0202 09:08:40.468647 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:40Z","lastTransitionTime":"2026-02-02T09:08:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:40 crc kubenswrapper[4764]: I0202 09:08:40.479708 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d66ab3c0-9b5b-41d5-b6c0-91ccddccd7e9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d92ac486f5dade46bc75de32099859a71fc2ed25a7abbe6ae8937df128e7df30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a4ac273a4f127d79e5fad7d4c30cd2044a346f5734dba5d33d68cd092c0c2d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d38c7cb5dcd22fb07524494ebe77aa1a7f4f808f566043563817e47e7ec12a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://408722c7f83d28919431532f8c797a1c795fd5d96736d14e7f06a05744899702\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f473b18cd8003f75bb1dafb82bcec45745a73ccd10434297356dac185a3d15c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e229f484d19cbdabedf53d90a822a4434db9e00b1301538e885e26acc19d25f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e229f484d19cbdabedf53d90a822a4434db9e00b1301538e885e26acc19d25f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a9cc673be8ff245dba81278e8cd34cfe45ed36ff0b8bd5b30a0ddf333ec6371\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a9cc673be8ff245dba81278e8cd34cfe45ed36ff0b8bd5b30a0ddf333ec6371\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bc7e49c7a5ae66ea939060192fa6a346dc0b0b9aa9750b29879d3c426cc0d276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc7e49c7a5ae66ea939060192fa6a346dc0b0b9aa9750b29879d3c426cc0d276\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:40Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:40 crc kubenswrapper[4764]: I0202 09:08:40.496768 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa9d037e66cb53b9dd698eff46284c9e718fd842cda6f7b525f0c79062d70bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:40Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:40 crc kubenswrapper[4764]: I0202 09:08:40.510332 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:40Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:40 crc kubenswrapper[4764]: I0202 09:08:40.535300 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24632dda-6100-4ab6-a28e-214ddae4360c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a6c67e9b7708cc0d85b0b6ffbbade0d34867e159a9c5487ca2606475c667a50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bffda48d7c881c1470251b501661324859373246cc1a21811bd36ca24887c7aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a989645cf485cd8ff80ffff1a86cd7808e1564eba5667535e161cc68db13b26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20b71d3ee7a381baa51b0b0d0470283022cfede30f491135737b5e82d631d763\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c380383b6a014d1b694a13ddc22f54b20b9e4f9d5c9b9000d76187621dce0a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5529f2bff87753428e9173826fc1880acf96a4c134fa9ecad195b15697cad73a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cb255e14eb02f7c08364c41fb9906e58b2155eb507f694b77ec8eb1f4251343\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3cb255e14eb02f7c08364c41fb9906e58b2155eb507f694b77ec8eb1f4251343\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-02T09:08:38Z\\\",\\\"message\\\":\\\":38.747848 6731 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-diagnostics/network-check-target-xd92c\\\\nI0202 09:08:38.747852 6731 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf in node crc\\\\nI0202 09:08:38.747854 6731 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-target-xd92c\\\\nI0202 09:08:38.747780 6731 obj_retry.go:386] Retry successful for *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc after 0 failed attempt(s)\\\\nI0202 09:08:38.747862 6731 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-target-xd92c in node crc\\\\nI0202 09:08:38.747864 6731 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf after 0 failed attempt(s)\\\\nI0202 09:08:38.747869 6731 default_network_controller.go:776] Recording success event on pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI0202 09:08:38.747837 6731 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-additional-cni-plugins-9p5dc in node crc\\\\nI0202 09:08:38.747884 6731 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-target-xd92c] creating logical port openshift-network-diagnostics_network-check-target-\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T09:08:37Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-zhn7j_openshift-ovn-kubernetes(24632dda-6100-4ab6-a28e-214ddae4360c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b4cfbd4e682bcdcaff785013bdb72580ce36fced733e235538536827036b93d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://579903eadb183466a2564737e6b678e04308239c0cd75b6808892053b97afbc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://579903eadb183466a2564737e6b678e04308239c0cd75b6808892053b97afbc0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zhn7j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:40Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:40 crc kubenswrapper[4764]: I0202 09:08:40.546908 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pcjmn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6e5ccd3b-b657-47a2-9c4d-bc9adb126533\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72863eef568d92a1abeb3cde2e89630a4f5fb7626c6949448ff5195774c98eab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2f7zl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pcjmn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:40Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:40 crc kubenswrapper[4764]: I0202 09:08:40.560835 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed747fd8-e3ce-4319-a6da-bda0fd51b85a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6f064bdc66470aace3180d4988244026b54009a63cff199e41c5d02a1d18479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://60258f84a1d8dcfe48b7e9688ed50f077e201971f20f3f6663fbf1ba58a4611a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f84be2e08217a2ac3edde858d78b4410376ed9bae3bcabe2a7f9b2fbe2f69c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a69d25d352d7ceac5c735bc1fcf446a1fea6dc22c65a5296cfab828cea17377\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:40Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:40 crc kubenswrapper[4764]: I0202 09:08:40.571030 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:40 crc kubenswrapper[4764]: I0202 09:08:40.571074 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:40 crc kubenswrapper[4764]: I0202 09:08:40.571090 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:40 crc kubenswrapper[4764]: I0202 09:08:40.571114 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:40 crc kubenswrapper[4764]: I0202 09:08:40.571129 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:40Z","lastTransitionTime":"2026-02-02T09:08:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:40 crc kubenswrapper[4764]: I0202 09:08:40.574675 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f6904228e052fff69ad1fd66df97758b6b60ba4351e97a7fbd03eee4df1b653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:40Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:40 crc kubenswrapper[4764]: I0202 09:08:40.587389 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6zxff" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aaba93da-12a5-43df-b7c7-4c8b988163c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3db293678a907cd55b31b1d0c0bd7c30e1ba75ecd5e13cbc68ab98fbf0fbd9d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqwb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6zxff\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:40Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:40 crc kubenswrapper[4764]: I0202 09:08:40.602442 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d192f670-9f9d-4539-9641-e4bed73acdd4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://040c5e8a5659f85f0d52d1a1e064154d998ae104f711101d2204c0b1caebbb30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbksf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fdf06410349b7c99aa166e87fbe80bf68e2ae923635a6a978b7e2cd50383f7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbksf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4ndm4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:40Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:40 crc kubenswrapper[4764]: I0202 09:08:40.673514 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:40 crc kubenswrapper[4764]: I0202 09:08:40.673554 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:40 crc kubenswrapper[4764]: I0202 09:08:40.673565 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:40 crc kubenswrapper[4764]: I0202 09:08:40.673579 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:40 crc kubenswrapper[4764]: I0202 09:08:40.673589 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:40Z","lastTransitionTime":"2026-02-02T09:08:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:40 crc kubenswrapper[4764]: I0202 09:08:40.777038 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:40 crc kubenswrapper[4764]: I0202 09:08:40.777163 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:40 crc kubenswrapper[4764]: I0202 09:08:40.777186 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:40 crc kubenswrapper[4764]: I0202 09:08:40.777210 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:40 crc kubenswrapper[4764]: I0202 09:08:40.777229 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:40Z","lastTransitionTime":"2026-02-02T09:08:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:40 crc kubenswrapper[4764]: I0202 09:08:40.824751 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-20 21:19:36.512082021 +0000 UTC Feb 02 09:08:40 crc kubenswrapper[4764]: I0202 09:08:40.825081 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 09:08:40 crc kubenswrapper[4764]: I0202 09:08:40.825081 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 09:08:40 crc kubenswrapper[4764]: I0202 09:08:40.825177 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 09:08:40 crc kubenswrapper[4764]: E0202 09:08:40.825540 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 09:08:40 crc kubenswrapper[4764]: E0202 09:08:40.826552 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 09:08:40 crc kubenswrapper[4764]: E0202 09:08:40.826837 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 09:08:40 crc kubenswrapper[4764]: I0202 09:08:40.879858 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:40 crc kubenswrapper[4764]: I0202 09:08:40.879892 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:40 crc kubenswrapper[4764]: I0202 09:08:40.879901 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:40 crc kubenswrapper[4764]: I0202 09:08:40.879913 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:40 crc kubenswrapper[4764]: I0202 09:08:40.879921 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:40Z","lastTransitionTime":"2026-02-02T09:08:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:40 crc kubenswrapper[4764]: I0202 09:08:40.982180 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:40 crc kubenswrapper[4764]: I0202 09:08:40.982224 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:40 crc kubenswrapper[4764]: I0202 09:08:40.982237 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:40 crc kubenswrapper[4764]: I0202 09:08:40.982254 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:40 crc kubenswrapper[4764]: I0202 09:08:40.982267 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:40Z","lastTransitionTime":"2026-02-02T09:08:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:41 crc kubenswrapper[4764]: I0202 09:08:41.085236 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:41 crc kubenswrapper[4764]: I0202 09:08:41.085306 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:41 crc kubenswrapper[4764]: I0202 09:08:41.085323 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:41 crc kubenswrapper[4764]: I0202 09:08:41.085350 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:41 crc kubenswrapper[4764]: I0202 09:08:41.085367 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:41Z","lastTransitionTime":"2026-02-02T09:08:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:41 crc kubenswrapper[4764]: I0202 09:08:41.188520 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:41 crc kubenswrapper[4764]: I0202 09:08:41.188583 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:41 crc kubenswrapper[4764]: I0202 09:08:41.188602 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:41 crc kubenswrapper[4764]: I0202 09:08:41.188629 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:41 crc kubenswrapper[4764]: I0202 09:08:41.188649 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:41Z","lastTransitionTime":"2026-02-02T09:08:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:41 crc kubenswrapper[4764]: I0202 09:08:41.290809 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:41 crc kubenswrapper[4764]: I0202 09:08:41.290864 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:41 crc kubenswrapper[4764]: I0202 09:08:41.290873 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:41 crc kubenswrapper[4764]: I0202 09:08:41.290887 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:41 crc kubenswrapper[4764]: I0202 09:08:41.290898 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:41Z","lastTransitionTime":"2026-02-02T09:08:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:41 crc kubenswrapper[4764]: I0202 09:08:41.393480 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:41 crc kubenswrapper[4764]: I0202 09:08:41.393520 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:41 crc kubenswrapper[4764]: I0202 09:08:41.393531 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:41 crc kubenswrapper[4764]: I0202 09:08:41.393547 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:41 crc kubenswrapper[4764]: I0202 09:08:41.393560 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:41Z","lastTransitionTime":"2026-02-02T09:08:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:41 crc kubenswrapper[4764]: I0202 09:08:41.497228 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:41 crc kubenswrapper[4764]: I0202 09:08:41.497592 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:41 crc kubenswrapper[4764]: I0202 09:08:41.497768 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:41 crc kubenswrapper[4764]: I0202 09:08:41.497964 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:41 crc kubenswrapper[4764]: I0202 09:08:41.498158 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:41Z","lastTransitionTime":"2026-02-02T09:08:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:41 crc kubenswrapper[4764]: I0202 09:08:41.601137 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:41 crc kubenswrapper[4764]: I0202 09:08:41.601230 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:41 crc kubenswrapper[4764]: I0202 09:08:41.601258 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:41 crc kubenswrapper[4764]: I0202 09:08:41.601312 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:41 crc kubenswrapper[4764]: I0202 09:08:41.601343 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:41Z","lastTransitionTime":"2026-02-02T09:08:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:41 crc kubenswrapper[4764]: I0202 09:08:41.704375 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:41 crc kubenswrapper[4764]: I0202 09:08:41.704424 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:41 crc kubenswrapper[4764]: I0202 09:08:41.704436 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:41 crc kubenswrapper[4764]: I0202 09:08:41.704458 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:41 crc kubenswrapper[4764]: I0202 09:08:41.704470 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:41Z","lastTransitionTime":"2026-02-02T09:08:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:41 crc kubenswrapper[4764]: I0202 09:08:41.807902 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:41 crc kubenswrapper[4764]: I0202 09:08:41.808012 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:41 crc kubenswrapper[4764]: I0202 09:08:41.808038 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:41 crc kubenswrapper[4764]: I0202 09:08:41.808069 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:41 crc kubenswrapper[4764]: I0202 09:08:41.808092 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:41Z","lastTransitionTime":"2026-02-02T09:08:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:41 crc kubenswrapper[4764]: I0202 09:08:41.825048 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-15 05:36:45.912331115 +0000 UTC Feb 02 09:08:41 crc kubenswrapper[4764]: I0202 09:08:41.825380 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kwtmr" Feb 02 09:08:41 crc kubenswrapper[4764]: E0202 09:08:41.826150 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kwtmr" podUID="991faa9a-dd25-4f49-82bd-ce60cefd4af2" Feb 02 09:08:41 crc kubenswrapper[4764]: I0202 09:08:41.911162 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:41 crc kubenswrapper[4764]: I0202 09:08:41.911244 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:41 crc kubenswrapper[4764]: I0202 09:08:41.911262 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:41 crc kubenswrapper[4764]: I0202 09:08:41.911290 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:41 crc kubenswrapper[4764]: I0202 09:08:41.911311 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:41Z","lastTransitionTime":"2026-02-02T09:08:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:42 crc kubenswrapper[4764]: I0202 09:08:42.014671 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:42 crc kubenswrapper[4764]: I0202 09:08:42.014749 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:42 crc kubenswrapper[4764]: I0202 09:08:42.014767 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:42 crc kubenswrapper[4764]: I0202 09:08:42.014793 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:42 crc kubenswrapper[4764]: I0202 09:08:42.014811 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:42Z","lastTransitionTime":"2026-02-02T09:08:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:42 crc kubenswrapper[4764]: I0202 09:08:42.119390 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:42 crc kubenswrapper[4764]: I0202 09:08:42.119455 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:42 crc kubenswrapper[4764]: I0202 09:08:42.119474 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:42 crc kubenswrapper[4764]: I0202 09:08:42.119500 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:42 crc kubenswrapper[4764]: I0202 09:08:42.119518 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:42Z","lastTransitionTime":"2026-02-02T09:08:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:42 crc kubenswrapper[4764]: I0202 09:08:42.224907 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:42 crc kubenswrapper[4764]: I0202 09:08:42.225206 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:42 crc kubenswrapper[4764]: I0202 09:08:42.225287 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:42 crc kubenswrapper[4764]: I0202 09:08:42.225356 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:42 crc kubenswrapper[4764]: I0202 09:08:42.225426 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:42Z","lastTransitionTime":"2026-02-02T09:08:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:42 crc kubenswrapper[4764]: I0202 09:08:42.327031 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:42 crc kubenswrapper[4764]: I0202 09:08:42.327340 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:42 crc kubenswrapper[4764]: I0202 09:08:42.327353 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:42 crc kubenswrapper[4764]: I0202 09:08:42.327366 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:42 crc kubenswrapper[4764]: I0202 09:08:42.327375 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:42Z","lastTransitionTime":"2026-02-02T09:08:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:42 crc kubenswrapper[4764]: I0202 09:08:42.429862 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:42 crc kubenswrapper[4764]: I0202 09:08:42.429905 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:42 crc kubenswrapper[4764]: I0202 09:08:42.429973 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:42 crc kubenswrapper[4764]: I0202 09:08:42.429990 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:42 crc kubenswrapper[4764]: I0202 09:08:42.430002 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:42Z","lastTransitionTime":"2026-02-02T09:08:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:42 crc kubenswrapper[4764]: I0202 09:08:42.532744 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:42 crc kubenswrapper[4764]: I0202 09:08:42.533276 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:42 crc kubenswrapper[4764]: I0202 09:08:42.533676 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:42 crc kubenswrapper[4764]: I0202 09:08:42.533903 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:42 crc kubenswrapper[4764]: I0202 09:08:42.534085 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:42Z","lastTransitionTime":"2026-02-02T09:08:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:42 crc kubenswrapper[4764]: I0202 09:08:42.592905 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 09:08:42 crc kubenswrapper[4764]: E0202 09:08:42.593189 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 09:09:46.593141111 +0000 UTC m=+149.526865249 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:08:42 crc kubenswrapper[4764]: I0202 09:08:42.636472 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:42 crc kubenswrapper[4764]: I0202 09:08:42.636535 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:42 crc kubenswrapper[4764]: I0202 09:08:42.636552 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:42 crc kubenswrapper[4764]: I0202 09:08:42.636628 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:42 crc kubenswrapper[4764]: I0202 09:08:42.636648 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:42Z","lastTransitionTime":"2026-02-02T09:08:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:42 crc kubenswrapper[4764]: I0202 09:08:42.694600 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 09:08:42 crc kubenswrapper[4764]: I0202 09:08:42.694648 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 09:08:42 crc kubenswrapper[4764]: I0202 09:08:42.694673 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 09:08:42 crc kubenswrapper[4764]: I0202 09:08:42.694702 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 09:08:42 crc kubenswrapper[4764]: E0202 09:08:42.694811 4764 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 02 09:08:42 crc kubenswrapper[4764]: E0202 09:08:42.694842 4764 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 02 09:08:42 crc kubenswrapper[4764]: E0202 09:08:42.694865 4764 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 02 09:08:42 crc kubenswrapper[4764]: E0202 09:08:42.694876 4764 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 02 09:08:42 crc kubenswrapper[4764]: E0202 09:08:42.694908 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-02 09:09:46.694888657 +0000 UTC m=+149.628612745 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 02 09:08:42 crc kubenswrapper[4764]: E0202 09:08:42.694947 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-02 09:09:46.694921148 +0000 UTC m=+149.628645376 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 02 09:08:42 crc kubenswrapper[4764]: E0202 09:08:42.694924 4764 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 02 09:08:42 crc kubenswrapper[4764]: E0202 09:08:42.694984 4764 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 02 09:08:42 crc kubenswrapper[4764]: E0202 09:08:42.695025 4764 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 02 09:08:42 crc kubenswrapper[4764]: E0202 09:08:42.695042 4764 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 02 09:08:42 crc kubenswrapper[4764]: E0202 09:08:42.695071 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-02 09:09:46.695045231 +0000 UTC m=+149.628769329 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 02 09:08:42 crc kubenswrapper[4764]: E0202 09:08:42.695105 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-02 09:09:46.695083162 +0000 UTC m=+149.628807320 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 02 09:08:42 crc kubenswrapper[4764]: I0202 09:08:42.738689 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:42 crc kubenswrapper[4764]: I0202 09:08:42.738738 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:42 crc kubenswrapper[4764]: I0202 09:08:42.738751 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:42 crc kubenswrapper[4764]: I0202 09:08:42.738771 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:42 crc kubenswrapper[4764]: I0202 09:08:42.738783 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:42Z","lastTransitionTime":"2026-02-02T09:08:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:42 crc kubenswrapper[4764]: I0202 09:08:42.824977 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 09:08:42 crc kubenswrapper[4764]: I0202 09:08:42.825069 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 09:08:42 crc kubenswrapper[4764]: E0202 09:08:42.825180 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 09:08:42 crc kubenswrapper[4764]: E0202 09:08:42.825638 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 09:08:42 crc kubenswrapper[4764]: I0202 09:08:42.825787 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 09:08:42 crc kubenswrapper[4764]: E0202 09:08:42.825865 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 09:08:42 crc kubenswrapper[4764]: I0202 09:08:42.826041 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-31 00:27:46.582526276 +0000 UTC Feb 02 09:08:42 crc kubenswrapper[4764]: I0202 09:08:42.841644 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:42 crc kubenswrapper[4764]: I0202 09:08:42.841709 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:42 crc kubenswrapper[4764]: I0202 09:08:42.841729 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:42 crc kubenswrapper[4764]: I0202 09:08:42.841755 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:42 crc kubenswrapper[4764]: I0202 09:08:42.841773 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:42Z","lastTransitionTime":"2026-02-02T09:08:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:42 crc kubenswrapper[4764]: I0202 09:08:42.944707 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:42 crc kubenswrapper[4764]: I0202 09:08:42.944746 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:42 crc kubenswrapper[4764]: I0202 09:08:42.944756 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:42 crc kubenswrapper[4764]: I0202 09:08:42.944770 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:42 crc kubenswrapper[4764]: I0202 09:08:42.944781 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:42Z","lastTransitionTime":"2026-02-02T09:08:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:43 crc kubenswrapper[4764]: I0202 09:08:43.047066 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:43 crc kubenswrapper[4764]: I0202 09:08:43.047141 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:43 crc kubenswrapper[4764]: I0202 09:08:43.047190 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:43 crc kubenswrapper[4764]: I0202 09:08:43.047222 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:43 crc kubenswrapper[4764]: I0202 09:08:43.047242 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:43Z","lastTransitionTime":"2026-02-02T09:08:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:43 crc kubenswrapper[4764]: I0202 09:08:43.150349 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:43 crc kubenswrapper[4764]: I0202 09:08:43.150623 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:43 crc kubenswrapper[4764]: I0202 09:08:43.150693 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:43 crc kubenswrapper[4764]: I0202 09:08:43.150763 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:43 crc kubenswrapper[4764]: I0202 09:08:43.150837 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:43Z","lastTransitionTime":"2026-02-02T09:08:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:43 crc kubenswrapper[4764]: I0202 09:08:43.253228 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:43 crc kubenswrapper[4764]: I0202 09:08:43.253267 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:43 crc kubenswrapper[4764]: I0202 09:08:43.253280 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:43 crc kubenswrapper[4764]: I0202 09:08:43.253297 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:43 crc kubenswrapper[4764]: I0202 09:08:43.253308 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:43Z","lastTransitionTime":"2026-02-02T09:08:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:43 crc kubenswrapper[4764]: I0202 09:08:43.356154 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:43 crc kubenswrapper[4764]: I0202 09:08:43.356195 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:43 crc kubenswrapper[4764]: I0202 09:08:43.356205 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:43 crc kubenswrapper[4764]: I0202 09:08:43.356221 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:43 crc kubenswrapper[4764]: I0202 09:08:43.356232 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:43Z","lastTransitionTime":"2026-02-02T09:08:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:43 crc kubenswrapper[4764]: I0202 09:08:43.460223 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:43 crc kubenswrapper[4764]: I0202 09:08:43.460269 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:43 crc kubenswrapper[4764]: I0202 09:08:43.460284 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:43 crc kubenswrapper[4764]: I0202 09:08:43.460302 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:43 crc kubenswrapper[4764]: I0202 09:08:43.460314 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:43Z","lastTransitionTime":"2026-02-02T09:08:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:43 crc kubenswrapper[4764]: I0202 09:08:43.563009 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:43 crc kubenswrapper[4764]: I0202 09:08:43.563058 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:43 crc kubenswrapper[4764]: I0202 09:08:43.563074 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:43 crc kubenswrapper[4764]: I0202 09:08:43.563092 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:43 crc kubenswrapper[4764]: I0202 09:08:43.563103 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:43Z","lastTransitionTime":"2026-02-02T09:08:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:43 crc kubenswrapper[4764]: I0202 09:08:43.665801 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:43 crc kubenswrapper[4764]: I0202 09:08:43.665875 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:43 crc kubenswrapper[4764]: I0202 09:08:43.665900 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:43 crc kubenswrapper[4764]: I0202 09:08:43.665994 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:43 crc kubenswrapper[4764]: I0202 09:08:43.666027 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:43Z","lastTransitionTime":"2026-02-02T09:08:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:43 crc kubenswrapper[4764]: I0202 09:08:43.768171 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:43 crc kubenswrapper[4764]: I0202 09:08:43.768215 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:43 crc kubenswrapper[4764]: I0202 09:08:43.768226 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:43 crc kubenswrapper[4764]: I0202 09:08:43.768242 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:43 crc kubenswrapper[4764]: I0202 09:08:43.768252 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:43Z","lastTransitionTime":"2026-02-02T09:08:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:43 crc kubenswrapper[4764]: I0202 09:08:43.825640 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kwtmr" Feb 02 09:08:43 crc kubenswrapper[4764]: E0202 09:08:43.825827 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kwtmr" podUID="991faa9a-dd25-4f49-82bd-ce60cefd4af2" Feb 02 09:08:43 crc kubenswrapper[4764]: I0202 09:08:43.826702 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-15 11:23:43.077947156 +0000 UTC Feb 02 09:08:43 crc kubenswrapper[4764]: I0202 09:08:43.870692 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:43 crc kubenswrapper[4764]: I0202 09:08:43.870733 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:43 crc kubenswrapper[4764]: I0202 09:08:43.870744 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:43 crc kubenswrapper[4764]: I0202 09:08:43.870769 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:43 crc kubenswrapper[4764]: I0202 09:08:43.870781 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:43Z","lastTransitionTime":"2026-02-02T09:08:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:43 crc kubenswrapper[4764]: I0202 09:08:43.973380 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:43 crc kubenswrapper[4764]: I0202 09:08:43.973414 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:43 crc kubenswrapper[4764]: I0202 09:08:43.973431 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:43 crc kubenswrapper[4764]: I0202 09:08:43.973451 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:43 crc kubenswrapper[4764]: I0202 09:08:43.973463 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:43Z","lastTransitionTime":"2026-02-02T09:08:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:44 crc kubenswrapper[4764]: I0202 09:08:44.076137 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:44 crc kubenswrapper[4764]: I0202 09:08:44.076578 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:44 crc kubenswrapper[4764]: I0202 09:08:44.076775 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:44 crc kubenswrapper[4764]: I0202 09:08:44.077151 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:44 crc kubenswrapper[4764]: I0202 09:08:44.077459 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:44Z","lastTransitionTime":"2026-02-02T09:08:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:44 crc kubenswrapper[4764]: I0202 09:08:44.180086 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:44 crc kubenswrapper[4764]: I0202 09:08:44.180166 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:44 crc kubenswrapper[4764]: I0202 09:08:44.180184 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:44 crc kubenswrapper[4764]: I0202 09:08:44.180205 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:44 crc kubenswrapper[4764]: I0202 09:08:44.180252 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:44Z","lastTransitionTime":"2026-02-02T09:08:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:44 crc kubenswrapper[4764]: I0202 09:08:44.282449 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:44 crc kubenswrapper[4764]: I0202 09:08:44.282782 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:44 crc kubenswrapper[4764]: I0202 09:08:44.282900 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:44 crc kubenswrapper[4764]: I0202 09:08:44.283041 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:44 crc kubenswrapper[4764]: I0202 09:08:44.283142 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:44Z","lastTransitionTime":"2026-02-02T09:08:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:44 crc kubenswrapper[4764]: I0202 09:08:44.385316 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:44 crc kubenswrapper[4764]: I0202 09:08:44.385361 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:44 crc kubenswrapper[4764]: I0202 09:08:44.385373 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:44 crc kubenswrapper[4764]: I0202 09:08:44.385391 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:44 crc kubenswrapper[4764]: I0202 09:08:44.385404 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:44Z","lastTransitionTime":"2026-02-02T09:08:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:44 crc kubenswrapper[4764]: I0202 09:08:44.487613 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:44 crc kubenswrapper[4764]: I0202 09:08:44.487641 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:44 crc kubenswrapper[4764]: I0202 09:08:44.487649 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:44 crc kubenswrapper[4764]: I0202 09:08:44.487661 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:44 crc kubenswrapper[4764]: I0202 09:08:44.487673 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:44Z","lastTransitionTime":"2026-02-02T09:08:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:44 crc kubenswrapper[4764]: I0202 09:08:44.590815 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:44 crc kubenswrapper[4764]: I0202 09:08:44.590917 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:44 crc kubenswrapper[4764]: I0202 09:08:44.590990 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:44 crc kubenswrapper[4764]: I0202 09:08:44.591022 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:44 crc kubenswrapper[4764]: I0202 09:08:44.591046 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:44Z","lastTransitionTime":"2026-02-02T09:08:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:44 crc kubenswrapper[4764]: I0202 09:08:44.694294 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:44 crc kubenswrapper[4764]: I0202 09:08:44.694348 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:44 crc kubenswrapper[4764]: I0202 09:08:44.694365 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:44 crc kubenswrapper[4764]: I0202 09:08:44.694387 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:44 crc kubenswrapper[4764]: I0202 09:08:44.694405 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:44Z","lastTransitionTime":"2026-02-02T09:08:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:44 crc kubenswrapper[4764]: I0202 09:08:44.796856 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:44 crc kubenswrapper[4764]: I0202 09:08:44.796920 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:44 crc kubenswrapper[4764]: I0202 09:08:44.796963 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:44 crc kubenswrapper[4764]: I0202 09:08:44.796990 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:44 crc kubenswrapper[4764]: I0202 09:08:44.797007 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:44Z","lastTransitionTime":"2026-02-02T09:08:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:44 crc kubenswrapper[4764]: I0202 09:08:44.824576 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 09:08:44 crc kubenswrapper[4764]: E0202 09:08:44.824731 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 09:08:44 crc kubenswrapper[4764]: I0202 09:08:44.824980 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 09:08:44 crc kubenswrapper[4764]: E0202 09:08:44.825049 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 09:08:44 crc kubenswrapper[4764]: I0202 09:08:44.825142 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 09:08:44 crc kubenswrapper[4764]: E0202 09:08:44.825295 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 09:08:44 crc kubenswrapper[4764]: I0202 09:08:44.827692 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-18 05:20:34.088236541 +0000 UTC Feb 02 09:08:44 crc kubenswrapper[4764]: I0202 09:08:44.900238 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:44 crc kubenswrapper[4764]: I0202 09:08:44.900318 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:44 crc kubenswrapper[4764]: I0202 09:08:44.900336 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:44 crc kubenswrapper[4764]: I0202 09:08:44.900362 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:44 crc kubenswrapper[4764]: I0202 09:08:44.900381 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:44Z","lastTransitionTime":"2026-02-02T09:08:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:45 crc kubenswrapper[4764]: I0202 09:08:45.003315 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:45 crc kubenswrapper[4764]: I0202 09:08:45.003354 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:45 crc kubenswrapper[4764]: I0202 09:08:45.003367 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:45 crc kubenswrapper[4764]: I0202 09:08:45.003384 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:45 crc kubenswrapper[4764]: I0202 09:08:45.003398 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:45Z","lastTransitionTime":"2026-02-02T09:08:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:45 crc kubenswrapper[4764]: I0202 09:08:45.106089 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:45 crc kubenswrapper[4764]: I0202 09:08:45.106139 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:45 crc kubenswrapper[4764]: I0202 09:08:45.106152 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:45 crc kubenswrapper[4764]: I0202 09:08:45.106173 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:45 crc kubenswrapper[4764]: I0202 09:08:45.106194 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:45Z","lastTransitionTime":"2026-02-02T09:08:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:45 crc kubenswrapper[4764]: I0202 09:08:45.209294 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:45 crc kubenswrapper[4764]: I0202 09:08:45.209633 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:45 crc kubenswrapper[4764]: I0202 09:08:45.209736 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:45 crc kubenswrapper[4764]: I0202 09:08:45.209832 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:45 crc kubenswrapper[4764]: I0202 09:08:45.209926 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:45Z","lastTransitionTime":"2026-02-02T09:08:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:45 crc kubenswrapper[4764]: I0202 09:08:45.313438 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:45 crc kubenswrapper[4764]: I0202 09:08:45.313495 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:45 crc kubenswrapper[4764]: I0202 09:08:45.313509 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:45 crc kubenswrapper[4764]: I0202 09:08:45.313576 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:45 crc kubenswrapper[4764]: I0202 09:08:45.313599 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:45Z","lastTransitionTime":"2026-02-02T09:08:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:45 crc kubenswrapper[4764]: I0202 09:08:45.416636 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:45 crc kubenswrapper[4764]: I0202 09:08:45.416687 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:45 crc kubenswrapper[4764]: I0202 09:08:45.416700 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:45 crc kubenswrapper[4764]: I0202 09:08:45.416719 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:45 crc kubenswrapper[4764]: I0202 09:08:45.416733 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:45Z","lastTransitionTime":"2026-02-02T09:08:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:45 crc kubenswrapper[4764]: I0202 09:08:45.519079 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:45 crc kubenswrapper[4764]: I0202 09:08:45.519134 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:45 crc kubenswrapper[4764]: I0202 09:08:45.519150 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:45 crc kubenswrapper[4764]: I0202 09:08:45.519170 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:45 crc kubenswrapper[4764]: I0202 09:08:45.519185 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:45Z","lastTransitionTime":"2026-02-02T09:08:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:45 crc kubenswrapper[4764]: I0202 09:08:45.621830 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:45 crc kubenswrapper[4764]: I0202 09:08:45.621893 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:45 crc kubenswrapper[4764]: I0202 09:08:45.621912 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:45 crc kubenswrapper[4764]: I0202 09:08:45.621968 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:45 crc kubenswrapper[4764]: I0202 09:08:45.621986 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:45Z","lastTransitionTime":"2026-02-02T09:08:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:45 crc kubenswrapper[4764]: I0202 09:08:45.725358 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:45 crc kubenswrapper[4764]: I0202 09:08:45.725418 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:45 crc kubenswrapper[4764]: I0202 09:08:45.725431 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:45 crc kubenswrapper[4764]: I0202 09:08:45.725452 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:45 crc kubenswrapper[4764]: I0202 09:08:45.725466 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:45Z","lastTransitionTime":"2026-02-02T09:08:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:45 crc kubenswrapper[4764]: I0202 09:08:45.824809 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kwtmr" Feb 02 09:08:45 crc kubenswrapper[4764]: E0202 09:08:45.825048 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kwtmr" podUID="991faa9a-dd25-4f49-82bd-ce60cefd4af2" Feb 02 09:08:45 crc kubenswrapper[4764]: I0202 09:08:45.829349 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-07 08:07:00.747934013 +0000 UTC Feb 02 09:08:45 crc kubenswrapper[4764]: I0202 09:08:45.829849 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:45 crc kubenswrapper[4764]: I0202 09:08:45.829881 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:45 crc kubenswrapper[4764]: I0202 09:08:45.829894 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:45 crc kubenswrapper[4764]: I0202 09:08:45.829911 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:45 crc kubenswrapper[4764]: I0202 09:08:45.829925 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:45Z","lastTransitionTime":"2026-02-02T09:08:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:45 crc kubenswrapper[4764]: I0202 09:08:45.932535 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:45 crc kubenswrapper[4764]: I0202 09:08:45.932606 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:45 crc kubenswrapper[4764]: I0202 09:08:45.932618 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:45 crc kubenswrapper[4764]: I0202 09:08:45.932640 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:45 crc kubenswrapper[4764]: I0202 09:08:45.932654 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:45Z","lastTransitionTime":"2026-02-02T09:08:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:46 crc kubenswrapper[4764]: I0202 09:08:46.035364 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:46 crc kubenswrapper[4764]: I0202 09:08:46.035407 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:46 crc kubenswrapper[4764]: I0202 09:08:46.035421 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:46 crc kubenswrapper[4764]: I0202 09:08:46.035437 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:46 crc kubenswrapper[4764]: I0202 09:08:46.035449 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:46Z","lastTransitionTime":"2026-02-02T09:08:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:46 crc kubenswrapper[4764]: I0202 09:08:46.137970 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:46 crc kubenswrapper[4764]: I0202 09:08:46.138011 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:46 crc kubenswrapper[4764]: I0202 09:08:46.138022 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:46 crc kubenswrapper[4764]: I0202 09:08:46.138040 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:46 crc kubenswrapper[4764]: I0202 09:08:46.138051 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:46Z","lastTransitionTime":"2026-02-02T09:08:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:46 crc kubenswrapper[4764]: I0202 09:08:46.240237 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:46 crc kubenswrapper[4764]: I0202 09:08:46.240589 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:46 crc kubenswrapper[4764]: I0202 09:08:46.240743 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:46 crc kubenswrapper[4764]: I0202 09:08:46.240891 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:46 crc kubenswrapper[4764]: I0202 09:08:46.241184 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:46Z","lastTransitionTime":"2026-02-02T09:08:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:46 crc kubenswrapper[4764]: I0202 09:08:46.344723 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:46 crc kubenswrapper[4764]: I0202 09:08:46.344781 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:46 crc kubenswrapper[4764]: I0202 09:08:46.344798 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:46 crc kubenswrapper[4764]: I0202 09:08:46.344849 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:46 crc kubenswrapper[4764]: I0202 09:08:46.344868 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:46Z","lastTransitionTime":"2026-02-02T09:08:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:46 crc kubenswrapper[4764]: I0202 09:08:46.447524 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:46 crc kubenswrapper[4764]: I0202 09:08:46.448007 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:46 crc kubenswrapper[4764]: I0202 09:08:46.448200 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:46 crc kubenswrapper[4764]: I0202 09:08:46.448368 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:46 crc kubenswrapper[4764]: I0202 09:08:46.448514 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:46Z","lastTransitionTime":"2026-02-02T09:08:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:46 crc kubenswrapper[4764]: I0202 09:08:46.551187 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:46 crc kubenswrapper[4764]: I0202 09:08:46.551243 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:46 crc kubenswrapper[4764]: I0202 09:08:46.551259 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:46 crc kubenswrapper[4764]: I0202 09:08:46.551282 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:46 crc kubenswrapper[4764]: I0202 09:08:46.551299 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:46Z","lastTransitionTime":"2026-02-02T09:08:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:46 crc kubenswrapper[4764]: I0202 09:08:46.654409 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:46 crc kubenswrapper[4764]: I0202 09:08:46.654444 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:46 crc kubenswrapper[4764]: I0202 09:08:46.654454 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:46 crc kubenswrapper[4764]: I0202 09:08:46.654471 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:46 crc kubenswrapper[4764]: I0202 09:08:46.654483 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:46Z","lastTransitionTime":"2026-02-02T09:08:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:46 crc kubenswrapper[4764]: I0202 09:08:46.739168 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:46 crc kubenswrapper[4764]: I0202 09:08:46.739242 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:46 crc kubenswrapper[4764]: I0202 09:08:46.739258 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:46 crc kubenswrapper[4764]: I0202 09:08:46.739276 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:46 crc kubenswrapper[4764]: I0202 09:08:46.739313 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:46Z","lastTransitionTime":"2026-02-02T09:08:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:46 crc kubenswrapper[4764]: E0202 09:08:46.759850 4764 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:08:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:08:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:08:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:08:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"31973a70-ce17-4811-9a14-f5476c6979db\\\",\\\"systemUUID\\\":\\\"ca92679c-8580-43c3-be51-62441e93d22b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:46Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:46 crc kubenswrapper[4764]: I0202 09:08:46.764112 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:46 crc kubenswrapper[4764]: I0202 09:08:46.764458 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:46 crc kubenswrapper[4764]: I0202 09:08:46.764577 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:46 crc kubenswrapper[4764]: I0202 09:08:46.764680 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:46 crc kubenswrapper[4764]: I0202 09:08:46.764774 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:46Z","lastTransitionTime":"2026-02-02T09:08:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:46 crc kubenswrapper[4764]: E0202 09:08:46.780240 4764 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:08:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:08:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:08:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:08:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"31973a70-ce17-4811-9a14-f5476c6979db\\\",\\\"systemUUID\\\":\\\"ca92679c-8580-43c3-be51-62441e93d22b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:46Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:46 crc kubenswrapper[4764]: I0202 09:08:46.784386 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:46 crc kubenswrapper[4764]: I0202 09:08:46.784428 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:46 crc kubenswrapper[4764]: I0202 09:08:46.784442 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:46 crc kubenswrapper[4764]: I0202 09:08:46.784468 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:46 crc kubenswrapper[4764]: I0202 09:08:46.784480 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:46Z","lastTransitionTime":"2026-02-02T09:08:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:46 crc kubenswrapper[4764]: E0202 09:08:46.800903 4764 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:08:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:08:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:08:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:08:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"31973a70-ce17-4811-9a14-f5476c6979db\\\",\\\"systemUUID\\\":\\\"ca92679c-8580-43c3-be51-62441e93d22b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:46Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:46 crc kubenswrapper[4764]: I0202 09:08:46.804923 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:46 crc kubenswrapper[4764]: I0202 09:08:46.804980 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:46 crc kubenswrapper[4764]: I0202 09:08:46.804992 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:46 crc kubenswrapper[4764]: I0202 09:08:46.805012 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:46 crc kubenswrapper[4764]: I0202 09:08:46.805024 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:46Z","lastTransitionTime":"2026-02-02T09:08:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:46 crc kubenswrapper[4764]: E0202 09:08:46.821886 4764 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:08:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:08:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:08:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:08:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"31973a70-ce17-4811-9a14-f5476c6979db\\\",\\\"systemUUID\\\":\\\"ca92679c-8580-43c3-be51-62441e93d22b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:46Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:46 crc kubenswrapper[4764]: I0202 09:08:46.824981 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 09:08:46 crc kubenswrapper[4764]: I0202 09:08:46.825067 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 09:08:46 crc kubenswrapper[4764]: E0202 09:08:46.825114 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 09:08:46 crc kubenswrapper[4764]: E0202 09:08:46.825256 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 09:08:46 crc kubenswrapper[4764]: I0202 09:08:46.825332 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 09:08:46 crc kubenswrapper[4764]: E0202 09:08:46.825398 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 09:08:46 crc kubenswrapper[4764]: I0202 09:08:46.827870 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:46 crc kubenswrapper[4764]: I0202 09:08:46.827951 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:46 crc kubenswrapper[4764]: I0202 09:08:46.827964 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:46 crc kubenswrapper[4764]: I0202 09:08:46.827978 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:46 crc kubenswrapper[4764]: I0202 09:08:46.827989 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:46Z","lastTransitionTime":"2026-02-02T09:08:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:46 crc kubenswrapper[4764]: I0202 09:08:46.829660 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-16 09:36:38.675550975 +0000 UTC Feb 02 09:08:46 crc kubenswrapper[4764]: E0202 09:08:46.840927 4764 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:08:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:08:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:08:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:08:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"31973a70-ce17-4811-9a14-f5476c6979db\\\",\\\"systemUUID\\\":\\\"ca92679c-8580-43c3-be51-62441e93d22b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:46Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:46 crc kubenswrapper[4764]: E0202 09:08:46.841140 4764 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 02 09:08:46 crc kubenswrapper[4764]: I0202 09:08:46.842853 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:46 crc kubenswrapper[4764]: I0202 09:08:46.842900 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:46 crc kubenswrapper[4764]: I0202 09:08:46.842915 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:46 crc kubenswrapper[4764]: I0202 09:08:46.842969 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:46 crc kubenswrapper[4764]: I0202 09:08:46.842983 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:46Z","lastTransitionTime":"2026-02-02T09:08:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:46 crc kubenswrapper[4764]: I0202 09:08:46.946727 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:46 crc kubenswrapper[4764]: I0202 09:08:46.946814 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:46 crc kubenswrapper[4764]: I0202 09:08:46.946865 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:46 crc kubenswrapper[4764]: I0202 09:08:46.946886 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:46 crc kubenswrapper[4764]: I0202 09:08:46.947103 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:46Z","lastTransitionTime":"2026-02-02T09:08:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:47 crc kubenswrapper[4764]: I0202 09:08:47.049489 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:47 crc kubenswrapper[4764]: I0202 09:08:47.050081 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:47 crc kubenswrapper[4764]: I0202 09:08:47.050325 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:47 crc kubenswrapper[4764]: I0202 09:08:47.050537 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:47 crc kubenswrapper[4764]: I0202 09:08:47.050729 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:47Z","lastTransitionTime":"2026-02-02T09:08:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:47 crc kubenswrapper[4764]: I0202 09:08:47.152881 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:47 crc kubenswrapper[4764]: I0202 09:08:47.153141 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:47 crc kubenswrapper[4764]: I0202 09:08:47.153225 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:47 crc kubenswrapper[4764]: I0202 09:08:47.153305 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:47 crc kubenswrapper[4764]: I0202 09:08:47.153398 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:47Z","lastTransitionTime":"2026-02-02T09:08:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:47 crc kubenswrapper[4764]: I0202 09:08:47.255185 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:47 crc kubenswrapper[4764]: I0202 09:08:47.255437 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:47 crc kubenswrapper[4764]: I0202 09:08:47.255628 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:47 crc kubenswrapper[4764]: I0202 09:08:47.255809 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:47 crc kubenswrapper[4764]: I0202 09:08:47.256037 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:47Z","lastTransitionTime":"2026-02-02T09:08:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:47 crc kubenswrapper[4764]: I0202 09:08:47.358102 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:47 crc kubenswrapper[4764]: I0202 09:08:47.358138 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:47 crc kubenswrapper[4764]: I0202 09:08:47.358147 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:47 crc kubenswrapper[4764]: I0202 09:08:47.358160 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:47 crc kubenswrapper[4764]: I0202 09:08:47.358170 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:47Z","lastTransitionTime":"2026-02-02T09:08:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:47 crc kubenswrapper[4764]: I0202 09:08:47.462139 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:47 crc kubenswrapper[4764]: I0202 09:08:47.462368 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:47 crc kubenswrapper[4764]: I0202 09:08:47.462473 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:47 crc kubenswrapper[4764]: I0202 09:08:47.462551 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:47 crc kubenswrapper[4764]: I0202 09:08:47.462618 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:47Z","lastTransitionTime":"2026-02-02T09:08:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:47 crc kubenswrapper[4764]: I0202 09:08:47.565669 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:47 crc kubenswrapper[4764]: I0202 09:08:47.565706 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:47 crc kubenswrapper[4764]: I0202 09:08:47.565715 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:47 crc kubenswrapper[4764]: I0202 09:08:47.565730 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:47 crc kubenswrapper[4764]: I0202 09:08:47.565742 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:47Z","lastTransitionTime":"2026-02-02T09:08:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:47 crc kubenswrapper[4764]: I0202 09:08:47.668078 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:47 crc kubenswrapper[4764]: I0202 09:08:47.668328 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:47 crc kubenswrapper[4764]: I0202 09:08:47.668420 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:47 crc kubenswrapper[4764]: I0202 09:08:47.668508 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:47 crc kubenswrapper[4764]: I0202 09:08:47.668580 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:47Z","lastTransitionTime":"2026-02-02T09:08:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:47 crc kubenswrapper[4764]: I0202 09:08:47.771790 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:47 crc kubenswrapper[4764]: I0202 09:08:47.771852 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:47 crc kubenswrapper[4764]: I0202 09:08:47.771870 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:47 crc kubenswrapper[4764]: I0202 09:08:47.771894 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:47 crc kubenswrapper[4764]: I0202 09:08:47.771911 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:47Z","lastTransitionTime":"2026-02-02T09:08:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:47 crc kubenswrapper[4764]: I0202 09:08:47.825600 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kwtmr" Feb 02 09:08:47 crc kubenswrapper[4764]: E0202 09:08:47.826159 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kwtmr" podUID="991faa9a-dd25-4f49-82bd-ce60cefd4af2" Feb 02 09:08:47 crc kubenswrapper[4764]: I0202 09:08:47.830659 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-06 23:57:39.757262834 +0000 UTC Feb 02 09:08:47 crc kubenswrapper[4764]: I0202 09:08:47.844282 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:47Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:47 crc kubenswrapper[4764]: I0202 09:08:47.868681 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24632dda-6100-4ab6-a28e-214ddae4360c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a6c67e9b7708cc0d85b0b6ffbbade0d34867e159a9c5487ca2606475c667a50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bffda48d7c881c1470251b501661324859373246cc1a21811bd36ca24887c7aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a989645cf485cd8ff80ffff1a86cd7808e1564eba5667535e161cc68db13b26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20b71d3ee7a381baa51b0b0d0470283022cfede30f491135737b5e82d631d763\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c380383b6a014d1b694a13ddc22f54b20b9e4f9d5c9b9000d76187621dce0a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5529f2bff87753428e9173826fc1880acf96a4c134fa9ecad195b15697cad73a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cb255e14eb02f7c08364c41fb9906e58b2155eb507f694b77ec8eb1f4251343\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3cb255e14eb02f7c08364c41fb9906e58b2155eb507f694b77ec8eb1f4251343\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-02T09:08:38Z\\\",\\\"message\\\":\\\":38.747848 6731 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-diagnostics/network-check-target-xd92c\\\\nI0202 09:08:38.747852 6731 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf in node crc\\\\nI0202 09:08:38.747854 6731 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-target-xd92c\\\\nI0202 09:08:38.747780 6731 obj_retry.go:386] Retry successful for *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc after 0 failed attempt(s)\\\\nI0202 09:08:38.747862 6731 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-target-xd92c in node crc\\\\nI0202 09:08:38.747864 6731 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf after 0 failed attempt(s)\\\\nI0202 09:08:38.747869 6731 default_network_controller.go:776] Recording success event on pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI0202 09:08:38.747837 6731 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-additional-cni-plugins-9p5dc in node crc\\\\nI0202 09:08:38.747884 6731 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-target-xd92c] creating logical port openshift-network-diagnostics_network-check-target-\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T09:08:37Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-zhn7j_openshift-ovn-kubernetes(24632dda-6100-4ab6-a28e-214ddae4360c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b4cfbd4e682bcdcaff785013bdb72580ce36fced733e235538536827036b93d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://579903eadb183466a2564737e6b678e04308239c0cd75b6808892053b97afbc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://579903eadb183466a2564737e6b678e04308239c0cd75b6808892053b97afbc0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zhn7j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:47Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:47 crc kubenswrapper[4764]: I0202 09:08:47.874088 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:47 crc kubenswrapper[4764]: I0202 09:08:47.874431 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:47 crc kubenswrapper[4764]: I0202 09:08:47.874619 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:47 crc kubenswrapper[4764]: I0202 09:08:47.874861 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:47 crc kubenswrapper[4764]: I0202 09:08:47.875067 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:47Z","lastTransitionTime":"2026-02-02T09:08:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:47 crc kubenswrapper[4764]: I0202 09:08:47.895625 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d66ab3c0-9b5b-41d5-b6c0-91ccddccd7e9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d92ac486f5dade46bc75de32099859a71fc2ed25a7abbe6ae8937df128e7df30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a4ac273a4f127d79e5fad7d4c30cd2044a346f5734dba5d33d68cd092c0c2d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d38c7cb5dcd22fb07524494ebe77aa1a7f4f808f566043563817e47e7ec12a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://408722c7f83d28919431532f8c797a1c795fd5d96736d14e7f06a05744899702\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f473b18cd8003f75bb1dafb82bcec45745a73ccd10434297356dac185a3d15c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e229f484d19cbdabedf53d90a822a4434db9e00b1301538e885e26acc19d25f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e229f484d19cbdabedf53d90a822a4434db9e00b1301538e885e26acc19d25f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a9cc673be8ff245dba81278e8cd34cfe45ed36ff0b8bd5b30a0ddf333ec6371\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a9cc673be8ff245dba81278e8cd34cfe45ed36ff0b8bd5b30a0ddf333ec6371\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bc7e49c7a5ae66ea939060192fa6a346dc0b0b9aa9750b29879d3c426cc0d276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc7e49c7a5ae66ea939060192fa6a346dc0b0b9aa9750b29879d3c426cc0d276\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:47Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:47 crc kubenswrapper[4764]: I0202 09:08:47.910663 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa9d037e66cb53b9dd698eff46284c9e718fd842cda6f7b525f0c79062d70bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:47Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:47 crc kubenswrapper[4764]: I0202 09:08:47.921667 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6zxff" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aaba93da-12a5-43df-b7c7-4c8b988163c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3db293678a907cd55b31b1d0c0bd7c30e1ba75ecd5e13cbc68ab98fbf0fbd9d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqwb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6zxff\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:47Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:47 crc kubenswrapper[4764]: I0202 09:08:47.935498 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d192f670-9f9d-4539-9641-e4bed73acdd4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://040c5e8a5659f85f0d52d1a1e064154d998ae104f711101d2204c0b1caebbb30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbksf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fdf06410349b7c99aa166e87fbe80bf68e2ae923635a6a978b7e2cd50383f7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbksf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4ndm4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:47Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:47 crc kubenswrapper[4764]: I0202 09:08:47.946736 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pcjmn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6e5ccd3b-b657-47a2-9c4d-bc9adb126533\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72863eef568d92a1abeb3cde2e89630a4f5fb7626c6949448ff5195774c98eab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2f7zl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pcjmn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:47Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:47 crc kubenswrapper[4764]: I0202 09:08:47.958829 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed747fd8-e3ce-4319-a6da-bda0fd51b85a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6f064bdc66470aace3180d4988244026b54009a63cff199e41c5d02a1d18479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://60258f84a1d8dcfe48b7e9688ed50f077e201971f20f3f6663fbf1ba58a4611a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f84be2e08217a2ac3edde858d78b4410376ed9bae3bcabe2a7f9b2fbe2f69c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a69d25d352d7ceac5c735bc1fcf446a1fea6dc22c65a5296cfab828cea17377\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:47Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:47 crc kubenswrapper[4764]: I0202 09:08:47.969858 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f6904228e052fff69ad1fd66df97758b6b60ba4351e97a7fbd03eee4df1b653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:47Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:47 crc kubenswrapper[4764]: I0202 09:08:47.978294 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:47 crc kubenswrapper[4764]: I0202 09:08:47.978373 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:47 crc kubenswrapper[4764]: I0202 09:08:47.978413 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:47 crc kubenswrapper[4764]: I0202 09:08:47.978463 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:47 crc kubenswrapper[4764]: I0202 09:08:47.978484 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:47Z","lastTransitionTime":"2026-02-02T09:08:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:47 crc kubenswrapper[4764]: I0202 09:08:47.980286 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lh6tm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2152e8a-f456-47b5-a547-b65ec837ecbc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38e176a00b07f2960fdd7ba2aadd1468a5e083b8d154e9d118d13f2585da81e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bb2sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://996e649011d5255783284421ddea8e0a5c383ef675b0920c3fdc0af6503d80e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bb2sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lh6tm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:47Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:47 crc kubenswrapper[4764]: I0202 09:08:47.992549 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fae745b-ea49-42f3-a486-dbc3dc2fb9ad\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://081d7f76d937ff71e35945f9b1d893888bb649d15bd84312fd4b61c9520ea71f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://429ead0993f5e648f6256d249ee12aea03514d3108618840552b672fb419cd74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fad17b01545eb20785f03e3c61bd5c982f6217e38176d7a3b23ac194298644a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85e2f75c56d63cacc93b9195c1e8bf46eb7a7234c45a024bd95a460db90edb6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85e2f75c56d63cacc93b9195c1e8bf46eb7a7234c45a024bd95a460db90edb6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:47Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:48 crc kubenswrapper[4764]: I0202 09:08:48.005815 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:48Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:48 crc kubenswrapper[4764]: I0202 09:08:48.019294 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5528225d91ac2778bef4554abb0bd036ac5d3b6cdac7357bab69e70c3894b6f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://071bbb80d0747cf3e109ba626d47f7d05f85f4cc0b3aba77276c795741f47ac6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:48Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:48 crc kubenswrapper[4764]: I0202 09:08:48.031687 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:48Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:48 crc kubenswrapper[4764]: I0202 09:08:48.044428 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nc7qt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b52b499d17b4e2e359475bd13730c72ef5cedb052e13def6dfe5bc5fd7487d2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1a2c7b22068b7165927fc158885d48bf9952b9eba521d8ae9a599eee53dc305\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-02T09:08:28Z\\\",\\\"message\\\":\\\"2026-02-02T09:07:42+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_32cbaf6a-865c-4787-8f89-a40249873b40\\\\n2026-02-02T09:07:42+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_32cbaf6a-865c-4787-8f89-a40249873b40 to /host/opt/cni/bin/\\\\n2026-02-02T09:07:43Z [verbose] multus-daemon started\\\\n2026-02-02T09:07:43Z [verbose] Readiness Indicator file check\\\\n2026-02-02T09:08:28Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:08:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2hz4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nc7qt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:48Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:48 crc kubenswrapper[4764]: I0202 09:08:48.060160 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9p5dc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"660ac006-2495-45c4-9fb3-e9c1dddcf7a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da812b124c97aa8ad1563f4bcfac8f3bfa025096fa4a2561e5847b1bf9cb30b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23e57e65a6f5c3059506b9db21e9ac2492c30bd9191a98db0281915e9ce39797\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23e57e65a6f5c3059506b9db21e9ac2492c30bd9191a98db0281915e9ce39797\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2417b66816da1a6cd1791a4b91eca4a3c4f3694462e066e7f0722f1b9c2105c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2417b66816da1a6cd1791a4b91eca4a3c4f3694462e066e7f0722f1b9c2105c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://883641f29e9120c2cea6b272bacf1dc63c752f55ab0408cdaf73db85c85ac369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://883641f29e9120c2cea6b272bacf1dc63c752f55ab0408cdaf73db85c85ac369\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2745baba1188ba5e8f820616e4117a12c50ecee5b68e71ad9cabbb621b6e131f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2745baba1188ba5e8f820616e4117a12c50ecee5b68e71ad9cabbb621b6e131f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8925e435be4d48d86824fc35f8b8819fb35d63e385164c4a20edb3bf3035c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a8925e435be4d48d86824fc35f8b8819fb35d63e385164c4a20edb3bf3035c3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d3fbf668ae7d04c8d9e574f43c2fe90c5604204ba38a3c60847afdfbceb7a57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d3fbf668ae7d04c8d9e574f43c2fe90c5604204ba38a3c60847afdfbceb7a57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9p5dc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:48Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:48 crc kubenswrapper[4764]: I0202 09:08:48.072615 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kwtmr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"991faa9a-dd25-4f49-82bd-ce60cefd4af2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h865b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h865b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:53Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kwtmr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:48Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:48 crc kubenswrapper[4764]: I0202 09:08:48.081396 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:48 crc kubenswrapper[4764]: I0202 09:08:48.081430 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:48 crc kubenswrapper[4764]: I0202 09:08:48.081439 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:48 crc kubenswrapper[4764]: I0202 09:08:48.081451 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:48 crc kubenswrapper[4764]: I0202 09:08:48.081461 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:48Z","lastTransitionTime":"2026-02-02T09:08:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:48 crc kubenswrapper[4764]: I0202 09:08:48.091202 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87d61aea-46f6-4ee6-a7a4-767a6a67fea4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e78b6d08e4d616cefbb73e3fd43b3c6886dcd8aa997c7b4948bf1ff63401293\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d468715c2da33996561c40043216233fd12e21539d0c58dd1f3b892fd9f91a9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b184f7c936dc4046f74c0018269cbc2c4b7ceddb650b4e871d3f8a707e0df88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://269c5074371bd00f8363fb3888c11f1df383081a61497aea6578f08700881d50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b25bc9395831524c53f9e565b4efbe74f21792efafeb1dd507161bc91aa0ecdf\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T09:07:22Z\\\",\\\"message\\\":\\\"W0202 09:07:21.597488 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0202 09:07:21.598117 1 crypto.go:601] Generating new CA for check-endpoints-signer@1770023241 cert, and key in /tmp/serving-cert-3856183724/serving-signer.crt, /tmp/serving-cert-3856183724/serving-signer.key\\\\nI0202 09:07:22.012184 1 observer_polling.go:159] Starting file observer\\\\nW0202 09:07:22.015493 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0202 09:07:22.015797 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 09:07:22.016596 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3856183724/tls.crt::/tmp/serving-cert-3856183724/tls.key\\\\\\\"\\\\nF0202 09:07:22.270605 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b44e17b8ca01f549bc10081dd5f109cb1b9fff364bf5bd4fd90d9f0e3eeb04db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfb06a5ec9262cf39b108a66014d15e325cdbd5441a2d9e7a811dd57265c99a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfb06a5ec9262cf39b108a66014d15e325cdbd5441a2d9e7a811dd57265c99a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:48Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:48 crc kubenswrapper[4764]: I0202 09:08:48.107780 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fea07bdc-9134-4993-934c-316c19d0dc76\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://664e0519c2a2c60904dfcc8deebaf5d638dac0bedb13d870c0aaea5b4f5428f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db5a54d3a2117252a1ff50884683f99185874d2904c0960161642280f614679d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db5a54d3a2117252a1ff50884683f99185874d2904c0960161642280f614679d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:48Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:48 crc kubenswrapper[4764]: I0202 09:08:48.183665 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:48 crc kubenswrapper[4764]: I0202 09:08:48.183705 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:48 crc kubenswrapper[4764]: I0202 09:08:48.183718 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:48 crc kubenswrapper[4764]: I0202 09:08:48.183734 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:48 crc kubenswrapper[4764]: I0202 09:08:48.183745 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:48Z","lastTransitionTime":"2026-02-02T09:08:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:48 crc kubenswrapper[4764]: I0202 09:08:48.286129 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:48 crc kubenswrapper[4764]: I0202 09:08:48.286186 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:48 crc kubenswrapper[4764]: I0202 09:08:48.286209 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:48 crc kubenswrapper[4764]: I0202 09:08:48.286238 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:48 crc kubenswrapper[4764]: I0202 09:08:48.286304 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:48Z","lastTransitionTime":"2026-02-02T09:08:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:48 crc kubenswrapper[4764]: I0202 09:08:48.388924 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:48 crc kubenswrapper[4764]: I0202 09:08:48.389023 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:48 crc kubenswrapper[4764]: I0202 09:08:48.389036 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:48 crc kubenswrapper[4764]: I0202 09:08:48.389051 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:48 crc kubenswrapper[4764]: I0202 09:08:48.389062 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:48Z","lastTransitionTime":"2026-02-02T09:08:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:48 crc kubenswrapper[4764]: I0202 09:08:48.493008 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:48 crc kubenswrapper[4764]: I0202 09:08:48.493055 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:48 crc kubenswrapper[4764]: I0202 09:08:48.493067 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:48 crc kubenswrapper[4764]: I0202 09:08:48.493088 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:48 crc kubenswrapper[4764]: I0202 09:08:48.493101 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:48Z","lastTransitionTime":"2026-02-02T09:08:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:48 crc kubenswrapper[4764]: I0202 09:08:48.596249 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:48 crc kubenswrapper[4764]: I0202 09:08:48.596596 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:48 crc kubenswrapper[4764]: I0202 09:08:48.596743 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:48 crc kubenswrapper[4764]: I0202 09:08:48.596825 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:48 crc kubenswrapper[4764]: I0202 09:08:48.596896 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:48Z","lastTransitionTime":"2026-02-02T09:08:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:48 crc kubenswrapper[4764]: I0202 09:08:48.699766 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:48 crc kubenswrapper[4764]: I0202 09:08:48.699802 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:48 crc kubenswrapper[4764]: I0202 09:08:48.699817 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:48 crc kubenswrapper[4764]: I0202 09:08:48.699837 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:48 crc kubenswrapper[4764]: I0202 09:08:48.699851 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:48Z","lastTransitionTime":"2026-02-02T09:08:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:48 crc kubenswrapper[4764]: I0202 09:08:48.802860 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:48 crc kubenswrapper[4764]: I0202 09:08:48.802908 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:48 crc kubenswrapper[4764]: I0202 09:08:48.802924 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:48 crc kubenswrapper[4764]: I0202 09:08:48.802981 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:48 crc kubenswrapper[4764]: I0202 09:08:48.802998 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:48Z","lastTransitionTime":"2026-02-02T09:08:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:48 crc kubenswrapper[4764]: I0202 09:08:48.825579 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 09:08:48 crc kubenswrapper[4764]: E0202 09:08:48.825724 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 09:08:48 crc kubenswrapper[4764]: I0202 09:08:48.825769 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 09:08:48 crc kubenswrapper[4764]: I0202 09:08:48.825859 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 09:08:48 crc kubenswrapper[4764]: E0202 09:08:48.826099 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 09:08:48 crc kubenswrapper[4764]: E0202 09:08:48.825929 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 09:08:48 crc kubenswrapper[4764]: I0202 09:08:48.831900 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-12 17:12:59.227314888 +0000 UTC Feb 02 09:08:48 crc kubenswrapper[4764]: I0202 09:08:48.905522 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:48 crc kubenswrapper[4764]: I0202 09:08:48.905595 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:48 crc kubenswrapper[4764]: I0202 09:08:48.905607 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:48 crc kubenswrapper[4764]: I0202 09:08:48.905631 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:48 crc kubenswrapper[4764]: I0202 09:08:48.905646 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:48Z","lastTransitionTime":"2026-02-02T09:08:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:49 crc kubenswrapper[4764]: I0202 09:08:49.009476 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:49 crc kubenswrapper[4764]: I0202 09:08:49.009546 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:49 crc kubenswrapper[4764]: I0202 09:08:49.009562 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:49 crc kubenswrapper[4764]: I0202 09:08:49.009588 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:49 crc kubenswrapper[4764]: I0202 09:08:49.009611 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:49Z","lastTransitionTime":"2026-02-02T09:08:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:49 crc kubenswrapper[4764]: I0202 09:08:49.112375 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:49 crc kubenswrapper[4764]: I0202 09:08:49.112421 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:49 crc kubenswrapper[4764]: I0202 09:08:49.112437 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:49 crc kubenswrapper[4764]: I0202 09:08:49.112458 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:49 crc kubenswrapper[4764]: I0202 09:08:49.112475 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:49Z","lastTransitionTime":"2026-02-02T09:08:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:49 crc kubenswrapper[4764]: I0202 09:08:49.215629 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:49 crc kubenswrapper[4764]: I0202 09:08:49.215725 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:49 crc kubenswrapper[4764]: I0202 09:08:49.215756 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:49 crc kubenswrapper[4764]: I0202 09:08:49.215791 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:49 crc kubenswrapper[4764]: I0202 09:08:49.215815 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:49Z","lastTransitionTime":"2026-02-02T09:08:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:49 crc kubenswrapper[4764]: I0202 09:08:49.320073 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:49 crc kubenswrapper[4764]: I0202 09:08:49.320543 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:49 crc kubenswrapper[4764]: I0202 09:08:49.320757 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:49 crc kubenswrapper[4764]: I0202 09:08:49.321028 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:49 crc kubenswrapper[4764]: I0202 09:08:49.321228 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:49Z","lastTransitionTime":"2026-02-02T09:08:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:49 crc kubenswrapper[4764]: I0202 09:08:49.425069 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:49 crc kubenswrapper[4764]: I0202 09:08:49.425114 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:49 crc kubenswrapper[4764]: I0202 09:08:49.425127 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:49 crc kubenswrapper[4764]: I0202 09:08:49.425149 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:49 crc kubenswrapper[4764]: I0202 09:08:49.425165 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:49Z","lastTransitionTime":"2026-02-02T09:08:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:49 crc kubenswrapper[4764]: I0202 09:08:49.529455 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:49 crc kubenswrapper[4764]: I0202 09:08:49.529504 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:49 crc kubenswrapper[4764]: I0202 09:08:49.529517 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:49 crc kubenswrapper[4764]: I0202 09:08:49.529542 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:49 crc kubenswrapper[4764]: I0202 09:08:49.529556 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:49Z","lastTransitionTime":"2026-02-02T09:08:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:49 crc kubenswrapper[4764]: I0202 09:08:49.632522 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:49 crc kubenswrapper[4764]: I0202 09:08:49.632782 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:49 crc kubenswrapper[4764]: I0202 09:08:49.632880 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:49 crc kubenswrapper[4764]: I0202 09:08:49.632972 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:49 crc kubenswrapper[4764]: I0202 09:08:49.633046 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:49Z","lastTransitionTime":"2026-02-02T09:08:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:49 crc kubenswrapper[4764]: I0202 09:08:49.737073 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:49 crc kubenswrapper[4764]: I0202 09:08:49.737148 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:49 crc kubenswrapper[4764]: I0202 09:08:49.737170 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:49 crc kubenswrapper[4764]: I0202 09:08:49.737200 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:49 crc kubenswrapper[4764]: I0202 09:08:49.737222 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:49Z","lastTransitionTime":"2026-02-02T09:08:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:49 crc kubenswrapper[4764]: I0202 09:08:49.825612 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kwtmr" Feb 02 09:08:49 crc kubenswrapper[4764]: E0202 09:08:49.825921 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kwtmr" podUID="991faa9a-dd25-4f49-82bd-ce60cefd4af2" Feb 02 09:08:49 crc kubenswrapper[4764]: I0202 09:08:49.833073 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-01 14:42:48.521557416 +0000 UTC Feb 02 09:08:49 crc kubenswrapper[4764]: I0202 09:08:49.839782 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:49 crc kubenswrapper[4764]: I0202 09:08:49.839813 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:49 crc kubenswrapper[4764]: I0202 09:08:49.839958 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:49 crc kubenswrapper[4764]: I0202 09:08:49.839977 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:49 crc kubenswrapper[4764]: I0202 09:08:49.839991 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:49Z","lastTransitionTime":"2026-02-02T09:08:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:49 crc kubenswrapper[4764]: I0202 09:08:49.942772 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:49 crc kubenswrapper[4764]: I0202 09:08:49.942845 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:49 crc kubenswrapper[4764]: I0202 09:08:49.942858 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:49 crc kubenswrapper[4764]: I0202 09:08:49.942879 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:49 crc kubenswrapper[4764]: I0202 09:08:49.942894 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:49Z","lastTransitionTime":"2026-02-02T09:08:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:50 crc kubenswrapper[4764]: I0202 09:08:50.045198 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:50 crc kubenswrapper[4764]: I0202 09:08:50.045536 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:50 crc kubenswrapper[4764]: I0202 09:08:50.045633 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:50 crc kubenswrapper[4764]: I0202 09:08:50.045752 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:50 crc kubenswrapper[4764]: I0202 09:08:50.045836 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:50Z","lastTransitionTime":"2026-02-02T09:08:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:50 crc kubenswrapper[4764]: I0202 09:08:50.148579 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:50 crc kubenswrapper[4764]: I0202 09:08:50.148626 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:50 crc kubenswrapper[4764]: I0202 09:08:50.148643 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:50 crc kubenswrapper[4764]: I0202 09:08:50.148667 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:50 crc kubenswrapper[4764]: I0202 09:08:50.148685 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:50Z","lastTransitionTime":"2026-02-02T09:08:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:50 crc kubenswrapper[4764]: I0202 09:08:50.250812 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:50 crc kubenswrapper[4764]: I0202 09:08:50.250863 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:50 crc kubenswrapper[4764]: I0202 09:08:50.250881 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:50 crc kubenswrapper[4764]: I0202 09:08:50.250903 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:50 crc kubenswrapper[4764]: I0202 09:08:50.250922 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:50Z","lastTransitionTime":"2026-02-02T09:08:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:50 crc kubenswrapper[4764]: I0202 09:08:50.353685 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:50 crc kubenswrapper[4764]: I0202 09:08:50.353714 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:50 crc kubenswrapper[4764]: I0202 09:08:50.353724 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:50 crc kubenswrapper[4764]: I0202 09:08:50.353740 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:50 crc kubenswrapper[4764]: I0202 09:08:50.353750 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:50Z","lastTransitionTime":"2026-02-02T09:08:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:50 crc kubenswrapper[4764]: I0202 09:08:50.456383 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:50 crc kubenswrapper[4764]: I0202 09:08:50.456444 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:50 crc kubenswrapper[4764]: I0202 09:08:50.456468 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:50 crc kubenswrapper[4764]: I0202 09:08:50.456502 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:50 crc kubenswrapper[4764]: I0202 09:08:50.456523 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:50Z","lastTransitionTime":"2026-02-02T09:08:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:50 crc kubenswrapper[4764]: I0202 09:08:50.563364 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:50 crc kubenswrapper[4764]: I0202 09:08:50.563417 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:50 crc kubenswrapper[4764]: I0202 09:08:50.563427 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:50 crc kubenswrapper[4764]: I0202 09:08:50.563440 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:50 crc kubenswrapper[4764]: I0202 09:08:50.563455 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:50Z","lastTransitionTime":"2026-02-02T09:08:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:50 crc kubenswrapper[4764]: I0202 09:08:50.667108 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:50 crc kubenswrapper[4764]: I0202 09:08:50.667158 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:50 crc kubenswrapper[4764]: I0202 09:08:50.667174 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:50 crc kubenswrapper[4764]: I0202 09:08:50.667196 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:50 crc kubenswrapper[4764]: I0202 09:08:50.667212 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:50Z","lastTransitionTime":"2026-02-02T09:08:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:50 crc kubenswrapper[4764]: I0202 09:08:50.769921 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:50 crc kubenswrapper[4764]: I0202 09:08:50.769970 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:50 crc kubenswrapper[4764]: I0202 09:08:50.769979 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:50 crc kubenswrapper[4764]: I0202 09:08:50.769992 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:50 crc kubenswrapper[4764]: I0202 09:08:50.770002 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:50Z","lastTransitionTime":"2026-02-02T09:08:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:50 crc kubenswrapper[4764]: I0202 09:08:50.825509 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 09:08:50 crc kubenswrapper[4764]: E0202 09:08:50.825892 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 09:08:50 crc kubenswrapper[4764]: I0202 09:08:50.825544 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 09:08:50 crc kubenswrapper[4764]: E0202 09:08:50.826146 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 09:08:50 crc kubenswrapper[4764]: I0202 09:08:50.825523 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 09:08:50 crc kubenswrapper[4764]: E0202 09:08:50.826338 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 09:08:50 crc kubenswrapper[4764]: I0202 09:08:50.833950 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-11 16:53:48.955067514 +0000 UTC Feb 02 09:08:50 crc kubenswrapper[4764]: I0202 09:08:50.872687 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:50 crc kubenswrapper[4764]: I0202 09:08:50.872725 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:50 crc kubenswrapper[4764]: I0202 09:08:50.872783 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:50 crc kubenswrapper[4764]: I0202 09:08:50.872805 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:50 crc kubenswrapper[4764]: I0202 09:08:50.872816 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:50Z","lastTransitionTime":"2026-02-02T09:08:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:50 crc kubenswrapper[4764]: I0202 09:08:50.975201 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:50 crc kubenswrapper[4764]: I0202 09:08:50.975255 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:50 crc kubenswrapper[4764]: I0202 09:08:50.975335 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:50 crc kubenswrapper[4764]: I0202 09:08:50.975352 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:50 crc kubenswrapper[4764]: I0202 09:08:50.975363 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:50Z","lastTransitionTime":"2026-02-02T09:08:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:51 crc kubenswrapper[4764]: I0202 09:08:51.077384 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:51 crc kubenswrapper[4764]: I0202 09:08:51.077424 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:51 crc kubenswrapper[4764]: I0202 09:08:51.077434 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:51 crc kubenswrapper[4764]: I0202 09:08:51.077449 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:51 crc kubenswrapper[4764]: I0202 09:08:51.077460 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:51Z","lastTransitionTime":"2026-02-02T09:08:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:51 crc kubenswrapper[4764]: I0202 09:08:51.179347 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:51 crc kubenswrapper[4764]: I0202 09:08:51.179575 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:51 crc kubenswrapper[4764]: I0202 09:08:51.179698 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:51 crc kubenswrapper[4764]: I0202 09:08:51.179818 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:51 crc kubenswrapper[4764]: I0202 09:08:51.179901 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:51Z","lastTransitionTime":"2026-02-02T09:08:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:51 crc kubenswrapper[4764]: I0202 09:08:51.282524 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:51 crc kubenswrapper[4764]: I0202 09:08:51.282558 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:51 crc kubenswrapper[4764]: I0202 09:08:51.282566 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:51 crc kubenswrapper[4764]: I0202 09:08:51.282579 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:51 crc kubenswrapper[4764]: I0202 09:08:51.282589 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:51Z","lastTransitionTime":"2026-02-02T09:08:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:51 crc kubenswrapper[4764]: I0202 09:08:51.386219 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:51 crc kubenswrapper[4764]: I0202 09:08:51.386292 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:51 crc kubenswrapper[4764]: I0202 09:08:51.386314 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:51 crc kubenswrapper[4764]: I0202 09:08:51.386341 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:51 crc kubenswrapper[4764]: I0202 09:08:51.386362 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:51Z","lastTransitionTime":"2026-02-02T09:08:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:51 crc kubenswrapper[4764]: I0202 09:08:51.489584 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:51 crc kubenswrapper[4764]: I0202 09:08:51.489623 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:51 crc kubenswrapper[4764]: I0202 09:08:51.489634 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:51 crc kubenswrapper[4764]: I0202 09:08:51.489649 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:51 crc kubenswrapper[4764]: I0202 09:08:51.489660 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:51Z","lastTransitionTime":"2026-02-02T09:08:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:51 crc kubenswrapper[4764]: I0202 09:08:51.592708 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:51 crc kubenswrapper[4764]: I0202 09:08:51.592748 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:51 crc kubenswrapper[4764]: I0202 09:08:51.592762 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:51 crc kubenswrapper[4764]: I0202 09:08:51.592778 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:51 crc kubenswrapper[4764]: I0202 09:08:51.592790 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:51Z","lastTransitionTime":"2026-02-02T09:08:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:51 crc kubenswrapper[4764]: I0202 09:08:51.695140 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:51 crc kubenswrapper[4764]: I0202 09:08:51.695182 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:51 crc kubenswrapper[4764]: I0202 09:08:51.695194 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:51 crc kubenswrapper[4764]: I0202 09:08:51.695211 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:51 crc kubenswrapper[4764]: I0202 09:08:51.695222 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:51Z","lastTransitionTime":"2026-02-02T09:08:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:51 crc kubenswrapper[4764]: I0202 09:08:51.798352 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:51 crc kubenswrapper[4764]: I0202 09:08:51.798394 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:51 crc kubenswrapper[4764]: I0202 09:08:51.798405 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:51 crc kubenswrapper[4764]: I0202 09:08:51.798422 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:51 crc kubenswrapper[4764]: I0202 09:08:51.798434 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:51Z","lastTransitionTime":"2026-02-02T09:08:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:51 crc kubenswrapper[4764]: I0202 09:08:51.824915 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kwtmr" Feb 02 09:08:51 crc kubenswrapper[4764]: E0202 09:08:51.825333 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kwtmr" podUID="991faa9a-dd25-4f49-82bd-ce60cefd4af2" Feb 02 09:08:51 crc kubenswrapper[4764]: I0202 09:08:51.835023 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-19 15:49:20.736692187 +0000 UTC Feb 02 09:08:51 crc kubenswrapper[4764]: I0202 09:08:51.901566 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:51 crc kubenswrapper[4764]: I0202 09:08:51.901627 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:51 crc kubenswrapper[4764]: I0202 09:08:51.901652 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:51 crc kubenswrapper[4764]: I0202 09:08:51.901680 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:51 crc kubenswrapper[4764]: I0202 09:08:51.901697 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:51Z","lastTransitionTime":"2026-02-02T09:08:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:52 crc kubenswrapper[4764]: I0202 09:08:52.004549 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:52 crc kubenswrapper[4764]: I0202 09:08:52.004584 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:52 crc kubenswrapper[4764]: I0202 09:08:52.004595 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:52 crc kubenswrapper[4764]: I0202 09:08:52.004608 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:52 crc kubenswrapper[4764]: I0202 09:08:52.004617 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:52Z","lastTransitionTime":"2026-02-02T09:08:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:52 crc kubenswrapper[4764]: I0202 09:08:52.106698 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:52 crc kubenswrapper[4764]: I0202 09:08:52.106749 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:52 crc kubenswrapper[4764]: I0202 09:08:52.106760 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:52 crc kubenswrapper[4764]: I0202 09:08:52.106777 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:52 crc kubenswrapper[4764]: I0202 09:08:52.106790 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:52Z","lastTransitionTime":"2026-02-02T09:08:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:52 crc kubenswrapper[4764]: I0202 09:08:52.209263 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:52 crc kubenswrapper[4764]: I0202 09:08:52.209293 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:52 crc kubenswrapper[4764]: I0202 09:08:52.209303 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:52 crc kubenswrapper[4764]: I0202 09:08:52.209318 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:52 crc kubenswrapper[4764]: I0202 09:08:52.209329 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:52Z","lastTransitionTime":"2026-02-02T09:08:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:52 crc kubenswrapper[4764]: I0202 09:08:52.312093 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:52 crc kubenswrapper[4764]: I0202 09:08:52.312126 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:52 crc kubenswrapper[4764]: I0202 09:08:52.312138 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:52 crc kubenswrapper[4764]: I0202 09:08:52.312153 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:52 crc kubenswrapper[4764]: I0202 09:08:52.312163 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:52Z","lastTransitionTime":"2026-02-02T09:08:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:52 crc kubenswrapper[4764]: I0202 09:08:52.415630 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:52 crc kubenswrapper[4764]: I0202 09:08:52.415678 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:52 crc kubenswrapper[4764]: I0202 09:08:52.415690 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:52 crc kubenswrapper[4764]: I0202 09:08:52.415707 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:52 crc kubenswrapper[4764]: I0202 09:08:52.415720 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:52Z","lastTransitionTime":"2026-02-02T09:08:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:52 crc kubenswrapper[4764]: I0202 09:08:52.518397 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:52 crc kubenswrapper[4764]: I0202 09:08:52.518435 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:52 crc kubenswrapper[4764]: I0202 09:08:52.518445 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:52 crc kubenswrapper[4764]: I0202 09:08:52.518459 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:52 crc kubenswrapper[4764]: I0202 09:08:52.518468 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:52Z","lastTransitionTime":"2026-02-02T09:08:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:52 crc kubenswrapper[4764]: I0202 09:08:52.621231 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:52 crc kubenswrapper[4764]: I0202 09:08:52.621269 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:52 crc kubenswrapper[4764]: I0202 09:08:52.621277 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:52 crc kubenswrapper[4764]: I0202 09:08:52.621291 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:52 crc kubenswrapper[4764]: I0202 09:08:52.621300 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:52Z","lastTransitionTime":"2026-02-02T09:08:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:52 crc kubenswrapper[4764]: I0202 09:08:52.724156 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:52 crc kubenswrapper[4764]: I0202 09:08:52.724192 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:52 crc kubenswrapper[4764]: I0202 09:08:52.724202 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:52 crc kubenswrapper[4764]: I0202 09:08:52.724220 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:52 crc kubenswrapper[4764]: I0202 09:08:52.724230 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:52Z","lastTransitionTime":"2026-02-02T09:08:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:52 crc kubenswrapper[4764]: I0202 09:08:52.824747 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 09:08:52 crc kubenswrapper[4764]: E0202 09:08:52.825171 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 09:08:52 crc kubenswrapper[4764]: I0202 09:08:52.825507 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 09:08:52 crc kubenswrapper[4764]: E0202 09:08:52.825641 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 09:08:52 crc kubenswrapper[4764]: I0202 09:08:52.825882 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 09:08:52 crc kubenswrapper[4764]: E0202 09:08:52.826027 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 09:08:52 crc kubenswrapper[4764]: I0202 09:08:52.826196 4764 scope.go:117] "RemoveContainer" containerID="3cb255e14eb02f7c08364c41fb9906e58b2155eb507f694b77ec8eb1f4251343" Feb 02 09:08:52 crc kubenswrapper[4764]: E0202 09:08:52.826445 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-zhn7j_openshift-ovn-kubernetes(24632dda-6100-4ab6-a28e-214ddae4360c)\"" pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" podUID="24632dda-6100-4ab6-a28e-214ddae4360c" Feb 02 09:08:52 crc kubenswrapper[4764]: I0202 09:08:52.827663 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:52 crc kubenswrapper[4764]: I0202 09:08:52.827720 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:52 crc kubenswrapper[4764]: I0202 09:08:52.827754 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:52 crc kubenswrapper[4764]: I0202 09:08:52.827779 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:52 crc kubenswrapper[4764]: I0202 09:08:52.827800 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:52Z","lastTransitionTime":"2026-02-02T09:08:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:52 crc kubenswrapper[4764]: I0202 09:08:52.835766 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-11 04:51:04.136310749 +0000 UTC Feb 02 09:08:52 crc kubenswrapper[4764]: I0202 09:08:52.931371 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:52 crc kubenswrapper[4764]: I0202 09:08:52.931422 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:52 crc kubenswrapper[4764]: I0202 09:08:52.931438 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:52 crc kubenswrapper[4764]: I0202 09:08:52.931461 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:52 crc kubenswrapper[4764]: I0202 09:08:52.931478 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:52Z","lastTransitionTime":"2026-02-02T09:08:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:53 crc kubenswrapper[4764]: I0202 09:08:53.034458 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:53 crc kubenswrapper[4764]: I0202 09:08:53.034527 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:53 crc kubenswrapper[4764]: I0202 09:08:53.034544 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:53 crc kubenswrapper[4764]: I0202 09:08:53.034571 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:53 crc kubenswrapper[4764]: I0202 09:08:53.034596 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:53Z","lastTransitionTime":"2026-02-02T09:08:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:53 crc kubenswrapper[4764]: I0202 09:08:53.138481 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:53 crc kubenswrapper[4764]: I0202 09:08:53.138546 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:53 crc kubenswrapper[4764]: I0202 09:08:53.138566 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:53 crc kubenswrapper[4764]: I0202 09:08:53.138592 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:53 crc kubenswrapper[4764]: I0202 09:08:53.138620 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:53Z","lastTransitionTime":"2026-02-02T09:08:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:53 crc kubenswrapper[4764]: I0202 09:08:53.241541 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:53 crc kubenswrapper[4764]: I0202 09:08:53.241571 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:53 crc kubenswrapper[4764]: I0202 09:08:53.241579 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:53 crc kubenswrapper[4764]: I0202 09:08:53.241593 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:53 crc kubenswrapper[4764]: I0202 09:08:53.241601 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:53Z","lastTransitionTime":"2026-02-02T09:08:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:53 crc kubenswrapper[4764]: I0202 09:08:53.344324 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:53 crc kubenswrapper[4764]: I0202 09:08:53.344652 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:53 crc kubenswrapper[4764]: I0202 09:08:53.344664 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:53 crc kubenswrapper[4764]: I0202 09:08:53.344699 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:53 crc kubenswrapper[4764]: I0202 09:08:53.344715 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:53Z","lastTransitionTime":"2026-02-02T09:08:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:53 crc kubenswrapper[4764]: I0202 09:08:53.446761 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:53 crc kubenswrapper[4764]: I0202 09:08:53.446809 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:53 crc kubenswrapper[4764]: I0202 09:08:53.446820 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:53 crc kubenswrapper[4764]: I0202 09:08:53.446836 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:53 crc kubenswrapper[4764]: I0202 09:08:53.446847 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:53Z","lastTransitionTime":"2026-02-02T09:08:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:53 crc kubenswrapper[4764]: I0202 09:08:53.549197 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:53 crc kubenswrapper[4764]: I0202 09:08:53.549233 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:53 crc kubenswrapper[4764]: I0202 09:08:53.549242 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:53 crc kubenswrapper[4764]: I0202 09:08:53.549255 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:53 crc kubenswrapper[4764]: I0202 09:08:53.549264 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:53Z","lastTransitionTime":"2026-02-02T09:08:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:53 crc kubenswrapper[4764]: I0202 09:08:53.652284 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:53 crc kubenswrapper[4764]: I0202 09:08:53.652551 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:53 crc kubenswrapper[4764]: I0202 09:08:53.652710 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:53 crc kubenswrapper[4764]: I0202 09:08:53.652868 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:53 crc kubenswrapper[4764]: I0202 09:08:53.652986 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:53Z","lastTransitionTime":"2026-02-02T09:08:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:53 crc kubenswrapper[4764]: I0202 09:08:53.754528 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:53 crc kubenswrapper[4764]: I0202 09:08:53.754555 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:53 crc kubenswrapper[4764]: I0202 09:08:53.754563 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:53 crc kubenswrapper[4764]: I0202 09:08:53.754576 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:53 crc kubenswrapper[4764]: I0202 09:08:53.754584 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:53Z","lastTransitionTime":"2026-02-02T09:08:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:53 crc kubenswrapper[4764]: I0202 09:08:53.825281 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kwtmr" Feb 02 09:08:53 crc kubenswrapper[4764]: E0202 09:08:53.825560 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kwtmr" podUID="991faa9a-dd25-4f49-82bd-ce60cefd4af2" Feb 02 09:08:53 crc kubenswrapper[4764]: I0202 09:08:53.837261 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-05 00:58:58.102674233 +0000 UTC Feb 02 09:08:53 crc kubenswrapper[4764]: I0202 09:08:53.856739 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:53 crc kubenswrapper[4764]: I0202 09:08:53.856779 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:53 crc kubenswrapper[4764]: I0202 09:08:53.856789 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:53 crc kubenswrapper[4764]: I0202 09:08:53.856803 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:53 crc kubenswrapper[4764]: I0202 09:08:53.856814 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:53Z","lastTransitionTime":"2026-02-02T09:08:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:53 crc kubenswrapper[4764]: I0202 09:08:53.959305 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:53 crc kubenswrapper[4764]: I0202 09:08:53.959371 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:53 crc kubenswrapper[4764]: I0202 09:08:53.959391 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:53 crc kubenswrapper[4764]: I0202 09:08:53.959416 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:53 crc kubenswrapper[4764]: I0202 09:08:53.959435 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:53Z","lastTransitionTime":"2026-02-02T09:08:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:54 crc kubenswrapper[4764]: I0202 09:08:54.063265 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:54 crc kubenswrapper[4764]: I0202 09:08:54.063309 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:54 crc kubenswrapper[4764]: I0202 09:08:54.063319 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:54 crc kubenswrapper[4764]: I0202 09:08:54.063336 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:54 crc kubenswrapper[4764]: I0202 09:08:54.063345 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:54Z","lastTransitionTime":"2026-02-02T09:08:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:54 crc kubenswrapper[4764]: I0202 09:08:54.166035 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:54 crc kubenswrapper[4764]: I0202 09:08:54.166074 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:54 crc kubenswrapper[4764]: I0202 09:08:54.166087 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:54 crc kubenswrapper[4764]: I0202 09:08:54.166103 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:54 crc kubenswrapper[4764]: I0202 09:08:54.166116 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:54Z","lastTransitionTime":"2026-02-02T09:08:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:54 crc kubenswrapper[4764]: I0202 09:08:54.269739 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:54 crc kubenswrapper[4764]: I0202 09:08:54.269773 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:54 crc kubenswrapper[4764]: I0202 09:08:54.269784 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:54 crc kubenswrapper[4764]: I0202 09:08:54.269800 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:54 crc kubenswrapper[4764]: I0202 09:08:54.269812 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:54Z","lastTransitionTime":"2026-02-02T09:08:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:54 crc kubenswrapper[4764]: I0202 09:08:54.373681 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:54 crc kubenswrapper[4764]: I0202 09:08:54.374039 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:54 crc kubenswrapper[4764]: I0202 09:08:54.374328 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:54 crc kubenswrapper[4764]: I0202 09:08:54.374470 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:54 crc kubenswrapper[4764]: I0202 09:08:54.374570 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:54Z","lastTransitionTime":"2026-02-02T09:08:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:54 crc kubenswrapper[4764]: I0202 09:08:54.477465 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:54 crc kubenswrapper[4764]: I0202 09:08:54.477756 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:54 crc kubenswrapper[4764]: I0202 09:08:54.477827 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:54 crc kubenswrapper[4764]: I0202 09:08:54.477899 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:54 crc kubenswrapper[4764]: I0202 09:08:54.477976 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:54Z","lastTransitionTime":"2026-02-02T09:08:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:54 crc kubenswrapper[4764]: I0202 09:08:54.581232 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:54 crc kubenswrapper[4764]: I0202 09:08:54.581271 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:54 crc kubenswrapper[4764]: I0202 09:08:54.581279 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:54 crc kubenswrapper[4764]: I0202 09:08:54.581292 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:54 crc kubenswrapper[4764]: I0202 09:08:54.581302 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:54Z","lastTransitionTime":"2026-02-02T09:08:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:54 crc kubenswrapper[4764]: I0202 09:08:54.683921 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:54 crc kubenswrapper[4764]: I0202 09:08:54.684208 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:54 crc kubenswrapper[4764]: I0202 09:08:54.684442 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:54 crc kubenswrapper[4764]: I0202 09:08:54.684621 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:54 crc kubenswrapper[4764]: I0202 09:08:54.684782 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:54Z","lastTransitionTime":"2026-02-02T09:08:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:54 crc kubenswrapper[4764]: I0202 09:08:54.787763 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:54 crc kubenswrapper[4764]: I0202 09:08:54.787796 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:54 crc kubenswrapper[4764]: I0202 09:08:54.787804 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:54 crc kubenswrapper[4764]: I0202 09:08:54.787819 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:54 crc kubenswrapper[4764]: I0202 09:08:54.787832 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:54Z","lastTransitionTime":"2026-02-02T09:08:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:54 crc kubenswrapper[4764]: I0202 09:08:54.824734 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 09:08:54 crc kubenswrapper[4764]: E0202 09:08:54.825064 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 09:08:54 crc kubenswrapper[4764]: I0202 09:08:54.825415 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 09:08:54 crc kubenswrapper[4764]: I0202 09:08:54.825450 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 09:08:54 crc kubenswrapper[4764]: E0202 09:08:54.825741 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 09:08:54 crc kubenswrapper[4764]: E0202 09:08:54.825849 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 09:08:54 crc kubenswrapper[4764]: I0202 09:08:54.837905 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-06 18:09:13.859653481 +0000 UTC Feb 02 09:08:54 crc kubenswrapper[4764]: I0202 09:08:54.891955 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:54 crc kubenswrapper[4764]: I0202 09:08:54.892029 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:54 crc kubenswrapper[4764]: I0202 09:08:54.892046 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:54 crc kubenswrapper[4764]: I0202 09:08:54.892076 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:54 crc kubenswrapper[4764]: I0202 09:08:54.892097 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:54Z","lastTransitionTime":"2026-02-02T09:08:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:54 crc kubenswrapper[4764]: I0202 09:08:54.994397 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:54 crc kubenswrapper[4764]: I0202 09:08:54.994999 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:54 crc kubenswrapper[4764]: I0202 09:08:54.995030 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:54 crc kubenswrapper[4764]: I0202 09:08:54.995046 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:54 crc kubenswrapper[4764]: I0202 09:08:54.995057 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:54Z","lastTransitionTime":"2026-02-02T09:08:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:55 crc kubenswrapper[4764]: I0202 09:08:55.098061 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:55 crc kubenswrapper[4764]: I0202 09:08:55.098105 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:55 crc kubenswrapper[4764]: I0202 09:08:55.098116 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:55 crc kubenswrapper[4764]: I0202 09:08:55.098133 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:55 crc kubenswrapper[4764]: I0202 09:08:55.098144 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:55Z","lastTransitionTime":"2026-02-02T09:08:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:55 crc kubenswrapper[4764]: I0202 09:08:55.200863 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:55 crc kubenswrapper[4764]: I0202 09:08:55.200916 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:55 crc kubenswrapper[4764]: I0202 09:08:55.200961 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:55 crc kubenswrapper[4764]: I0202 09:08:55.200988 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:55 crc kubenswrapper[4764]: I0202 09:08:55.201007 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:55Z","lastTransitionTime":"2026-02-02T09:08:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:55 crc kubenswrapper[4764]: I0202 09:08:55.303844 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:55 crc kubenswrapper[4764]: I0202 09:08:55.303881 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:55 crc kubenswrapper[4764]: I0202 09:08:55.303895 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:55 crc kubenswrapper[4764]: I0202 09:08:55.303914 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:55 crc kubenswrapper[4764]: I0202 09:08:55.303925 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:55Z","lastTransitionTime":"2026-02-02T09:08:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:55 crc kubenswrapper[4764]: I0202 09:08:55.406071 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:55 crc kubenswrapper[4764]: I0202 09:08:55.406320 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:55 crc kubenswrapper[4764]: I0202 09:08:55.406505 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:55 crc kubenswrapper[4764]: I0202 09:08:55.406640 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:55 crc kubenswrapper[4764]: I0202 09:08:55.406754 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:55Z","lastTransitionTime":"2026-02-02T09:08:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:55 crc kubenswrapper[4764]: I0202 09:08:55.509415 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:55 crc kubenswrapper[4764]: I0202 09:08:55.509791 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:55 crc kubenswrapper[4764]: I0202 09:08:55.509988 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:55 crc kubenswrapper[4764]: I0202 09:08:55.510114 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:55 crc kubenswrapper[4764]: I0202 09:08:55.510195 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:55Z","lastTransitionTime":"2026-02-02T09:08:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:55 crc kubenswrapper[4764]: I0202 09:08:55.612688 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:55 crc kubenswrapper[4764]: I0202 09:08:55.612748 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:55 crc kubenswrapper[4764]: I0202 09:08:55.612757 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:55 crc kubenswrapper[4764]: I0202 09:08:55.612773 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:55 crc kubenswrapper[4764]: I0202 09:08:55.612784 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:55Z","lastTransitionTime":"2026-02-02T09:08:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:55 crc kubenswrapper[4764]: I0202 09:08:55.716062 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:55 crc kubenswrapper[4764]: I0202 09:08:55.716109 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:55 crc kubenswrapper[4764]: I0202 09:08:55.716125 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:55 crc kubenswrapper[4764]: I0202 09:08:55.716148 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:55 crc kubenswrapper[4764]: I0202 09:08:55.716167 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:55Z","lastTransitionTime":"2026-02-02T09:08:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:55 crc kubenswrapper[4764]: I0202 09:08:55.819342 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:55 crc kubenswrapper[4764]: I0202 09:08:55.819385 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:55 crc kubenswrapper[4764]: I0202 09:08:55.819398 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:55 crc kubenswrapper[4764]: I0202 09:08:55.819413 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:55 crc kubenswrapper[4764]: I0202 09:08:55.819423 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:55Z","lastTransitionTime":"2026-02-02T09:08:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:55 crc kubenswrapper[4764]: I0202 09:08:55.825957 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kwtmr" Feb 02 09:08:55 crc kubenswrapper[4764]: E0202 09:08:55.826100 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kwtmr" podUID="991faa9a-dd25-4f49-82bd-ce60cefd4af2" Feb 02 09:08:55 crc kubenswrapper[4764]: I0202 09:08:55.838782 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-13 15:55:56.931888222 +0000 UTC Feb 02 09:08:55 crc kubenswrapper[4764]: I0202 09:08:55.922529 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:55 crc kubenswrapper[4764]: I0202 09:08:55.922574 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:55 crc kubenswrapper[4764]: I0202 09:08:55.922586 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:55 crc kubenswrapper[4764]: I0202 09:08:55.922603 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:55 crc kubenswrapper[4764]: I0202 09:08:55.922614 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:55Z","lastTransitionTime":"2026-02-02T09:08:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:56 crc kubenswrapper[4764]: I0202 09:08:56.025550 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:56 crc kubenswrapper[4764]: I0202 09:08:56.025619 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:56 crc kubenswrapper[4764]: I0202 09:08:56.025639 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:56 crc kubenswrapper[4764]: I0202 09:08:56.025663 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:56 crc kubenswrapper[4764]: I0202 09:08:56.025680 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:56Z","lastTransitionTime":"2026-02-02T09:08:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:56 crc kubenswrapper[4764]: I0202 09:08:56.128637 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:56 crc kubenswrapper[4764]: I0202 09:08:56.128698 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:56 crc kubenswrapper[4764]: I0202 09:08:56.128717 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:56 crc kubenswrapper[4764]: I0202 09:08:56.128741 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:56 crc kubenswrapper[4764]: I0202 09:08:56.128758 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:56Z","lastTransitionTime":"2026-02-02T09:08:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:56 crc kubenswrapper[4764]: I0202 09:08:56.231259 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:56 crc kubenswrapper[4764]: I0202 09:08:56.231314 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:56 crc kubenswrapper[4764]: I0202 09:08:56.231327 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:56 crc kubenswrapper[4764]: I0202 09:08:56.231344 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:56 crc kubenswrapper[4764]: I0202 09:08:56.231357 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:56Z","lastTransitionTime":"2026-02-02T09:08:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:56 crc kubenswrapper[4764]: I0202 09:08:56.335267 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:56 crc kubenswrapper[4764]: I0202 09:08:56.335315 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:56 crc kubenswrapper[4764]: I0202 09:08:56.335326 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:56 crc kubenswrapper[4764]: I0202 09:08:56.335343 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:56 crc kubenswrapper[4764]: I0202 09:08:56.335355 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:56Z","lastTransitionTime":"2026-02-02T09:08:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:56 crc kubenswrapper[4764]: I0202 09:08:56.437137 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:56 crc kubenswrapper[4764]: I0202 09:08:56.437194 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:56 crc kubenswrapper[4764]: I0202 09:08:56.437211 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:56 crc kubenswrapper[4764]: I0202 09:08:56.437233 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:56 crc kubenswrapper[4764]: I0202 09:08:56.437250 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:56Z","lastTransitionTime":"2026-02-02T09:08:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:56 crc kubenswrapper[4764]: I0202 09:08:56.539768 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:56 crc kubenswrapper[4764]: I0202 09:08:56.539813 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:56 crc kubenswrapper[4764]: I0202 09:08:56.539823 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:56 crc kubenswrapper[4764]: I0202 09:08:56.539837 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:56 crc kubenswrapper[4764]: I0202 09:08:56.539852 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:56Z","lastTransitionTime":"2026-02-02T09:08:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:56 crc kubenswrapper[4764]: I0202 09:08:56.642916 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:56 crc kubenswrapper[4764]: I0202 09:08:56.643074 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:56 crc kubenswrapper[4764]: I0202 09:08:56.643100 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:56 crc kubenswrapper[4764]: I0202 09:08:56.643123 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:56 crc kubenswrapper[4764]: I0202 09:08:56.643141 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:56Z","lastTransitionTime":"2026-02-02T09:08:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:56 crc kubenswrapper[4764]: I0202 09:08:56.746030 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:56 crc kubenswrapper[4764]: I0202 09:08:56.746090 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:56 crc kubenswrapper[4764]: I0202 09:08:56.746112 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:56 crc kubenswrapper[4764]: I0202 09:08:56.746149 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:56 crc kubenswrapper[4764]: I0202 09:08:56.746174 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:56Z","lastTransitionTime":"2026-02-02T09:08:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:56 crc kubenswrapper[4764]: I0202 09:08:56.825263 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 09:08:56 crc kubenswrapper[4764]: I0202 09:08:56.825303 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 09:08:56 crc kubenswrapper[4764]: I0202 09:08:56.825303 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 09:08:56 crc kubenswrapper[4764]: E0202 09:08:56.825478 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 09:08:56 crc kubenswrapper[4764]: E0202 09:08:56.825569 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 09:08:56 crc kubenswrapper[4764]: E0202 09:08:56.825656 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 09:08:56 crc kubenswrapper[4764]: I0202 09:08:56.839310 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-16 16:50:01.720058949 +0000 UTC Feb 02 09:08:56 crc kubenswrapper[4764]: I0202 09:08:56.848023 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:56 crc kubenswrapper[4764]: I0202 09:08:56.848078 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:56 crc kubenswrapper[4764]: I0202 09:08:56.848089 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:56 crc kubenswrapper[4764]: I0202 09:08:56.848106 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:56 crc kubenswrapper[4764]: I0202 09:08:56.848118 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:56Z","lastTransitionTime":"2026-02-02T09:08:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:56 crc kubenswrapper[4764]: I0202 09:08:56.877770 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:56 crc kubenswrapper[4764]: I0202 09:08:56.877817 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:56 crc kubenswrapper[4764]: I0202 09:08:56.877829 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:56 crc kubenswrapper[4764]: I0202 09:08:56.877845 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:56 crc kubenswrapper[4764]: I0202 09:08:56.877857 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:56Z","lastTransitionTime":"2026-02-02T09:08:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:56 crc kubenswrapper[4764]: E0202 09:08:56.892023 4764 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:08:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:08:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:08:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:08:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"31973a70-ce17-4811-9a14-f5476c6979db\\\",\\\"systemUUID\\\":\\\"ca92679c-8580-43c3-be51-62441e93d22b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:56Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:56 crc kubenswrapper[4764]: I0202 09:08:56.896112 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:56 crc kubenswrapper[4764]: I0202 09:08:56.896154 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:56 crc kubenswrapper[4764]: I0202 09:08:56.896167 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:56 crc kubenswrapper[4764]: I0202 09:08:56.896182 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:56 crc kubenswrapper[4764]: I0202 09:08:56.896198 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:56Z","lastTransitionTime":"2026-02-02T09:08:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:56 crc kubenswrapper[4764]: E0202 09:08:56.909920 4764 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:08:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:08:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:08:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:08:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"31973a70-ce17-4811-9a14-f5476c6979db\\\",\\\"systemUUID\\\":\\\"ca92679c-8580-43c3-be51-62441e93d22b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:56Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:56 crc kubenswrapper[4764]: I0202 09:08:56.914918 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:56 crc kubenswrapper[4764]: I0202 09:08:56.914987 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:56 crc kubenswrapper[4764]: I0202 09:08:56.914999 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:56 crc kubenswrapper[4764]: I0202 09:08:56.915015 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:56 crc kubenswrapper[4764]: I0202 09:08:56.915028 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:56Z","lastTransitionTime":"2026-02-02T09:08:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:56 crc kubenswrapper[4764]: E0202 09:08:56.926858 4764 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:08:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:08:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:08:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:08:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"31973a70-ce17-4811-9a14-f5476c6979db\\\",\\\"systemUUID\\\":\\\"ca92679c-8580-43c3-be51-62441e93d22b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:56Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:56 crc kubenswrapper[4764]: I0202 09:08:56.930310 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:56 crc kubenswrapper[4764]: I0202 09:08:56.930334 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:56 crc kubenswrapper[4764]: I0202 09:08:56.930344 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:56 crc kubenswrapper[4764]: I0202 09:08:56.930359 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:56 crc kubenswrapper[4764]: I0202 09:08:56.930369 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:56Z","lastTransitionTime":"2026-02-02T09:08:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:56 crc kubenswrapper[4764]: E0202 09:08:56.949534 4764 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:08:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:08:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:08:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:08:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"31973a70-ce17-4811-9a14-f5476c6979db\\\",\\\"systemUUID\\\":\\\"ca92679c-8580-43c3-be51-62441e93d22b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:56Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:56 crc kubenswrapper[4764]: I0202 09:08:56.953049 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:56 crc kubenswrapper[4764]: I0202 09:08:56.953113 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:56 crc kubenswrapper[4764]: I0202 09:08:56.953123 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:56 crc kubenswrapper[4764]: I0202 09:08:56.953152 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:56 crc kubenswrapper[4764]: I0202 09:08:56.953162 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:56Z","lastTransitionTime":"2026-02-02T09:08:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:56 crc kubenswrapper[4764]: E0202 09:08:56.965130 4764 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:08:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:08:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:08:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:08:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"31973a70-ce17-4811-9a14-f5476c6979db\\\",\\\"systemUUID\\\":\\\"ca92679c-8580-43c3-be51-62441e93d22b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:56Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:56 crc kubenswrapper[4764]: E0202 09:08:56.965245 4764 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 02 09:08:56 crc kubenswrapper[4764]: I0202 09:08:56.966692 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:56 crc kubenswrapper[4764]: I0202 09:08:56.966724 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:56 crc kubenswrapper[4764]: I0202 09:08:56.966738 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:56 crc kubenswrapper[4764]: I0202 09:08:56.966753 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:56 crc kubenswrapper[4764]: I0202 09:08:56.966762 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:56Z","lastTransitionTime":"2026-02-02T09:08:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:57 crc kubenswrapper[4764]: I0202 09:08:57.069060 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:57 crc kubenswrapper[4764]: I0202 09:08:57.069100 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:57 crc kubenswrapper[4764]: I0202 09:08:57.069109 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:57 crc kubenswrapper[4764]: I0202 09:08:57.069121 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:57 crc kubenswrapper[4764]: I0202 09:08:57.069133 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:57Z","lastTransitionTime":"2026-02-02T09:08:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:57 crc kubenswrapper[4764]: I0202 09:08:57.171703 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:57 crc kubenswrapper[4764]: I0202 09:08:57.171751 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:57 crc kubenswrapper[4764]: I0202 09:08:57.171768 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:57 crc kubenswrapper[4764]: I0202 09:08:57.171788 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:57 crc kubenswrapper[4764]: I0202 09:08:57.171797 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:57Z","lastTransitionTime":"2026-02-02T09:08:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:57 crc kubenswrapper[4764]: I0202 09:08:57.274320 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:57 crc kubenswrapper[4764]: I0202 09:08:57.274369 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:57 crc kubenswrapper[4764]: I0202 09:08:57.274381 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:57 crc kubenswrapper[4764]: I0202 09:08:57.274404 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:57 crc kubenswrapper[4764]: I0202 09:08:57.274417 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:57Z","lastTransitionTime":"2026-02-02T09:08:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:57 crc kubenswrapper[4764]: I0202 09:08:57.376672 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:57 crc kubenswrapper[4764]: I0202 09:08:57.376704 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:57 crc kubenswrapper[4764]: I0202 09:08:57.376715 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:57 crc kubenswrapper[4764]: I0202 09:08:57.376732 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:57 crc kubenswrapper[4764]: I0202 09:08:57.376742 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:57Z","lastTransitionTime":"2026-02-02T09:08:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:57 crc kubenswrapper[4764]: I0202 09:08:57.479185 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:57 crc kubenswrapper[4764]: I0202 09:08:57.479245 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:57 crc kubenswrapper[4764]: I0202 09:08:57.479258 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:57 crc kubenswrapper[4764]: I0202 09:08:57.479285 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:57 crc kubenswrapper[4764]: I0202 09:08:57.479299 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:57Z","lastTransitionTime":"2026-02-02T09:08:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:57 crc kubenswrapper[4764]: I0202 09:08:57.581438 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:57 crc kubenswrapper[4764]: I0202 09:08:57.581476 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:57 crc kubenswrapper[4764]: I0202 09:08:57.581485 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:57 crc kubenswrapper[4764]: I0202 09:08:57.581497 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:57 crc kubenswrapper[4764]: I0202 09:08:57.581506 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:57Z","lastTransitionTime":"2026-02-02T09:08:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:57 crc kubenswrapper[4764]: I0202 09:08:57.687621 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:57 crc kubenswrapper[4764]: I0202 09:08:57.687695 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:57 crc kubenswrapper[4764]: I0202 09:08:57.687721 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:57 crc kubenswrapper[4764]: I0202 09:08:57.687752 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:57 crc kubenswrapper[4764]: I0202 09:08:57.688067 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:57Z","lastTransitionTime":"2026-02-02T09:08:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:57 crc kubenswrapper[4764]: I0202 09:08:57.791016 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:57 crc kubenswrapper[4764]: I0202 09:08:57.791051 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:57 crc kubenswrapper[4764]: I0202 09:08:57.791062 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:57 crc kubenswrapper[4764]: I0202 09:08:57.791077 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:57 crc kubenswrapper[4764]: I0202 09:08:57.791089 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:57Z","lastTransitionTime":"2026-02-02T09:08:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:57 crc kubenswrapper[4764]: I0202 09:08:57.825213 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kwtmr" Feb 02 09:08:57 crc kubenswrapper[4764]: E0202 09:08:57.825328 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kwtmr" podUID="991faa9a-dd25-4f49-82bd-ce60cefd4af2" Feb 02 09:08:57 crc kubenswrapper[4764]: I0202 09:08:57.836329 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fea07bdc-9134-4993-934c-316c19d0dc76\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://664e0519c2a2c60904dfcc8deebaf5d638dac0bedb13d870c0aaea5b4f5428f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db5a54d3a2117252a1ff50884683f99185874d2904c0960161642280f614679d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db5a54d3a2117252a1ff50884683f99185874d2904c0960161642280f614679d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:57Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:57 crc kubenswrapper[4764]: I0202 09:08:57.839913 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-21 02:10:17.496457207 +0000 UTC Feb 02 09:08:57 crc kubenswrapper[4764]: I0202 09:08:57.852261 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5528225d91ac2778bef4554abb0bd036ac5d3b6cdac7357bab69e70c3894b6f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://071bbb80d0747cf3e109ba626d47f7d05f85f4cc0b3aba77276c795741f47ac6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:57Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:57 crc kubenswrapper[4764]: I0202 09:08:57.863019 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:57Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:57 crc kubenswrapper[4764]: I0202 09:08:57.874033 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nc7qt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b52b499d17b4e2e359475bd13730c72ef5cedb052e13def6dfe5bc5fd7487d2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1a2c7b22068b7165927fc158885d48bf9952b9eba521d8ae9a599eee53dc305\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-02T09:08:28Z\\\",\\\"message\\\":\\\"2026-02-02T09:07:42+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_32cbaf6a-865c-4787-8f89-a40249873b40\\\\n2026-02-02T09:07:42+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_32cbaf6a-865c-4787-8f89-a40249873b40 to /host/opt/cni/bin/\\\\n2026-02-02T09:07:43Z [verbose] multus-daemon started\\\\n2026-02-02T09:07:43Z [verbose] Readiness Indicator file check\\\\n2026-02-02T09:08:28Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:08:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2hz4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nc7qt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:57Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:57 crc kubenswrapper[4764]: I0202 09:08:57.887378 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9p5dc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"660ac006-2495-45c4-9fb3-e9c1dddcf7a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da812b124c97aa8ad1563f4bcfac8f3bfa025096fa4a2561e5847b1bf9cb30b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23e57e65a6f5c3059506b9db21e9ac2492c30bd9191a98db0281915e9ce39797\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23e57e65a6f5c3059506b9db21e9ac2492c30bd9191a98db0281915e9ce39797\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2417b66816da1a6cd1791a4b91eca4a3c4f3694462e066e7f0722f1b9c2105c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2417b66816da1a6cd1791a4b91eca4a3c4f3694462e066e7f0722f1b9c2105c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://883641f29e9120c2cea6b272bacf1dc63c752f55ab0408cdaf73db85c85ac369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://883641f29e9120c2cea6b272bacf1dc63c752f55ab0408cdaf73db85c85ac369\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2745baba1188ba5e8f820616e4117a12c50ecee5b68e71ad9cabbb621b6e131f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2745baba1188ba5e8f820616e4117a12c50ecee5b68e71ad9cabbb621b6e131f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8925e435be4d48d86824fc35f8b8819fb35d63e385164c4a20edb3bf3035c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a8925e435be4d48d86824fc35f8b8819fb35d63e385164c4a20edb3bf3035c3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d3fbf668ae7d04c8d9e574f43c2fe90c5604204ba38a3c60847afdfbceb7a57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d3fbf668ae7d04c8d9e574f43c2fe90c5604204ba38a3c60847afdfbceb7a57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9p5dc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:57Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:57 crc kubenswrapper[4764]: I0202 09:08:57.894175 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:57 crc kubenswrapper[4764]: I0202 09:08:57.894213 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:57 crc kubenswrapper[4764]: I0202 09:08:57.894223 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:57 crc kubenswrapper[4764]: I0202 09:08:57.894238 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:57 crc kubenswrapper[4764]: I0202 09:08:57.894249 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:57Z","lastTransitionTime":"2026-02-02T09:08:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:57 crc kubenswrapper[4764]: I0202 09:08:57.899973 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kwtmr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"991faa9a-dd25-4f49-82bd-ce60cefd4af2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h865b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h865b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:53Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kwtmr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:57Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:57 crc kubenswrapper[4764]: I0202 09:08:57.912767 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87d61aea-46f6-4ee6-a7a4-767a6a67fea4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e78b6d08e4d616cefbb73e3fd43b3c6886dcd8aa997c7b4948bf1ff63401293\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d468715c2da33996561c40043216233fd12e21539d0c58dd1f3b892fd9f91a9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b184f7c936dc4046f74c0018269cbc2c4b7ceddb650b4e871d3f8a707e0df88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://269c5074371bd00f8363fb3888c11f1df383081a61497aea6578f08700881d50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b25bc9395831524c53f9e565b4efbe74f21792efafeb1dd507161bc91aa0ecdf\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T09:07:22Z\\\",\\\"message\\\":\\\"W0202 09:07:21.597488 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0202 09:07:21.598117 1 crypto.go:601] Generating new CA for check-endpoints-signer@1770023241 cert, and key in /tmp/serving-cert-3856183724/serving-signer.crt, /tmp/serving-cert-3856183724/serving-signer.key\\\\nI0202 09:07:22.012184 1 observer_polling.go:159] Starting file observer\\\\nW0202 09:07:22.015493 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0202 09:07:22.015797 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 09:07:22.016596 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3856183724/tls.crt::/tmp/serving-cert-3856183724/tls.key\\\\\\\"\\\\nF0202 09:07:22.270605 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b44e17b8ca01f549bc10081dd5f109cb1b9fff364bf5bd4fd90d9f0e3eeb04db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfb06a5ec9262cf39b108a66014d15e325cdbd5441a2d9e7a811dd57265c99a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfb06a5ec9262cf39b108a66014d15e325cdbd5441a2d9e7a811dd57265c99a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:57Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:57 crc kubenswrapper[4764]: I0202 09:08:57.929257 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa9d037e66cb53b9dd698eff46284c9e718fd842cda6f7b525f0c79062d70bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:57Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:57 crc kubenswrapper[4764]: I0202 09:08:57.940995 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:57Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:57 crc kubenswrapper[4764]: I0202 09:08:57.960124 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24632dda-6100-4ab6-a28e-214ddae4360c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a6c67e9b7708cc0d85b0b6ffbbade0d34867e159a9c5487ca2606475c667a50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bffda48d7c881c1470251b501661324859373246cc1a21811bd36ca24887c7aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a989645cf485cd8ff80ffff1a86cd7808e1564eba5667535e161cc68db13b26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20b71d3ee7a381baa51b0b0d0470283022cfede30f491135737b5e82d631d763\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c380383b6a014d1b694a13ddc22f54b20b9e4f9d5c9b9000d76187621dce0a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5529f2bff87753428e9173826fc1880acf96a4c134fa9ecad195b15697cad73a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cb255e14eb02f7c08364c41fb9906e58b2155eb507f694b77ec8eb1f4251343\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3cb255e14eb02f7c08364c41fb9906e58b2155eb507f694b77ec8eb1f4251343\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-02T09:08:38Z\\\",\\\"message\\\":\\\":38.747848 6731 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-diagnostics/network-check-target-xd92c\\\\nI0202 09:08:38.747852 6731 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf in node crc\\\\nI0202 09:08:38.747854 6731 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-target-xd92c\\\\nI0202 09:08:38.747780 6731 obj_retry.go:386] Retry successful for *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc after 0 failed attempt(s)\\\\nI0202 09:08:38.747862 6731 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-target-xd92c in node crc\\\\nI0202 09:08:38.747864 6731 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf after 0 failed attempt(s)\\\\nI0202 09:08:38.747869 6731 default_network_controller.go:776] Recording success event on pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI0202 09:08:38.747837 6731 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-additional-cni-plugins-9p5dc in node crc\\\\nI0202 09:08:38.747884 6731 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-target-xd92c] creating logical port openshift-network-diagnostics_network-check-target-\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T09:08:37Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-zhn7j_openshift-ovn-kubernetes(24632dda-6100-4ab6-a28e-214ddae4360c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b4cfbd4e682bcdcaff785013bdb72580ce36fced733e235538536827036b93d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://579903eadb183466a2564737e6b678e04308239c0cd75b6808892053b97afbc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://579903eadb183466a2564737e6b678e04308239c0cd75b6808892053b97afbc0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zhn7j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:57Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:57 crc kubenswrapper[4764]: I0202 09:08:57.962789 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/991faa9a-dd25-4f49-82bd-ce60cefd4af2-metrics-certs\") pod \"network-metrics-daemon-kwtmr\" (UID: \"991faa9a-dd25-4f49-82bd-ce60cefd4af2\") " pod="openshift-multus/network-metrics-daemon-kwtmr" Feb 02 09:08:57 crc kubenswrapper[4764]: E0202 09:08:57.962923 4764 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 02 09:08:57 crc kubenswrapper[4764]: E0202 09:08:57.962994 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/991faa9a-dd25-4f49-82bd-ce60cefd4af2-metrics-certs podName:991faa9a-dd25-4f49-82bd-ce60cefd4af2 nodeName:}" failed. No retries permitted until 2026-02-02 09:10:01.962980205 +0000 UTC m=+164.896704293 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/991faa9a-dd25-4f49-82bd-ce60cefd4af2-metrics-certs") pod "network-metrics-daemon-kwtmr" (UID: "991faa9a-dd25-4f49-82bd-ce60cefd4af2") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 02 09:08:57 crc kubenswrapper[4764]: I0202 09:08:57.979139 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d66ab3c0-9b5b-41d5-b6c0-91ccddccd7e9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d92ac486f5dade46bc75de32099859a71fc2ed25a7abbe6ae8937df128e7df30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a4ac273a4f127d79e5fad7d4c30cd2044a346f5734dba5d33d68cd092c0c2d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d38c7cb5dcd22fb07524494ebe77aa1a7f4f808f566043563817e47e7ec12a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://408722c7f83d28919431532f8c797a1c795fd5d96736d14e7f06a05744899702\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f473b18cd8003f75bb1dafb82bcec45745a73ccd10434297356dac185a3d15c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e229f484d19cbdabedf53d90a822a4434db9e00b1301538e885e26acc19d25f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e229f484d19cbdabedf53d90a822a4434db9e00b1301538e885e26acc19d25f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a9cc673be8ff245dba81278e8cd34cfe45ed36ff0b8bd5b30a0ddf333ec6371\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a9cc673be8ff245dba81278e8cd34cfe45ed36ff0b8bd5b30a0ddf333ec6371\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bc7e49c7a5ae66ea939060192fa6a346dc0b0b9aa9750b29879d3c426cc0d276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc7e49c7a5ae66ea939060192fa6a346dc0b0b9aa9750b29879d3c426cc0d276\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:57Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:57 crc kubenswrapper[4764]: I0202 09:08:57.990519 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f6904228e052fff69ad1fd66df97758b6b60ba4351e97a7fbd03eee4df1b653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:57Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:57 crc kubenswrapper[4764]: I0202 09:08:57.996600 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:57 crc kubenswrapper[4764]: I0202 09:08:57.996631 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:57 crc kubenswrapper[4764]: I0202 09:08:57.996640 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:57 crc kubenswrapper[4764]: I0202 09:08:57.996653 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:57 crc kubenswrapper[4764]: I0202 09:08:57.996662 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:57Z","lastTransitionTime":"2026-02-02T09:08:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:58 crc kubenswrapper[4764]: I0202 09:08:58.000464 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6zxff" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aaba93da-12a5-43df-b7c7-4c8b988163c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3db293678a907cd55b31b1d0c0bd7c30e1ba75ecd5e13cbc68ab98fbf0fbd9d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqwb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6zxff\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:57Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:58 crc kubenswrapper[4764]: I0202 09:08:58.011054 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d192f670-9f9d-4539-9641-e4bed73acdd4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://040c5e8a5659f85f0d52d1a1e064154d998ae104f711101d2204c0b1caebbb30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbksf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fdf06410349b7c99aa166e87fbe80bf68e2ae923635a6a978b7e2cd50383f7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbksf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4ndm4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:58Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:58 crc kubenswrapper[4764]: I0202 09:08:58.020280 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pcjmn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6e5ccd3b-b657-47a2-9c4d-bc9adb126533\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72863eef568d92a1abeb3cde2e89630a4f5fb7626c6949448ff5195774c98eab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2f7zl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pcjmn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:58Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:58 crc kubenswrapper[4764]: I0202 09:08:58.030669 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed747fd8-e3ce-4319-a6da-bda0fd51b85a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6f064bdc66470aace3180d4988244026b54009a63cff199e41c5d02a1d18479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://60258f84a1d8dcfe48b7e9688ed50f077e201971f20f3f6663fbf1ba58a4611a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f84be2e08217a2ac3edde858d78b4410376ed9bae3bcabe2a7f9b2fbe2f69c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a69d25d352d7ceac5c735bc1fcf446a1fea6dc22c65a5296cfab828cea17377\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:58Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:58 crc kubenswrapper[4764]: I0202 09:08:58.041149 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:58Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:58 crc kubenswrapper[4764]: I0202 09:08:58.051327 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lh6tm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2152e8a-f456-47b5-a547-b65ec837ecbc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38e176a00b07f2960fdd7ba2aadd1468a5e083b8d154e9d118d13f2585da81e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bb2sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://996e649011d5255783284421ddea8e0a5c383ef675b0920c3fdc0af6503d80e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bb2sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lh6tm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:58Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:58 crc kubenswrapper[4764]: I0202 09:08:58.061748 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fae745b-ea49-42f3-a486-dbc3dc2fb9ad\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://081d7f76d937ff71e35945f9b1d893888bb649d15bd84312fd4b61c9520ea71f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://429ead0993f5e648f6256d249ee12aea03514d3108618840552b672fb419cd74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fad17b01545eb20785f03e3c61bd5c982f6217e38176d7a3b23ac194298644a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85e2f75c56d63cacc93b9195c1e8bf46eb7a7234c45a024bd95a460db90edb6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85e2f75c56d63cacc93b9195c1e8bf46eb7a7234c45a024bd95a460db90edb6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:08:58Z is after 2025-08-24T17:21:41Z" Feb 02 09:08:58 crc kubenswrapper[4764]: I0202 09:08:58.099093 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:58 crc kubenswrapper[4764]: I0202 09:08:58.099134 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:58 crc kubenswrapper[4764]: I0202 09:08:58.099147 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:58 crc kubenswrapper[4764]: I0202 09:08:58.099164 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:58 crc kubenswrapper[4764]: I0202 09:08:58.099176 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:58Z","lastTransitionTime":"2026-02-02T09:08:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:58 crc kubenswrapper[4764]: I0202 09:08:58.201052 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:58 crc kubenswrapper[4764]: I0202 09:08:58.201109 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:58 crc kubenswrapper[4764]: I0202 09:08:58.201119 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:58 crc kubenswrapper[4764]: I0202 09:08:58.201133 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:58 crc kubenswrapper[4764]: I0202 09:08:58.201145 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:58Z","lastTransitionTime":"2026-02-02T09:08:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:58 crc kubenswrapper[4764]: I0202 09:08:58.304054 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:58 crc kubenswrapper[4764]: I0202 09:08:58.304085 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:58 crc kubenswrapper[4764]: I0202 09:08:58.304092 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:58 crc kubenswrapper[4764]: I0202 09:08:58.304123 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:58 crc kubenswrapper[4764]: I0202 09:08:58.304131 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:58Z","lastTransitionTime":"2026-02-02T09:08:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:58 crc kubenswrapper[4764]: I0202 09:08:58.406713 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:58 crc kubenswrapper[4764]: I0202 09:08:58.406742 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:58 crc kubenswrapper[4764]: I0202 09:08:58.406751 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:58 crc kubenswrapper[4764]: I0202 09:08:58.406764 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:58 crc kubenswrapper[4764]: I0202 09:08:58.406797 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:58Z","lastTransitionTime":"2026-02-02T09:08:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:58 crc kubenswrapper[4764]: I0202 09:08:58.509385 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:58 crc kubenswrapper[4764]: I0202 09:08:58.509466 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:58 crc kubenswrapper[4764]: I0202 09:08:58.509483 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:58 crc kubenswrapper[4764]: I0202 09:08:58.509505 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:58 crc kubenswrapper[4764]: I0202 09:08:58.509520 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:58Z","lastTransitionTime":"2026-02-02T09:08:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:58 crc kubenswrapper[4764]: I0202 09:08:58.612905 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:58 crc kubenswrapper[4764]: I0202 09:08:58.613081 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:58 crc kubenswrapper[4764]: I0202 09:08:58.613116 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:58 crc kubenswrapper[4764]: I0202 09:08:58.613142 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:58 crc kubenswrapper[4764]: I0202 09:08:58.613160 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:58Z","lastTransitionTime":"2026-02-02T09:08:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:58 crc kubenswrapper[4764]: I0202 09:08:58.714876 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:58 crc kubenswrapper[4764]: I0202 09:08:58.714918 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:58 crc kubenswrapper[4764]: I0202 09:08:58.714947 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:58 crc kubenswrapper[4764]: I0202 09:08:58.714988 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:58 crc kubenswrapper[4764]: I0202 09:08:58.715001 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:58Z","lastTransitionTime":"2026-02-02T09:08:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:58 crc kubenswrapper[4764]: I0202 09:08:58.817834 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:58 crc kubenswrapper[4764]: I0202 09:08:58.818031 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:58 crc kubenswrapper[4764]: I0202 09:08:58.818065 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:58 crc kubenswrapper[4764]: I0202 09:08:58.818143 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:58 crc kubenswrapper[4764]: I0202 09:08:58.818169 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:58Z","lastTransitionTime":"2026-02-02T09:08:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:58 crc kubenswrapper[4764]: I0202 09:08:58.825142 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 09:08:58 crc kubenswrapper[4764]: I0202 09:08:58.825203 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 09:08:58 crc kubenswrapper[4764]: E0202 09:08:58.825230 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 09:08:58 crc kubenswrapper[4764]: E0202 09:08:58.825398 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 09:08:58 crc kubenswrapper[4764]: I0202 09:08:58.825791 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 09:08:58 crc kubenswrapper[4764]: E0202 09:08:58.826151 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 09:08:58 crc kubenswrapper[4764]: I0202 09:08:58.840528 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-08 11:54:34.082045725 +0000 UTC Feb 02 09:08:58 crc kubenswrapper[4764]: I0202 09:08:58.921154 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:58 crc kubenswrapper[4764]: I0202 09:08:58.921375 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:58 crc kubenswrapper[4764]: I0202 09:08:58.921545 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:58 crc kubenswrapper[4764]: I0202 09:08:58.921627 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:58 crc kubenswrapper[4764]: I0202 09:08:58.921697 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:58Z","lastTransitionTime":"2026-02-02T09:08:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:59 crc kubenswrapper[4764]: I0202 09:08:59.024122 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:59 crc kubenswrapper[4764]: I0202 09:08:59.024176 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:59 crc kubenswrapper[4764]: I0202 09:08:59.024184 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:59 crc kubenswrapper[4764]: I0202 09:08:59.024197 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:59 crc kubenswrapper[4764]: I0202 09:08:59.024205 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:59Z","lastTransitionTime":"2026-02-02T09:08:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:59 crc kubenswrapper[4764]: I0202 09:08:59.126448 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:59 crc kubenswrapper[4764]: I0202 09:08:59.126492 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:59 crc kubenswrapper[4764]: I0202 09:08:59.126503 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:59 crc kubenswrapper[4764]: I0202 09:08:59.126520 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:59 crc kubenswrapper[4764]: I0202 09:08:59.126533 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:59Z","lastTransitionTime":"2026-02-02T09:08:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:59 crc kubenswrapper[4764]: I0202 09:08:59.228529 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:59 crc kubenswrapper[4764]: I0202 09:08:59.228598 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:59 crc kubenswrapper[4764]: I0202 09:08:59.228616 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:59 crc kubenswrapper[4764]: I0202 09:08:59.228639 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:59 crc kubenswrapper[4764]: I0202 09:08:59.228656 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:59Z","lastTransitionTime":"2026-02-02T09:08:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:59 crc kubenswrapper[4764]: I0202 09:08:59.331106 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:59 crc kubenswrapper[4764]: I0202 09:08:59.331148 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:59 crc kubenswrapper[4764]: I0202 09:08:59.331161 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:59 crc kubenswrapper[4764]: I0202 09:08:59.331179 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:59 crc kubenswrapper[4764]: I0202 09:08:59.331191 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:59Z","lastTransitionTime":"2026-02-02T09:08:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:59 crc kubenswrapper[4764]: I0202 09:08:59.434096 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:59 crc kubenswrapper[4764]: I0202 09:08:59.434145 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:59 crc kubenswrapper[4764]: I0202 09:08:59.434162 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:59 crc kubenswrapper[4764]: I0202 09:08:59.434189 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:59 crc kubenswrapper[4764]: I0202 09:08:59.434207 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:59Z","lastTransitionTime":"2026-02-02T09:08:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:59 crc kubenswrapper[4764]: I0202 09:08:59.536739 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:59 crc kubenswrapper[4764]: I0202 09:08:59.536780 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:59 crc kubenswrapper[4764]: I0202 09:08:59.536790 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:59 crc kubenswrapper[4764]: I0202 09:08:59.536805 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:59 crc kubenswrapper[4764]: I0202 09:08:59.536815 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:59Z","lastTransitionTime":"2026-02-02T09:08:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:59 crc kubenswrapper[4764]: I0202 09:08:59.639431 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:59 crc kubenswrapper[4764]: I0202 09:08:59.639720 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:59 crc kubenswrapper[4764]: I0202 09:08:59.639808 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:59 crc kubenswrapper[4764]: I0202 09:08:59.639890 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:59 crc kubenswrapper[4764]: I0202 09:08:59.640004 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:59Z","lastTransitionTime":"2026-02-02T09:08:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:59 crc kubenswrapper[4764]: I0202 09:08:59.743347 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:59 crc kubenswrapper[4764]: I0202 09:08:59.743768 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:59 crc kubenswrapper[4764]: I0202 09:08:59.743861 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:59 crc kubenswrapper[4764]: I0202 09:08:59.744007 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:59 crc kubenswrapper[4764]: I0202 09:08:59.744116 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:59Z","lastTransitionTime":"2026-02-02T09:08:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:59 crc kubenswrapper[4764]: I0202 09:08:59.824630 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kwtmr" Feb 02 09:08:59 crc kubenswrapper[4764]: E0202 09:08:59.824779 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kwtmr" podUID="991faa9a-dd25-4f49-82bd-ce60cefd4af2" Feb 02 09:08:59 crc kubenswrapper[4764]: I0202 09:08:59.841671 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-23 23:37:25.850456377 +0000 UTC Feb 02 09:08:59 crc kubenswrapper[4764]: I0202 09:08:59.847383 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:59 crc kubenswrapper[4764]: I0202 09:08:59.847420 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:59 crc kubenswrapper[4764]: I0202 09:08:59.847432 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:59 crc kubenswrapper[4764]: I0202 09:08:59.847448 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:59 crc kubenswrapper[4764]: I0202 09:08:59.847460 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:59Z","lastTransitionTime":"2026-02-02T09:08:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:08:59 crc kubenswrapper[4764]: I0202 09:08:59.950228 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:08:59 crc kubenswrapper[4764]: I0202 09:08:59.950875 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:08:59 crc kubenswrapper[4764]: I0202 09:08:59.950916 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:08:59 crc kubenswrapper[4764]: I0202 09:08:59.950969 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:08:59 crc kubenswrapper[4764]: I0202 09:08:59.950988 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:08:59Z","lastTransitionTime":"2026-02-02T09:08:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:00 crc kubenswrapper[4764]: I0202 09:09:00.053558 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:00 crc kubenswrapper[4764]: I0202 09:09:00.053611 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:00 crc kubenswrapper[4764]: I0202 09:09:00.053621 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:00 crc kubenswrapper[4764]: I0202 09:09:00.053638 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:00 crc kubenswrapper[4764]: I0202 09:09:00.053647 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:00Z","lastTransitionTime":"2026-02-02T09:09:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:00 crc kubenswrapper[4764]: I0202 09:09:00.157374 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:00 crc kubenswrapper[4764]: I0202 09:09:00.157480 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:00 crc kubenswrapper[4764]: I0202 09:09:00.157498 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:00 crc kubenswrapper[4764]: I0202 09:09:00.157522 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:00 crc kubenswrapper[4764]: I0202 09:09:00.157544 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:00Z","lastTransitionTime":"2026-02-02T09:09:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:00 crc kubenswrapper[4764]: I0202 09:09:00.260206 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:00 crc kubenswrapper[4764]: I0202 09:09:00.260237 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:00 crc kubenswrapper[4764]: I0202 09:09:00.260329 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:00 crc kubenswrapper[4764]: I0202 09:09:00.260352 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:00 crc kubenswrapper[4764]: I0202 09:09:00.260363 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:00Z","lastTransitionTime":"2026-02-02T09:09:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:00 crc kubenswrapper[4764]: I0202 09:09:00.362782 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:00 crc kubenswrapper[4764]: I0202 09:09:00.362830 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:00 crc kubenswrapper[4764]: I0202 09:09:00.362845 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:00 crc kubenswrapper[4764]: I0202 09:09:00.362864 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:00 crc kubenswrapper[4764]: I0202 09:09:00.362880 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:00Z","lastTransitionTime":"2026-02-02T09:09:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:00 crc kubenswrapper[4764]: I0202 09:09:00.465269 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:00 crc kubenswrapper[4764]: I0202 09:09:00.465327 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:00 crc kubenswrapper[4764]: I0202 09:09:00.465350 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:00 crc kubenswrapper[4764]: I0202 09:09:00.465369 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:00 crc kubenswrapper[4764]: I0202 09:09:00.465385 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:00Z","lastTransitionTime":"2026-02-02T09:09:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:00 crc kubenswrapper[4764]: I0202 09:09:00.568340 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:00 crc kubenswrapper[4764]: I0202 09:09:00.568399 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:00 crc kubenswrapper[4764]: I0202 09:09:00.568416 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:00 crc kubenswrapper[4764]: I0202 09:09:00.568441 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:00 crc kubenswrapper[4764]: I0202 09:09:00.568460 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:00Z","lastTransitionTime":"2026-02-02T09:09:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:00 crc kubenswrapper[4764]: I0202 09:09:00.671231 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:00 crc kubenswrapper[4764]: I0202 09:09:00.671294 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:00 crc kubenswrapper[4764]: I0202 09:09:00.671316 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:00 crc kubenswrapper[4764]: I0202 09:09:00.671349 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:00 crc kubenswrapper[4764]: I0202 09:09:00.671371 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:00Z","lastTransitionTime":"2026-02-02T09:09:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:00 crc kubenswrapper[4764]: I0202 09:09:00.774248 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:00 crc kubenswrapper[4764]: I0202 09:09:00.774297 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:00 crc kubenswrapper[4764]: I0202 09:09:00.774316 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:00 crc kubenswrapper[4764]: I0202 09:09:00.774338 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:00 crc kubenswrapper[4764]: I0202 09:09:00.774354 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:00Z","lastTransitionTime":"2026-02-02T09:09:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:00 crc kubenswrapper[4764]: I0202 09:09:00.825118 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 09:09:00 crc kubenswrapper[4764]: I0202 09:09:00.825164 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 09:09:00 crc kubenswrapper[4764]: I0202 09:09:00.825235 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 09:09:00 crc kubenswrapper[4764]: E0202 09:09:00.825417 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 09:09:00 crc kubenswrapper[4764]: E0202 09:09:00.825495 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 09:09:00 crc kubenswrapper[4764]: E0202 09:09:00.825517 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 09:09:00 crc kubenswrapper[4764]: I0202 09:09:00.842203 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-02 23:34:26.381188114 +0000 UTC Feb 02 09:09:00 crc kubenswrapper[4764]: I0202 09:09:00.878433 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:00 crc kubenswrapper[4764]: I0202 09:09:00.878486 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:00 crc kubenswrapper[4764]: I0202 09:09:00.878506 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:00 crc kubenswrapper[4764]: I0202 09:09:00.878539 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:00 crc kubenswrapper[4764]: I0202 09:09:00.878565 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:00Z","lastTransitionTime":"2026-02-02T09:09:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:00 crc kubenswrapper[4764]: I0202 09:09:00.981362 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:00 crc kubenswrapper[4764]: I0202 09:09:00.981411 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:00 crc kubenswrapper[4764]: I0202 09:09:00.981428 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:00 crc kubenswrapper[4764]: I0202 09:09:00.981451 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:00 crc kubenswrapper[4764]: I0202 09:09:00.981466 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:00Z","lastTransitionTime":"2026-02-02T09:09:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:01 crc kubenswrapper[4764]: I0202 09:09:01.085108 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:01 crc kubenswrapper[4764]: I0202 09:09:01.085165 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:01 crc kubenswrapper[4764]: I0202 09:09:01.085177 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:01 crc kubenswrapper[4764]: I0202 09:09:01.085195 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:01 crc kubenswrapper[4764]: I0202 09:09:01.085207 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:01Z","lastTransitionTime":"2026-02-02T09:09:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:01 crc kubenswrapper[4764]: I0202 09:09:01.188777 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:01 crc kubenswrapper[4764]: I0202 09:09:01.188836 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:01 crc kubenswrapper[4764]: I0202 09:09:01.188852 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:01 crc kubenswrapper[4764]: I0202 09:09:01.188875 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:01 crc kubenswrapper[4764]: I0202 09:09:01.188898 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:01Z","lastTransitionTime":"2026-02-02T09:09:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:01 crc kubenswrapper[4764]: I0202 09:09:01.292515 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:01 crc kubenswrapper[4764]: I0202 09:09:01.292559 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:01 crc kubenswrapper[4764]: I0202 09:09:01.292572 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:01 crc kubenswrapper[4764]: I0202 09:09:01.292618 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:01 crc kubenswrapper[4764]: I0202 09:09:01.292634 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:01Z","lastTransitionTime":"2026-02-02T09:09:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:01 crc kubenswrapper[4764]: I0202 09:09:01.398170 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:01 crc kubenswrapper[4764]: I0202 09:09:01.398247 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:01 crc kubenswrapper[4764]: I0202 09:09:01.398265 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:01 crc kubenswrapper[4764]: I0202 09:09:01.398287 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:01 crc kubenswrapper[4764]: I0202 09:09:01.398304 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:01Z","lastTransitionTime":"2026-02-02T09:09:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:01 crc kubenswrapper[4764]: I0202 09:09:01.500375 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:01 crc kubenswrapper[4764]: I0202 09:09:01.500413 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:01 crc kubenswrapper[4764]: I0202 09:09:01.500423 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:01 crc kubenswrapper[4764]: I0202 09:09:01.500437 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:01 crc kubenswrapper[4764]: I0202 09:09:01.500447 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:01Z","lastTransitionTime":"2026-02-02T09:09:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:01 crc kubenswrapper[4764]: I0202 09:09:01.602677 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:01 crc kubenswrapper[4764]: I0202 09:09:01.602714 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:01 crc kubenswrapper[4764]: I0202 09:09:01.602738 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:01 crc kubenswrapper[4764]: I0202 09:09:01.602750 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:01 crc kubenswrapper[4764]: I0202 09:09:01.602759 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:01Z","lastTransitionTime":"2026-02-02T09:09:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:01 crc kubenswrapper[4764]: I0202 09:09:01.705102 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:01 crc kubenswrapper[4764]: I0202 09:09:01.705320 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:01 crc kubenswrapper[4764]: I0202 09:09:01.705341 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:01 crc kubenswrapper[4764]: I0202 09:09:01.705363 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:01 crc kubenswrapper[4764]: I0202 09:09:01.705380 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:01Z","lastTransitionTime":"2026-02-02T09:09:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:01 crc kubenswrapper[4764]: I0202 09:09:01.808279 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:01 crc kubenswrapper[4764]: I0202 09:09:01.808316 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:01 crc kubenswrapper[4764]: I0202 09:09:01.808327 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:01 crc kubenswrapper[4764]: I0202 09:09:01.808345 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:01 crc kubenswrapper[4764]: I0202 09:09:01.808356 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:01Z","lastTransitionTime":"2026-02-02T09:09:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:01 crc kubenswrapper[4764]: I0202 09:09:01.824853 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kwtmr" Feb 02 09:09:01 crc kubenswrapper[4764]: E0202 09:09:01.825097 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kwtmr" podUID="991faa9a-dd25-4f49-82bd-ce60cefd4af2" Feb 02 09:09:01 crc kubenswrapper[4764]: I0202 09:09:01.843167 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-03 23:30:27.590238729 +0000 UTC Feb 02 09:09:01 crc kubenswrapper[4764]: I0202 09:09:01.911715 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:01 crc kubenswrapper[4764]: I0202 09:09:01.911757 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:01 crc kubenswrapper[4764]: I0202 09:09:01.911768 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:01 crc kubenswrapper[4764]: I0202 09:09:01.911783 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:01 crc kubenswrapper[4764]: I0202 09:09:01.911792 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:01Z","lastTransitionTime":"2026-02-02T09:09:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:02 crc kubenswrapper[4764]: I0202 09:09:02.014337 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:02 crc kubenswrapper[4764]: I0202 09:09:02.014422 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:02 crc kubenswrapper[4764]: I0202 09:09:02.014455 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:02 crc kubenswrapper[4764]: I0202 09:09:02.014487 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:02 crc kubenswrapper[4764]: I0202 09:09:02.014505 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:02Z","lastTransitionTime":"2026-02-02T09:09:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:02 crc kubenswrapper[4764]: I0202 09:09:02.117210 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:02 crc kubenswrapper[4764]: I0202 09:09:02.117286 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:02 crc kubenswrapper[4764]: I0202 09:09:02.117311 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:02 crc kubenswrapper[4764]: I0202 09:09:02.117346 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:02 crc kubenswrapper[4764]: I0202 09:09:02.117371 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:02Z","lastTransitionTime":"2026-02-02T09:09:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:02 crc kubenswrapper[4764]: I0202 09:09:02.219810 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:02 crc kubenswrapper[4764]: I0202 09:09:02.219865 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:02 crc kubenswrapper[4764]: I0202 09:09:02.219883 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:02 crc kubenswrapper[4764]: I0202 09:09:02.219907 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:02 crc kubenswrapper[4764]: I0202 09:09:02.219923 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:02Z","lastTransitionTime":"2026-02-02T09:09:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:02 crc kubenswrapper[4764]: I0202 09:09:02.323023 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:02 crc kubenswrapper[4764]: I0202 09:09:02.323065 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:02 crc kubenswrapper[4764]: I0202 09:09:02.323095 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:02 crc kubenswrapper[4764]: I0202 09:09:02.323113 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:02 crc kubenswrapper[4764]: I0202 09:09:02.323124 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:02Z","lastTransitionTime":"2026-02-02T09:09:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:02 crc kubenswrapper[4764]: I0202 09:09:02.427215 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:02 crc kubenswrapper[4764]: I0202 09:09:02.427257 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:02 crc kubenswrapper[4764]: I0202 09:09:02.427267 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:02 crc kubenswrapper[4764]: I0202 09:09:02.427281 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:02 crc kubenswrapper[4764]: I0202 09:09:02.427290 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:02Z","lastTransitionTime":"2026-02-02T09:09:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:02 crc kubenswrapper[4764]: I0202 09:09:02.529059 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:02 crc kubenswrapper[4764]: I0202 09:09:02.529275 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:02 crc kubenswrapper[4764]: I0202 09:09:02.529287 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:02 crc kubenswrapper[4764]: I0202 09:09:02.529302 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:02 crc kubenswrapper[4764]: I0202 09:09:02.529313 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:02Z","lastTransitionTime":"2026-02-02T09:09:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:02 crc kubenswrapper[4764]: I0202 09:09:02.633048 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:02 crc kubenswrapper[4764]: I0202 09:09:02.633120 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:02 crc kubenswrapper[4764]: I0202 09:09:02.633139 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:02 crc kubenswrapper[4764]: I0202 09:09:02.633346 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:02 crc kubenswrapper[4764]: I0202 09:09:02.633364 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:02Z","lastTransitionTime":"2026-02-02T09:09:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:02 crc kubenswrapper[4764]: I0202 09:09:02.736566 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:02 crc kubenswrapper[4764]: I0202 09:09:02.736608 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:02 crc kubenswrapper[4764]: I0202 09:09:02.736617 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:02 crc kubenswrapper[4764]: I0202 09:09:02.736632 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:02 crc kubenswrapper[4764]: I0202 09:09:02.736640 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:02Z","lastTransitionTime":"2026-02-02T09:09:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:02 crc kubenswrapper[4764]: I0202 09:09:02.824735 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 09:09:02 crc kubenswrapper[4764]: I0202 09:09:02.824808 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 09:09:02 crc kubenswrapper[4764]: I0202 09:09:02.824826 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 09:09:02 crc kubenswrapper[4764]: E0202 09:09:02.824892 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 09:09:02 crc kubenswrapper[4764]: E0202 09:09:02.825170 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 09:09:02 crc kubenswrapper[4764]: E0202 09:09:02.825219 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 09:09:02 crc kubenswrapper[4764]: I0202 09:09:02.840999 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:02 crc kubenswrapper[4764]: I0202 09:09:02.841063 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:02 crc kubenswrapper[4764]: I0202 09:09:02.841085 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:02 crc kubenswrapper[4764]: I0202 09:09:02.841113 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:02 crc kubenswrapper[4764]: I0202 09:09:02.841135 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:02Z","lastTransitionTime":"2026-02-02T09:09:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:02 crc kubenswrapper[4764]: I0202 09:09:02.844315 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-02 03:47:03.955520843 +0000 UTC Feb 02 09:09:02 crc kubenswrapper[4764]: I0202 09:09:02.944229 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:02 crc kubenswrapper[4764]: I0202 09:09:02.944461 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:02 crc kubenswrapper[4764]: I0202 09:09:02.944534 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:02 crc kubenswrapper[4764]: I0202 09:09:02.944642 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:02 crc kubenswrapper[4764]: I0202 09:09:02.944737 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:02Z","lastTransitionTime":"2026-02-02T09:09:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:03 crc kubenswrapper[4764]: I0202 09:09:03.047677 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:03 crc kubenswrapper[4764]: I0202 09:09:03.048006 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:03 crc kubenswrapper[4764]: I0202 09:09:03.048191 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:03 crc kubenswrapper[4764]: I0202 09:09:03.048336 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:03 crc kubenswrapper[4764]: I0202 09:09:03.048444 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:03Z","lastTransitionTime":"2026-02-02T09:09:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:03 crc kubenswrapper[4764]: I0202 09:09:03.151583 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:03 crc kubenswrapper[4764]: I0202 09:09:03.151622 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:03 crc kubenswrapper[4764]: I0202 09:09:03.151632 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:03 crc kubenswrapper[4764]: I0202 09:09:03.151647 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:03 crc kubenswrapper[4764]: I0202 09:09:03.151656 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:03Z","lastTransitionTime":"2026-02-02T09:09:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:03 crc kubenswrapper[4764]: I0202 09:09:03.255152 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:03 crc kubenswrapper[4764]: I0202 09:09:03.255213 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:03 crc kubenswrapper[4764]: I0202 09:09:03.255222 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:03 crc kubenswrapper[4764]: I0202 09:09:03.255236 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:03 crc kubenswrapper[4764]: I0202 09:09:03.255245 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:03Z","lastTransitionTime":"2026-02-02T09:09:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:03 crc kubenswrapper[4764]: I0202 09:09:03.357578 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:03 crc kubenswrapper[4764]: I0202 09:09:03.357610 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:03 crc kubenswrapper[4764]: I0202 09:09:03.357620 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:03 crc kubenswrapper[4764]: I0202 09:09:03.357634 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:03 crc kubenswrapper[4764]: I0202 09:09:03.357647 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:03Z","lastTransitionTime":"2026-02-02T09:09:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:03 crc kubenswrapper[4764]: I0202 09:09:03.460145 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:03 crc kubenswrapper[4764]: I0202 09:09:03.460193 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:03 crc kubenswrapper[4764]: I0202 09:09:03.460209 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:03 crc kubenswrapper[4764]: I0202 09:09:03.460231 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:03 crc kubenswrapper[4764]: I0202 09:09:03.460249 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:03Z","lastTransitionTime":"2026-02-02T09:09:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:03 crc kubenswrapper[4764]: I0202 09:09:03.570702 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:03 crc kubenswrapper[4764]: I0202 09:09:03.570753 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:03 crc kubenswrapper[4764]: I0202 09:09:03.570764 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:03 crc kubenswrapper[4764]: I0202 09:09:03.570783 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:03 crc kubenswrapper[4764]: I0202 09:09:03.570796 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:03Z","lastTransitionTime":"2026-02-02T09:09:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:03 crc kubenswrapper[4764]: I0202 09:09:03.673544 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:03 crc kubenswrapper[4764]: I0202 09:09:03.673574 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:03 crc kubenswrapper[4764]: I0202 09:09:03.673582 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:03 crc kubenswrapper[4764]: I0202 09:09:03.673595 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:03 crc kubenswrapper[4764]: I0202 09:09:03.673604 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:03Z","lastTransitionTime":"2026-02-02T09:09:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:03 crc kubenswrapper[4764]: I0202 09:09:03.776049 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:03 crc kubenswrapper[4764]: I0202 09:09:03.776404 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:03 crc kubenswrapper[4764]: I0202 09:09:03.776616 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:03 crc kubenswrapper[4764]: I0202 09:09:03.776790 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:03 crc kubenswrapper[4764]: I0202 09:09:03.776927 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:03Z","lastTransitionTime":"2026-02-02T09:09:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:03 crc kubenswrapper[4764]: I0202 09:09:03.825519 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kwtmr" Feb 02 09:09:03 crc kubenswrapper[4764]: E0202 09:09:03.826588 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kwtmr" podUID="991faa9a-dd25-4f49-82bd-ce60cefd4af2" Feb 02 09:09:03 crc kubenswrapper[4764]: I0202 09:09:03.827021 4764 scope.go:117] "RemoveContainer" containerID="3cb255e14eb02f7c08364c41fb9906e58b2155eb507f694b77ec8eb1f4251343" Feb 02 09:09:03 crc kubenswrapper[4764]: E0202 09:09:03.827477 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-zhn7j_openshift-ovn-kubernetes(24632dda-6100-4ab6-a28e-214ddae4360c)\"" pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" podUID="24632dda-6100-4ab6-a28e-214ddae4360c" Feb 02 09:09:03 crc kubenswrapper[4764]: I0202 09:09:03.845751 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-14 12:56:34.404992118 +0000 UTC Feb 02 09:09:03 crc kubenswrapper[4764]: I0202 09:09:03.879798 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:03 crc kubenswrapper[4764]: I0202 09:09:03.879842 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:03 crc kubenswrapper[4764]: I0202 09:09:03.879854 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:03 crc kubenswrapper[4764]: I0202 09:09:03.879873 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:03 crc kubenswrapper[4764]: I0202 09:09:03.879889 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:03Z","lastTransitionTime":"2026-02-02T09:09:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:03 crc kubenswrapper[4764]: I0202 09:09:03.982366 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:03 crc kubenswrapper[4764]: I0202 09:09:03.982401 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:03 crc kubenswrapper[4764]: I0202 09:09:03.982411 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:03 crc kubenswrapper[4764]: I0202 09:09:03.982427 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:03 crc kubenswrapper[4764]: I0202 09:09:03.982439 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:03Z","lastTransitionTime":"2026-02-02T09:09:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:04 crc kubenswrapper[4764]: I0202 09:09:04.084774 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:04 crc kubenswrapper[4764]: I0202 09:09:04.085207 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:04 crc kubenswrapper[4764]: I0202 09:09:04.085340 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:04 crc kubenswrapper[4764]: I0202 09:09:04.085478 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:04 crc kubenswrapper[4764]: I0202 09:09:04.085607 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:04Z","lastTransitionTime":"2026-02-02T09:09:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:04 crc kubenswrapper[4764]: I0202 09:09:04.188574 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:04 crc kubenswrapper[4764]: I0202 09:09:04.188660 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:04 crc kubenswrapper[4764]: I0202 09:09:04.188682 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:04 crc kubenswrapper[4764]: I0202 09:09:04.188711 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:04 crc kubenswrapper[4764]: I0202 09:09:04.188734 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:04Z","lastTransitionTime":"2026-02-02T09:09:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:04 crc kubenswrapper[4764]: I0202 09:09:04.292461 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:04 crc kubenswrapper[4764]: I0202 09:09:04.292515 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:04 crc kubenswrapper[4764]: I0202 09:09:04.292531 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:04 crc kubenswrapper[4764]: I0202 09:09:04.292554 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:04 crc kubenswrapper[4764]: I0202 09:09:04.292571 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:04Z","lastTransitionTime":"2026-02-02T09:09:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:04 crc kubenswrapper[4764]: I0202 09:09:04.395320 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:04 crc kubenswrapper[4764]: I0202 09:09:04.395377 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:04 crc kubenswrapper[4764]: I0202 09:09:04.395394 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:04 crc kubenswrapper[4764]: I0202 09:09:04.395418 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:04 crc kubenswrapper[4764]: I0202 09:09:04.395435 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:04Z","lastTransitionTime":"2026-02-02T09:09:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:04 crc kubenswrapper[4764]: I0202 09:09:04.498340 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:04 crc kubenswrapper[4764]: I0202 09:09:04.498409 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:04 crc kubenswrapper[4764]: I0202 09:09:04.498430 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:04 crc kubenswrapper[4764]: I0202 09:09:04.498453 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:04 crc kubenswrapper[4764]: I0202 09:09:04.498470 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:04Z","lastTransitionTime":"2026-02-02T09:09:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:04 crc kubenswrapper[4764]: I0202 09:09:04.602320 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:04 crc kubenswrapper[4764]: I0202 09:09:04.602399 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:04 crc kubenswrapper[4764]: I0202 09:09:04.602420 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:04 crc kubenswrapper[4764]: I0202 09:09:04.602443 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:04 crc kubenswrapper[4764]: I0202 09:09:04.602459 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:04Z","lastTransitionTime":"2026-02-02T09:09:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:04 crc kubenswrapper[4764]: I0202 09:09:04.706552 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:04 crc kubenswrapper[4764]: I0202 09:09:04.706609 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:04 crc kubenswrapper[4764]: I0202 09:09:04.706627 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:04 crc kubenswrapper[4764]: I0202 09:09:04.706649 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:04 crc kubenswrapper[4764]: I0202 09:09:04.706667 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:04Z","lastTransitionTime":"2026-02-02T09:09:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:04 crc kubenswrapper[4764]: I0202 09:09:04.810333 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:04 crc kubenswrapper[4764]: I0202 09:09:04.810402 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:04 crc kubenswrapper[4764]: I0202 09:09:04.810423 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:04 crc kubenswrapper[4764]: I0202 09:09:04.810453 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:04 crc kubenswrapper[4764]: I0202 09:09:04.810474 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:04Z","lastTransitionTime":"2026-02-02T09:09:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:04 crc kubenswrapper[4764]: I0202 09:09:04.824998 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 09:09:04 crc kubenswrapper[4764]: I0202 09:09:04.825063 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 09:09:04 crc kubenswrapper[4764]: E0202 09:09:04.825195 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 09:09:04 crc kubenswrapper[4764]: E0202 09:09:04.825290 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 09:09:04 crc kubenswrapper[4764]: I0202 09:09:04.825046 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 09:09:04 crc kubenswrapper[4764]: E0202 09:09:04.825745 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 09:09:04 crc kubenswrapper[4764]: I0202 09:09:04.846781 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-01 05:00:08.646614365 +0000 UTC Feb 02 09:09:04 crc kubenswrapper[4764]: I0202 09:09:04.912651 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:04 crc kubenswrapper[4764]: I0202 09:09:04.912685 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:04 crc kubenswrapper[4764]: I0202 09:09:04.912697 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:04 crc kubenswrapper[4764]: I0202 09:09:04.912715 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:04 crc kubenswrapper[4764]: I0202 09:09:04.912726 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:04Z","lastTransitionTime":"2026-02-02T09:09:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:05 crc kubenswrapper[4764]: I0202 09:09:05.015396 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:05 crc kubenswrapper[4764]: I0202 09:09:05.015487 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:05 crc kubenswrapper[4764]: I0202 09:09:05.015508 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:05 crc kubenswrapper[4764]: I0202 09:09:05.015534 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:05 crc kubenswrapper[4764]: I0202 09:09:05.015552 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:05Z","lastTransitionTime":"2026-02-02T09:09:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:05 crc kubenswrapper[4764]: I0202 09:09:05.118333 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:05 crc kubenswrapper[4764]: I0202 09:09:05.118388 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:05 crc kubenswrapper[4764]: I0202 09:09:05.118402 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:05 crc kubenswrapper[4764]: I0202 09:09:05.118421 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:05 crc kubenswrapper[4764]: I0202 09:09:05.118434 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:05Z","lastTransitionTime":"2026-02-02T09:09:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:05 crc kubenswrapper[4764]: I0202 09:09:05.220290 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:05 crc kubenswrapper[4764]: I0202 09:09:05.220610 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:05 crc kubenswrapper[4764]: I0202 09:09:05.220691 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:05 crc kubenswrapper[4764]: I0202 09:09:05.220764 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:05 crc kubenswrapper[4764]: I0202 09:09:05.220863 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:05Z","lastTransitionTime":"2026-02-02T09:09:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:05 crc kubenswrapper[4764]: I0202 09:09:05.324791 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:05 crc kubenswrapper[4764]: I0202 09:09:05.325337 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:05 crc kubenswrapper[4764]: I0202 09:09:05.325418 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:05 crc kubenswrapper[4764]: I0202 09:09:05.325527 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:05 crc kubenswrapper[4764]: I0202 09:09:05.325634 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:05Z","lastTransitionTime":"2026-02-02T09:09:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:05 crc kubenswrapper[4764]: I0202 09:09:05.428869 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:05 crc kubenswrapper[4764]: I0202 09:09:05.428921 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:05 crc kubenswrapper[4764]: I0202 09:09:05.428972 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:05 crc kubenswrapper[4764]: I0202 09:09:05.428993 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:05 crc kubenswrapper[4764]: I0202 09:09:05.429005 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:05Z","lastTransitionTime":"2026-02-02T09:09:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:05 crc kubenswrapper[4764]: I0202 09:09:05.532040 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:05 crc kubenswrapper[4764]: I0202 09:09:05.532083 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:05 crc kubenswrapper[4764]: I0202 09:09:05.532096 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:05 crc kubenswrapper[4764]: I0202 09:09:05.532111 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:05 crc kubenswrapper[4764]: I0202 09:09:05.532163 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:05Z","lastTransitionTime":"2026-02-02T09:09:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:05 crc kubenswrapper[4764]: I0202 09:09:05.633983 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:05 crc kubenswrapper[4764]: I0202 09:09:05.634272 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:05 crc kubenswrapper[4764]: I0202 09:09:05.634446 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:05 crc kubenswrapper[4764]: I0202 09:09:05.634611 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:05 crc kubenswrapper[4764]: I0202 09:09:05.634742 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:05Z","lastTransitionTime":"2026-02-02T09:09:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:05 crc kubenswrapper[4764]: I0202 09:09:05.737889 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:05 crc kubenswrapper[4764]: I0202 09:09:05.738154 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:05 crc kubenswrapper[4764]: I0202 09:09:05.738218 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:05 crc kubenswrapper[4764]: I0202 09:09:05.738278 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:05 crc kubenswrapper[4764]: I0202 09:09:05.738334 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:05Z","lastTransitionTime":"2026-02-02T09:09:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:05 crc kubenswrapper[4764]: I0202 09:09:05.824693 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kwtmr" Feb 02 09:09:05 crc kubenswrapper[4764]: E0202 09:09:05.825125 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kwtmr" podUID="991faa9a-dd25-4f49-82bd-ce60cefd4af2" Feb 02 09:09:05 crc kubenswrapper[4764]: I0202 09:09:05.841136 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:05 crc kubenswrapper[4764]: I0202 09:09:05.841218 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:05 crc kubenswrapper[4764]: I0202 09:09:05.841244 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:05 crc kubenswrapper[4764]: I0202 09:09:05.841271 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:05 crc kubenswrapper[4764]: I0202 09:09:05.841291 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:05Z","lastTransitionTime":"2026-02-02T09:09:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:05 crc kubenswrapper[4764]: I0202 09:09:05.847053 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-27 18:16:41.878840238 +0000 UTC Feb 02 09:09:05 crc kubenswrapper[4764]: I0202 09:09:05.950167 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:05 crc kubenswrapper[4764]: I0202 09:09:05.950203 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:05 crc kubenswrapper[4764]: I0202 09:09:05.950212 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:05 crc kubenswrapper[4764]: I0202 09:09:05.950230 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:05 crc kubenswrapper[4764]: I0202 09:09:05.950239 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:05Z","lastTransitionTime":"2026-02-02T09:09:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:06 crc kubenswrapper[4764]: I0202 09:09:06.054064 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:06 crc kubenswrapper[4764]: I0202 09:09:06.054434 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:06 crc kubenswrapper[4764]: I0202 09:09:06.054970 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:06 crc kubenswrapper[4764]: I0202 09:09:06.055182 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:06 crc kubenswrapper[4764]: I0202 09:09:06.055373 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:06Z","lastTransitionTime":"2026-02-02T09:09:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:06 crc kubenswrapper[4764]: I0202 09:09:06.159099 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:06 crc kubenswrapper[4764]: I0202 09:09:06.159447 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:06 crc kubenswrapper[4764]: I0202 09:09:06.159607 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:06 crc kubenswrapper[4764]: I0202 09:09:06.159776 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:06 crc kubenswrapper[4764]: I0202 09:09:06.160041 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:06Z","lastTransitionTime":"2026-02-02T09:09:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:06 crc kubenswrapper[4764]: I0202 09:09:06.263320 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:06 crc kubenswrapper[4764]: I0202 09:09:06.263375 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:06 crc kubenswrapper[4764]: I0202 09:09:06.263392 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:06 crc kubenswrapper[4764]: I0202 09:09:06.263415 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:06 crc kubenswrapper[4764]: I0202 09:09:06.263431 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:06Z","lastTransitionTime":"2026-02-02T09:09:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:06 crc kubenswrapper[4764]: I0202 09:09:06.366032 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:06 crc kubenswrapper[4764]: I0202 09:09:06.366093 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:06 crc kubenswrapper[4764]: I0202 09:09:06.366112 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:06 crc kubenswrapper[4764]: I0202 09:09:06.366135 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:06 crc kubenswrapper[4764]: I0202 09:09:06.366152 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:06Z","lastTransitionTime":"2026-02-02T09:09:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:06 crc kubenswrapper[4764]: I0202 09:09:06.468874 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:06 crc kubenswrapper[4764]: I0202 09:09:06.468922 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:06 crc kubenswrapper[4764]: I0202 09:09:06.469193 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:06 crc kubenswrapper[4764]: I0202 09:09:06.469222 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:06 crc kubenswrapper[4764]: I0202 09:09:06.469235 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:06Z","lastTransitionTime":"2026-02-02T09:09:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:06 crc kubenswrapper[4764]: I0202 09:09:06.572456 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:06 crc kubenswrapper[4764]: I0202 09:09:06.572491 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:06 crc kubenswrapper[4764]: I0202 09:09:06.572500 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:06 crc kubenswrapper[4764]: I0202 09:09:06.572536 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:06 crc kubenswrapper[4764]: I0202 09:09:06.572547 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:06Z","lastTransitionTime":"2026-02-02T09:09:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:06 crc kubenswrapper[4764]: I0202 09:09:06.675711 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:06 crc kubenswrapper[4764]: I0202 09:09:06.675767 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:06 crc kubenswrapper[4764]: I0202 09:09:06.675788 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:06 crc kubenswrapper[4764]: I0202 09:09:06.675819 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:06 crc kubenswrapper[4764]: I0202 09:09:06.675839 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:06Z","lastTransitionTime":"2026-02-02T09:09:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:06 crc kubenswrapper[4764]: I0202 09:09:06.779744 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:06 crc kubenswrapper[4764]: I0202 09:09:06.779794 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:06 crc kubenswrapper[4764]: I0202 09:09:06.779811 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:06 crc kubenswrapper[4764]: I0202 09:09:06.779838 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:06 crc kubenswrapper[4764]: I0202 09:09:06.779857 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:06Z","lastTransitionTime":"2026-02-02T09:09:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:06 crc kubenswrapper[4764]: I0202 09:09:06.825025 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 09:09:06 crc kubenswrapper[4764]: I0202 09:09:06.825190 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 09:09:06 crc kubenswrapper[4764]: E0202 09:09:06.825159 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 09:09:06 crc kubenswrapper[4764]: I0202 09:09:06.825289 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 09:09:06 crc kubenswrapper[4764]: E0202 09:09:06.825397 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 09:09:06 crc kubenswrapper[4764]: E0202 09:09:06.825469 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 09:09:06 crc kubenswrapper[4764]: I0202 09:09:06.847240 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-14 07:28:50.420778696 +0000 UTC Feb 02 09:09:06 crc kubenswrapper[4764]: I0202 09:09:06.882718 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:06 crc kubenswrapper[4764]: I0202 09:09:06.882779 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:06 crc kubenswrapper[4764]: I0202 09:09:06.882806 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:06 crc kubenswrapper[4764]: I0202 09:09:06.882838 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:06 crc kubenswrapper[4764]: I0202 09:09:06.882861 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:06Z","lastTransitionTime":"2026-02-02T09:09:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:06 crc kubenswrapper[4764]: I0202 09:09:06.985265 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:06 crc kubenswrapper[4764]: I0202 09:09:06.985365 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:06 crc kubenswrapper[4764]: I0202 09:09:06.985449 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:06 crc kubenswrapper[4764]: I0202 09:09:06.985481 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:06 crc kubenswrapper[4764]: I0202 09:09:06.985499 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:06Z","lastTransitionTime":"2026-02-02T09:09:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:07 crc kubenswrapper[4764]: I0202 09:09:07.087778 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:07 crc kubenswrapper[4764]: I0202 09:09:07.087809 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:07 crc kubenswrapper[4764]: I0202 09:09:07.087833 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:07 crc kubenswrapper[4764]: I0202 09:09:07.087849 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:07 crc kubenswrapper[4764]: I0202 09:09:07.087873 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:07Z","lastTransitionTime":"2026-02-02T09:09:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:07 crc kubenswrapper[4764]: I0202 09:09:07.188380 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:07 crc kubenswrapper[4764]: I0202 09:09:07.188411 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:07 crc kubenswrapper[4764]: I0202 09:09:07.188419 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:07 crc kubenswrapper[4764]: I0202 09:09:07.188433 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:07 crc kubenswrapper[4764]: I0202 09:09:07.188444 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:07Z","lastTransitionTime":"2026-02-02T09:09:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:07 crc kubenswrapper[4764]: E0202 09:09:07.205241 4764 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:09:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:09:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:09:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:09:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:09:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:09:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:09:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:09:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"31973a70-ce17-4811-9a14-f5476c6979db\\\",\\\"systemUUID\\\":\\\"ca92679c-8580-43c3-be51-62441e93d22b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:09:07Z is after 2025-08-24T17:21:41Z" Feb 02 09:09:07 crc kubenswrapper[4764]: I0202 09:09:07.210955 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:07 crc kubenswrapper[4764]: I0202 09:09:07.210988 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:07 crc kubenswrapper[4764]: I0202 09:09:07.210996 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:07 crc kubenswrapper[4764]: I0202 09:09:07.211010 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:07 crc kubenswrapper[4764]: I0202 09:09:07.211020 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:07Z","lastTransitionTime":"2026-02-02T09:09:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:07 crc kubenswrapper[4764]: E0202 09:09:07.225176 4764 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:09:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:09:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:09:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:09:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:09:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:09:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:09:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:09:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"31973a70-ce17-4811-9a14-f5476c6979db\\\",\\\"systemUUID\\\":\\\"ca92679c-8580-43c3-be51-62441e93d22b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:09:07Z is after 2025-08-24T17:21:41Z" Feb 02 09:09:07 crc kubenswrapper[4764]: I0202 09:09:07.230043 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:07 crc kubenswrapper[4764]: I0202 09:09:07.230232 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:07 crc kubenswrapper[4764]: I0202 09:09:07.230373 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:07 crc kubenswrapper[4764]: I0202 09:09:07.230541 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:07 crc kubenswrapper[4764]: I0202 09:09:07.230652 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:07Z","lastTransitionTime":"2026-02-02T09:09:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:07 crc kubenswrapper[4764]: E0202 09:09:07.243749 4764 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:09:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:09:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:09:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:09:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:09:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:09:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:09:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:09:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"31973a70-ce17-4811-9a14-f5476c6979db\\\",\\\"systemUUID\\\":\\\"ca92679c-8580-43c3-be51-62441e93d22b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:09:07Z is after 2025-08-24T17:21:41Z" Feb 02 09:09:07 crc kubenswrapper[4764]: I0202 09:09:07.247689 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:07 crc kubenswrapper[4764]: I0202 09:09:07.248023 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:07 crc kubenswrapper[4764]: I0202 09:09:07.248159 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:07 crc kubenswrapper[4764]: I0202 09:09:07.248377 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:07 crc kubenswrapper[4764]: I0202 09:09:07.248620 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:07Z","lastTransitionTime":"2026-02-02T09:09:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:07 crc kubenswrapper[4764]: E0202 09:09:07.261147 4764 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:09:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:09:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:09:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:09:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:09:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:09:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:09:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:09:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"31973a70-ce17-4811-9a14-f5476c6979db\\\",\\\"systemUUID\\\":\\\"ca92679c-8580-43c3-be51-62441e93d22b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:09:07Z is after 2025-08-24T17:21:41Z" Feb 02 09:09:07 crc kubenswrapper[4764]: I0202 09:09:07.265610 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:07 crc kubenswrapper[4764]: I0202 09:09:07.265687 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:07 crc kubenswrapper[4764]: I0202 09:09:07.265705 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:07 crc kubenswrapper[4764]: I0202 09:09:07.265729 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:07 crc kubenswrapper[4764]: I0202 09:09:07.265750 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:07Z","lastTransitionTime":"2026-02-02T09:09:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:07 crc kubenswrapper[4764]: E0202 09:09:07.283876 4764 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:09:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:09:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:09:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:09:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:09:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:09:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T09:09:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T09:09:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"31973a70-ce17-4811-9a14-f5476c6979db\\\",\\\"systemUUID\\\":\\\"ca92679c-8580-43c3-be51-62441e93d22b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:09:07Z is after 2025-08-24T17:21:41Z" Feb 02 09:09:07 crc kubenswrapper[4764]: E0202 09:09:07.284052 4764 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 02 09:09:07 crc kubenswrapper[4764]: I0202 09:09:07.285890 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:07 crc kubenswrapper[4764]: I0202 09:09:07.285956 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:07 crc kubenswrapper[4764]: I0202 09:09:07.285973 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:07 crc kubenswrapper[4764]: I0202 09:09:07.285994 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:07 crc kubenswrapper[4764]: I0202 09:09:07.286009 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:07Z","lastTransitionTime":"2026-02-02T09:09:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:07 crc kubenswrapper[4764]: I0202 09:09:07.388641 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:07 crc kubenswrapper[4764]: I0202 09:09:07.389209 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:07 crc kubenswrapper[4764]: I0202 09:09:07.389345 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:07 crc kubenswrapper[4764]: I0202 09:09:07.389449 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:07 crc kubenswrapper[4764]: I0202 09:09:07.389536 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:07Z","lastTransitionTime":"2026-02-02T09:09:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:07 crc kubenswrapper[4764]: I0202 09:09:07.492281 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:07 crc kubenswrapper[4764]: I0202 09:09:07.492626 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:07 crc kubenswrapper[4764]: I0202 09:09:07.492730 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:07 crc kubenswrapper[4764]: I0202 09:09:07.492806 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:07 crc kubenswrapper[4764]: I0202 09:09:07.492873 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:07Z","lastTransitionTime":"2026-02-02T09:09:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:07 crc kubenswrapper[4764]: I0202 09:09:07.595703 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:07 crc kubenswrapper[4764]: I0202 09:09:07.595748 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:07 crc kubenswrapper[4764]: I0202 09:09:07.595758 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:07 crc kubenswrapper[4764]: I0202 09:09:07.595772 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:07 crc kubenswrapper[4764]: I0202 09:09:07.595780 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:07Z","lastTransitionTime":"2026-02-02T09:09:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:07 crc kubenswrapper[4764]: I0202 09:09:07.698547 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:07 crc kubenswrapper[4764]: I0202 09:09:07.698597 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:07 crc kubenswrapper[4764]: I0202 09:09:07.698607 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:07 crc kubenswrapper[4764]: I0202 09:09:07.698627 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:07 crc kubenswrapper[4764]: I0202 09:09:07.698641 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:07Z","lastTransitionTime":"2026-02-02T09:09:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:07 crc kubenswrapper[4764]: I0202 09:09:07.801760 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:07 crc kubenswrapper[4764]: I0202 09:09:07.801834 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:07 crc kubenswrapper[4764]: I0202 09:09:07.801981 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:07 crc kubenswrapper[4764]: I0202 09:09:07.802019 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:07 crc kubenswrapper[4764]: I0202 09:09:07.802038 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:07Z","lastTransitionTime":"2026-02-02T09:09:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:07 crc kubenswrapper[4764]: I0202 09:09:07.825233 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kwtmr" Feb 02 09:09:07 crc kubenswrapper[4764]: E0202 09:09:07.825425 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kwtmr" podUID="991faa9a-dd25-4f49-82bd-ce60cefd4af2" Feb 02 09:09:07 crc kubenswrapper[4764]: I0202 09:09:07.847704 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-31 18:10:44.491604085 +0000 UTC Feb 02 09:09:07 crc kubenswrapper[4764]: I0202 09:09:07.853170 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d66ab3c0-9b5b-41d5-b6c0-91ccddccd7e9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d92ac486f5dade46bc75de32099859a71fc2ed25a7abbe6ae8937df128e7df30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a4ac273a4f127d79e5fad7d4c30cd2044a346f5734dba5d33d68cd092c0c2d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d38c7cb5dcd22fb07524494ebe77aa1a7f4f808f566043563817e47e7ec12a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://408722c7f83d28919431532f8c797a1c795fd5d96736d14e7f06a05744899702\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f473b18cd8003f75bb1dafb82bcec45745a73ccd10434297356dac185a3d15c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e229f484d19cbdabedf53d90a822a4434db9e00b1301538e885e26acc19d25f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e229f484d19cbdabedf53d90a822a4434db9e00b1301538e885e26acc19d25f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a9cc673be8ff245dba81278e8cd34cfe45ed36ff0b8bd5b30a0ddf333ec6371\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a9cc673be8ff245dba81278e8cd34cfe45ed36ff0b8bd5b30a0ddf333ec6371\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bc7e49c7a5ae66ea939060192fa6a346dc0b0b9aa9750b29879d3c426cc0d276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc7e49c7a5ae66ea939060192fa6a346dc0b0b9aa9750b29879d3c426cc0d276\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:09:07Z is after 2025-08-24T17:21:41Z" Feb 02 09:09:07 crc kubenswrapper[4764]: I0202 09:09:07.874205 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa9d037e66cb53b9dd698eff46284c9e718fd842cda6f7b525f0c79062d70bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:09:07Z is after 2025-08-24T17:21:41Z" Feb 02 09:09:07 crc kubenswrapper[4764]: I0202 09:09:07.892134 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:09:07Z is after 2025-08-24T17:21:41Z" Feb 02 09:09:07 crc kubenswrapper[4764]: I0202 09:09:07.905279 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:07 crc kubenswrapper[4764]: I0202 09:09:07.905333 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:07 crc kubenswrapper[4764]: I0202 09:09:07.905351 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:07 crc kubenswrapper[4764]: I0202 09:09:07.905376 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:07 crc kubenswrapper[4764]: I0202 09:09:07.905394 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:07Z","lastTransitionTime":"2026-02-02T09:09:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:07 crc kubenswrapper[4764]: I0202 09:09:07.916502 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24632dda-6100-4ab6-a28e-214ddae4360c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a6c67e9b7708cc0d85b0b6ffbbade0d34867e159a9c5487ca2606475c667a50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bffda48d7c881c1470251b501661324859373246cc1a21811bd36ca24887c7aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a989645cf485cd8ff80ffff1a86cd7808e1564eba5667535e161cc68db13b26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20b71d3ee7a381baa51b0b0d0470283022cfede30f491135737b5e82d631d763\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c380383b6a014d1b694a13ddc22f54b20b9e4f9d5c9b9000d76187621dce0a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5529f2bff87753428e9173826fc1880acf96a4c134fa9ecad195b15697cad73a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cb255e14eb02f7c08364c41fb9906e58b2155eb507f694b77ec8eb1f4251343\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3cb255e14eb02f7c08364c41fb9906e58b2155eb507f694b77ec8eb1f4251343\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-02T09:08:38Z\\\",\\\"message\\\":\\\":38.747848 6731 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-diagnostics/network-check-target-xd92c\\\\nI0202 09:08:38.747852 6731 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf in node crc\\\\nI0202 09:08:38.747854 6731 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-target-xd92c\\\\nI0202 09:08:38.747780 6731 obj_retry.go:386] Retry successful for *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc after 0 failed attempt(s)\\\\nI0202 09:08:38.747862 6731 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-target-xd92c in node crc\\\\nI0202 09:08:38.747864 6731 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf after 0 failed attempt(s)\\\\nI0202 09:08:38.747869 6731 default_network_controller.go:776] Recording success event on pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI0202 09:08:38.747837 6731 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-additional-cni-plugins-9p5dc in node crc\\\\nI0202 09:08:38.747884 6731 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-target-xd92c] creating logical port openshift-network-diagnostics_network-check-target-\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T09:08:37Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-zhn7j_openshift-ovn-kubernetes(24632dda-6100-4ab6-a28e-214ddae4360c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b4cfbd4e682bcdcaff785013bdb72580ce36fced733e235538536827036b93d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://579903eadb183466a2564737e6b678e04308239c0cd75b6808892053b97afbc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://579903eadb183466a2564737e6b678e04308239c0cd75b6808892053b97afbc0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7r4sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zhn7j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:09:07Z is after 2025-08-24T17:21:41Z" Feb 02 09:09:07 crc kubenswrapper[4764]: I0202 09:09:07.935608 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed747fd8-e3ce-4319-a6da-bda0fd51b85a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6f064bdc66470aace3180d4988244026b54009a63cff199e41c5d02a1d18479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://60258f84a1d8dcfe48b7e9688ed50f077e201971f20f3f6663fbf1ba58a4611a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f84be2e08217a2ac3edde858d78b4410376ed9bae3bcabe2a7f9b2fbe2f69c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a69d25d352d7ceac5c735bc1fcf446a1fea6dc22c65a5296cfab828cea17377\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:09:07Z is after 2025-08-24T17:21:41Z" Feb 02 09:09:07 crc kubenswrapper[4764]: I0202 09:09:07.953198 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f6904228e052fff69ad1fd66df97758b6b60ba4351e97a7fbd03eee4df1b653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:09:07Z is after 2025-08-24T17:21:41Z" Feb 02 09:09:07 crc kubenswrapper[4764]: I0202 09:09:07.966952 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6zxff" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aaba93da-12a5-43df-b7c7-4c8b988163c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3db293678a907cd55b31b1d0c0bd7c30e1ba75ecd5e13cbc68ab98fbf0fbd9d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqwb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6zxff\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:09:07Z is after 2025-08-24T17:21:41Z" Feb 02 09:09:07 crc kubenswrapper[4764]: I0202 09:09:07.983678 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d192f670-9f9d-4539-9641-e4bed73acdd4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://040c5e8a5659f85f0d52d1a1e064154d998ae104f711101d2204c0b1caebbb30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbksf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fdf06410349b7c99aa166e87fbe80bf68e2ae923635a6a978b7e2cd50383f7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fbksf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4ndm4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:09:07Z is after 2025-08-24T17:21:41Z" Feb 02 09:09:07 crc kubenswrapper[4764]: I0202 09:09:07.995384 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pcjmn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6e5ccd3b-b657-47a2-9c4d-bc9adb126533\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72863eef568d92a1abeb3cde2e89630a4f5fb7626c6949448ff5195774c98eab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2f7zl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pcjmn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:09:07Z is after 2025-08-24T17:21:41Z" Feb 02 09:09:08 crc kubenswrapper[4764]: I0202 09:09:08.008449 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:08 crc kubenswrapper[4764]: I0202 09:09:08.008494 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:08 crc kubenswrapper[4764]: I0202 09:09:08.008507 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:08 crc kubenswrapper[4764]: I0202 09:09:08.008526 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:08 crc kubenswrapper[4764]: I0202 09:09:08.008537 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:08Z","lastTransitionTime":"2026-02-02T09:09:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:08 crc kubenswrapper[4764]: I0202 09:09:08.012843 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fae745b-ea49-42f3-a486-dbc3dc2fb9ad\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://081d7f76d937ff71e35945f9b1d893888bb649d15bd84312fd4b61c9520ea71f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://429ead0993f5e648f6256d249ee12aea03514d3108618840552b672fb419cd74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fad17b01545eb20785f03e3c61bd5c982f6217e38176d7a3b23ac194298644a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85e2f75c56d63cacc93b9195c1e8bf46eb7a7234c45a024bd95a460db90edb6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85e2f75c56d63cacc93b9195c1e8bf46eb7a7234c45a024bd95a460db90edb6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:09:08Z is after 2025-08-24T17:21:41Z" Feb 02 09:09:08 crc kubenswrapper[4764]: I0202 09:09:08.027062 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:09:08Z is after 2025-08-24T17:21:41Z" Feb 02 09:09:08 crc kubenswrapper[4764]: I0202 09:09:08.040369 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lh6tm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2152e8a-f456-47b5-a547-b65ec837ecbc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38e176a00b07f2960fdd7ba2aadd1468a5e083b8d154e9d118d13f2585da81e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bb2sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://996e649011d5255783284421ddea8e0a5c383ef675b0920c3fdc0af6503d80e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bb2sk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lh6tm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:09:08Z is after 2025-08-24T17:21:41Z" Feb 02 09:09:08 crc kubenswrapper[4764]: I0202 09:09:08.053416 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87d61aea-46f6-4ee6-a7a4-767a6a67fea4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e78b6d08e4d616cefbb73e3fd43b3c6886dcd8aa997c7b4948bf1ff63401293\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d468715c2da33996561c40043216233fd12e21539d0c58dd1f3b892fd9f91a9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b184f7c936dc4046f74c0018269cbc2c4b7ceddb650b4e871d3f8a707e0df88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://269c5074371bd00f8363fb3888c11f1df383081a61497aea6578f08700881d50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b25bc9395831524c53f9e565b4efbe74f21792efafeb1dd507161bc91aa0ecdf\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T09:07:22Z\\\",\\\"message\\\":\\\"W0202 09:07:21.597488 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0202 09:07:21.598117 1 crypto.go:601] Generating new CA for check-endpoints-signer@1770023241 cert, and key in /tmp/serving-cert-3856183724/serving-signer.crt, /tmp/serving-cert-3856183724/serving-signer.key\\\\nI0202 09:07:22.012184 1 observer_polling.go:159] Starting file observer\\\\nW0202 09:07:22.015493 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0202 09:07:22.015797 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 09:07:22.016596 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3856183724/tls.crt::/tmp/serving-cert-3856183724/tls.key\\\\\\\"\\\\nF0202 09:07:22.270605 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b44e17b8ca01f549bc10081dd5f109cb1b9fff364bf5bd4fd90d9f0e3eeb04db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfb06a5ec9262cf39b108a66014d15e325cdbd5441a2d9e7a811dd57265c99a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfb06a5ec9262cf39b108a66014d15e325cdbd5441a2d9e7a811dd57265c99a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:09:08Z is after 2025-08-24T17:21:41Z" Feb 02 09:09:08 crc kubenswrapper[4764]: I0202 09:09:08.065898 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fea07bdc-9134-4993-934c-316c19d0dc76\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://664e0519c2a2c60904dfcc8deebaf5d638dac0bedb13d870c0aaea5b4f5428f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db5a54d3a2117252a1ff50884683f99185874d2904c0960161642280f614679d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db5a54d3a2117252a1ff50884683f99185874d2904c0960161642280f614679d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:09:08Z is after 2025-08-24T17:21:41Z" Feb 02 09:09:08 crc kubenswrapper[4764]: I0202 09:09:08.084793 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5528225d91ac2778bef4554abb0bd036ac5d3b6cdac7357bab69e70c3894b6f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://071bbb80d0747cf3e109ba626d47f7d05f85f4cc0b3aba77276c795741f47ac6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:09:08Z is after 2025-08-24T17:21:41Z" Feb 02 09:09:08 crc kubenswrapper[4764]: I0202 09:09:08.102881 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:09:08Z is after 2025-08-24T17:21:41Z" Feb 02 09:09:08 crc kubenswrapper[4764]: I0202 09:09:08.111183 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:08 crc kubenswrapper[4764]: I0202 09:09:08.111237 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:08 crc kubenswrapper[4764]: I0202 09:09:08.111248 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:08 crc kubenswrapper[4764]: I0202 09:09:08.111269 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:08 crc kubenswrapper[4764]: I0202 09:09:08.111280 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:08Z","lastTransitionTime":"2026-02-02T09:09:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:08 crc kubenswrapper[4764]: I0202 09:09:08.120414 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nc7qt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:08:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b52b499d17b4e2e359475bd13730c72ef5cedb052e13def6dfe5bc5fd7487d2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1a2c7b22068b7165927fc158885d48bf9952b9eba521d8ae9a599eee53dc305\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-02T09:08:28Z\\\",\\\"message\\\":\\\"2026-02-02T09:07:42+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_32cbaf6a-865c-4787-8f89-a40249873b40\\\\n2026-02-02T09:07:42+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_32cbaf6a-865c-4787-8f89-a40249873b40 to /host/opt/cni/bin/\\\\n2026-02-02T09:07:43Z [verbose] multus-daemon started\\\\n2026-02-02T09:07:43Z [verbose] Readiness Indicator file check\\\\n2026-02-02T09:08:28Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:08:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n2hz4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nc7qt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:09:08Z is after 2025-08-24T17:21:41Z" Feb 02 09:09:08 crc kubenswrapper[4764]: I0202 09:09:08.137918 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9p5dc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"660ac006-2495-45c4-9fb3-e9c1dddcf7a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da812b124c97aa8ad1563f4bcfac8f3bfa025096fa4a2561e5847b1bf9cb30b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T09:07:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23e57e65a6f5c3059506b9db21e9ac2492c30bd9191a98db0281915e9ce39797\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23e57e65a6f5c3059506b9db21e9ac2492c30bd9191a98db0281915e9ce39797\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2417b66816da1a6cd1791a4b91eca4a3c4f3694462e066e7f0722f1b9c2105c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2417b66816da1a6cd1791a4b91eca4a3c4f3694462e066e7f0722f1b9c2105c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://883641f29e9120c2cea6b272bacf1dc63c752f55ab0408cdaf73db85c85ac369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://883641f29e9120c2cea6b272bacf1dc63c752f55ab0408cdaf73db85c85ac369\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2745baba1188ba5e8f820616e4117a12c50ecee5b68e71ad9cabbb621b6e131f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2745baba1188ba5e8f820616e4117a12c50ecee5b68e71ad9cabbb621b6e131f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8925e435be4d48d86824fc35f8b8819fb35d63e385164c4a20edb3bf3035c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a8925e435be4d48d86824fc35f8b8819fb35d63e385164c4a20edb3bf3035c3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d3fbf668ae7d04c8d9e574f43c2fe90c5604204ba38a3c60847afdfbceb7a57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d3fbf668ae7d04c8d9e574f43c2fe90c5604204ba38a3c60847afdfbceb7a57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T09:07:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T09:07:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ch5jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9p5dc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:09:08Z is after 2025-08-24T17:21:41Z" Feb 02 09:09:08 crc kubenswrapper[4764]: I0202 09:09:08.153911 4764 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kwtmr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"991faa9a-dd25-4f49-82bd-ce60cefd4af2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T09:07:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h865b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h865b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T09:07:53Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kwtmr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T09:09:08Z is after 2025-08-24T17:21:41Z" Feb 02 09:09:08 crc kubenswrapper[4764]: I0202 09:09:08.214499 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:08 crc kubenswrapper[4764]: I0202 09:09:08.214554 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:08 crc kubenswrapper[4764]: I0202 09:09:08.214567 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:08 crc kubenswrapper[4764]: I0202 09:09:08.214587 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:08 crc kubenswrapper[4764]: I0202 09:09:08.214599 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:08Z","lastTransitionTime":"2026-02-02T09:09:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:08 crc kubenswrapper[4764]: I0202 09:09:08.323056 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:08 crc kubenswrapper[4764]: I0202 09:09:08.323103 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:08 crc kubenswrapper[4764]: I0202 09:09:08.323129 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:08 crc kubenswrapper[4764]: I0202 09:09:08.323147 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:08 crc kubenswrapper[4764]: I0202 09:09:08.323158 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:08Z","lastTransitionTime":"2026-02-02T09:09:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:08 crc kubenswrapper[4764]: I0202 09:09:08.425323 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:08 crc kubenswrapper[4764]: I0202 09:09:08.425389 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:08 crc kubenswrapper[4764]: I0202 09:09:08.425407 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:08 crc kubenswrapper[4764]: I0202 09:09:08.425430 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:08 crc kubenswrapper[4764]: I0202 09:09:08.425446 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:08Z","lastTransitionTime":"2026-02-02T09:09:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:08 crc kubenswrapper[4764]: I0202 09:09:08.528896 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:08 crc kubenswrapper[4764]: I0202 09:09:08.528984 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:08 crc kubenswrapper[4764]: I0202 09:09:08.529002 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:08 crc kubenswrapper[4764]: I0202 09:09:08.529023 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:08 crc kubenswrapper[4764]: I0202 09:09:08.529037 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:08Z","lastTransitionTime":"2026-02-02T09:09:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:08 crc kubenswrapper[4764]: I0202 09:09:08.632186 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:08 crc kubenswrapper[4764]: I0202 09:09:08.632223 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:08 crc kubenswrapper[4764]: I0202 09:09:08.632232 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:08 crc kubenswrapper[4764]: I0202 09:09:08.632247 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:08 crc kubenswrapper[4764]: I0202 09:09:08.632256 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:08Z","lastTransitionTime":"2026-02-02T09:09:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:08 crc kubenswrapper[4764]: I0202 09:09:08.734880 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:08 crc kubenswrapper[4764]: I0202 09:09:08.734962 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:08 crc kubenswrapper[4764]: I0202 09:09:08.734987 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:08 crc kubenswrapper[4764]: I0202 09:09:08.735005 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:08 crc kubenswrapper[4764]: I0202 09:09:08.735019 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:08Z","lastTransitionTime":"2026-02-02T09:09:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:08 crc kubenswrapper[4764]: I0202 09:09:08.824800 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 09:09:08 crc kubenswrapper[4764]: E0202 09:09:08.824928 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 09:09:08 crc kubenswrapper[4764]: I0202 09:09:08.825012 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 09:09:08 crc kubenswrapper[4764]: I0202 09:09:08.825032 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 09:09:08 crc kubenswrapper[4764]: E0202 09:09:08.825457 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 09:09:08 crc kubenswrapper[4764]: E0202 09:09:08.825585 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 09:09:08 crc kubenswrapper[4764]: I0202 09:09:08.837531 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:08 crc kubenswrapper[4764]: I0202 09:09:08.837578 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:08 crc kubenswrapper[4764]: I0202 09:09:08.837588 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:08 crc kubenswrapper[4764]: I0202 09:09:08.837602 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:08 crc kubenswrapper[4764]: I0202 09:09:08.837613 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:08Z","lastTransitionTime":"2026-02-02T09:09:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:08 crc kubenswrapper[4764]: I0202 09:09:08.848789 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-17 05:09:45.316690791 +0000 UTC Feb 02 09:09:08 crc kubenswrapper[4764]: I0202 09:09:08.940696 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:08 crc kubenswrapper[4764]: I0202 09:09:08.940746 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:08 crc kubenswrapper[4764]: I0202 09:09:08.940758 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:08 crc kubenswrapper[4764]: I0202 09:09:08.940813 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:08 crc kubenswrapper[4764]: I0202 09:09:08.940829 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:08Z","lastTransitionTime":"2026-02-02T09:09:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:09 crc kubenswrapper[4764]: I0202 09:09:09.043575 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:09 crc kubenswrapper[4764]: I0202 09:09:09.043625 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:09 crc kubenswrapper[4764]: I0202 09:09:09.043638 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:09 crc kubenswrapper[4764]: I0202 09:09:09.043656 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:09 crc kubenswrapper[4764]: I0202 09:09:09.043670 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:09Z","lastTransitionTime":"2026-02-02T09:09:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:09 crc kubenswrapper[4764]: I0202 09:09:09.146285 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:09 crc kubenswrapper[4764]: I0202 09:09:09.146541 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:09 crc kubenswrapper[4764]: I0202 09:09:09.146642 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:09 crc kubenswrapper[4764]: I0202 09:09:09.146731 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:09 crc kubenswrapper[4764]: I0202 09:09:09.146884 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:09Z","lastTransitionTime":"2026-02-02T09:09:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:09 crc kubenswrapper[4764]: I0202 09:09:09.249636 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:09 crc kubenswrapper[4764]: I0202 09:09:09.250004 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:09 crc kubenswrapper[4764]: I0202 09:09:09.250111 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:09 crc kubenswrapper[4764]: I0202 09:09:09.250210 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:09 crc kubenswrapper[4764]: I0202 09:09:09.250295 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:09Z","lastTransitionTime":"2026-02-02T09:09:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:09 crc kubenswrapper[4764]: I0202 09:09:09.353307 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:09 crc kubenswrapper[4764]: I0202 09:09:09.353355 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:09 crc kubenswrapper[4764]: I0202 09:09:09.353365 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:09 crc kubenswrapper[4764]: I0202 09:09:09.353384 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:09 crc kubenswrapper[4764]: I0202 09:09:09.353394 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:09Z","lastTransitionTime":"2026-02-02T09:09:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:09 crc kubenswrapper[4764]: I0202 09:09:09.456152 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:09 crc kubenswrapper[4764]: I0202 09:09:09.456196 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:09 crc kubenswrapper[4764]: I0202 09:09:09.456207 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:09 crc kubenswrapper[4764]: I0202 09:09:09.456226 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:09 crc kubenswrapper[4764]: I0202 09:09:09.456240 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:09Z","lastTransitionTime":"2026-02-02T09:09:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:09 crc kubenswrapper[4764]: I0202 09:09:09.559277 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:09 crc kubenswrapper[4764]: I0202 09:09:09.559358 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:09 crc kubenswrapper[4764]: I0202 09:09:09.559383 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:09 crc kubenswrapper[4764]: I0202 09:09:09.559415 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:09 crc kubenswrapper[4764]: I0202 09:09:09.559437 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:09Z","lastTransitionTime":"2026-02-02T09:09:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:09 crc kubenswrapper[4764]: I0202 09:09:09.661959 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:09 crc kubenswrapper[4764]: I0202 09:09:09.661995 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:09 crc kubenswrapper[4764]: I0202 09:09:09.662004 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:09 crc kubenswrapper[4764]: I0202 09:09:09.662018 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:09 crc kubenswrapper[4764]: I0202 09:09:09.662029 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:09Z","lastTransitionTime":"2026-02-02T09:09:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:09 crc kubenswrapper[4764]: I0202 09:09:09.764049 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:09 crc kubenswrapper[4764]: I0202 09:09:09.764079 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:09 crc kubenswrapper[4764]: I0202 09:09:09.764087 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:09 crc kubenswrapper[4764]: I0202 09:09:09.764100 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:09 crc kubenswrapper[4764]: I0202 09:09:09.764109 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:09Z","lastTransitionTime":"2026-02-02T09:09:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:09 crc kubenswrapper[4764]: I0202 09:09:09.825321 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kwtmr" Feb 02 09:09:09 crc kubenswrapper[4764]: E0202 09:09:09.825447 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kwtmr" podUID="991faa9a-dd25-4f49-82bd-ce60cefd4af2" Feb 02 09:09:09 crc kubenswrapper[4764]: I0202 09:09:09.849999 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-29 08:18:43.801957753 +0000 UTC Feb 02 09:09:09 crc kubenswrapper[4764]: I0202 09:09:09.866561 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:09 crc kubenswrapper[4764]: I0202 09:09:09.866596 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:09 crc kubenswrapper[4764]: I0202 09:09:09.866608 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:09 crc kubenswrapper[4764]: I0202 09:09:09.866624 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:09 crc kubenswrapper[4764]: I0202 09:09:09.866635 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:09Z","lastTransitionTime":"2026-02-02T09:09:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:09 crc kubenswrapper[4764]: I0202 09:09:09.968325 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:09 crc kubenswrapper[4764]: I0202 09:09:09.968865 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:09 crc kubenswrapper[4764]: I0202 09:09:09.969037 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:09 crc kubenswrapper[4764]: I0202 09:09:09.969110 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:09 crc kubenswrapper[4764]: I0202 09:09:09.969184 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:09Z","lastTransitionTime":"2026-02-02T09:09:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:10 crc kubenswrapper[4764]: I0202 09:09:10.071299 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:10 crc kubenswrapper[4764]: I0202 09:09:10.071568 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:10 crc kubenswrapper[4764]: I0202 09:09:10.071651 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:10 crc kubenswrapper[4764]: I0202 09:09:10.071728 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:10 crc kubenswrapper[4764]: I0202 09:09:10.071804 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:10Z","lastTransitionTime":"2026-02-02T09:09:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:10 crc kubenswrapper[4764]: I0202 09:09:10.174627 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:10 crc kubenswrapper[4764]: I0202 09:09:10.174688 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:10 crc kubenswrapper[4764]: I0202 09:09:10.174702 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:10 crc kubenswrapper[4764]: I0202 09:09:10.174716 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:10 crc kubenswrapper[4764]: I0202 09:09:10.174726 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:10Z","lastTransitionTime":"2026-02-02T09:09:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:10 crc kubenswrapper[4764]: I0202 09:09:10.277618 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:10 crc kubenswrapper[4764]: I0202 09:09:10.277663 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:10 crc kubenswrapper[4764]: I0202 09:09:10.277675 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:10 crc kubenswrapper[4764]: I0202 09:09:10.277694 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:10 crc kubenswrapper[4764]: I0202 09:09:10.277708 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:10Z","lastTransitionTime":"2026-02-02T09:09:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:10 crc kubenswrapper[4764]: I0202 09:09:10.380515 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:10 crc kubenswrapper[4764]: I0202 09:09:10.380833 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:10 crc kubenswrapper[4764]: I0202 09:09:10.380925 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:10 crc kubenswrapper[4764]: I0202 09:09:10.381082 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:10 crc kubenswrapper[4764]: I0202 09:09:10.381166 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:10Z","lastTransitionTime":"2026-02-02T09:09:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:10 crc kubenswrapper[4764]: I0202 09:09:10.484805 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:10 crc kubenswrapper[4764]: I0202 09:09:10.484877 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:10 crc kubenswrapper[4764]: I0202 09:09:10.484895 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:10 crc kubenswrapper[4764]: I0202 09:09:10.484918 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:10 crc kubenswrapper[4764]: I0202 09:09:10.484988 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:10Z","lastTransitionTime":"2026-02-02T09:09:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:10 crc kubenswrapper[4764]: I0202 09:09:10.587120 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:10 crc kubenswrapper[4764]: I0202 09:09:10.587152 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:10 crc kubenswrapper[4764]: I0202 09:09:10.587163 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:10 crc kubenswrapper[4764]: I0202 09:09:10.587179 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:10 crc kubenswrapper[4764]: I0202 09:09:10.587188 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:10Z","lastTransitionTime":"2026-02-02T09:09:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:10 crc kubenswrapper[4764]: I0202 09:09:10.689769 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:10 crc kubenswrapper[4764]: I0202 09:09:10.689833 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:10 crc kubenswrapper[4764]: I0202 09:09:10.689851 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:10 crc kubenswrapper[4764]: I0202 09:09:10.689876 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:10 crc kubenswrapper[4764]: I0202 09:09:10.689894 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:10Z","lastTransitionTime":"2026-02-02T09:09:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:10 crc kubenswrapper[4764]: I0202 09:09:10.791859 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:10 crc kubenswrapper[4764]: I0202 09:09:10.792128 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:10 crc kubenswrapper[4764]: I0202 09:09:10.792278 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:10 crc kubenswrapper[4764]: I0202 09:09:10.792384 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:10 crc kubenswrapper[4764]: I0202 09:09:10.792463 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:10Z","lastTransitionTime":"2026-02-02T09:09:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:10 crc kubenswrapper[4764]: I0202 09:09:10.824830 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 09:09:10 crc kubenswrapper[4764]: I0202 09:09:10.824953 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 09:09:10 crc kubenswrapper[4764]: E0202 09:09:10.825000 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 09:09:10 crc kubenswrapper[4764]: E0202 09:09:10.825079 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 09:09:10 crc kubenswrapper[4764]: I0202 09:09:10.825244 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 09:09:10 crc kubenswrapper[4764]: E0202 09:09:10.825372 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 09:09:10 crc kubenswrapper[4764]: I0202 09:09:10.850655 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-25 04:56:03.812814655 +0000 UTC Feb 02 09:09:10 crc kubenswrapper[4764]: I0202 09:09:10.894871 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:10 crc kubenswrapper[4764]: I0202 09:09:10.895180 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:10 crc kubenswrapper[4764]: I0202 09:09:10.895292 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:10 crc kubenswrapper[4764]: I0202 09:09:10.895406 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:10 crc kubenswrapper[4764]: I0202 09:09:10.895508 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:10Z","lastTransitionTime":"2026-02-02T09:09:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:10 crc kubenswrapper[4764]: I0202 09:09:10.997760 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:10 crc kubenswrapper[4764]: I0202 09:09:10.997791 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:10 crc kubenswrapper[4764]: I0202 09:09:10.997801 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:10 crc kubenswrapper[4764]: I0202 09:09:10.997816 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:10 crc kubenswrapper[4764]: I0202 09:09:10.997827 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:10Z","lastTransitionTime":"2026-02-02T09:09:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:11 crc kubenswrapper[4764]: I0202 09:09:11.100559 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:11 crc kubenswrapper[4764]: I0202 09:09:11.100599 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:11 crc kubenswrapper[4764]: I0202 09:09:11.100608 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:11 crc kubenswrapper[4764]: I0202 09:09:11.100625 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:11 crc kubenswrapper[4764]: I0202 09:09:11.100634 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:11Z","lastTransitionTime":"2026-02-02T09:09:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:11 crc kubenswrapper[4764]: I0202 09:09:11.202599 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:11 crc kubenswrapper[4764]: I0202 09:09:11.202639 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:11 crc kubenswrapper[4764]: I0202 09:09:11.202650 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:11 crc kubenswrapper[4764]: I0202 09:09:11.202666 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:11 crc kubenswrapper[4764]: I0202 09:09:11.202679 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:11Z","lastTransitionTime":"2026-02-02T09:09:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:11 crc kubenswrapper[4764]: I0202 09:09:11.305322 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:11 crc kubenswrapper[4764]: I0202 09:09:11.305609 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:11 crc kubenswrapper[4764]: I0202 09:09:11.305687 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:11 crc kubenswrapper[4764]: I0202 09:09:11.305752 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:11 crc kubenswrapper[4764]: I0202 09:09:11.305820 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:11Z","lastTransitionTime":"2026-02-02T09:09:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:11 crc kubenswrapper[4764]: I0202 09:09:11.407728 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:11 crc kubenswrapper[4764]: I0202 09:09:11.407764 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:11 crc kubenswrapper[4764]: I0202 09:09:11.407833 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:11 crc kubenswrapper[4764]: I0202 09:09:11.407876 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:11 crc kubenswrapper[4764]: I0202 09:09:11.407884 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:11Z","lastTransitionTime":"2026-02-02T09:09:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:11 crc kubenswrapper[4764]: I0202 09:09:11.511001 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:11 crc kubenswrapper[4764]: I0202 09:09:11.511060 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:11 crc kubenswrapper[4764]: I0202 09:09:11.511076 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:11 crc kubenswrapper[4764]: I0202 09:09:11.511104 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:11 crc kubenswrapper[4764]: I0202 09:09:11.511124 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:11Z","lastTransitionTime":"2026-02-02T09:09:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:11 crc kubenswrapper[4764]: I0202 09:09:11.613580 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:11 crc kubenswrapper[4764]: I0202 09:09:11.613646 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:11 crc kubenswrapper[4764]: I0202 09:09:11.613669 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:11 crc kubenswrapper[4764]: I0202 09:09:11.613697 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:11 crc kubenswrapper[4764]: I0202 09:09:11.613721 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:11Z","lastTransitionTime":"2026-02-02T09:09:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:11 crc kubenswrapper[4764]: I0202 09:09:11.716696 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:11 crc kubenswrapper[4764]: I0202 09:09:11.717111 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:11 crc kubenswrapper[4764]: I0202 09:09:11.717333 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:11 crc kubenswrapper[4764]: I0202 09:09:11.717578 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:11 crc kubenswrapper[4764]: I0202 09:09:11.717788 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:11Z","lastTransitionTime":"2026-02-02T09:09:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:11 crc kubenswrapper[4764]: I0202 09:09:11.822484 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:11 crc kubenswrapper[4764]: I0202 09:09:11.822544 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:11 crc kubenswrapper[4764]: I0202 09:09:11.822566 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:11 crc kubenswrapper[4764]: I0202 09:09:11.822593 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:11 crc kubenswrapper[4764]: I0202 09:09:11.822614 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:11Z","lastTransitionTime":"2026-02-02T09:09:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:11 crc kubenswrapper[4764]: I0202 09:09:11.826169 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kwtmr" Feb 02 09:09:11 crc kubenswrapper[4764]: E0202 09:09:11.826393 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kwtmr" podUID="991faa9a-dd25-4f49-82bd-ce60cefd4af2" Feb 02 09:09:11 crc kubenswrapper[4764]: I0202 09:09:11.851809 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-04 04:28:19.636822418 +0000 UTC Feb 02 09:09:11 crc kubenswrapper[4764]: I0202 09:09:11.925166 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:11 crc kubenswrapper[4764]: I0202 09:09:11.925496 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:11 crc kubenswrapper[4764]: I0202 09:09:11.925636 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:11 crc kubenswrapper[4764]: I0202 09:09:11.925777 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:11 crc kubenswrapper[4764]: I0202 09:09:11.925920 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:11Z","lastTransitionTime":"2026-02-02T09:09:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:12 crc kubenswrapper[4764]: I0202 09:09:12.029061 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:12 crc kubenswrapper[4764]: I0202 09:09:12.029237 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:12 crc kubenswrapper[4764]: I0202 09:09:12.029268 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:12 crc kubenswrapper[4764]: I0202 09:09:12.029301 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:12 crc kubenswrapper[4764]: I0202 09:09:12.029322 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:12Z","lastTransitionTime":"2026-02-02T09:09:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:12 crc kubenswrapper[4764]: I0202 09:09:12.132529 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:12 crc kubenswrapper[4764]: I0202 09:09:12.132598 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:12 crc kubenswrapper[4764]: I0202 09:09:12.132621 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:12 crc kubenswrapper[4764]: I0202 09:09:12.132652 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:12 crc kubenswrapper[4764]: I0202 09:09:12.132674 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:12Z","lastTransitionTime":"2026-02-02T09:09:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:12 crc kubenswrapper[4764]: I0202 09:09:12.235594 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:12 crc kubenswrapper[4764]: I0202 09:09:12.235661 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:12 crc kubenswrapper[4764]: I0202 09:09:12.235680 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:12 crc kubenswrapper[4764]: I0202 09:09:12.235707 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:12 crc kubenswrapper[4764]: I0202 09:09:12.235728 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:12Z","lastTransitionTime":"2026-02-02T09:09:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:12 crc kubenswrapper[4764]: I0202 09:09:12.339249 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:12 crc kubenswrapper[4764]: I0202 09:09:12.339302 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:12 crc kubenswrapper[4764]: I0202 09:09:12.339320 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:12 crc kubenswrapper[4764]: I0202 09:09:12.339388 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:12 crc kubenswrapper[4764]: I0202 09:09:12.339410 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:12Z","lastTransitionTime":"2026-02-02T09:09:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:12 crc kubenswrapper[4764]: I0202 09:09:12.442323 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:12 crc kubenswrapper[4764]: I0202 09:09:12.442451 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:12 crc kubenswrapper[4764]: I0202 09:09:12.442471 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:12 crc kubenswrapper[4764]: I0202 09:09:12.442497 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:12 crc kubenswrapper[4764]: I0202 09:09:12.442514 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:12Z","lastTransitionTime":"2026-02-02T09:09:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:12 crc kubenswrapper[4764]: I0202 09:09:12.545619 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:12 crc kubenswrapper[4764]: I0202 09:09:12.545691 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:12 crc kubenswrapper[4764]: I0202 09:09:12.545744 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:12 crc kubenswrapper[4764]: I0202 09:09:12.545775 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:12 crc kubenswrapper[4764]: I0202 09:09:12.545796 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:12Z","lastTransitionTime":"2026-02-02T09:09:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:12 crc kubenswrapper[4764]: I0202 09:09:12.648421 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:12 crc kubenswrapper[4764]: I0202 09:09:12.648519 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:12 crc kubenswrapper[4764]: I0202 09:09:12.648586 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:12 crc kubenswrapper[4764]: I0202 09:09:12.648611 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:12 crc kubenswrapper[4764]: I0202 09:09:12.648647 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:12Z","lastTransitionTime":"2026-02-02T09:09:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:12 crc kubenswrapper[4764]: I0202 09:09:12.751025 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:12 crc kubenswrapper[4764]: I0202 09:09:12.751089 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:12 crc kubenswrapper[4764]: I0202 09:09:12.751142 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:12 crc kubenswrapper[4764]: I0202 09:09:12.751170 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:12 crc kubenswrapper[4764]: I0202 09:09:12.751188 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:12Z","lastTransitionTime":"2026-02-02T09:09:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:12 crc kubenswrapper[4764]: I0202 09:09:12.825400 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 09:09:12 crc kubenswrapper[4764]: I0202 09:09:12.825445 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 09:09:12 crc kubenswrapper[4764]: I0202 09:09:12.825469 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 09:09:12 crc kubenswrapper[4764]: E0202 09:09:12.825548 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 09:09:12 crc kubenswrapper[4764]: E0202 09:09:12.825684 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 09:09:12 crc kubenswrapper[4764]: E0202 09:09:12.825796 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 09:09:12 crc kubenswrapper[4764]: I0202 09:09:12.852393 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-15 15:43:42.727195682 +0000 UTC Feb 02 09:09:12 crc kubenswrapper[4764]: I0202 09:09:12.854370 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:12 crc kubenswrapper[4764]: I0202 09:09:12.854415 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:12 crc kubenswrapper[4764]: I0202 09:09:12.854426 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:12 crc kubenswrapper[4764]: I0202 09:09:12.854443 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:12 crc kubenswrapper[4764]: I0202 09:09:12.854454 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:12Z","lastTransitionTime":"2026-02-02T09:09:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:12 crc kubenswrapper[4764]: I0202 09:09:12.957072 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:12 crc kubenswrapper[4764]: I0202 09:09:12.957154 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:12 crc kubenswrapper[4764]: I0202 09:09:12.957173 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:12 crc kubenswrapper[4764]: I0202 09:09:12.957197 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:12 crc kubenswrapper[4764]: I0202 09:09:12.957217 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:12Z","lastTransitionTime":"2026-02-02T09:09:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:13 crc kubenswrapper[4764]: I0202 09:09:13.061337 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:13 crc kubenswrapper[4764]: I0202 09:09:13.061429 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:13 crc kubenswrapper[4764]: I0202 09:09:13.061467 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:13 crc kubenswrapper[4764]: I0202 09:09:13.061492 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:13 crc kubenswrapper[4764]: I0202 09:09:13.061532 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:13Z","lastTransitionTime":"2026-02-02T09:09:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:13 crc kubenswrapper[4764]: I0202 09:09:13.164989 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:13 crc kubenswrapper[4764]: I0202 09:09:13.165064 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:13 crc kubenswrapper[4764]: I0202 09:09:13.165082 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:13 crc kubenswrapper[4764]: I0202 09:09:13.165106 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:13 crc kubenswrapper[4764]: I0202 09:09:13.165123 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:13Z","lastTransitionTime":"2026-02-02T09:09:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:13 crc kubenswrapper[4764]: I0202 09:09:13.267907 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:13 crc kubenswrapper[4764]: I0202 09:09:13.268053 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:13 crc kubenswrapper[4764]: I0202 09:09:13.268128 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:13 crc kubenswrapper[4764]: I0202 09:09:13.268170 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:13 crc kubenswrapper[4764]: I0202 09:09:13.268187 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:13Z","lastTransitionTime":"2026-02-02T09:09:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:13 crc kubenswrapper[4764]: I0202 09:09:13.372299 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:13 crc kubenswrapper[4764]: I0202 09:09:13.372385 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:13 crc kubenswrapper[4764]: I0202 09:09:13.372401 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:13 crc kubenswrapper[4764]: I0202 09:09:13.372452 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:13 crc kubenswrapper[4764]: I0202 09:09:13.372464 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:13Z","lastTransitionTime":"2026-02-02T09:09:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:13 crc kubenswrapper[4764]: I0202 09:09:13.476516 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:13 crc kubenswrapper[4764]: I0202 09:09:13.476598 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:13 crc kubenswrapper[4764]: I0202 09:09:13.476618 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:13 crc kubenswrapper[4764]: I0202 09:09:13.476640 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:13 crc kubenswrapper[4764]: I0202 09:09:13.476656 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:13Z","lastTransitionTime":"2026-02-02T09:09:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:13 crc kubenswrapper[4764]: I0202 09:09:13.581751 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:13 crc kubenswrapper[4764]: I0202 09:09:13.581810 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:13 crc kubenswrapper[4764]: I0202 09:09:13.581826 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:13 crc kubenswrapper[4764]: I0202 09:09:13.581852 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:13 crc kubenswrapper[4764]: I0202 09:09:13.581868 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:13Z","lastTransitionTime":"2026-02-02T09:09:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:13 crc kubenswrapper[4764]: I0202 09:09:13.684527 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:13 crc kubenswrapper[4764]: I0202 09:09:13.684579 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:13 crc kubenswrapper[4764]: I0202 09:09:13.684600 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:13 crc kubenswrapper[4764]: I0202 09:09:13.684631 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:13 crc kubenswrapper[4764]: I0202 09:09:13.684653 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:13Z","lastTransitionTime":"2026-02-02T09:09:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:13 crc kubenswrapper[4764]: I0202 09:09:13.789021 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:13 crc kubenswrapper[4764]: I0202 09:09:13.789075 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:13 crc kubenswrapper[4764]: I0202 09:09:13.789088 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:13 crc kubenswrapper[4764]: I0202 09:09:13.789105 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:13 crc kubenswrapper[4764]: I0202 09:09:13.789124 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:13Z","lastTransitionTime":"2026-02-02T09:09:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:13 crc kubenswrapper[4764]: I0202 09:09:13.825513 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kwtmr" Feb 02 09:09:13 crc kubenswrapper[4764]: E0202 09:09:13.825677 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kwtmr" podUID="991faa9a-dd25-4f49-82bd-ce60cefd4af2" Feb 02 09:09:13 crc kubenswrapper[4764]: I0202 09:09:13.853463 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-29 17:22:30.305758545 +0000 UTC Feb 02 09:09:13 crc kubenswrapper[4764]: I0202 09:09:13.891476 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:13 crc kubenswrapper[4764]: I0202 09:09:13.891532 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:13 crc kubenswrapper[4764]: I0202 09:09:13.891550 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:13 crc kubenswrapper[4764]: I0202 09:09:13.891567 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:13 crc kubenswrapper[4764]: I0202 09:09:13.891578 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:13Z","lastTransitionTime":"2026-02-02T09:09:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:13 crc kubenswrapper[4764]: I0202 09:09:13.993878 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:13 crc kubenswrapper[4764]: I0202 09:09:13.993927 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:13 crc kubenswrapper[4764]: I0202 09:09:13.993995 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:13 crc kubenswrapper[4764]: I0202 09:09:13.994021 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:13 crc kubenswrapper[4764]: I0202 09:09:13.994040 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:13Z","lastTransitionTime":"2026-02-02T09:09:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:14 crc kubenswrapper[4764]: I0202 09:09:14.097532 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:14 crc kubenswrapper[4764]: I0202 09:09:14.097609 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:14 crc kubenswrapper[4764]: I0202 09:09:14.097649 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:14 crc kubenswrapper[4764]: I0202 09:09:14.097681 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:14 crc kubenswrapper[4764]: I0202 09:09:14.097706 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:14Z","lastTransitionTime":"2026-02-02T09:09:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:14 crc kubenswrapper[4764]: I0202 09:09:14.200432 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:14 crc kubenswrapper[4764]: I0202 09:09:14.200460 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:14 crc kubenswrapper[4764]: I0202 09:09:14.200468 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:14 crc kubenswrapper[4764]: I0202 09:09:14.200479 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:14 crc kubenswrapper[4764]: I0202 09:09:14.200487 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:14Z","lastTransitionTime":"2026-02-02T09:09:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:14 crc kubenswrapper[4764]: I0202 09:09:14.302991 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:14 crc kubenswrapper[4764]: I0202 09:09:14.303039 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:14 crc kubenswrapper[4764]: I0202 09:09:14.303051 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:14 crc kubenswrapper[4764]: I0202 09:09:14.303068 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:14 crc kubenswrapper[4764]: I0202 09:09:14.303081 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:14Z","lastTransitionTime":"2026-02-02T09:09:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:14 crc kubenswrapper[4764]: I0202 09:09:14.398684 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-nc7qt_6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e/kube-multus/1.log" Feb 02 09:09:14 crc kubenswrapper[4764]: I0202 09:09:14.399389 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-nc7qt_6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e/kube-multus/0.log" Feb 02 09:09:14 crc kubenswrapper[4764]: I0202 09:09:14.399445 4764 generic.go:334] "Generic (PLEG): container finished" podID="6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e" containerID="b52b499d17b4e2e359475bd13730c72ef5cedb052e13def6dfe5bc5fd7487d2d" exitCode=1 Feb 02 09:09:14 crc kubenswrapper[4764]: I0202 09:09:14.399477 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-nc7qt" event={"ID":"6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e","Type":"ContainerDied","Data":"b52b499d17b4e2e359475bd13730c72ef5cedb052e13def6dfe5bc5fd7487d2d"} Feb 02 09:09:14 crc kubenswrapper[4764]: I0202 09:09:14.399514 4764 scope.go:117] "RemoveContainer" containerID="b1a2c7b22068b7165927fc158885d48bf9952b9eba521d8ae9a599eee53dc305" Feb 02 09:09:14 crc kubenswrapper[4764]: I0202 09:09:14.400061 4764 scope.go:117] "RemoveContainer" containerID="b52b499d17b4e2e359475bd13730c72ef5cedb052e13def6dfe5bc5fd7487d2d" Feb 02 09:09:14 crc kubenswrapper[4764]: E0202 09:09:14.400206 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-nc7qt_openshift-multus(6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e)\"" pod="openshift-multus/multus-nc7qt" podUID="6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e" Feb 02 09:09:14 crc kubenswrapper[4764]: I0202 09:09:14.405476 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:14 crc kubenswrapper[4764]: I0202 09:09:14.405560 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:14 crc kubenswrapper[4764]: I0202 09:09:14.405574 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:14 crc kubenswrapper[4764]: I0202 09:09:14.405627 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:14 crc kubenswrapper[4764]: I0202 09:09:14.405646 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:14Z","lastTransitionTime":"2026-02-02T09:09:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:14 crc kubenswrapper[4764]: I0202 09:09:14.497267 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-9p5dc" podStartSLOduration=96.497238229 podStartE2EDuration="1m36.497238229s" podCreationTimestamp="2026-02-02 09:07:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:09:14.484243132 +0000 UTC m=+117.417967270" watchObservedRunningTime="2026-02-02 09:09:14.497238229 +0000 UTC m=+117.430962357" Feb 02 09:09:14 crc kubenswrapper[4764]: I0202 09:09:14.507969 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:14 crc kubenswrapper[4764]: I0202 09:09:14.508033 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:14 crc kubenswrapper[4764]: I0202 09:09:14.508059 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:14 crc kubenswrapper[4764]: I0202 09:09:14.508093 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:14 crc kubenswrapper[4764]: I0202 09:09:14.508116 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:14Z","lastTransitionTime":"2026-02-02T09:09:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:14 crc kubenswrapper[4764]: I0202 09:09:14.525120 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=95.525092994 podStartE2EDuration="1m35.525092994s" podCreationTimestamp="2026-02-02 09:07:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:09:14.512502418 +0000 UTC m=+117.446226506" watchObservedRunningTime="2026-02-02 09:09:14.525092994 +0000 UTC m=+117.458817152" Feb 02 09:09:14 crc kubenswrapper[4764]: I0202 09:09:14.542741 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=44.542720268 podStartE2EDuration="44.542720268s" podCreationTimestamp="2026-02-02 09:08:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:09:14.525567147 +0000 UTC m=+117.459291235" watchObservedRunningTime="2026-02-02 09:09:14.542720268 +0000 UTC m=+117.476444356" Feb 02 09:09:14 crc kubenswrapper[4764]: I0202 09:09:14.590572 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=95.590554462 podStartE2EDuration="1m35.590554462s" podCreationTimestamp="2026-02-02 09:07:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:09:14.589243326 +0000 UTC m=+117.522967404" watchObservedRunningTime="2026-02-02 09:09:14.590554462 +0000 UTC m=+117.524278570" Feb 02 09:09:14 crc kubenswrapper[4764]: I0202 09:09:14.609876 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:14 crc kubenswrapper[4764]: I0202 09:09:14.609904 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:14 crc kubenswrapper[4764]: I0202 09:09:14.609912 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:14 crc kubenswrapper[4764]: I0202 09:09:14.609924 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:14 crc kubenswrapper[4764]: I0202 09:09:14.609950 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:14Z","lastTransitionTime":"2026-02-02T09:09:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:14 crc kubenswrapper[4764]: I0202 09:09:14.632645 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podStartSLOduration=96.632623208 podStartE2EDuration="1m36.632623208s" podCreationTimestamp="2026-02-02 09:07:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:09:14.632351541 +0000 UTC m=+117.566075629" watchObservedRunningTime="2026-02-02 09:09:14.632623208 +0000 UTC m=+117.566347296" Feb 02 09:09:14 crc kubenswrapper[4764]: I0202 09:09:14.632828 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-6zxff" podStartSLOduration=96.632825194 podStartE2EDuration="1m36.632825194s" podCreationTimestamp="2026-02-02 09:07:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:09:14.618260263 +0000 UTC m=+117.551984361" watchObservedRunningTime="2026-02-02 09:09:14.632825194 +0000 UTC m=+117.566549282" Feb 02 09:09:14 crc kubenswrapper[4764]: I0202 09:09:14.642228 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-pcjmn" podStartSLOduration=96.642212082 podStartE2EDuration="1m36.642212082s" podCreationTimestamp="2026-02-02 09:07:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:09:14.641772589 +0000 UTC m=+117.575496697" watchObservedRunningTime="2026-02-02 09:09:14.642212082 +0000 UTC m=+117.575936170" Feb 02 09:09:14 crc kubenswrapper[4764]: I0202 09:09:14.654878 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=92.654859399 podStartE2EDuration="1m32.654859399s" podCreationTimestamp="2026-02-02 09:07:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:09:14.654669294 +0000 UTC m=+117.588393382" watchObservedRunningTime="2026-02-02 09:09:14.654859399 +0000 UTC m=+117.588583497" Feb 02 09:09:14 crc kubenswrapper[4764]: I0202 09:09:14.676561 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lh6tm" podStartSLOduration=95.676545735 podStartE2EDuration="1m35.676545735s" podCreationTimestamp="2026-02-02 09:07:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:09:14.676332299 +0000 UTC m=+117.610056387" watchObservedRunningTime="2026-02-02 09:09:14.676545735 +0000 UTC m=+117.610269823" Feb 02 09:09:14 crc kubenswrapper[4764]: I0202 09:09:14.704966 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=66.704949316 podStartE2EDuration="1m6.704949316s" podCreationTimestamp="2026-02-02 09:08:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:09:14.69382983 +0000 UTC m=+117.627553928" watchObservedRunningTime="2026-02-02 09:09:14.704949316 +0000 UTC m=+117.638673404" Feb 02 09:09:14 crc kubenswrapper[4764]: I0202 09:09:14.712008 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:14 crc kubenswrapper[4764]: I0202 09:09:14.712041 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:14 crc kubenswrapper[4764]: I0202 09:09:14.712049 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:14 crc kubenswrapper[4764]: I0202 09:09:14.712061 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:14 crc kubenswrapper[4764]: I0202 09:09:14.712069 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:14Z","lastTransitionTime":"2026-02-02T09:09:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:14 crc kubenswrapper[4764]: I0202 09:09:14.813753 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:14 crc kubenswrapper[4764]: I0202 09:09:14.813790 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:14 crc kubenswrapper[4764]: I0202 09:09:14.813798 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:14 crc kubenswrapper[4764]: I0202 09:09:14.813811 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:14 crc kubenswrapper[4764]: I0202 09:09:14.813820 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:14Z","lastTransitionTime":"2026-02-02T09:09:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:14 crc kubenswrapper[4764]: I0202 09:09:14.824977 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 09:09:14 crc kubenswrapper[4764]: I0202 09:09:14.824997 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 09:09:14 crc kubenswrapper[4764]: I0202 09:09:14.825026 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 09:09:14 crc kubenswrapper[4764]: E0202 09:09:14.825099 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 09:09:14 crc kubenswrapper[4764]: E0202 09:09:14.825150 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 09:09:14 crc kubenswrapper[4764]: E0202 09:09:14.825217 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 09:09:14 crc kubenswrapper[4764]: I0202 09:09:14.854598 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-07 06:16:45.019861199 +0000 UTC Feb 02 09:09:14 crc kubenswrapper[4764]: I0202 09:09:14.916357 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:14 crc kubenswrapper[4764]: I0202 09:09:14.916380 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:14 crc kubenswrapper[4764]: I0202 09:09:14.916388 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:14 crc kubenswrapper[4764]: I0202 09:09:14.916400 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:14 crc kubenswrapper[4764]: I0202 09:09:14.916409 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:14Z","lastTransitionTime":"2026-02-02T09:09:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:15 crc kubenswrapper[4764]: I0202 09:09:15.018519 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:15 crc kubenswrapper[4764]: I0202 09:09:15.018578 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:15 crc kubenswrapper[4764]: I0202 09:09:15.018595 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:15 crc kubenswrapper[4764]: I0202 09:09:15.018617 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:15 crc kubenswrapper[4764]: I0202 09:09:15.018632 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:15Z","lastTransitionTime":"2026-02-02T09:09:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:15 crc kubenswrapper[4764]: I0202 09:09:15.121030 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:15 crc kubenswrapper[4764]: I0202 09:09:15.121074 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:15 crc kubenswrapper[4764]: I0202 09:09:15.121085 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:15 crc kubenswrapper[4764]: I0202 09:09:15.121100 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:15 crc kubenswrapper[4764]: I0202 09:09:15.121112 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:15Z","lastTransitionTime":"2026-02-02T09:09:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:15 crc kubenswrapper[4764]: I0202 09:09:15.223770 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:15 crc kubenswrapper[4764]: I0202 09:09:15.223818 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:15 crc kubenswrapper[4764]: I0202 09:09:15.223828 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:15 crc kubenswrapper[4764]: I0202 09:09:15.223842 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:15 crc kubenswrapper[4764]: I0202 09:09:15.223853 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:15Z","lastTransitionTime":"2026-02-02T09:09:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:15 crc kubenswrapper[4764]: I0202 09:09:15.326140 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:15 crc kubenswrapper[4764]: I0202 09:09:15.326220 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:15 crc kubenswrapper[4764]: I0202 09:09:15.326243 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:15 crc kubenswrapper[4764]: I0202 09:09:15.326275 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:15 crc kubenswrapper[4764]: I0202 09:09:15.326303 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:15Z","lastTransitionTime":"2026-02-02T09:09:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:15 crc kubenswrapper[4764]: I0202 09:09:15.403970 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-nc7qt_6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e/kube-multus/1.log" Feb 02 09:09:15 crc kubenswrapper[4764]: I0202 09:09:15.429156 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:15 crc kubenswrapper[4764]: I0202 09:09:15.429224 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:15 crc kubenswrapper[4764]: I0202 09:09:15.429248 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:15 crc kubenswrapper[4764]: I0202 09:09:15.429277 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:15 crc kubenswrapper[4764]: I0202 09:09:15.429301 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:15Z","lastTransitionTime":"2026-02-02T09:09:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:15 crc kubenswrapper[4764]: I0202 09:09:15.532454 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:15 crc kubenswrapper[4764]: I0202 09:09:15.532497 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:15 crc kubenswrapper[4764]: I0202 09:09:15.532508 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:15 crc kubenswrapper[4764]: I0202 09:09:15.532525 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:15 crc kubenswrapper[4764]: I0202 09:09:15.532534 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:15Z","lastTransitionTime":"2026-02-02T09:09:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:15 crc kubenswrapper[4764]: I0202 09:09:15.635312 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:15 crc kubenswrapper[4764]: I0202 09:09:15.635364 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:15 crc kubenswrapper[4764]: I0202 09:09:15.635380 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:15 crc kubenswrapper[4764]: I0202 09:09:15.635402 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:15 crc kubenswrapper[4764]: I0202 09:09:15.635418 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:15Z","lastTransitionTime":"2026-02-02T09:09:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:15 crc kubenswrapper[4764]: I0202 09:09:15.738108 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:15 crc kubenswrapper[4764]: I0202 09:09:15.738140 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:15 crc kubenswrapper[4764]: I0202 09:09:15.738148 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:15 crc kubenswrapper[4764]: I0202 09:09:15.738161 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:15 crc kubenswrapper[4764]: I0202 09:09:15.738184 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:15Z","lastTransitionTime":"2026-02-02T09:09:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:15 crc kubenswrapper[4764]: I0202 09:09:15.824817 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kwtmr" Feb 02 09:09:15 crc kubenswrapper[4764]: E0202 09:09:15.824993 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kwtmr" podUID="991faa9a-dd25-4f49-82bd-ce60cefd4af2" Feb 02 09:09:15 crc kubenswrapper[4764]: I0202 09:09:15.840772 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:15 crc kubenswrapper[4764]: I0202 09:09:15.840854 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:15 crc kubenswrapper[4764]: I0202 09:09:15.840877 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:15 crc kubenswrapper[4764]: I0202 09:09:15.840907 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:15 crc kubenswrapper[4764]: I0202 09:09:15.840927 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:15Z","lastTransitionTime":"2026-02-02T09:09:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:15 crc kubenswrapper[4764]: I0202 09:09:15.855264 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-28 22:26:50.166617997 +0000 UTC Feb 02 09:09:15 crc kubenswrapper[4764]: I0202 09:09:15.943335 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:15 crc kubenswrapper[4764]: I0202 09:09:15.943375 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:15 crc kubenswrapper[4764]: I0202 09:09:15.943383 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:15 crc kubenswrapper[4764]: I0202 09:09:15.943396 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:15 crc kubenswrapper[4764]: I0202 09:09:15.943405 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:15Z","lastTransitionTime":"2026-02-02T09:09:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:16 crc kubenswrapper[4764]: I0202 09:09:16.045530 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:16 crc kubenswrapper[4764]: I0202 09:09:16.045590 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:16 crc kubenswrapper[4764]: I0202 09:09:16.045612 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:16 crc kubenswrapper[4764]: I0202 09:09:16.045639 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:16 crc kubenswrapper[4764]: I0202 09:09:16.045662 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:16Z","lastTransitionTime":"2026-02-02T09:09:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:16 crc kubenswrapper[4764]: I0202 09:09:16.148801 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:16 crc kubenswrapper[4764]: I0202 09:09:16.148873 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:16 crc kubenswrapper[4764]: I0202 09:09:16.148898 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:16 crc kubenswrapper[4764]: I0202 09:09:16.148927 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:16 crc kubenswrapper[4764]: I0202 09:09:16.148996 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:16Z","lastTransitionTime":"2026-02-02T09:09:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:16 crc kubenswrapper[4764]: I0202 09:09:16.251309 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:16 crc kubenswrapper[4764]: I0202 09:09:16.251400 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:16 crc kubenswrapper[4764]: I0202 09:09:16.251414 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:16 crc kubenswrapper[4764]: I0202 09:09:16.251429 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:16 crc kubenswrapper[4764]: I0202 09:09:16.251440 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:16Z","lastTransitionTime":"2026-02-02T09:09:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:16 crc kubenswrapper[4764]: I0202 09:09:16.354004 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:16 crc kubenswrapper[4764]: I0202 09:09:16.354057 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:16 crc kubenswrapper[4764]: I0202 09:09:16.354068 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:16 crc kubenswrapper[4764]: I0202 09:09:16.354085 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:16 crc kubenswrapper[4764]: I0202 09:09:16.354096 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:16Z","lastTransitionTime":"2026-02-02T09:09:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:16 crc kubenswrapper[4764]: I0202 09:09:16.457369 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:16 crc kubenswrapper[4764]: I0202 09:09:16.457495 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:16 crc kubenswrapper[4764]: I0202 09:09:16.457512 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:16 crc kubenswrapper[4764]: I0202 09:09:16.457536 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:16 crc kubenswrapper[4764]: I0202 09:09:16.457553 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:16Z","lastTransitionTime":"2026-02-02T09:09:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:16 crc kubenswrapper[4764]: I0202 09:09:16.560506 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:16 crc kubenswrapper[4764]: I0202 09:09:16.560553 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:16 crc kubenswrapper[4764]: I0202 09:09:16.560574 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:16 crc kubenswrapper[4764]: I0202 09:09:16.560599 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:16 crc kubenswrapper[4764]: I0202 09:09:16.560616 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:16Z","lastTransitionTime":"2026-02-02T09:09:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:16 crc kubenswrapper[4764]: I0202 09:09:16.662672 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:16 crc kubenswrapper[4764]: I0202 09:09:16.662720 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:16 crc kubenswrapper[4764]: I0202 09:09:16.662731 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:16 crc kubenswrapper[4764]: I0202 09:09:16.662747 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:16 crc kubenswrapper[4764]: I0202 09:09:16.662758 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:16Z","lastTransitionTime":"2026-02-02T09:09:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:16 crc kubenswrapper[4764]: I0202 09:09:16.765180 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:16 crc kubenswrapper[4764]: I0202 09:09:16.765241 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:16 crc kubenswrapper[4764]: I0202 09:09:16.765258 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:16 crc kubenswrapper[4764]: I0202 09:09:16.765281 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:16 crc kubenswrapper[4764]: I0202 09:09:16.765298 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:16Z","lastTransitionTime":"2026-02-02T09:09:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:16 crc kubenswrapper[4764]: I0202 09:09:16.824834 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 09:09:16 crc kubenswrapper[4764]: I0202 09:09:16.824840 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 09:09:16 crc kubenswrapper[4764]: I0202 09:09:16.824866 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 09:09:16 crc kubenswrapper[4764]: E0202 09:09:16.825125 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 09:09:16 crc kubenswrapper[4764]: E0202 09:09:16.825302 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 09:09:16 crc kubenswrapper[4764]: E0202 09:09:16.825403 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 09:09:16 crc kubenswrapper[4764]: I0202 09:09:16.855616 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-24 20:14:58.345507417 +0000 UTC Feb 02 09:09:16 crc kubenswrapper[4764]: I0202 09:09:16.868182 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:16 crc kubenswrapper[4764]: I0202 09:09:16.868216 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:16 crc kubenswrapper[4764]: I0202 09:09:16.868228 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:16 crc kubenswrapper[4764]: I0202 09:09:16.868243 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:16 crc kubenswrapper[4764]: I0202 09:09:16.868255 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:16Z","lastTransitionTime":"2026-02-02T09:09:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:16 crc kubenswrapper[4764]: I0202 09:09:16.970565 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:16 crc kubenswrapper[4764]: I0202 09:09:16.971033 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:16 crc kubenswrapper[4764]: I0202 09:09:16.971188 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:16 crc kubenswrapper[4764]: I0202 09:09:16.971356 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:16 crc kubenswrapper[4764]: I0202 09:09:16.971811 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:16Z","lastTransitionTime":"2026-02-02T09:09:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:17 crc kubenswrapper[4764]: I0202 09:09:17.075057 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:17 crc kubenswrapper[4764]: I0202 09:09:17.075150 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:17 crc kubenswrapper[4764]: I0202 09:09:17.075173 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:17 crc kubenswrapper[4764]: I0202 09:09:17.075200 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:17 crc kubenswrapper[4764]: I0202 09:09:17.075221 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:17Z","lastTransitionTime":"2026-02-02T09:09:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:17 crc kubenswrapper[4764]: I0202 09:09:17.178878 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:17 crc kubenswrapper[4764]: I0202 09:09:17.178977 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:17 crc kubenswrapper[4764]: I0202 09:09:17.178998 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:17 crc kubenswrapper[4764]: I0202 09:09:17.179023 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:17 crc kubenswrapper[4764]: I0202 09:09:17.179044 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:17Z","lastTransitionTime":"2026-02-02T09:09:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:17 crc kubenswrapper[4764]: I0202 09:09:17.282069 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:17 crc kubenswrapper[4764]: I0202 09:09:17.282132 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:17 crc kubenswrapper[4764]: I0202 09:09:17.282155 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:17 crc kubenswrapper[4764]: I0202 09:09:17.282184 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:17 crc kubenswrapper[4764]: I0202 09:09:17.282205 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:17Z","lastTransitionTime":"2026-02-02T09:09:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:17 crc kubenswrapper[4764]: I0202 09:09:17.386271 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:17 crc kubenswrapper[4764]: I0202 09:09:17.386323 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:17 crc kubenswrapper[4764]: I0202 09:09:17.386334 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:17 crc kubenswrapper[4764]: I0202 09:09:17.386355 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:17 crc kubenswrapper[4764]: I0202 09:09:17.386366 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:17Z","lastTransitionTime":"2026-02-02T09:09:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:17 crc kubenswrapper[4764]: I0202 09:09:17.446024 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 09:09:17 crc kubenswrapper[4764]: I0202 09:09:17.446452 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 09:09:17 crc kubenswrapper[4764]: I0202 09:09:17.446650 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 09:09:17 crc kubenswrapper[4764]: I0202 09:09:17.446894 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 09:09:17 crc kubenswrapper[4764]: I0202 09:09:17.447181 4764 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T09:09:17Z","lastTransitionTime":"2026-02-02T09:09:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 09:09:17 crc kubenswrapper[4764]: I0202 09:09:17.496112 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-mdsp2"] Feb 02 09:09:17 crc kubenswrapper[4764]: I0202 09:09:17.496924 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mdsp2" Feb 02 09:09:17 crc kubenswrapper[4764]: I0202 09:09:17.498839 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Feb 02 09:09:17 crc kubenswrapper[4764]: I0202 09:09:17.499010 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Feb 02 09:09:17 crc kubenswrapper[4764]: I0202 09:09:17.498901 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Feb 02 09:09:17 crc kubenswrapper[4764]: I0202 09:09:17.501400 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Feb 02 09:09:17 crc kubenswrapper[4764]: I0202 09:09:17.566591 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/373124b6-c842-47ab-8bdf-9a02e594c009-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-mdsp2\" (UID: \"373124b6-c842-47ab-8bdf-9a02e594c009\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mdsp2" Feb 02 09:09:17 crc kubenswrapper[4764]: I0202 09:09:17.566632 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/373124b6-c842-47ab-8bdf-9a02e594c009-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-mdsp2\" (UID: \"373124b6-c842-47ab-8bdf-9a02e594c009\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mdsp2" Feb 02 09:09:17 crc kubenswrapper[4764]: I0202 09:09:17.566682 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/373124b6-c842-47ab-8bdf-9a02e594c009-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-mdsp2\" (UID: \"373124b6-c842-47ab-8bdf-9a02e594c009\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mdsp2" Feb 02 09:09:17 crc kubenswrapper[4764]: I0202 09:09:17.566714 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/373124b6-c842-47ab-8bdf-9a02e594c009-service-ca\") pod \"cluster-version-operator-5c965bbfc6-mdsp2\" (UID: \"373124b6-c842-47ab-8bdf-9a02e594c009\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mdsp2" Feb 02 09:09:17 crc kubenswrapper[4764]: I0202 09:09:17.566753 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/373124b6-c842-47ab-8bdf-9a02e594c009-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-mdsp2\" (UID: \"373124b6-c842-47ab-8bdf-9a02e594c009\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mdsp2" Feb 02 09:09:17 crc kubenswrapper[4764]: I0202 09:09:17.667549 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/373124b6-c842-47ab-8bdf-9a02e594c009-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-mdsp2\" (UID: \"373124b6-c842-47ab-8bdf-9a02e594c009\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mdsp2" Feb 02 09:09:17 crc kubenswrapper[4764]: I0202 09:09:17.667843 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/373124b6-c842-47ab-8bdf-9a02e594c009-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-mdsp2\" (UID: \"373124b6-c842-47ab-8bdf-9a02e594c009\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mdsp2" Feb 02 09:09:17 crc kubenswrapper[4764]: I0202 09:09:17.668001 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/373124b6-c842-47ab-8bdf-9a02e594c009-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-mdsp2\" (UID: \"373124b6-c842-47ab-8bdf-9a02e594c009\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mdsp2" Feb 02 09:09:17 crc kubenswrapper[4764]: I0202 09:09:17.668200 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/373124b6-c842-47ab-8bdf-9a02e594c009-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-mdsp2\" (UID: \"373124b6-c842-47ab-8bdf-9a02e594c009\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mdsp2" Feb 02 09:09:17 crc kubenswrapper[4764]: I0202 09:09:17.668385 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/373124b6-c842-47ab-8bdf-9a02e594c009-service-ca\") pod \"cluster-version-operator-5c965bbfc6-mdsp2\" (UID: \"373124b6-c842-47ab-8bdf-9a02e594c009\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mdsp2" Feb 02 09:09:17 crc kubenswrapper[4764]: I0202 09:09:17.668328 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/373124b6-c842-47ab-8bdf-9a02e594c009-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-mdsp2\" (UID: \"373124b6-c842-47ab-8bdf-9a02e594c009\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mdsp2" Feb 02 09:09:17 crc kubenswrapper[4764]: I0202 09:09:17.668138 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/373124b6-c842-47ab-8bdf-9a02e594c009-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-mdsp2\" (UID: \"373124b6-c842-47ab-8bdf-9a02e594c009\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mdsp2" Feb 02 09:09:17 crc kubenswrapper[4764]: I0202 09:09:17.669518 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/373124b6-c842-47ab-8bdf-9a02e594c009-service-ca\") pod \"cluster-version-operator-5c965bbfc6-mdsp2\" (UID: \"373124b6-c842-47ab-8bdf-9a02e594c009\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mdsp2" Feb 02 09:09:17 crc kubenswrapper[4764]: I0202 09:09:17.673154 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/373124b6-c842-47ab-8bdf-9a02e594c009-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-mdsp2\" (UID: \"373124b6-c842-47ab-8bdf-9a02e594c009\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mdsp2" Feb 02 09:09:17 crc kubenswrapper[4764]: I0202 09:09:17.686673 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/373124b6-c842-47ab-8bdf-9a02e594c009-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-mdsp2\" (UID: \"373124b6-c842-47ab-8bdf-9a02e594c009\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mdsp2" Feb 02 09:09:17 crc kubenswrapper[4764]: E0202 09:09:17.793579 4764 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Feb 02 09:09:17 crc kubenswrapper[4764]: I0202 09:09:17.812926 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Feb 02 09:09:17 crc kubenswrapper[4764]: I0202 09:09:17.821762 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mdsp2" Feb 02 09:09:17 crc kubenswrapper[4764]: I0202 09:09:17.825122 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kwtmr" Feb 02 09:09:17 crc kubenswrapper[4764]: E0202 09:09:17.826325 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kwtmr" podUID="991faa9a-dd25-4f49-82bd-ce60cefd4af2" Feb 02 09:09:17 crc kubenswrapper[4764]: W0202 09:09:17.844584 4764 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod373124b6_c842_47ab_8bdf_9a02e594c009.slice/crio-0da536c1f0c8f45a403660ef8e19926dde2108549d41c52155f0a773823630c0 WatchSource:0}: Error finding container 0da536c1f0c8f45a403660ef8e19926dde2108549d41c52155f0a773823630c0: Status 404 returned error can't find the container with id 0da536c1f0c8f45a403660ef8e19926dde2108549d41c52155f0a773823630c0 Feb 02 09:09:17 crc kubenswrapper[4764]: I0202 09:09:17.856488 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-26 01:20:06.781827653 +0000 UTC Feb 02 09:09:17 crc kubenswrapper[4764]: I0202 09:09:17.856798 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Rotating certificates Feb 02 09:09:17 crc kubenswrapper[4764]: I0202 09:09:17.865116 4764 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Feb 02 09:09:17 crc kubenswrapper[4764]: E0202 09:09:17.911415 4764 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Feb 02 09:09:18 crc kubenswrapper[4764]: I0202 09:09:18.417919 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mdsp2" event={"ID":"373124b6-c842-47ab-8bdf-9a02e594c009","Type":"ContainerStarted","Data":"9b188e8fc31054123983c1ff38e7630d7b7c832ac3458f74efc2ebe9761a67ac"} Feb 02 09:09:18 crc kubenswrapper[4764]: I0202 09:09:18.418006 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mdsp2" event={"ID":"373124b6-c842-47ab-8bdf-9a02e594c009","Type":"ContainerStarted","Data":"0da536c1f0c8f45a403660ef8e19926dde2108549d41c52155f0a773823630c0"} Feb 02 09:09:18 crc kubenswrapper[4764]: I0202 09:09:18.825467 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 09:09:18 crc kubenswrapper[4764]: I0202 09:09:18.825500 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 09:09:18 crc kubenswrapper[4764]: I0202 09:09:18.825588 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 09:09:18 crc kubenswrapper[4764]: E0202 09:09:18.825732 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 09:09:18 crc kubenswrapper[4764]: E0202 09:09:18.826046 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 09:09:18 crc kubenswrapper[4764]: E0202 09:09:18.826590 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 09:09:18 crc kubenswrapper[4764]: I0202 09:09:18.826776 4764 scope.go:117] "RemoveContainer" containerID="3cb255e14eb02f7c08364c41fb9906e58b2155eb507f694b77ec8eb1f4251343" Feb 02 09:09:19 crc kubenswrapper[4764]: I0202 09:09:19.423860 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zhn7j_24632dda-6100-4ab6-a28e-214ddae4360c/ovnkube-controller/3.log" Feb 02 09:09:19 crc kubenswrapper[4764]: I0202 09:09:19.426981 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" event={"ID":"24632dda-6100-4ab6-a28e-214ddae4360c","Type":"ContainerStarted","Data":"39395584eacf319103cbe03b647c7c7672646b5619a631dffd7f115435591187"} Feb 02 09:09:19 crc kubenswrapper[4764]: I0202 09:09:19.427453 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" Feb 02 09:09:19 crc kubenswrapper[4764]: I0202 09:09:19.459075 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" podStartSLOduration=101.459056215 podStartE2EDuration="1m41.459056215s" podCreationTimestamp="2026-02-02 09:07:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:09:19.458092659 +0000 UTC m=+122.391816767" watchObservedRunningTime="2026-02-02 09:09:19.459056215 +0000 UTC m=+122.392780313" Feb 02 09:09:19 crc kubenswrapper[4764]: I0202 09:09:19.460209 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mdsp2" podStartSLOduration=101.460197867 podStartE2EDuration="1m41.460197867s" podCreationTimestamp="2026-02-02 09:07:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:09:18.430924823 +0000 UTC m=+121.364648901" watchObservedRunningTime="2026-02-02 09:09:19.460197867 +0000 UTC m=+122.393921965" Feb 02 09:09:19 crc kubenswrapper[4764]: I0202 09:09:19.746162 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-kwtmr"] Feb 02 09:09:19 crc kubenswrapper[4764]: I0202 09:09:19.746325 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kwtmr" Feb 02 09:09:19 crc kubenswrapper[4764]: E0202 09:09:19.746440 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kwtmr" podUID="991faa9a-dd25-4f49-82bd-ce60cefd4af2" Feb 02 09:09:20 crc kubenswrapper[4764]: I0202 09:09:20.825370 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 09:09:20 crc kubenswrapper[4764]: I0202 09:09:20.825408 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 09:09:20 crc kubenswrapper[4764]: E0202 09:09:20.825749 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 09:09:20 crc kubenswrapper[4764]: E0202 09:09:20.825887 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 09:09:20 crc kubenswrapper[4764]: I0202 09:09:20.825408 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 09:09:20 crc kubenswrapper[4764]: E0202 09:09:20.826010 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 09:09:21 crc kubenswrapper[4764]: I0202 09:09:21.824734 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kwtmr" Feb 02 09:09:21 crc kubenswrapper[4764]: E0202 09:09:21.824907 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kwtmr" podUID="991faa9a-dd25-4f49-82bd-ce60cefd4af2" Feb 02 09:09:22 crc kubenswrapper[4764]: I0202 09:09:22.825343 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 09:09:22 crc kubenswrapper[4764]: I0202 09:09:22.825415 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 09:09:22 crc kubenswrapper[4764]: I0202 09:09:22.825343 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 09:09:22 crc kubenswrapper[4764]: E0202 09:09:22.825475 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 09:09:22 crc kubenswrapper[4764]: E0202 09:09:22.825522 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 09:09:22 crc kubenswrapper[4764]: E0202 09:09:22.825576 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 09:09:22 crc kubenswrapper[4764]: E0202 09:09:22.912621 4764 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Feb 02 09:09:23 crc kubenswrapper[4764]: I0202 09:09:23.824871 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kwtmr" Feb 02 09:09:23 crc kubenswrapper[4764]: E0202 09:09:23.825118 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kwtmr" podUID="991faa9a-dd25-4f49-82bd-ce60cefd4af2" Feb 02 09:09:24 crc kubenswrapper[4764]: I0202 09:09:24.824997 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 09:09:24 crc kubenswrapper[4764]: I0202 09:09:24.825084 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 09:09:24 crc kubenswrapper[4764]: E0202 09:09:24.825393 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 09:09:24 crc kubenswrapper[4764]: E0202 09:09:24.825331 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 09:09:24 crc kubenswrapper[4764]: I0202 09:09:24.825111 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 09:09:24 crc kubenswrapper[4764]: E0202 09:09:24.825467 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 09:09:24 crc kubenswrapper[4764]: I0202 09:09:24.841567 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" Feb 02 09:09:25 crc kubenswrapper[4764]: I0202 09:09:25.824671 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kwtmr" Feb 02 09:09:25 crc kubenswrapper[4764]: E0202 09:09:25.824828 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kwtmr" podUID="991faa9a-dd25-4f49-82bd-ce60cefd4af2" Feb 02 09:09:26 crc kubenswrapper[4764]: I0202 09:09:26.825097 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 09:09:26 crc kubenswrapper[4764]: I0202 09:09:26.825245 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 09:09:26 crc kubenswrapper[4764]: I0202 09:09:26.825266 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 09:09:26 crc kubenswrapper[4764]: I0202 09:09:26.825671 4764 scope.go:117] "RemoveContainer" containerID="b52b499d17b4e2e359475bd13730c72ef5cedb052e13def6dfe5bc5fd7487d2d" Feb 02 09:09:26 crc kubenswrapper[4764]: E0202 09:09:26.826294 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 09:09:26 crc kubenswrapper[4764]: E0202 09:09:26.826356 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 09:09:26 crc kubenswrapper[4764]: E0202 09:09:26.826378 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 09:09:27 crc kubenswrapper[4764]: I0202 09:09:27.453023 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-nc7qt_6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e/kube-multus/1.log" Feb 02 09:09:27 crc kubenswrapper[4764]: I0202 09:09:27.453343 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-nc7qt" event={"ID":"6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e","Type":"ContainerStarted","Data":"b7882e3ed0730ea5035949695b7b94a50060918bee89edc807fd67484b957ee5"} Feb 02 09:09:27 crc kubenswrapper[4764]: I0202 09:09:27.825083 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kwtmr" Feb 02 09:09:27 crc kubenswrapper[4764]: E0202 09:09:27.826378 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kwtmr" podUID="991faa9a-dd25-4f49-82bd-ce60cefd4af2" Feb 02 09:09:27 crc kubenswrapper[4764]: E0202 09:09:27.913195 4764 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Feb 02 09:09:28 crc kubenswrapper[4764]: I0202 09:09:28.825009 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 09:09:28 crc kubenswrapper[4764]: I0202 09:09:28.825074 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 09:09:28 crc kubenswrapper[4764]: E0202 09:09:28.825133 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 09:09:28 crc kubenswrapper[4764]: I0202 09:09:28.825015 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 09:09:28 crc kubenswrapper[4764]: E0202 09:09:28.825253 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 09:09:28 crc kubenswrapper[4764]: E0202 09:09:28.825316 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 09:09:29 crc kubenswrapper[4764]: I0202 09:09:29.825477 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kwtmr" Feb 02 09:09:29 crc kubenswrapper[4764]: E0202 09:09:29.825618 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kwtmr" podUID="991faa9a-dd25-4f49-82bd-ce60cefd4af2" Feb 02 09:09:30 crc kubenswrapper[4764]: I0202 09:09:30.825390 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 09:09:30 crc kubenswrapper[4764]: I0202 09:09:30.825470 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 09:09:30 crc kubenswrapper[4764]: E0202 09:09:30.825496 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 09:09:30 crc kubenswrapper[4764]: I0202 09:09:30.825502 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 09:09:30 crc kubenswrapper[4764]: E0202 09:09:30.825605 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 09:09:30 crc kubenswrapper[4764]: E0202 09:09:30.825687 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 09:09:31 crc kubenswrapper[4764]: I0202 09:09:31.824946 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kwtmr" Feb 02 09:09:31 crc kubenswrapper[4764]: E0202 09:09:31.825100 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kwtmr" podUID="991faa9a-dd25-4f49-82bd-ce60cefd4af2" Feb 02 09:09:32 crc kubenswrapper[4764]: I0202 09:09:32.824545 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 09:09:32 crc kubenswrapper[4764]: I0202 09:09:32.824588 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 09:09:32 crc kubenswrapper[4764]: I0202 09:09:32.824547 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 09:09:32 crc kubenswrapper[4764]: E0202 09:09:32.824679 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 09:09:32 crc kubenswrapper[4764]: E0202 09:09:32.825159 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 09:09:32 crc kubenswrapper[4764]: E0202 09:09:32.825254 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 09:09:33 crc kubenswrapper[4764]: I0202 09:09:33.825388 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kwtmr" Feb 02 09:09:33 crc kubenswrapper[4764]: I0202 09:09:33.827898 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Feb 02 09:09:33 crc kubenswrapper[4764]: I0202 09:09:33.829089 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Feb 02 09:09:34 crc kubenswrapper[4764]: I0202 09:09:34.824774 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 09:09:34 crc kubenswrapper[4764]: I0202 09:09:34.824872 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 09:09:34 crc kubenswrapper[4764]: I0202 09:09:34.827309 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 09:09:34 crc kubenswrapper[4764]: I0202 09:09:34.829770 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Feb 02 09:09:34 crc kubenswrapper[4764]: I0202 09:09:34.830106 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Feb 02 09:09:34 crc kubenswrapper[4764]: I0202 09:09:34.830366 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Feb 02 09:09:34 crc kubenswrapper[4764]: I0202 09:09:34.834257 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.129125 4764 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.181219 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-nc7qt" podStartSLOduration=120.181192704 podStartE2EDuration="2m0.181192704s" podCreationTimestamp="2026-02-02 09:07:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:09:27.479256925 +0000 UTC m=+130.412981013" watchObservedRunningTime="2026-02-02 09:09:38.181192704 +0000 UTC m=+141.114916832" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.182621 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-966nh"] Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.183279 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-966nh" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.191989 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-h5vrh"] Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.192726 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-h5vrh" Feb 02 09:09:38 crc kubenswrapper[4764]: W0202 09:09:38.194175 4764 reflector.go:561] object-"openshift-controller-manager"/"config": failed to list *v1.ConfigMap: configmaps "config" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-controller-manager": no relationship found between node 'crc' and this object Feb 02 09:09:38 crc kubenswrapper[4764]: E0202 09:09:38.194237 4764 reflector.go:158] "Unhandled Error" err="object-\"openshift-controller-manager\"/\"config\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"config\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-controller-manager\": no relationship found between node 'crc' and this object" logger="UnhandledError" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.194595 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.195150 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.195187 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.195346 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.195382 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.197157 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-98wxj"] Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.197888 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-98wxj" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.202453 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.202963 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.203096 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.203565 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.203716 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.203846 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.203995 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.204088 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.204535 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.204802 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.205173 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.205413 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-zh65k"] Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.206173 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-zh65k" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.206479 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-7k5qb"] Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.207134 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-7k5qb" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.207645 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.208119 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.208234 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.208720 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-gzpjj"] Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.209198 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gzpjj" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.210264 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-4h9v4"] Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.210755 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-4h9v4" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.211315 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-ngtsq"] Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.211895 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ngtsq" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.212978 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-cnkss"] Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.213346 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-cnkss" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.223412 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-44gqd"] Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.223881 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-44gqd" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.224795 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.224812 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.226133 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-sxb6c"] Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.226274 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.226645 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-sxb6c" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.226845 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.227524 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-scdc2"] Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.228079 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-scdc2" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.228754 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-tkx8j"] Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.238512 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-gp4jg"] Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.238910 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.251508 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-9mdwc"] Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.261888 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-gp4jg" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.262156 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.262070 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-9mdwc" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.262021 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-6dljc"] Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.265375 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.265417 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.265378 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.265607 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.265827 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.265994 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.266114 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.266252 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.266434 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.266510 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.266571 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.266681 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.266809 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.266912 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.267013 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.267704 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.267990 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.268107 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.268202 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.268311 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.268317 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.268561 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.268709 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.269210 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.271277 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.273771 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.274069 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.274353 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.274543 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.274849 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.274973 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.276137 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.276407 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-df5xv"] Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.276637 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-9nkvv"] Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.276740 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.276867 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-njspk"] Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.276925 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.277014 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.277117 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.277195 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.277329 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-tkx8j" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.277632 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-fdcsz"] Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.277989 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-6dljc" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.278239 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-df5xv" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.278421 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-fdcsz" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.278591 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-9nkvv" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.278606 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-njspk" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.278991 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/40fcb5fe-ee7b-4696-a775-efc703af9b61-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-ngtsq\" (UID: \"40fcb5fe-ee7b-4696-a775-efc703af9b61\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ngtsq" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.279018 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/d874a84d-48e3-4cb2-ac74-4f0db52e9d75-console-serving-cert\") pod \"console-f9d7485db-cnkss\" (UID: \"d874a84d-48e3-4cb2-ac74-4f0db52e9d75\") " pod="openshift-console/console-f9d7485db-cnkss" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.279035 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8617aa45-3d34-4ab9-9870-67f8ccaeec16-config\") pod \"machine-approver-56656f9798-gzpjj\" (UID: \"8617aa45-3d34-4ab9-9870-67f8ccaeec16\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gzpjj" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.279052 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/40fcb5fe-ee7b-4696-a775-efc703af9b61-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-ngtsq\" (UID: \"40fcb5fe-ee7b-4696-a775-efc703af9b61\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ngtsq" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.279067 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tr7xh\" (UniqueName: \"kubernetes.io/projected/e4968496-d03e-4450-b85d-240cd75bd900-kube-api-access-tr7xh\") pod \"openshift-apiserver-operator-796bbdcf4f-sxb6c\" (UID: \"e4968496-d03e-4450-b85d-240cd75bd900\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-sxb6c" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.279083 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/d874a84d-48e3-4cb2-ac74-4f0db52e9d75-oauth-serving-cert\") pod \"console-f9d7485db-cnkss\" (UID: \"d874a84d-48e3-4cb2-ac74-4f0db52e9d75\") " pod="openshift-console/console-f9d7485db-cnkss" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.279104 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7drzs\" (UniqueName: \"kubernetes.io/projected/f02fb0c2-81fc-45ed-8742-f2f8a51a5437-kube-api-access-7drzs\") pod \"openshift-controller-manager-operator-756b6f6bc6-44gqd\" (UID: \"f02fb0c2-81fc-45ed-8742-f2f8a51a5437\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-44gqd" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.279123 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/40fcb5fe-ee7b-4696-a775-efc703af9b61-audit-dir\") pod \"apiserver-7bbb656c7d-ngtsq\" (UID: \"40fcb5fe-ee7b-4696-a775-efc703af9b61\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ngtsq" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.279139 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k9qt7\" (UniqueName: \"kubernetes.io/projected/6c0cd275-710b-43d8-a9b9-b46bc379454b-kube-api-access-k9qt7\") pod \"apiserver-76f77b778f-h5vrh\" (UID: \"6c0cd275-710b-43d8-a9b9-b46bc379454b\") " pod="openshift-apiserver/apiserver-76f77b778f-h5vrh" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.279155 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zkpf8\" (UniqueName: \"kubernetes.io/projected/348075cc-bd79-42f2-bbe2-0e5367092963-kube-api-access-zkpf8\") pod \"cluster-samples-operator-665b6dd947-zh65k\" (UID: \"348075cc-bd79-42f2-bbe2-0e5367092963\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-zh65k" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.279171 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/d874a84d-48e3-4cb2-ac74-4f0db52e9d75-console-oauth-config\") pod \"console-f9d7485db-cnkss\" (UID: \"d874a84d-48e3-4cb2-ac74-4f0db52e9d75\") " pod="openshift-console/console-f9d7485db-cnkss" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.279186 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/8617aa45-3d34-4ab9-9870-67f8ccaeec16-auth-proxy-config\") pod \"machine-approver-56656f9798-gzpjj\" (UID: \"8617aa45-3d34-4ab9-9870-67f8ccaeec16\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gzpjj" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.279202 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lpkq9\" (UniqueName: \"kubernetes.io/projected/d874a84d-48e3-4cb2-ac74-4f0db52e9d75-kube-api-access-lpkq9\") pod \"console-f9d7485db-cnkss\" (UID: \"d874a84d-48e3-4cb2-ac74-4f0db52e9d75\") " pod="openshift-console/console-f9d7485db-cnkss" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.279217 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nkmcl\" (UniqueName: \"kubernetes.io/projected/1eaf78eb-9719-4737-9f33-a8ab1dcb5dcf-kube-api-access-nkmcl\") pod \"openshift-config-operator-7777fb866f-4h9v4\" (UID: \"1eaf78eb-9719-4737-9f33-a8ab1dcb5dcf\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-4h9v4" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.279232 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jlhqp\" (UniqueName: \"kubernetes.io/projected/8617aa45-3d34-4ab9-9870-67f8ccaeec16-kube-api-access-jlhqp\") pod \"machine-approver-56656f9798-gzpjj\" (UID: \"8617aa45-3d34-4ab9-9870-67f8ccaeec16\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gzpjj" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.279248 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/40fcb5fe-ee7b-4696-a775-efc703af9b61-serving-cert\") pod \"apiserver-7bbb656c7d-ngtsq\" (UID: \"40fcb5fe-ee7b-4696-a775-efc703af9b61\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ngtsq" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.279274 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/6c0cd275-710b-43d8-a9b9-b46bc379454b-audit\") pod \"apiserver-76f77b778f-h5vrh\" (UID: \"6c0cd275-710b-43d8-a9b9-b46bc379454b\") " pod="openshift-apiserver/apiserver-76f77b778f-h5vrh" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.279292 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6c0cd275-710b-43d8-a9b9-b46bc379454b-serving-cert\") pod \"apiserver-76f77b778f-h5vrh\" (UID: \"6c0cd275-710b-43d8-a9b9-b46bc379454b\") " pod="openshift-apiserver/apiserver-76f77b778f-h5vrh" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.279306 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d874a84d-48e3-4cb2-ac74-4f0db52e9d75-trusted-ca-bundle\") pod \"console-f9d7485db-cnkss\" (UID: \"d874a84d-48e3-4cb2-ac74-4f0db52e9d75\") " pod="openshift-console/console-f9d7485db-cnkss" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.279322 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e4968496-d03e-4450-b85d-240cd75bd900-config\") pod \"openshift-apiserver-operator-796bbdcf4f-sxb6c\" (UID: \"e4968496-d03e-4450-b85d-240cd75bd900\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-sxb6c" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.279349 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4w4ht\" (UniqueName: \"kubernetes.io/projected/c4b729bd-782f-4393-b162-16c02eb3cb71-kube-api-access-4w4ht\") pod \"downloads-7954f5f757-7k5qb\" (UID: \"c4b729bd-782f-4393-b162-16c02eb3cb71\") " pod="openshift-console/downloads-7954f5f757-7k5qb" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.279374 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/40fcb5fe-ee7b-4696-a775-efc703af9b61-etcd-client\") pod \"apiserver-7bbb656c7d-ngtsq\" (UID: \"40fcb5fe-ee7b-4696-a775-efc703af9b61\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ngtsq" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.279390 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/d874a84d-48e3-4cb2-ac74-4f0db52e9d75-console-config\") pod \"console-f9d7485db-cnkss\" (UID: \"d874a84d-48e3-4cb2-ac74-4f0db52e9d75\") " pod="openshift-console/console-f9d7485db-cnkss" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.279404 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6c0cd275-710b-43d8-a9b9-b46bc379454b-config\") pod \"apiserver-76f77b778f-h5vrh\" (UID: \"6c0cd275-710b-43d8-a9b9-b46bc379454b\") " pod="openshift-apiserver/apiserver-76f77b778f-h5vrh" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.279420 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/6c0cd275-710b-43d8-a9b9-b46bc379454b-audit-dir\") pod \"apiserver-76f77b778f-h5vrh\" (UID: \"6c0cd275-710b-43d8-a9b9-b46bc379454b\") " pod="openshift-apiserver/apiserver-76f77b778f-h5vrh" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.279441 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kc4gx\" (UniqueName: \"kubernetes.io/projected/40fcb5fe-ee7b-4696-a775-efc703af9b61-kube-api-access-kc4gx\") pod \"apiserver-7bbb656c7d-ngtsq\" (UID: \"40fcb5fe-ee7b-4696-a775-efc703af9b61\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ngtsq" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.279471 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/08bc3386-ee9a-4a3c-aa52-22d667971076-images\") pod \"machine-api-operator-5694c8668f-98wxj\" (UID: \"08bc3386-ee9a-4a3c-aa52-22d667971076\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-98wxj" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.279499 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/1eaf78eb-9719-4737-9f33-a8ab1dcb5dcf-available-featuregates\") pod \"openshift-config-operator-7777fb866f-4h9v4\" (UID: \"1eaf78eb-9719-4737-9f33-a8ab1dcb5dcf\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-4h9v4" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.279514 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f02fb0c2-81fc-45ed-8742-f2f8a51a5437-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-44gqd\" (UID: \"f02fb0c2-81fc-45ed-8742-f2f8a51a5437\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-44gqd" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.279531 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/40fcb5fe-ee7b-4696-a775-efc703af9b61-encryption-config\") pod \"apiserver-7bbb656c7d-ngtsq\" (UID: \"40fcb5fe-ee7b-4696-a775-efc703af9b61\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ngtsq" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.279546 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/6c0cd275-710b-43d8-a9b9-b46bc379454b-node-pullsecrets\") pod \"apiserver-76f77b778f-h5vrh\" (UID: \"6c0cd275-710b-43d8-a9b9-b46bc379454b\") " pod="openshift-apiserver/apiserver-76f77b778f-h5vrh" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.279565 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/8617aa45-3d34-4ab9-9870-67f8ccaeec16-machine-approver-tls\") pod \"machine-approver-56656f9798-gzpjj\" (UID: \"8617aa45-3d34-4ab9-9870-67f8ccaeec16\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gzpjj" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.279595 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/40fcb5fe-ee7b-4696-a775-efc703af9b61-audit-policies\") pod \"apiserver-7bbb656c7d-ngtsq\" (UID: \"40fcb5fe-ee7b-4696-a775-efc703af9b61\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ngtsq" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.279618 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/08bc3386-ee9a-4a3c-aa52-22d667971076-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-98wxj\" (UID: \"08bc3386-ee9a-4a3c-aa52-22d667971076\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-98wxj" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.279659 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f02fb0c2-81fc-45ed-8742-f2f8a51a5437-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-44gqd\" (UID: \"f02fb0c2-81fc-45ed-8742-f2f8a51a5437\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-44gqd" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.279682 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/d874a84d-48e3-4cb2-ac74-4f0db52e9d75-service-ca\") pod \"console-f9d7485db-cnkss\" (UID: \"d874a84d-48e3-4cb2-ac74-4f0db52e9d75\") " pod="openshift-console/console-f9d7485db-cnkss" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.279698 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/08bc3386-ee9a-4a3c-aa52-22d667971076-config\") pod \"machine-api-operator-5694c8668f-98wxj\" (UID: \"08bc3386-ee9a-4a3c-aa52-22d667971076\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-98wxj" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.279712 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/70342a10-423d-4657-9fc0-062ab9ab8902-client-ca\") pod \"controller-manager-879f6c89f-966nh\" (UID: \"70342a10-423d-4657-9fc0-062ab9ab8902\") " pod="openshift-controller-manager/controller-manager-879f6c89f-966nh" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.279727 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/70342a10-423d-4657-9fc0-062ab9ab8902-serving-cert\") pod \"controller-manager-879f6c89f-966nh\" (UID: \"70342a10-423d-4657-9fc0-062ab9ab8902\") " pod="openshift-controller-manager/controller-manager-879f6c89f-966nh" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.279745 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/6c0cd275-710b-43d8-a9b9-b46bc379454b-etcd-serving-ca\") pod \"apiserver-76f77b778f-h5vrh\" (UID: \"6c0cd275-710b-43d8-a9b9-b46bc379454b\") " pod="openshift-apiserver/apiserver-76f77b778f-h5vrh" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.279768 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e4968496-d03e-4450-b85d-240cd75bd900-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-sxb6c\" (UID: \"e4968496-d03e-4450-b85d-240cd75bd900\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-sxb6c" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.279783 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/6c0cd275-710b-43d8-a9b9-b46bc379454b-etcd-client\") pod \"apiserver-76f77b778f-h5vrh\" (UID: \"6c0cd275-710b-43d8-a9b9-b46bc379454b\") " pod="openshift-apiserver/apiserver-76f77b778f-h5vrh" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.279796 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/6c0cd275-710b-43d8-a9b9-b46bc379454b-image-import-ca\") pod \"apiserver-76f77b778f-h5vrh\" (UID: \"6c0cd275-710b-43d8-a9b9-b46bc379454b\") " pod="openshift-apiserver/apiserver-76f77b778f-h5vrh" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.279827 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/6c0cd275-710b-43d8-a9b9-b46bc379454b-encryption-config\") pod \"apiserver-76f77b778f-h5vrh\" (UID: \"6c0cd275-710b-43d8-a9b9-b46bc379454b\") " pod="openshift-apiserver/apiserver-76f77b778f-h5vrh" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.279859 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/70342a10-423d-4657-9fc0-062ab9ab8902-config\") pod \"controller-manager-879f6c89f-966nh\" (UID: \"70342a10-423d-4657-9fc0-062ab9ab8902\") " pod="openshift-controller-manager/controller-manager-879f6c89f-966nh" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.279879 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/70342a10-423d-4657-9fc0-062ab9ab8902-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-966nh\" (UID: \"70342a10-423d-4657-9fc0-062ab9ab8902\") " pod="openshift-controller-manager/controller-manager-879f6c89f-966nh" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.279894 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1ceb5fce-77f5-46be-80ff-be6e932bfb35-serving-cert\") pod \"console-operator-58897d9998-scdc2\" (UID: \"1ceb5fce-77f5-46be-80ff-be6e932bfb35\") " pod="openshift-console-operator/console-operator-58897d9998-scdc2" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.279909 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lmxq7\" (UniqueName: \"kubernetes.io/projected/1ceb5fce-77f5-46be-80ff-be6e932bfb35-kube-api-access-lmxq7\") pod \"console-operator-58897d9998-scdc2\" (UID: \"1ceb5fce-77f5-46be-80ff-be6e932bfb35\") " pod="openshift-console-operator/console-operator-58897d9998-scdc2" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.279957 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1ceb5fce-77f5-46be-80ff-be6e932bfb35-config\") pod \"console-operator-58897d9998-scdc2\" (UID: \"1ceb5fce-77f5-46be-80ff-be6e932bfb35\") " pod="openshift-console-operator/console-operator-58897d9998-scdc2" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.279971 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1ceb5fce-77f5-46be-80ff-be6e932bfb35-trusted-ca\") pod \"console-operator-58897d9998-scdc2\" (UID: \"1ceb5fce-77f5-46be-80ff-be6e932bfb35\") " pod="openshift-console-operator/console-operator-58897d9998-scdc2" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.279987 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kcl59\" (UniqueName: \"kubernetes.io/projected/08bc3386-ee9a-4a3c-aa52-22d667971076-kube-api-access-kcl59\") pod \"machine-api-operator-5694c8668f-98wxj\" (UID: \"08bc3386-ee9a-4a3c-aa52-22d667971076\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-98wxj" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.280004 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1eaf78eb-9719-4737-9f33-a8ab1dcb5dcf-serving-cert\") pod \"openshift-config-operator-7777fb866f-4h9v4\" (UID: \"1eaf78eb-9719-4737-9f33-a8ab1dcb5dcf\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-4h9v4" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.280018 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5jxh9\" (UniqueName: \"kubernetes.io/projected/70342a10-423d-4657-9fc0-062ab9ab8902-kube-api-access-5jxh9\") pod \"controller-manager-879f6c89f-966nh\" (UID: \"70342a10-423d-4657-9fc0-062ab9ab8902\") " pod="openshift-controller-manager/controller-manager-879f6c89f-966nh" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.280033 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6c0cd275-710b-43d8-a9b9-b46bc379454b-trusted-ca-bundle\") pod \"apiserver-76f77b778f-h5vrh\" (UID: \"6c0cd275-710b-43d8-a9b9-b46bc379454b\") " pod="openshift-apiserver/apiserver-76f77b778f-h5vrh" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.280050 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/348075cc-bd79-42f2-bbe2-0e5367092963-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-zh65k\" (UID: \"348075cc-bd79-42f2-bbe2-0e5367092963\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-zh65k" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.280797 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.281005 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.281759 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.282098 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.282604 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.283988 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.284121 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.284209 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.285525 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-4298w"] Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.286148 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-966nh"] Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.286247 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4298w" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.290951 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.291201 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.292024 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-2m47l"] Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.296970 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.298648 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.298827 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.298973 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.298993 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.299005 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.299070 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.299251 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.299268 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.299318 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.299772 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.301388 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2m47l" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.302501 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-mkwq2"] Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.303314 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-mkwq2" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.304231 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-n59jp"] Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.304566 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-n59jp" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.309057 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-dlpkk"] Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.309445 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29500380-2gwrp"] Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.309683 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-52th4"] Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.309985 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-52th4" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.310410 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-dlpkk" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.310576 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29500380-2gwrp" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.312339 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.314242 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.315637 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.315701 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.324065 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.326379 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qq96b"] Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.327222 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qq96b" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.327531 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.328203 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-2cdx8"] Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.332266 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-drqfg"] Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.336039 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.338499 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2cdx8" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.338721 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-drqfg" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.338736 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mmcq4"] Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.354339 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9flm5"] Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.354397 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.354639 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-p66d5"] Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.355116 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-p66d5" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.355260 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mmcq4" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.355278 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9flm5" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.355678 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.355773 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.356099 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.356278 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.357242 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-wqbtk"] Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.357503 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.358287 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.359222 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.359359 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.359465 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.360339 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.361038 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-fffd9"] Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.365685 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-l5jcj"] Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.366065 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-h5vrh"] Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.366371 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-fffd9" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.366476 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-wqbtk" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.366448 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-l5jcj" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.367403 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-tpldd"] Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.368166 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-7mtl6"] Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.368704 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-7mtl6" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.369300 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-tpldd" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.384808 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/08bc3386-ee9a-4a3c-aa52-22d667971076-images\") pod \"machine-api-operator-5694c8668f-98wxj\" (UID: \"08bc3386-ee9a-4a3c-aa52-22d667971076\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-98wxj" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.384839 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/1eaf78eb-9719-4737-9f33-a8ab1dcb5dcf-available-featuregates\") pod \"openshift-config-operator-7777fb866f-4h9v4\" (UID: \"1eaf78eb-9719-4737-9f33-a8ab1dcb5dcf\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-4h9v4" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.384857 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f02fb0c2-81fc-45ed-8742-f2f8a51a5437-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-44gqd\" (UID: \"f02fb0c2-81fc-45ed-8742-f2f8a51a5437\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-44gqd" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.384874 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/40fcb5fe-ee7b-4696-a775-efc703af9b61-encryption-config\") pod \"apiserver-7bbb656c7d-ngtsq\" (UID: \"40fcb5fe-ee7b-4696-a775-efc703af9b61\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ngtsq" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.384889 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/08bc3386-ee9a-4a3c-aa52-22d667971076-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-98wxj\" (UID: \"08bc3386-ee9a-4a3c-aa52-22d667971076\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-98wxj" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.384903 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/6c0cd275-710b-43d8-a9b9-b46bc379454b-node-pullsecrets\") pod \"apiserver-76f77b778f-h5vrh\" (UID: \"6c0cd275-710b-43d8-a9b9-b46bc379454b\") " pod="openshift-apiserver/apiserver-76f77b778f-h5vrh" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.384917 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/8617aa45-3d34-4ab9-9870-67f8ccaeec16-machine-approver-tls\") pod \"machine-approver-56656f9798-gzpjj\" (UID: \"8617aa45-3d34-4ab9-9870-67f8ccaeec16\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gzpjj" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.384957 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/40fcb5fe-ee7b-4696-a775-efc703af9b61-audit-policies\") pod \"apiserver-7bbb656c7d-ngtsq\" (UID: \"40fcb5fe-ee7b-4696-a775-efc703af9b61\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ngtsq" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.384978 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f02fb0c2-81fc-45ed-8742-f2f8a51a5437-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-44gqd\" (UID: \"f02fb0c2-81fc-45ed-8742-f2f8a51a5437\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-44gqd" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.384993 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/6c0cd275-710b-43d8-a9b9-b46bc379454b-etcd-serving-ca\") pod \"apiserver-76f77b778f-h5vrh\" (UID: \"6c0cd275-710b-43d8-a9b9-b46bc379454b\") " pod="openshift-apiserver/apiserver-76f77b778f-h5vrh" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.385008 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/d874a84d-48e3-4cb2-ac74-4f0db52e9d75-service-ca\") pod \"console-f9d7485db-cnkss\" (UID: \"d874a84d-48e3-4cb2-ac74-4f0db52e9d75\") " pod="openshift-console/console-f9d7485db-cnkss" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.385021 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/08bc3386-ee9a-4a3c-aa52-22d667971076-config\") pod \"machine-api-operator-5694c8668f-98wxj\" (UID: \"08bc3386-ee9a-4a3c-aa52-22d667971076\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-98wxj" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.385035 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/70342a10-423d-4657-9fc0-062ab9ab8902-client-ca\") pod \"controller-manager-879f6c89f-966nh\" (UID: \"70342a10-423d-4657-9fc0-062ab9ab8902\") " pod="openshift-controller-manager/controller-manager-879f6c89f-966nh" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.385047 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/70342a10-423d-4657-9fc0-062ab9ab8902-serving-cert\") pod \"controller-manager-879f6c89f-966nh\" (UID: \"70342a10-423d-4657-9fc0-062ab9ab8902\") " pod="openshift-controller-manager/controller-manager-879f6c89f-966nh" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.385064 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e4968496-d03e-4450-b85d-240cd75bd900-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-sxb6c\" (UID: \"e4968496-d03e-4450-b85d-240cd75bd900\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-sxb6c" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.385078 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/6c0cd275-710b-43d8-a9b9-b46bc379454b-etcd-client\") pod \"apiserver-76f77b778f-h5vrh\" (UID: \"6c0cd275-710b-43d8-a9b9-b46bc379454b\") " pod="openshift-apiserver/apiserver-76f77b778f-h5vrh" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.385091 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/70342a10-423d-4657-9fc0-062ab9ab8902-config\") pod \"controller-manager-879f6c89f-966nh\" (UID: \"70342a10-423d-4657-9fc0-062ab9ab8902\") " pod="openshift-controller-manager/controller-manager-879f6c89f-966nh" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.385104 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/6c0cd275-710b-43d8-a9b9-b46bc379454b-image-import-ca\") pod \"apiserver-76f77b778f-h5vrh\" (UID: \"6c0cd275-710b-43d8-a9b9-b46bc379454b\") " pod="openshift-apiserver/apiserver-76f77b778f-h5vrh" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.385120 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/6c0cd275-710b-43d8-a9b9-b46bc379454b-encryption-config\") pod \"apiserver-76f77b778f-h5vrh\" (UID: \"6c0cd275-710b-43d8-a9b9-b46bc379454b\") " pod="openshift-apiserver/apiserver-76f77b778f-h5vrh" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.385135 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/70342a10-423d-4657-9fc0-062ab9ab8902-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-966nh\" (UID: \"70342a10-423d-4657-9fc0-062ab9ab8902\") " pod="openshift-controller-manager/controller-manager-879f6c89f-966nh" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.385149 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1ceb5fce-77f5-46be-80ff-be6e932bfb35-serving-cert\") pod \"console-operator-58897d9998-scdc2\" (UID: \"1ceb5fce-77f5-46be-80ff-be6e932bfb35\") " pod="openshift-console-operator/console-operator-58897d9998-scdc2" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.385165 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lmxq7\" (UniqueName: \"kubernetes.io/projected/1ceb5fce-77f5-46be-80ff-be6e932bfb35-kube-api-access-lmxq7\") pod \"console-operator-58897d9998-scdc2\" (UID: \"1ceb5fce-77f5-46be-80ff-be6e932bfb35\") " pod="openshift-console-operator/console-operator-58897d9998-scdc2" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.385185 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1ceb5fce-77f5-46be-80ff-be6e932bfb35-config\") pod \"console-operator-58897d9998-scdc2\" (UID: \"1ceb5fce-77f5-46be-80ff-be6e932bfb35\") " pod="openshift-console-operator/console-operator-58897d9998-scdc2" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.385198 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1ceb5fce-77f5-46be-80ff-be6e932bfb35-trusted-ca\") pod \"console-operator-58897d9998-scdc2\" (UID: \"1ceb5fce-77f5-46be-80ff-be6e932bfb35\") " pod="openshift-console-operator/console-operator-58897d9998-scdc2" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.385215 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5jxh9\" (UniqueName: \"kubernetes.io/projected/70342a10-423d-4657-9fc0-062ab9ab8902-kube-api-access-5jxh9\") pod \"controller-manager-879f6c89f-966nh\" (UID: \"70342a10-423d-4657-9fc0-062ab9ab8902\") " pod="openshift-controller-manager/controller-manager-879f6c89f-966nh" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.385229 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kcl59\" (UniqueName: \"kubernetes.io/projected/08bc3386-ee9a-4a3c-aa52-22d667971076-kube-api-access-kcl59\") pod \"machine-api-operator-5694c8668f-98wxj\" (UID: \"08bc3386-ee9a-4a3c-aa52-22d667971076\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-98wxj" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.385243 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1eaf78eb-9719-4737-9f33-a8ab1dcb5dcf-serving-cert\") pod \"openshift-config-operator-7777fb866f-4h9v4\" (UID: \"1eaf78eb-9719-4737-9f33-a8ab1dcb5dcf\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-4h9v4" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.385257 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6c0cd275-710b-43d8-a9b9-b46bc379454b-trusted-ca-bundle\") pod \"apiserver-76f77b778f-h5vrh\" (UID: \"6c0cd275-710b-43d8-a9b9-b46bc379454b\") " pod="openshift-apiserver/apiserver-76f77b778f-h5vrh" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.385274 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/348075cc-bd79-42f2-bbe2-0e5367092963-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-zh65k\" (UID: \"348075cc-bd79-42f2-bbe2-0e5367092963\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-zh65k" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.385287 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/40fcb5fe-ee7b-4696-a775-efc703af9b61-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-ngtsq\" (UID: \"40fcb5fe-ee7b-4696-a775-efc703af9b61\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ngtsq" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.385300 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/d874a84d-48e3-4cb2-ac74-4f0db52e9d75-console-serving-cert\") pod \"console-f9d7485db-cnkss\" (UID: \"d874a84d-48e3-4cb2-ac74-4f0db52e9d75\") " pod="openshift-console/console-f9d7485db-cnkss" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.385398 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8617aa45-3d34-4ab9-9870-67f8ccaeec16-config\") pod \"machine-approver-56656f9798-gzpjj\" (UID: \"8617aa45-3d34-4ab9-9870-67f8ccaeec16\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gzpjj" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.385416 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7drzs\" (UniqueName: \"kubernetes.io/projected/f02fb0c2-81fc-45ed-8742-f2f8a51a5437-kube-api-access-7drzs\") pod \"openshift-controller-manager-operator-756b6f6bc6-44gqd\" (UID: \"f02fb0c2-81fc-45ed-8742-f2f8a51a5437\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-44gqd" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.385432 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/40fcb5fe-ee7b-4696-a775-efc703af9b61-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-ngtsq\" (UID: \"40fcb5fe-ee7b-4696-a775-efc703af9b61\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ngtsq" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.385447 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tr7xh\" (UniqueName: \"kubernetes.io/projected/e4968496-d03e-4450-b85d-240cd75bd900-kube-api-access-tr7xh\") pod \"openshift-apiserver-operator-796bbdcf4f-sxb6c\" (UID: \"e4968496-d03e-4450-b85d-240cd75bd900\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-sxb6c" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.385470 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/d874a84d-48e3-4cb2-ac74-4f0db52e9d75-oauth-serving-cert\") pod \"console-f9d7485db-cnkss\" (UID: \"d874a84d-48e3-4cb2-ac74-4f0db52e9d75\") " pod="openshift-console/console-f9d7485db-cnkss" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.385484 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/40fcb5fe-ee7b-4696-a775-efc703af9b61-audit-dir\") pod \"apiserver-7bbb656c7d-ngtsq\" (UID: \"40fcb5fe-ee7b-4696-a775-efc703af9b61\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ngtsq" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.385500 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k9qt7\" (UniqueName: \"kubernetes.io/projected/6c0cd275-710b-43d8-a9b9-b46bc379454b-kube-api-access-k9qt7\") pod \"apiserver-76f77b778f-h5vrh\" (UID: \"6c0cd275-710b-43d8-a9b9-b46bc379454b\") " pod="openshift-apiserver/apiserver-76f77b778f-h5vrh" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.385515 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zkpf8\" (UniqueName: \"kubernetes.io/projected/348075cc-bd79-42f2-bbe2-0e5367092963-kube-api-access-zkpf8\") pod \"cluster-samples-operator-665b6dd947-zh65k\" (UID: \"348075cc-bd79-42f2-bbe2-0e5367092963\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-zh65k" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.385575 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/d874a84d-48e3-4cb2-ac74-4f0db52e9d75-console-oauth-config\") pod \"console-f9d7485db-cnkss\" (UID: \"d874a84d-48e3-4cb2-ac74-4f0db52e9d75\") " pod="openshift-console/console-f9d7485db-cnkss" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.385596 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/8617aa45-3d34-4ab9-9870-67f8ccaeec16-auth-proxy-config\") pod \"machine-approver-56656f9798-gzpjj\" (UID: \"8617aa45-3d34-4ab9-9870-67f8ccaeec16\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gzpjj" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.385612 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lpkq9\" (UniqueName: \"kubernetes.io/projected/d874a84d-48e3-4cb2-ac74-4f0db52e9d75-kube-api-access-lpkq9\") pod \"console-f9d7485db-cnkss\" (UID: \"d874a84d-48e3-4cb2-ac74-4f0db52e9d75\") " pod="openshift-console/console-f9d7485db-cnkss" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.385631 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jlhqp\" (UniqueName: \"kubernetes.io/projected/8617aa45-3d34-4ab9-9870-67f8ccaeec16-kube-api-access-jlhqp\") pod \"machine-approver-56656f9798-gzpjj\" (UID: \"8617aa45-3d34-4ab9-9870-67f8ccaeec16\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gzpjj" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.385648 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nkmcl\" (UniqueName: \"kubernetes.io/projected/1eaf78eb-9719-4737-9f33-a8ab1dcb5dcf-kube-api-access-nkmcl\") pod \"openshift-config-operator-7777fb866f-4h9v4\" (UID: \"1eaf78eb-9719-4737-9f33-a8ab1dcb5dcf\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-4h9v4" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.385662 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/40fcb5fe-ee7b-4696-a775-efc703af9b61-serving-cert\") pod \"apiserver-7bbb656c7d-ngtsq\" (UID: \"40fcb5fe-ee7b-4696-a775-efc703af9b61\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ngtsq" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.385677 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d874a84d-48e3-4cb2-ac74-4f0db52e9d75-trusted-ca-bundle\") pod \"console-f9d7485db-cnkss\" (UID: \"d874a84d-48e3-4cb2-ac74-4f0db52e9d75\") " pod="openshift-console/console-f9d7485db-cnkss" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.385691 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/6c0cd275-710b-43d8-a9b9-b46bc379454b-audit\") pod \"apiserver-76f77b778f-h5vrh\" (UID: \"6c0cd275-710b-43d8-a9b9-b46bc379454b\") " pod="openshift-apiserver/apiserver-76f77b778f-h5vrh" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.385705 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6c0cd275-710b-43d8-a9b9-b46bc379454b-serving-cert\") pod \"apiserver-76f77b778f-h5vrh\" (UID: \"6c0cd275-710b-43d8-a9b9-b46bc379454b\") " pod="openshift-apiserver/apiserver-76f77b778f-h5vrh" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.385719 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e4968496-d03e-4450-b85d-240cd75bd900-config\") pod \"openshift-apiserver-operator-796bbdcf4f-sxb6c\" (UID: \"e4968496-d03e-4450-b85d-240cd75bd900\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-sxb6c" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.385736 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4w4ht\" (UniqueName: \"kubernetes.io/projected/c4b729bd-782f-4393-b162-16c02eb3cb71-kube-api-access-4w4ht\") pod \"downloads-7954f5f757-7k5qb\" (UID: \"c4b729bd-782f-4393-b162-16c02eb3cb71\") " pod="openshift-console/downloads-7954f5f757-7k5qb" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.385758 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/40fcb5fe-ee7b-4696-a775-efc703af9b61-etcd-client\") pod \"apiserver-7bbb656c7d-ngtsq\" (UID: \"40fcb5fe-ee7b-4696-a775-efc703af9b61\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ngtsq" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.385772 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/d874a84d-48e3-4cb2-ac74-4f0db52e9d75-console-config\") pod \"console-f9d7485db-cnkss\" (UID: \"d874a84d-48e3-4cb2-ac74-4f0db52e9d75\") " pod="openshift-console/console-f9d7485db-cnkss" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.385785 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6c0cd275-710b-43d8-a9b9-b46bc379454b-config\") pod \"apiserver-76f77b778f-h5vrh\" (UID: \"6c0cd275-710b-43d8-a9b9-b46bc379454b\") " pod="openshift-apiserver/apiserver-76f77b778f-h5vrh" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.385798 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/6c0cd275-710b-43d8-a9b9-b46bc379454b-audit-dir\") pod \"apiserver-76f77b778f-h5vrh\" (UID: \"6c0cd275-710b-43d8-a9b9-b46bc379454b\") " pod="openshift-apiserver/apiserver-76f77b778f-h5vrh" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.385815 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kc4gx\" (UniqueName: \"kubernetes.io/projected/40fcb5fe-ee7b-4696-a775-efc703af9b61-kube-api-access-kc4gx\") pod \"apiserver-7bbb656c7d-ngtsq\" (UID: \"40fcb5fe-ee7b-4696-a775-efc703af9b61\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ngtsq" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.386634 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/08bc3386-ee9a-4a3c-aa52-22d667971076-images\") pod \"machine-api-operator-5694c8668f-98wxj\" (UID: \"08bc3386-ee9a-4a3c-aa52-22d667971076\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-98wxj" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.386866 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/1eaf78eb-9719-4737-9f33-a8ab1dcb5dcf-available-featuregates\") pod \"openshift-config-operator-7777fb866f-4h9v4\" (UID: \"1eaf78eb-9719-4737-9f33-a8ab1dcb5dcf\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-4h9v4" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.388203 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/40fcb5fe-ee7b-4696-a775-efc703af9b61-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-ngtsq\" (UID: \"40fcb5fe-ee7b-4696-a775-efc703af9b61\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ngtsq" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.388492 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/6c0cd275-710b-43d8-a9b9-b46bc379454b-image-import-ca\") pod \"apiserver-76f77b778f-h5vrh\" (UID: \"6c0cd275-710b-43d8-a9b9-b46bc379454b\") " pod="openshift-apiserver/apiserver-76f77b778f-h5vrh" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.388764 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/6c0cd275-710b-43d8-a9b9-b46bc379454b-etcd-serving-ca\") pod \"apiserver-76f77b778f-h5vrh\" (UID: \"6c0cd275-710b-43d8-a9b9-b46bc379454b\") " pod="openshift-apiserver/apiserver-76f77b778f-h5vrh" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.389914 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.390098 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/08bc3386-ee9a-4a3c-aa52-22d667971076-config\") pod \"machine-api-operator-5694c8668f-98wxj\" (UID: \"08bc3386-ee9a-4a3c-aa52-22d667971076\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-98wxj" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.390391 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f02fb0c2-81fc-45ed-8742-f2f8a51a5437-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-44gqd\" (UID: \"f02fb0c2-81fc-45ed-8742-f2f8a51a5437\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-44gqd" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.391734 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/d874a84d-48e3-4cb2-ac74-4f0db52e9d75-service-ca\") pod \"console-f9d7485db-cnkss\" (UID: \"d874a84d-48e3-4cb2-ac74-4f0db52e9d75\") " pod="openshift-console/console-f9d7485db-cnkss" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.394849 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/70342a10-423d-4657-9fc0-062ab9ab8902-client-ca\") pod \"controller-manager-879f6c89f-966nh\" (UID: \"70342a10-423d-4657-9fc0-062ab9ab8902\") " pod="openshift-controller-manager/controller-manager-879f6c89f-966nh" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.396804 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/70342a10-423d-4657-9fc0-062ab9ab8902-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-966nh\" (UID: \"70342a10-423d-4657-9fc0-062ab9ab8902\") " pod="openshift-controller-manager/controller-manager-879f6c89f-966nh" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.397558 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1ceb5fce-77f5-46be-80ff-be6e932bfb35-config\") pod \"console-operator-58897d9998-scdc2\" (UID: \"1ceb5fce-77f5-46be-80ff-be6e932bfb35\") " pod="openshift-console-operator/console-operator-58897d9998-scdc2" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.398234 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1ceb5fce-77f5-46be-80ff-be6e932bfb35-trusted-ca\") pod \"console-operator-58897d9998-scdc2\" (UID: \"1ceb5fce-77f5-46be-80ff-be6e932bfb35\") " pod="openshift-console-operator/console-operator-58897d9998-scdc2" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.398520 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-sxb6c"] Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.415285 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-t5q9w"] Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.416362 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-fdcsz"] Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.416384 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-ngtsq"] Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.400962 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/40fcb5fe-ee7b-4696-a775-efc703af9b61-audit-dir\") pod \"apiserver-7bbb656c7d-ngtsq\" (UID: \"40fcb5fe-ee7b-4696-a775-efc703af9b61\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ngtsq" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.402396 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.398592 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/6c0cd275-710b-43d8-a9b9-b46bc379454b-node-pullsecrets\") pod \"apiserver-76f77b778f-h5vrh\" (UID: \"6c0cd275-710b-43d8-a9b9-b46bc379454b\") " pod="openshift-apiserver/apiserver-76f77b778f-h5vrh" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.404473 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/40fcb5fe-ee7b-4696-a775-efc703af9b61-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-ngtsq\" (UID: \"40fcb5fe-ee7b-4696-a775-efc703af9b61\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ngtsq" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.408467 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/d874a84d-48e3-4cb2-ac74-4f0db52e9d75-console-oauth-config\") pod \"console-f9d7485db-cnkss\" (UID: \"d874a84d-48e3-4cb2-ac74-4f0db52e9d75\") " pod="openshift-console/console-f9d7485db-cnkss" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.408927 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/8617aa45-3d34-4ab9-9870-67f8ccaeec16-auth-proxy-config\") pod \"machine-approver-56656f9798-gzpjj\" (UID: \"8617aa45-3d34-4ab9-9870-67f8ccaeec16\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gzpjj" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.409535 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1ceb5fce-77f5-46be-80ff-be6e932bfb35-serving-cert\") pod \"console-operator-58897d9998-scdc2\" (UID: \"1ceb5fce-77f5-46be-80ff-be6e932bfb35\") " pod="openshift-console-operator/console-operator-58897d9998-scdc2" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.410434 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d874a84d-48e3-4cb2-ac74-4f0db52e9d75-trusted-ca-bundle\") pod \"console-f9d7485db-cnkss\" (UID: \"d874a84d-48e3-4cb2-ac74-4f0db52e9d75\") " pod="openshift-console/console-f9d7485db-cnkss" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.410794 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/6c0cd275-710b-43d8-a9b9-b46bc379454b-audit\") pod \"apiserver-76f77b778f-h5vrh\" (UID: \"6c0cd275-710b-43d8-a9b9-b46bc379454b\") " pod="openshift-apiserver/apiserver-76f77b778f-h5vrh" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.412521 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/40fcb5fe-ee7b-4696-a775-efc703af9b61-serving-cert\") pod \"apiserver-7bbb656c7d-ngtsq\" (UID: \"40fcb5fe-ee7b-4696-a775-efc703af9b61\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ngtsq" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.414116 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/8617aa45-3d34-4ab9-9870-67f8ccaeec16-machine-approver-tls\") pod \"machine-approver-56656f9798-gzpjj\" (UID: \"8617aa45-3d34-4ab9-9870-67f8ccaeec16\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gzpjj" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.414631 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/40fcb5fe-ee7b-4696-a775-efc703af9b61-audit-policies\") pod \"apiserver-7bbb656c7d-ngtsq\" (UID: \"40fcb5fe-ee7b-4696-a775-efc703af9b61\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ngtsq" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.415158 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/70342a10-423d-4657-9fc0-062ab9ab8902-serving-cert\") pod \"controller-manager-879f6c89f-966nh\" (UID: \"70342a10-423d-4657-9fc0-062ab9ab8902\") " pod="openshift-controller-manager/controller-manager-879f6c89f-966nh" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.399648 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8617aa45-3d34-4ab9-9870-67f8ccaeec16-config\") pod \"machine-approver-56656f9798-gzpjj\" (UID: \"8617aa45-3d34-4ab9-9870-67f8ccaeec16\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gzpjj" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.416854 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-t5q9w" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.400914 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/d874a84d-48e3-4cb2-ac74-4f0db52e9d75-oauth-serving-cert\") pod \"console-f9d7485db-cnkss\" (UID: \"d874a84d-48e3-4cb2-ac74-4f0db52e9d75\") " pod="openshift-console/console-f9d7485db-cnkss" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.403838 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/40fcb5fe-ee7b-4696-a775-efc703af9b61-encryption-config\") pod \"apiserver-7bbb656c7d-ngtsq\" (UID: \"40fcb5fe-ee7b-4696-a775-efc703af9b61\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ngtsq" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.402793 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.402878 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.409671 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.420716 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/40fcb5fe-ee7b-4696-a775-efc703af9b61-etcd-client\") pod \"apiserver-7bbb656c7d-ngtsq\" (UID: \"40fcb5fe-ee7b-4696-a775-efc703af9b61\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ngtsq" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.421316 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/d874a84d-48e3-4cb2-ac74-4f0db52e9d75-console-config\") pod \"console-f9d7485db-cnkss\" (UID: \"d874a84d-48e3-4cb2-ac74-4f0db52e9d75\") " pod="openshift-console/console-f9d7485db-cnkss" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.427021 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-4h9v4"] Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.427067 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-tkx8j"] Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.429610 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/6c0cd275-710b-43d8-a9b9-b46bc379454b-encryption-config\") pod \"apiserver-76f77b778f-h5vrh\" (UID: \"6c0cd275-710b-43d8-a9b9-b46bc379454b\") " pod="openshift-apiserver/apiserver-76f77b778f-h5vrh" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.429737 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.429888 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/6c0cd275-710b-43d8-a9b9-b46bc379454b-audit-dir\") pod \"apiserver-76f77b778f-h5vrh\" (UID: \"6c0cd275-710b-43d8-a9b9-b46bc379454b\") " pod="openshift-apiserver/apiserver-76f77b778f-h5vrh" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.433795 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1eaf78eb-9719-4737-9f33-a8ab1dcb5dcf-serving-cert\") pod \"openshift-config-operator-7777fb866f-4h9v4\" (UID: \"1eaf78eb-9719-4737-9f33-a8ab1dcb5dcf\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-4h9v4" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.434300 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e4968496-d03e-4450-b85d-240cd75bd900-config\") pod \"openshift-apiserver-operator-796bbdcf4f-sxb6c\" (UID: \"e4968496-d03e-4450-b85d-240cd75bd900\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-sxb6c" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.434599 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.434649 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6c0cd275-710b-43d8-a9b9-b46bc379454b-config\") pod \"apiserver-76f77b778f-h5vrh\" (UID: \"6c0cd275-710b-43d8-a9b9-b46bc379454b\") " pod="openshift-apiserver/apiserver-76f77b778f-h5vrh" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.434784 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-44gqd"] Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.434820 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-gp4jg"] Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.435130 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/d874a84d-48e3-4cb2-ac74-4f0db52e9d75-console-serving-cert\") pod \"console-f9d7485db-cnkss\" (UID: \"d874a84d-48e3-4cb2-ac74-4f0db52e9d75\") " pod="openshift-console/console-f9d7485db-cnkss" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.441026 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6c0cd275-710b-43d8-a9b9-b46bc379454b-serving-cert\") pod \"apiserver-76f77b778f-h5vrh\" (UID: \"6c0cd275-710b-43d8-a9b9-b46bc379454b\") " pod="openshift-apiserver/apiserver-76f77b778f-h5vrh" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.442281 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6c0cd275-710b-43d8-a9b9-b46bc379454b-trusted-ca-bundle\") pod \"apiserver-76f77b778f-h5vrh\" (UID: \"6c0cd275-710b-43d8-a9b9-b46bc379454b\") " pod="openshift-apiserver/apiserver-76f77b778f-h5vrh" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.449243 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.454221 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-7k5qb"] Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.459893 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-df5xv"] Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.469012 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/348075cc-bd79-42f2-bbe2-0e5367092963-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-zh65k\" (UID: \"348075cc-bd79-42f2-bbe2-0e5367092963\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-zh65k" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.469141 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/08bc3386-ee9a-4a3c-aa52-22d667971076-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-98wxj\" (UID: \"08bc3386-ee9a-4a3c-aa52-22d667971076\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-98wxj" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.469351 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e4968496-d03e-4450-b85d-240cd75bd900-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-sxb6c\" (UID: \"e4968496-d03e-4450-b85d-240cd75bd900\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-sxb6c" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.469468 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/6c0cd275-710b-43d8-a9b9-b46bc379454b-etcd-client\") pod \"apiserver-76f77b778f-h5vrh\" (UID: \"6c0cd275-710b-43d8-a9b9-b46bc379454b\") " pod="openshift-apiserver/apiserver-76f77b778f-h5vrh" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.469736 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f02fb0c2-81fc-45ed-8742-f2f8a51a5437-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-44gqd\" (UID: \"f02fb0c2-81fc-45ed-8742-f2f8a51a5437\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-44gqd" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.471688 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-98wxj"] Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.472766 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.485528 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-dlpkk"] Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.485572 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.492126 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-scdc2"] Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.495867 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-cnkss"] Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.495893 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-2cdx8"] Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.496355 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-7mtl6"] Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.497709 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29500380-2gwrp"] Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.500965 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-9mdwc"] Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.500985 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qq96b"] Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.501290 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-4298w"] Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.504286 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-njspk"] Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.505687 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-6dljc"] Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.507047 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mmcq4"] Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.507051 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.508207 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-b894n"] Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.508739 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-b894n" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.509870 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-hfwsg"] Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.510551 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-hfwsg" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.511670 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-52th4"] Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.513000 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-p66d5"] Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.514273 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-drqfg"] Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.515345 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-zh65k"] Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.516638 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-wqbtk"] Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.517860 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-n59jp"] Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.519545 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-2m47l"] Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.520436 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-t5q9w"] Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.522071 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-hfwsg"] Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.523405 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9flm5"] Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.524290 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-fffd9"] Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.524942 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.526413 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-9nkvv"] Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.527995 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-l5jcj"] Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.529226 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-tpldd"] Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.530442 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-7v5xl"] Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.531441 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-7v5xl" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.531618 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-7v5xl"] Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.545889 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.565272 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.585243 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.604821 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.625699 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.644996 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.665425 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.685742 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.705119 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.725301 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.745116 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.765063 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.786431 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.814063 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.825795 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.845722 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.866112 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.885893 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.905843 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.925821 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.944883 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.965772 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Feb 02 09:09:38 crc kubenswrapper[4764]: I0202 09:09:38.985712 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Feb 02 09:09:39 crc kubenswrapper[4764]: I0202 09:09:39.005837 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Feb 02 09:09:39 crc kubenswrapper[4764]: I0202 09:09:39.026483 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Feb 02 09:09:39 crc kubenswrapper[4764]: I0202 09:09:39.046708 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Feb 02 09:09:39 crc kubenswrapper[4764]: I0202 09:09:39.065706 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Feb 02 09:09:39 crc kubenswrapper[4764]: I0202 09:09:39.085371 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Feb 02 09:09:39 crc kubenswrapper[4764]: I0202 09:09:39.105080 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Feb 02 09:09:39 crc kubenswrapper[4764]: I0202 09:09:39.124894 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 02 09:09:39 crc kubenswrapper[4764]: I0202 09:09:39.159172 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Feb 02 09:09:39 crc kubenswrapper[4764]: I0202 09:09:39.166916 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 02 09:09:39 crc kubenswrapper[4764]: I0202 09:09:39.186034 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Feb 02 09:09:39 crc kubenswrapper[4764]: I0202 09:09:39.205851 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Feb 02 09:09:39 crc kubenswrapper[4764]: I0202 09:09:39.225345 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Feb 02 09:09:39 crc kubenswrapper[4764]: I0202 09:09:39.246088 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Feb 02 09:09:39 crc kubenswrapper[4764]: I0202 09:09:39.265512 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Feb 02 09:09:39 crc kubenswrapper[4764]: I0202 09:09:39.286041 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Feb 02 09:09:39 crc kubenswrapper[4764]: I0202 09:09:39.306237 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Feb 02 09:09:39 crc kubenswrapper[4764]: I0202 09:09:39.325486 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Feb 02 09:09:39 crc kubenswrapper[4764]: I0202 09:09:39.346015 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Feb 02 09:09:39 crc kubenswrapper[4764]: I0202 09:09:39.363846 4764 request.go:700] Waited for 1.01013697s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/secrets?fieldSelector=metadata.name%3Dmachine-config-operator-dockercfg-98p87&limit=500&resourceVersion=0 Feb 02 09:09:39 crc kubenswrapper[4764]: I0202 09:09:39.365598 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Feb 02 09:09:39 crc kubenswrapper[4764]: I0202 09:09:39.386413 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Feb 02 09:09:39 crc kubenswrapper[4764]: E0202 09:09:39.387808 4764 configmap.go:193] Couldn't get configMap openshift-controller-manager/config: failed to sync configmap cache: timed out waiting for the condition Feb 02 09:09:39 crc kubenswrapper[4764]: E0202 09:09:39.387924 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/70342a10-423d-4657-9fc0-062ab9ab8902-config podName:70342a10-423d-4657-9fc0-062ab9ab8902 nodeName:}" failed. No retries permitted until 2026-02-02 09:09:39.887894898 +0000 UTC m=+142.821619016 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/70342a10-423d-4657-9fc0-062ab9ab8902-config") pod "controller-manager-879f6c89f-966nh" (UID: "70342a10-423d-4657-9fc0-062ab9ab8902") : failed to sync configmap cache: timed out waiting for the condition Feb 02 09:09:39 crc kubenswrapper[4764]: I0202 09:09:39.425086 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Feb 02 09:09:39 crc kubenswrapper[4764]: I0202 09:09:39.446226 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Feb 02 09:09:39 crc kubenswrapper[4764]: I0202 09:09:39.485613 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Feb 02 09:09:39 crc kubenswrapper[4764]: I0202 09:09:39.505305 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Feb 02 09:09:39 crc kubenswrapper[4764]: I0202 09:09:39.525815 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Feb 02 09:09:39 crc kubenswrapper[4764]: I0202 09:09:39.546097 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Feb 02 09:09:39 crc kubenswrapper[4764]: I0202 09:09:39.564957 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Feb 02 09:09:39 crc kubenswrapper[4764]: I0202 09:09:39.586197 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Feb 02 09:09:39 crc kubenswrapper[4764]: I0202 09:09:39.607004 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Feb 02 09:09:39 crc kubenswrapper[4764]: I0202 09:09:39.625993 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Feb 02 09:09:39 crc kubenswrapper[4764]: I0202 09:09:39.646033 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Feb 02 09:09:39 crc kubenswrapper[4764]: I0202 09:09:39.664978 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Feb 02 09:09:39 crc kubenswrapper[4764]: I0202 09:09:39.686651 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Feb 02 09:09:39 crc kubenswrapper[4764]: I0202 09:09:39.705918 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Feb 02 09:09:39 crc kubenswrapper[4764]: I0202 09:09:39.725605 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Feb 02 09:09:39 crc kubenswrapper[4764]: I0202 09:09:39.746462 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Feb 02 09:09:39 crc kubenswrapper[4764]: I0202 09:09:39.765980 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Feb 02 09:09:39 crc kubenswrapper[4764]: I0202 09:09:39.786313 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Feb 02 09:09:39 crc kubenswrapper[4764]: I0202 09:09:39.805381 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Feb 02 09:09:39 crc kubenswrapper[4764]: I0202 09:09:39.832221 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Feb 02 09:09:39 crc kubenswrapper[4764]: I0202 09:09:39.845010 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Feb 02 09:09:39 crc kubenswrapper[4764]: I0202 09:09:39.866163 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Feb 02 09:09:39 crc kubenswrapper[4764]: I0202 09:09:39.886226 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Feb 02 09:09:39 crc kubenswrapper[4764]: I0202 09:09:39.906279 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Feb 02 09:09:39 crc kubenswrapper[4764]: I0202 09:09:39.907162 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/70342a10-423d-4657-9fc0-062ab9ab8902-config\") pod \"controller-manager-879f6c89f-966nh\" (UID: \"70342a10-423d-4657-9fc0-062ab9ab8902\") " pod="openshift-controller-manager/controller-manager-879f6c89f-966nh" Feb 02 09:09:39 crc kubenswrapper[4764]: I0202 09:09:39.926001 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Feb 02 09:09:39 crc kubenswrapper[4764]: I0202 09:09:39.945235 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Feb 02 09:09:39 crc kubenswrapper[4764]: I0202 09:09:39.965434 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Feb 02 09:09:39 crc kubenswrapper[4764]: I0202 09:09:39.985896 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.019010 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kc4gx\" (UniqueName: \"kubernetes.io/projected/40fcb5fe-ee7b-4696-a775-efc703af9b61-kube-api-access-kc4gx\") pod \"apiserver-7bbb656c7d-ngtsq\" (UID: \"40fcb5fe-ee7b-4696-a775-efc703af9b61\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ngtsq" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.045785 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lmxq7\" (UniqueName: \"kubernetes.io/projected/1ceb5fce-77f5-46be-80ff-be6e932bfb35-kube-api-access-lmxq7\") pod \"console-operator-58897d9998-scdc2\" (UID: \"1ceb5fce-77f5-46be-80ff-be6e932bfb35\") " pod="openshift-console-operator/console-operator-58897d9998-scdc2" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.061998 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5jxh9\" (UniqueName: \"kubernetes.io/projected/70342a10-423d-4657-9fc0-062ab9ab8902-kube-api-access-5jxh9\") pod \"controller-manager-879f6c89f-966nh\" (UID: \"70342a10-423d-4657-9fc0-062ab9ab8902\") " pod="openshift-controller-manager/controller-manager-879f6c89f-966nh" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.095700 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jlhqp\" (UniqueName: \"kubernetes.io/projected/8617aa45-3d34-4ab9-9870-67f8ccaeec16-kube-api-access-jlhqp\") pod \"machine-approver-56656f9798-gzpjj\" (UID: \"8617aa45-3d34-4ab9-9870-67f8ccaeec16\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gzpjj" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.098985 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gzpjj" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.105798 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7drzs\" (UniqueName: \"kubernetes.io/projected/f02fb0c2-81fc-45ed-8742-f2f8a51a5437-kube-api-access-7drzs\") pod \"openshift-controller-manager-operator-756b6f6bc6-44gqd\" (UID: \"f02fb0c2-81fc-45ed-8742-f2f8a51a5437\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-44gqd" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.130551 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tr7xh\" (UniqueName: \"kubernetes.io/projected/e4968496-d03e-4450-b85d-240cd75bd900-kube-api-access-tr7xh\") pod \"openshift-apiserver-operator-796bbdcf4f-sxb6c\" (UID: \"e4968496-d03e-4450-b85d-240cd75bd900\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-sxb6c" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.145122 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k9qt7\" (UniqueName: \"kubernetes.io/projected/6c0cd275-710b-43d8-a9b9-b46bc379454b-kube-api-access-k9qt7\") pod \"apiserver-76f77b778f-h5vrh\" (UID: \"6c0cd275-710b-43d8-a9b9-b46bc379454b\") " pod="openshift-apiserver/apiserver-76f77b778f-h5vrh" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.160788 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ngtsq" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.164203 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zkpf8\" (UniqueName: \"kubernetes.io/projected/348075cc-bd79-42f2-bbe2-0e5367092963-kube-api-access-zkpf8\") pod \"cluster-samples-operator-665b6dd947-zh65k\" (UID: \"348075cc-bd79-42f2-bbe2-0e5367092963\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-zh65k" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.167494 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-44gqd" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.179886 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nkmcl\" (UniqueName: \"kubernetes.io/projected/1eaf78eb-9719-4737-9f33-a8ab1dcb5dcf-kube-api-access-nkmcl\") pod \"openshift-config-operator-7777fb866f-4h9v4\" (UID: \"1eaf78eb-9719-4737-9f33-a8ab1dcb5dcf\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-4h9v4" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.194130 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-sxb6c" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.195054 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-scdc2" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.207586 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lpkq9\" (UniqueName: \"kubernetes.io/projected/d874a84d-48e3-4cb2-ac74-4f0db52e9d75-kube-api-access-lpkq9\") pod \"console-f9d7485db-cnkss\" (UID: \"d874a84d-48e3-4cb2-ac74-4f0db52e9d75\") " pod="openshift-console/console-f9d7485db-cnkss" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.223405 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4w4ht\" (UniqueName: \"kubernetes.io/projected/c4b729bd-782f-4393-b162-16c02eb3cb71-kube-api-access-4w4ht\") pod \"downloads-7954f5f757-7k5qb\" (UID: \"c4b729bd-782f-4393-b162-16c02eb3cb71\") " pod="openshift-console/downloads-7954f5f757-7k5qb" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.224998 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.269391 4764 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.271021 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kcl59\" (UniqueName: \"kubernetes.io/projected/08bc3386-ee9a-4a3c-aa52-22d667971076-kube-api-access-kcl59\") pod \"machine-api-operator-5694c8668f-98wxj\" (UID: \"08bc3386-ee9a-4a3c-aa52-22d667971076\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-98wxj" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.287579 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.306885 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.327199 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.332245 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-h5vrh" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.346619 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.364487 4764 request.go:700] Waited for 1.853780134s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-ingress-canary/secrets?fieldSelector=metadata.name%3Ddefault-dockercfg-2llfx&limit=500&resourceVersion=0 Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.367228 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.378917 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-98wxj" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.379182 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-zh65k" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.382139 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-7k5qb" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.385506 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.405065 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.408895 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-44gqd"] Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.418162 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-4h9v4" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.430415 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.451531 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.460588 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-cnkss" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.467301 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.485755 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.502247 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-sxb6c"] Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.505256 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.509418 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/70342a10-423d-4657-9fc0-062ab9ab8902-config\") pod \"controller-manager-879f6c89f-966nh\" (UID: \"70342a10-423d-4657-9fc0-062ab9ab8902\") " pod="openshift-controller-manager/controller-manager-879f6c89f-966nh" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.515915 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gzpjj" event={"ID":"8617aa45-3d34-4ab9-9870-67f8ccaeec16","Type":"ContainerStarted","Data":"3cbb8ecd9aa3f669d0a5a0bc595bd07c36246d78ae3ae788a41ef09e8a0b3b75"} Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.515982 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gzpjj" event={"ID":"8617aa45-3d34-4ab9-9870-67f8ccaeec16","Type":"ContainerStarted","Data":"8f0a4a18c75924ada1859348ef0426c18c6102038b3e5e833221fcb6dd309fed"} Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.527131 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-44gqd" event={"ID":"f02fb0c2-81fc-45ed-8742-f2f8a51a5437","Type":"ContainerStarted","Data":"f7e2af026d012c16b0f29c12ae3d1a15268e422404902685f4dae02b4b6da3a1"} Feb 02 09:09:40 crc kubenswrapper[4764]: W0202 09:09:40.573975 4764 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode4968496_d03e_4450_b85d_240cd75bd900.slice/crio-be225897b40b16ddaddf1f37c81f62152ea40d57355be05f48f0a87041236bb6 WatchSource:0}: Error finding container be225897b40b16ddaddf1f37c81f62152ea40d57355be05f48f0a87041236bb6: Status 404 returned error can't find the container with id be225897b40b16ddaddf1f37c81f62152ea40d57355be05f48f0a87041236bb6 Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.583234 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-h5vrh"] Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.598899 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-ngtsq"] Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.613793 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-966nh" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.621266 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/856a5985-578c-48e4-a535-01496383f3d4-srv-cert\") pod \"olm-operator-6b444d44fb-52th4\" (UID: \"856a5985-578c-48e4-a535-01496383f3d4\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-52th4" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.621315 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a1f840f6-138f-496d-b64d-fbe6d49e0bdd-metrics-tls\") pod \"ingress-operator-5b745b69d9-2m47l\" (UID: \"a1f840f6-138f-496d-b64d-fbe6d49e0bdd\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2m47l" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.621343 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pq5xc\" (UniqueName: \"kubernetes.io/projected/545e08e9-b2c4-493e-8fc7-8353fcab998d-kube-api-access-pq5xc\") pod \"control-plane-machine-set-operator-78cbb6b69f-qq96b\" (UID: \"545e08e9-b2c4-493e-8fc7-8353fcab998d\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qq96b" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.621387 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/37d3c5dc-0535-4f2e-aa2f-debd99e0f1e3-config\") pod \"route-controller-manager-6576b87f9c-4298w\" (UID: \"37d3c5dc-0535-4f2e-aa2f-debd99e0f1e3\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4298w" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.621401 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fp6f6\" (UniqueName: \"kubernetes.io/projected/856a5985-578c-48e4-a535-01496383f3d4-kube-api-access-fp6f6\") pod \"olm-operator-6b444d44fb-52th4\" (UID: \"856a5985-578c-48e4-a535-01496383f3d4\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-52th4" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.621417 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fb629\" (UniqueName: \"kubernetes.io/projected/dc0a29a2-27b1-4f7a-afaa-a0ae6cd290e9-kube-api-access-fb629\") pod \"authentication-operator-69f744f599-gp4jg\" (UID: \"dc0a29a2-27b1-4f7a-afaa-a0ae6cd290e9\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-gp4jg" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.621431 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/1ea380a7-6b81-4938-a6f0-0be18293846d-profile-collector-cert\") pod \"catalog-operator-68c6474976-mmcq4\" (UID: \"1ea380a7-6b81-4938-a6f0-0be18293846d\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mmcq4" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.621456 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/513ecde9-f189-4a22-8742-1b262523d462-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-9mdwc\" (UID: \"513ecde9-f189-4a22-8742-1b262523d462\") " pod="openshift-authentication/oauth-openshift-558db77b4-9mdwc" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.621473 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/91e66500-e26e-4421-83e2-62b738656d5d-signing-cabundle\") pod \"service-ca-9c57cc56f-drqfg\" (UID: \"91e66500-e26e-4421-83e2-62b738656d5d\") " pod="openshift-service-ca/service-ca-9c57cc56f-drqfg" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.621488 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/37d3c5dc-0535-4f2e-aa2f-debd99e0f1e3-client-ca\") pod \"route-controller-manager-6576b87f9c-4298w\" (UID: \"37d3c5dc-0535-4f2e-aa2f-debd99e0f1e3\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4298w" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.621503 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9q5cz\" (UniqueName: \"kubernetes.io/projected/37d3c5dc-0535-4f2e-aa2f-debd99e0f1e3-kube-api-access-9q5cz\") pod \"route-controller-manager-6576b87f9c-4298w\" (UID: \"37d3c5dc-0535-4f2e-aa2f-debd99e0f1e3\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4298w" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.621520 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/545e08e9-b2c4-493e-8fc7-8353fcab998d-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-qq96b\" (UID: \"545e08e9-b2c4-493e-8fc7-8353fcab998d\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qq96b" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.621545 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/cb1b5d6b-7d42-4b40-99e3-4827a5379133-tmpfs\") pod \"packageserver-d55dfcdfc-njspk\" (UID: \"cb1b5d6b-7d42-4b40-99e3-4827a5379133\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-njspk" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.621559 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/cb1b5d6b-7d42-4b40-99e3-4827a5379133-apiservice-cert\") pod \"packageserver-d55dfcdfc-njspk\" (UID: \"cb1b5d6b-7d42-4b40-99e3-4827a5379133\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-njspk" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.621573 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/513ecde9-f189-4a22-8742-1b262523d462-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-9mdwc\" (UID: \"513ecde9-f189-4a22-8742-1b262523d462\") " pod="openshift-authentication/oauth-openshift-558db77b4-9mdwc" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.621588 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dc0a29a2-27b1-4f7a-afaa-a0ae6cd290e9-config\") pod \"authentication-operator-69f744f599-gp4jg\" (UID: \"dc0a29a2-27b1-4f7a-afaa-a0ae6cd290e9\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-gp4jg" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.621612 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/527edab8-f88f-4d82-b42e-b8d52fcaa7d4-registry-certificates\") pod \"image-registry-697d97f7c8-9nkvv\" (UID: \"527edab8-f88f-4d82-b42e-b8d52fcaa7d4\") " pod="openshift-image-registry/image-registry-697d97f7c8-9nkvv" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.621630 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a1f840f6-138f-496d-b64d-fbe6d49e0bdd-trusted-ca\") pod \"ingress-operator-5b745b69d9-2m47l\" (UID: \"a1f840f6-138f-496d-b64d-fbe6d49e0bdd\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2m47l" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.621657 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/527edab8-f88f-4d82-b42e-b8d52fcaa7d4-registry-tls\") pod \"image-registry-697d97f7c8-9nkvv\" (UID: \"527edab8-f88f-4d82-b42e-b8d52fcaa7d4\") " pod="openshift-image-registry/image-registry-697d97f7c8-9nkvv" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.621683 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k27lp\" (UniqueName: \"kubernetes.io/projected/d7534d2e-fb32-4895-945b-ea5ccd35fcbb-kube-api-access-k27lp\") pod \"etcd-operator-b45778765-df5xv\" (UID: \"d7534d2e-fb32-4895-945b-ea5ccd35fcbb\") " pod="openshift-etcd-operator/etcd-operator-b45778765-df5xv" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.621699 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/513ecde9-f189-4a22-8742-1b262523d462-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-9mdwc\" (UID: \"513ecde9-f189-4a22-8742-1b262523d462\") " pod="openshift-authentication/oauth-openshift-558db77b4-9mdwc" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.621724 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a1f840f6-138f-496d-b64d-fbe6d49e0bdd-bound-sa-token\") pod \"ingress-operator-5b745b69d9-2m47l\" (UID: \"a1f840f6-138f-496d-b64d-fbe6d49e0bdd\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2m47l" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.621742 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/513ecde9-f189-4a22-8742-1b262523d462-audit-dir\") pod \"oauth-openshift-558db77b4-9mdwc\" (UID: \"513ecde9-f189-4a22-8742-1b262523d462\") " pod="openshift-authentication/oauth-openshift-558db77b4-9mdwc" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.621793 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f552c254-ade6-4fe4-baa7-04e53c67479b-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-6dljc\" (UID: \"f552c254-ade6-4fe4-baa7-04e53c67479b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-6dljc" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.621815 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dc0a29a2-27b1-4f7a-afaa-a0ae6cd290e9-serving-cert\") pod \"authentication-operator-69f744f599-gp4jg\" (UID: \"dc0a29a2-27b1-4f7a-afaa-a0ae6cd290e9\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-gp4jg" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.621832 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/83fed8a7-bd5f-463d-b70e-3a01700e8315-stats-auth\") pod \"router-default-5444994796-mkwq2\" (UID: \"83fed8a7-bd5f-463d-b70e-3a01700e8315\") " pod="openshift-ingress/router-default-5444994796-mkwq2" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.621848 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/88f8519e-b576-4bef-96d6-771b78c67c11-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-dlpkk\" (UID: \"88f8519e-b576-4bef-96d6-771b78c67c11\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-dlpkk" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.621869 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9nkvv\" (UID: \"527edab8-f88f-4d82-b42e-b8d52fcaa7d4\") " pod="openshift-image-registry/image-registry-697d97f7c8-9nkvv" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.621884 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/36de8e6a-591b-4365-b434-ebd8197abf41-proxy-tls\") pod \"machine-config-operator-74547568cd-2cdx8\" (UID: \"36de8e6a-591b-4365-b434-ebd8197abf41\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2cdx8" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.621899 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/83fed8a7-bd5f-463d-b70e-3a01700e8315-service-ca-bundle\") pod \"router-default-5444994796-mkwq2\" (UID: \"83fed8a7-bd5f-463d-b70e-3a01700e8315\") " pod="openshift-ingress/router-default-5444994796-mkwq2" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.621914 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dl7xr\" (UniqueName: \"kubernetes.io/projected/91e66500-e26e-4421-83e2-62b738656d5d-kube-api-access-dl7xr\") pod \"service-ca-9c57cc56f-drqfg\" (UID: \"91e66500-e26e-4421-83e2-62b738656d5d\") " pod="openshift-service-ca/service-ca-9c57cc56f-drqfg" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.621928 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/513ecde9-f189-4a22-8742-1b262523d462-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-9mdwc\" (UID: \"513ecde9-f189-4a22-8742-1b262523d462\") " pod="openshift-authentication/oauth-openshift-558db77b4-9mdwc" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.621985 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-95srz\" (UniqueName: \"kubernetes.io/projected/527edab8-f88f-4d82-b42e-b8d52fcaa7d4-kube-api-access-95srz\") pod \"image-registry-697d97f7c8-9nkvv\" (UID: \"527edab8-f88f-4d82-b42e-b8d52fcaa7d4\") " pod="openshift-image-registry/image-registry-697d97f7c8-9nkvv" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.622011 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/527edab8-f88f-4d82-b42e-b8d52fcaa7d4-ca-trust-extracted\") pod \"image-registry-697d97f7c8-9nkvv\" (UID: \"527edab8-f88f-4d82-b42e-b8d52fcaa7d4\") " pod="openshift-image-registry/image-registry-697d97f7c8-9nkvv" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.622033 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/1ea380a7-6b81-4938-a6f0-0be18293846d-srv-cert\") pod \"catalog-operator-68c6474976-mmcq4\" (UID: \"1ea380a7-6b81-4938-a6f0-0be18293846d\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mmcq4" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.622070 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/513ecde9-f189-4a22-8742-1b262523d462-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-9mdwc\" (UID: \"513ecde9-f189-4a22-8742-1b262523d462\") " pod="openshift-authentication/oauth-openshift-558db77b4-9mdwc" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.622086 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5l4gj\" (UniqueName: \"kubernetes.io/projected/36de8e6a-591b-4365-b434-ebd8197abf41-kube-api-access-5l4gj\") pod \"machine-config-operator-74547568cd-2cdx8\" (UID: \"36de8e6a-591b-4365-b434-ebd8197abf41\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2cdx8" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.622110 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/527edab8-f88f-4d82-b42e-b8d52fcaa7d4-bound-sa-token\") pod \"image-registry-697d97f7c8-9nkvv\" (UID: \"527edab8-f88f-4d82-b42e-b8d52fcaa7d4\") " pod="openshift-image-registry/image-registry-697d97f7c8-9nkvv" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.622125 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-srt2d\" (UniqueName: \"kubernetes.io/projected/83fed8a7-bd5f-463d-b70e-3a01700e8315-kube-api-access-srt2d\") pod \"router-default-5444994796-mkwq2\" (UID: \"83fed8a7-bd5f-463d-b70e-3a01700e8315\") " pod="openshift-ingress/router-default-5444994796-mkwq2" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.622174 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/513ecde9-f189-4a22-8742-1b262523d462-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-9mdwc\" (UID: \"513ecde9-f189-4a22-8742-1b262523d462\") " pod="openshift-authentication/oauth-openshift-558db77b4-9mdwc" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.622221 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6g6m8\" (UniqueName: \"kubernetes.io/projected/fea30e22-e331-4518-a777-b6258c053f06-kube-api-access-6g6m8\") pod \"dns-operator-744455d44c-tkx8j\" (UID: \"fea30e22-e331-4518-a777-b6258c053f06\") " pod="openshift-dns-operator/dns-operator-744455d44c-tkx8j" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.622236 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qx9cd\" (UniqueName: \"kubernetes.io/projected/1ea380a7-6b81-4938-a6f0-0be18293846d-kube-api-access-qx9cd\") pod \"catalog-operator-68c6474976-mmcq4\" (UID: \"1ea380a7-6b81-4938-a6f0-0be18293846d\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mmcq4" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.622270 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4btgj\" (UniqueName: \"kubernetes.io/projected/006ce277-7c6b-49a5-9e98-0bf7caee30f9-kube-api-access-4btgj\") pod \"collect-profiles-29500380-2gwrp\" (UID: \"006ce277-7c6b-49a5-9e98-0bf7caee30f9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500380-2gwrp" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.622286 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/36de8e6a-591b-4365-b434-ebd8197abf41-images\") pod \"machine-config-operator-74547568cd-2cdx8\" (UID: \"36de8e6a-591b-4365-b434-ebd8197abf41\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2cdx8" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.622302 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/006ce277-7c6b-49a5-9e98-0bf7caee30f9-config-volume\") pod \"collect-profiles-29500380-2gwrp\" (UID: \"006ce277-7c6b-49a5-9e98-0bf7caee30f9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500380-2gwrp" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.622326 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/cb1b5d6b-7d42-4b40-99e3-4827a5379133-webhook-cert\") pod \"packageserver-d55dfcdfc-njspk\" (UID: \"cb1b5d6b-7d42-4b40-99e3-4827a5379133\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-njspk" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.622354 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/d7534d2e-fb32-4895-945b-ea5ccd35fcbb-etcd-service-ca\") pod \"etcd-operator-b45778765-df5xv\" (UID: \"d7534d2e-fb32-4895-945b-ea5ccd35fcbb\") " pod="openshift-etcd-operator/etcd-operator-b45778765-df5xv" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.622369 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/f552c254-ade6-4fe4-baa7-04e53c67479b-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-6dljc\" (UID: \"f552c254-ade6-4fe4-baa7-04e53c67479b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-6dljc" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.622383 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/83fed8a7-bd5f-463d-b70e-3a01700e8315-default-certificate\") pod \"router-default-5444994796-mkwq2\" (UID: \"83fed8a7-bd5f-463d-b70e-3a01700e8315\") " pod="openshift-ingress/router-default-5444994796-mkwq2" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.622409 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/d7534d2e-fb32-4895-945b-ea5ccd35fcbb-etcd-ca\") pod \"etcd-operator-b45778765-df5xv\" (UID: \"d7534d2e-fb32-4895-945b-ea5ccd35fcbb\") " pod="openshift-etcd-operator/etcd-operator-b45778765-df5xv" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.622425 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f9cpr\" (UniqueName: \"kubernetes.io/projected/a1f840f6-138f-496d-b64d-fbe6d49e0bdd-kube-api-access-f9cpr\") pod \"ingress-operator-5b745b69d9-2m47l\" (UID: \"a1f840f6-138f-496d-b64d-fbe6d49e0bdd\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2m47l" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.622441 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/36de8e6a-591b-4365-b434-ebd8197abf41-auth-proxy-config\") pod \"machine-config-operator-74547568cd-2cdx8\" (UID: \"36de8e6a-591b-4365-b434-ebd8197abf41\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2cdx8" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.622456 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f552c254-ade6-4fe4-baa7-04e53c67479b-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-6dljc\" (UID: \"f552c254-ade6-4fe4-baa7-04e53c67479b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-6dljc" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.622481 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d7534d2e-fb32-4895-945b-ea5ccd35fcbb-config\") pod \"etcd-operator-b45778765-df5xv\" (UID: \"d7534d2e-fb32-4895-945b-ea5ccd35fcbb\") " pod="openshift-etcd-operator/etcd-operator-b45778765-df5xv" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.622506 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/513ecde9-f189-4a22-8742-1b262523d462-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-9mdwc\" (UID: \"513ecde9-f189-4a22-8742-1b262523d462\") " pod="openshift-authentication/oauth-openshift-558db77b4-9mdwc" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.622524 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/513ecde9-f189-4a22-8742-1b262523d462-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-9mdwc\" (UID: \"513ecde9-f189-4a22-8742-1b262523d462\") " pod="openshift-authentication/oauth-openshift-558db77b4-9mdwc" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.622541 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/dc0a29a2-27b1-4f7a-afaa-a0ae6cd290e9-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-gp4jg\" (UID: \"dc0a29a2-27b1-4f7a-afaa-a0ae6cd290e9\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-gp4jg" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.622574 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/527edab8-f88f-4d82-b42e-b8d52fcaa7d4-installation-pull-secrets\") pod \"image-registry-697d97f7c8-9nkvv\" (UID: \"527edab8-f88f-4d82-b42e-b8d52fcaa7d4\") " pod="openshift-image-registry/image-registry-697d97f7c8-9nkvv" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.622619 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/856a5985-578c-48e4-a535-01496383f3d4-profile-collector-cert\") pod \"olm-operator-6b444d44fb-52th4\" (UID: \"856a5985-578c-48e4-a535-01496383f3d4\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-52th4" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.622641 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/91e66500-e26e-4421-83e2-62b738656d5d-signing-key\") pod \"service-ca-9c57cc56f-drqfg\" (UID: \"91e66500-e26e-4421-83e2-62b738656d5d\") " pod="openshift-service-ca/service-ca-9c57cc56f-drqfg" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.622663 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d7534d2e-fb32-4895-945b-ea5ccd35fcbb-serving-cert\") pod \"etcd-operator-b45778765-df5xv\" (UID: \"d7534d2e-fb32-4895-945b-ea5ccd35fcbb\") " pod="openshift-etcd-operator/etcd-operator-b45778765-df5xv" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.622694 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/513ecde9-f189-4a22-8742-1b262523d462-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-9mdwc\" (UID: \"513ecde9-f189-4a22-8742-1b262523d462\") " pod="openshift-authentication/oauth-openshift-558db77b4-9mdwc" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.622713 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4t7sg\" (UniqueName: \"kubernetes.io/projected/513ecde9-f189-4a22-8742-1b262523d462-kube-api-access-4t7sg\") pod \"oauth-openshift-558db77b4-9mdwc\" (UID: \"513ecde9-f189-4a22-8742-1b262523d462\") " pod="openshift-authentication/oauth-openshift-558db77b4-9mdwc" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.622760 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/720e4205-b87d-4fc7-95df-3ad7c4849f97-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-n59jp\" (UID: \"720e4205-b87d-4fc7-95df-3ad7c4849f97\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-n59jp" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.624033 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/83fed8a7-bd5f-463d-b70e-3a01700e8315-metrics-certs\") pod \"router-default-5444994796-mkwq2\" (UID: \"83fed8a7-bd5f-463d-b70e-3a01700e8315\") " pod="openshift-ingress/router-default-5444994796-mkwq2" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.624058 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s2jjf\" (UniqueName: \"kubernetes.io/projected/f552c254-ade6-4fe4-baa7-04e53c67479b-kube-api-access-s2jjf\") pod \"cluster-image-registry-operator-dc59b4c8b-6dljc\" (UID: \"f552c254-ade6-4fe4-baa7-04e53c67479b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-6dljc" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.624077 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/fea30e22-e331-4518-a777-b6258c053f06-metrics-tls\") pod \"dns-operator-744455d44c-tkx8j\" (UID: \"fea30e22-e331-4518-a777-b6258c053f06\") " pod="openshift-dns-operator/dns-operator-744455d44c-tkx8j" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.624132 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/513ecde9-f189-4a22-8742-1b262523d462-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-9mdwc\" (UID: \"513ecde9-f189-4a22-8742-1b262523d462\") " pod="openshift-authentication/oauth-openshift-558db77b4-9mdwc" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.624150 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/720e4205-b87d-4fc7-95df-3ad7c4849f97-config\") pod \"kube-apiserver-operator-766d6c64bb-n59jp\" (UID: \"720e4205-b87d-4fc7-95df-3ad7c4849f97\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-n59jp" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.624169 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/513ecde9-f189-4a22-8742-1b262523d462-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-9mdwc\" (UID: \"513ecde9-f189-4a22-8742-1b262523d462\") " pod="openshift-authentication/oauth-openshift-558db77b4-9mdwc" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.624189 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s8z52\" (UniqueName: \"kubernetes.io/projected/f353fe2c-cc41-447a-b4b8-45eed47a29e2-kube-api-access-s8z52\") pod \"multus-admission-controller-857f4d67dd-fdcsz\" (UID: \"f353fe2c-cc41-447a-b4b8-45eed47a29e2\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-fdcsz" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.624234 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/37d3c5dc-0535-4f2e-aa2f-debd99e0f1e3-serving-cert\") pod \"route-controller-manager-6576b87f9c-4298w\" (UID: \"37d3c5dc-0535-4f2e-aa2f-debd99e0f1e3\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4298w" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.624264 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/f353fe2c-cc41-447a-b4b8-45eed47a29e2-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-fdcsz\" (UID: \"f353fe2c-cc41-447a-b4b8-45eed47a29e2\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-fdcsz" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.624279 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/720e4205-b87d-4fc7-95df-3ad7c4849f97-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-n59jp\" (UID: \"720e4205-b87d-4fc7-95df-3ad7c4849f97\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-n59jp" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.624294 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qr7gn\" (UniqueName: \"kubernetes.io/projected/cb1b5d6b-7d42-4b40-99e3-4827a5379133-kube-api-access-qr7gn\") pod \"packageserver-d55dfcdfc-njspk\" (UID: \"cb1b5d6b-7d42-4b40-99e3-4827a5379133\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-njspk" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.624312 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d7534d2e-fb32-4895-945b-ea5ccd35fcbb-etcd-client\") pod \"etcd-operator-b45778765-df5xv\" (UID: \"d7534d2e-fb32-4895-945b-ea5ccd35fcbb\") " pod="openshift-etcd-operator/etcd-operator-b45778765-df5xv" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.624327 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/513ecde9-f189-4a22-8742-1b262523d462-audit-policies\") pod \"oauth-openshift-558db77b4-9mdwc\" (UID: \"513ecde9-f189-4a22-8742-1b262523d462\") " pod="openshift-authentication/oauth-openshift-558db77b4-9mdwc" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.624346 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/006ce277-7c6b-49a5-9e98-0bf7caee30f9-secret-volume\") pod \"collect-profiles-29500380-2gwrp\" (UID: \"006ce277-7c6b-49a5-9e98-0bf7caee30f9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500380-2gwrp" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.624364 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/527edab8-f88f-4d82-b42e-b8d52fcaa7d4-trusted-ca\") pod \"image-registry-697d97f7c8-9nkvv\" (UID: \"527edab8-f88f-4d82-b42e-b8d52fcaa7d4\") " pod="openshift-image-registry/image-registry-697d97f7c8-9nkvv" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.624383 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/88f8519e-b576-4bef-96d6-771b78c67c11-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-dlpkk\" (UID: \"88f8519e-b576-4bef-96d6-771b78c67c11\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-dlpkk" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.624399 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/dc0a29a2-27b1-4f7a-afaa-a0ae6cd290e9-service-ca-bundle\") pod \"authentication-operator-69f744f599-gp4jg\" (UID: \"dc0a29a2-27b1-4f7a-afaa-a0ae6cd290e9\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-gp4jg" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.624435 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/88f8519e-b576-4bef-96d6-771b78c67c11-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-dlpkk\" (UID: \"88f8519e-b576-4bef-96d6-771b78c67c11\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-dlpkk" Feb 02 09:09:40 crc kubenswrapper[4764]: E0202 09:09:40.627656 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 09:09:41.127638571 +0000 UTC m=+144.061362759 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9nkvv" (UID: "527edab8-f88f-4d82-b42e-b8d52fcaa7d4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.675422 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-98wxj"] Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.716794 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-scdc2"] Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.725201 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.725377 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/37d3c5dc-0535-4f2e-aa2f-debd99e0f1e3-config\") pod \"route-controller-manager-6576b87f9c-4298w\" (UID: \"37d3c5dc-0535-4f2e-aa2f-debd99e0f1e3\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4298w" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.725398 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fp6f6\" (UniqueName: \"kubernetes.io/projected/856a5985-578c-48e4-a535-01496383f3d4-kube-api-access-fp6f6\") pod \"olm-operator-6b444d44fb-52th4\" (UID: \"856a5985-578c-48e4-a535-01496383f3d4\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-52th4" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.725421 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/d17edd1e-7d52-4c32-90a9-7d16723f3875-node-bootstrap-token\") pod \"machine-config-server-b894n\" (UID: \"d17edd1e-7d52-4c32-90a9-7d16723f3875\") " pod="openshift-machine-config-operator/machine-config-server-b894n" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.725445 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/513ecde9-f189-4a22-8742-1b262523d462-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-9mdwc\" (UID: \"513ecde9-f189-4a22-8742-1b262523d462\") " pod="openshift-authentication/oauth-openshift-558db77b4-9mdwc" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.725462 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fb629\" (UniqueName: \"kubernetes.io/projected/dc0a29a2-27b1-4f7a-afaa-a0ae6cd290e9-kube-api-access-fb629\") pod \"authentication-operator-69f744f599-gp4jg\" (UID: \"dc0a29a2-27b1-4f7a-afaa-a0ae6cd290e9\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-gp4jg" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.725479 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/1ea380a7-6b81-4938-a6f0-0be18293846d-profile-collector-cert\") pod \"catalog-operator-68c6474976-mmcq4\" (UID: \"1ea380a7-6b81-4938-a6f0-0be18293846d\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mmcq4" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.725494 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/91e66500-e26e-4421-83e2-62b738656d5d-signing-cabundle\") pod \"service-ca-9c57cc56f-drqfg\" (UID: \"91e66500-e26e-4421-83e2-62b738656d5d\") " pod="openshift-service-ca/service-ca-9c57cc56f-drqfg" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.725509 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/501c9952-ea7c-4f45-b181-e25abae6986b-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-9flm5\" (UID: \"501c9952-ea7c-4f45-b181-e25abae6986b\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9flm5" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.725525 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/37d3c5dc-0535-4f2e-aa2f-debd99e0f1e3-client-ca\") pod \"route-controller-manager-6576b87f9c-4298w\" (UID: \"37d3c5dc-0535-4f2e-aa2f-debd99e0f1e3\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4298w" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.725540 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9q5cz\" (UniqueName: \"kubernetes.io/projected/37d3c5dc-0535-4f2e-aa2f-debd99e0f1e3-kube-api-access-9q5cz\") pod \"route-controller-manager-6576b87f9c-4298w\" (UID: \"37d3c5dc-0535-4f2e-aa2f-debd99e0f1e3\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4298w" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.725557 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/545e08e9-b2c4-493e-8fc7-8353fcab998d-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-qq96b\" (UID: \"545e08e9-b2c4-493e-8fc7-8353fcab998d\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qq96b" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.725573 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c8a9e64e-fad9-466c-84b9-2d33a9f20a3b-config-volume\") pod \"dns-default-7v5xl\" (UID: \"c8a9e64e-fad9-466c-84b9-2d33a9f20a3b\") " pod="openshift-dns/dns-default-7v5xl" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.725591 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/cb1b5d6b-7d42-4b40-99e3-4827a5379133-tmpfs\") pod \"packageserver-d55dfcdfc-njspk\" (UID: \"cb1b5d6b-7d42-4b40-99e3-4827a5379133\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-njspk" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.725607 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/7fb1a6fb-ea84-468e-a6ca-02a3604f81a7-proxy-tls\") pod \"machine-config-controller-84d6567774-tpldd\" (UID: \"7fb1a6fb-ea84-468e-a6ca-02a3604f81a7\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-tpldd" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.725626 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/fac965f9-5e91-4c50-a538-48c1301fa072-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-wqbtk\" (UID: \"fac965f9-5e91-4c50-a538-48c1301fa072\") " pod="openshift-marketplace/marketplace-operator-79b997595-wqbtk" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.725641 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/502b6ff7-05d4-41c4-b41a-31b70fb6247c-cert\") pod \"ingress-canary-hfwsg\" (UID: \"502b6ff7-05d4-41c4-b41a-31b70fb6247c\") " pod="openshift-ingress-canary/ingress-canary-hfwsg" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.725654 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/cb1b5d6b-7d42-4b40-99e3-4827a5379133-apiservice-cert\") pod \"packageserver-d55dfcdfc-njspk\" (UID: \"cb1b5d6b-7d42-4b40-99e3-4827a5379133\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-njspk" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.725668 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/513ecde9-f189-4a22-8742-1b262523d462-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-9mdwc\" (UID: \"513ecde9-f189-4a22-8742-1b262523d462\") " pod="openshift-authentication/oauth-openshift-558db77b4-9mdwc" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.725685 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/527edab8-f88f-4d82-b42e-b8d52fcaa7d4-registry-certificates\") pod \"image-registry-697d97f7c8-9nkvv\" (UID: \"527edab8-f88f-4d82-b42e-b8d52fcaa7d4\") " pod="openshift-image-registry/image-registry-697d97f7c8-9nkvv" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.725699 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a1f840f6-138f-496d-b64d-fbe6d49e0bdd-trusted-ca\") pod \"ingress-operator-5b745b69d9-2m47l\" (UID: \"a1f840f6-138f-496d-b64d-fbe6d49e0bdd\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2m47l" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.725717 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dc0a29a2-27b1-4f7a-afaa-a0ae6cd290e9-config\") pod \"authentication-operator-69f744f599-gp4jg\" (UID: \"dc0a29a2-27b1-4f7a-afaa-a0ae6cd290e9\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-gp4jg" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.725733 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/d17edd1e-7d52-4c32-90a9-7d16723f3875-certs\") pod \"machine-config-server-b894n\" (UID: \"d17edd1e-7d52-4c32-90a9-7d16723f3875\") " pod="openshift-machine-config-operator/machine-config-server-b894n" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.725748 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/527edab8-f88f-4d82-b42e-b8d52fcaa7d4-registry-tls\") pod \"image-registry-697d97f7c8-9nkvv\" (UID: \"527edab8-f88f-4d82-b42e-b8d52fcaa7d4\") " pod="openshift-image-registry/image-registry-697d97f7c8-9nkvv" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.725764 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/513ecde9-f189-4a22-8742-1b262523d462-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-9mdwc\" (UID: \"513ecde9-f189-4a22-8742-1b262523d462\") " pod="openshift-authentication/oauth-openshift-558db77b4-9mdwc" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.725779 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/7fb1a6fb-ea84-468e-a6ca-02a3604f81a7-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-tpldd\" (UID: \"7fb1a6fb-ea84-468e-a6ca-02a3604f81a7\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-tpldd" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.725795 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k27lp\" (UniqueName: \"kubernetes.io/projected/d7534d2e-fb32-4895-945b-ea5ccd35fcbb-kube-api-access-k27lp\") pod \"etcd-operator-b45778765-df5xv\" (UID: \"d7534d2e-fb32-4895-945b-ea5ccd35fcbb\") " pod="openshift-etcd-operator/etcd-operator-b45778765-df5xv" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.725810 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/513ecde9-f189-4a22-8742-1b262523d462-audit-dir\") pod \"oauth-openshift-558db77b4-9mdwc\" (UID: \"513ecde9-f189-4a22-8742-1b262523d462\") " pod="openshift-authentication/oauth-openshift-558db77b4-9mdwc" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.725824 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a1f840f6-138f-496d-b64d-fbe6d49e0bdd-bound-sa-token\") pod \"ingress-operator-5b745b69d9-2m47l\" (UID: \"a1f840f6-138f-496d-b64d-fbe6d49e0bdd\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2m47l" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.725842 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/83fed8a7-bd5f-463d-b70e-3a01700e8315-stats-auth\") pod \"router-default-5444994796-mkwq2\" (UID: \"83fed8a7-bd5f-463d-b70e-3a01700e8315\") " pod="openshift-ingress/router-default-5444994796-mkwq2" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.725856 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f552c254-ade6-4fe4-baa7-04e53c67479b-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-6dljc\" (UID: \"f552c254-ade6-4fe4-baa7-04e53c67479b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-6dljc" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.725871 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dc0a29a2-27b1-4f7a-afaa-a0ae6cd290e9-serving-cert\") pod \"authentication-operator-69f744f599-gp4jg\" (UID: \"dc0a29a2-27b1-4f7a-afaa-a0ae6cd290e9\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-gp4jg" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.725886 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aec51473-5be4-4a0b-9ccc-7fcd912f9cf4-config\") pod \"service-ca-operator-777779d784-7mtl6\" (UID: \"aec51473-5be4-4a0b-9ccc-7fcd912f9cf4\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-7mtl6" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.725910 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/88f8519e-b576-4bef-96d6-771b78c67c11-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-dlpkk\" (UID: \"88f8519e-b576-4bef-96d6-771b78c67c11\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-dlpkk" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.725925 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dl7xr\" (UniqueName: \"kubernetes.io/projected/91e66500-e26e-4421-83e2-62b738656d5d-kube-api-access-dl7xr\") pod \"service-ca-9c57cc56f-drqfg\" (UID: \"91e66500-e26e-4421-83e2-62b738656d5d\") " pod="openshift-service-ca/service-ca-9c57cc56f-drqfg" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.725955 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/513ecde9-f189-4a22-8742-1b262523d462-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-9mdwc\" (UID: \"513ecde9-f189-4a22-8742-1b262523d462\") " pod="openshift-authentication/oauth-openshift-558db77b4-9mdwc" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.725970 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/36de8e6a-591b-4365-b434-ebd8197abf41-proxy-tls\") pod \"machine-config-operator-74547568cd-2cdx8\" (UID: \"36de8e6a-591b-4365-b434-ebd8197abf41\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2cdx8" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.725986 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/83fed8a7-bd5f-463d-b70e-3a01700e8315-service-ca-bundle\") pod \"router-default-5444994796-mkwq2\" (UID: \"83fed8a7-bd5f-463d-b70e-3a01700e8315\") " pod="openshift-ingress/router-default-5444994796-mkwq2" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.726017 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-95srz\" (UniqueName: \"kubernetes.io/projected/527edab8-f88f-4d82-b42e-b8d52fcaa7d4-kube-api-access-95srz\") pod \"image-registry-697d97f7c8-9nkvv\" (UID: \"527edab8-f88f-4d82-b42e-b8d52fcaa7d4\") " pod="openshift-image-registry/image-registry-697d97f7c8-9nkvv" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.726032 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/527edab8-f88f-4d82-b42e-b8d52fcaa7d4-ca-trust-extracted\") pod \"image-registry-697d97f7c8-9nkvv\" (UID: \"527edab8-f88f-4d82-b42e-b8d52fcaa7d4\") " pod="openshift-image-registry/image-registry-697d97f7c8-9nkvv" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.726047 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/1ea380a7-6b81-4938-a6f0-0be18293846d-srv-cert\") pod \"catalog-operator-68c6474976-mmcq4\" (UID: \"1ea380a7-6b81-4938-a6f0-0be18293846d\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mmcq4" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.726063 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9a4ebc92-5ef5-4088-a509-5399beef3c7d-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-l5jcj\" (UID: \"9a4ebc92-5ef5-4088-a509-5399beef3c7d\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-l5jcj" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.726079 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/3ebf456e-6cb3-4cd4-a20d-a189ce1b9333-csi-data-dir\") pod \"csi-hostpathplugin-t5q9w\" (UID: \"3ebf456e-6cb3-4cd4-a20d-a189ce1b9333\") " pod="hostpath-provisioner/csi-hostpathplugin-t5q9w" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.726108 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5l4gj\" (UniqueName: \"kubernetes.io/projected/36de8e6a-591b-4365-b434-ebd8197abf41-kube-api-access-5l4gj\") pod \"machine-config-operator-74547568cd-2cdx8\" (UID: \"36de8e6a-591b-4365-b434-ebd8197abf41\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2cdx8" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.726125 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9c69v\" (UniqueName: \"kubernetes.io/projected/d17edd1e-7d52-4c32-90a9-7d16723f3875-kube-api-access-9c69v\") pod \"machine-config-server-b894n\" (UID: \"d17edd1e-7d52-4c32-90a9-7d16723f3875\") " pod="openshift-machine-config-operator/machine-config-server-b894n" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.726142 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/513ecde9-f189-4a22-8742-1b262523d462-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-9mdwc\" (UID: \"513ecde9-f189-4a22-8742-1b262523d462\") " pod="openshift-authentication/oauth-openshift-558db77b4-9mdwc" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.726159 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/3ebf456e-6cb3-4cd4-a20d-a189ce1b9333-socket-dir\") pod \"csi-hostpathplugin-t5q9w\" (UID: \"3ebf456e-6cb3-4cd4-a20d-a189ce1b9333\") " pod="hostpath-provisioner/csi-hostpathplugin-t5q9w" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.726174 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qpvm6\" (UniqueName: \"kubernetes.io/projected/c8a9e64e-fad9-466c-84b9-2d33a9f20a3b-kube-api-access-qpvm6\") pod \"dns-default-7v5xl\" (UID: \"c8a9e64e-fad9-466c-84b9-2d33a9f20a3b\") " pod="openshift-dns/dns-default-7v5xl" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.726190 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/527edab8-f88f-4d82-b42e-b8d52fcaa7d4-bound-sa-token\") pod \"image-registry-697d97f7c8-9nkvv\" (UID: \"527edab8-f88f-4d82-b42e-b8d52fcaa7d4\") " pod="openshift-image-registry/image-registry-697d97f7c8-9nkvv" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.726206 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-srt2d\" (UniqueName: \"kubernetes.io/projected/83fed8a7-bd5f-463d-b70e-3a01700e8315-kube-api-access-srt2d\") pod \"router-default-5444994796-mkwq2\" (UID: \"83fed8a7-bd5f-463d-b70e-3a01700e8315\") " pod="openshift-ingress/router-default-5444994796-mkwq2" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.726229 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/42967bdb-847e-4823-b7b4-979844830e87-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-p66d5\" (UID: \"42967bdb-847e-4823-b7b4-979844830e87\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-p66d5" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.726245 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7rk5j\" (UniqueName: \"kubernetes.io/projected/3ebf456e-6cb3-4cd4-a20d-a189ce1b9333-kube-api-access-7rk5j\") pod \"csi-hostpathplugin-t5q9w\" (UID: \"3ebf456e-6cb3-4cd4-a20d-a189ce1b9333\") " pod="hostpath-provisioner/csi-hostpathplugin-t5q9w" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.726260 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9a4ebc92-5ef5-4088-a509-5399beef3c7d-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-l5jcj\" (UID: \"9a4ebc92-5ef5-4088-a509-5399beef3c7d\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-l5jcj" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.726277 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/fac965f9-5e91-4c50-a538-48c1301fa072-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-wqbtk\" (UID: \"fac965f9-5e91-4c50-a538-48c1301fa072\") " pod="openshift-marketplace/marketplace-operator-79b997595-wqbtk" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.726293 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-99z4n\" (UniqueName: \"kubernetes.io/projected/9a4ebc92-5ef5-4088-a509-5399beef3c7d-kube-api-access-99z4n\") pod \"kube-storage-version-migrator-operator-b67b599dd-l5jcj\" (UID: \"9a4ebc92-5ef5-4088-a509-5399beef3c7d\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-l5jcj" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.726309 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/513ecde9-f189-4a22-8742-1b262523d462-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-9mdwc\" (UID: \"513ecde9-f189-4a22-8742-1b262523d462\") " pod="openshift-authentication/oauth-openshift-558db77b4-9mdwc" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.726339 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6g6m8\" (UniqueName: \"kubernetes.io/projected/fea30e22-e331-4518-a777-b6258c053f06-kube-api-access-6g6m8\") pod \"dns-operator-744455d44c-tkx8j\" (UID: \"fea30e22-e331-4518-a777-b6258c053f06\") " pod="openshift-dns-operator/dns-operator-744455d44c-tkx8j" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.726354 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qx9cd\" (UniqueName: \"kubernetes.io/projected/1ea380a7-6b81-4938-a6f0-0be18293846d-kube-api-access-qx9cd\") pod \"catalog-operator-68c6474976-mmcq4\" (UID: \"1ea380a7-6b81-4938-a6f0-0be18293846d\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mmcq4" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.726370 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hq2fm\" (UniqueName: \"kubernetes.io/projected/7df533b7-06a3-4a8f-abfa-d7a45272d93d-kube-api-access-hq2fm\") pod \"migrator-59844c95c7-fffd9\" (UID: \"7df533b7-06a3-4a8f-abfa-d7a45272d93d\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-fffd9" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.726387 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4btgj\" (UniqueName: \"kubernetes.io/projected/006ce277-7c6b-49a5-9e98-0bf7caee30f9-kube-api-access-4btgj\") pod \"collect-profiles-29500380-2gwrp\" (UID: \"006ce277-7c6b-49a5-9e98-0bf7caee30f9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500380-2gwrp" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.726403 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/cb1b5d6b-7d42-4b40-99e3-4827a5379133-webhook-cert\") pod \"packageserver-d55dfcdfc-njspk\" (UID: \"cb1b5d6b-7d42-4b40-99e3-4827a5379133\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-njspk" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.726419 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/36de8e6a-591b-4365-b434-ebd8197abf41-images\") pod \"machine-config-operator-74547568cd-2cdx8\" (UID: \"36de8e6a-591b-4365-b434-ebd8197abf41\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2cdx8" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.726433 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/006ce277-7c6b-49a5-9e98-0bf7caee30f9-config-volume\") pod \"collect-profiles-29500380-2gwrp\" (UID: \"006ce277-7c6b-49a5-9e98-0bf7caee30f9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500380-2gwrp" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.726448 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/d7534d2e-fb32-4895-945b-ea5ccd35fcbb-etcd-service-ca\") pod \"etcd-operator-b45778765-df5xv\" (UID: \"d7534d2e-fb32-4895-945b-ea5ccd35fcbb\") " pod="openshift-etcd-operator/etcd-operator-b45778765-df5xv" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.726464 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/f552c254-ade6-4fe4-baa7-04e53c67479b-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-6dljc\" (UID: \"f552c254-ade6-4fe4-baa7-04e53c67479b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-6dljc" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.726481 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/501c9952-ea7c-4f45-b181-e25abae6986b-config\") pod \"kube-controller-manager-operator-78b949d7b-9flm5\" (UID: \"501c9952-ea7c-4f45-b181-e25abae6986b\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9flm5" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.726504 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/83fed8a7-bd5f-463d-b70e-3a01700e8315-default-certificate\") pod \"router-default-5444994796-mkwq2\" (UID: \"83fed8a7-bd5f-463d-b70e-3a01700e8315\") " pod="openshift-ingress/router-default-5444994796-mkwq2" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.726519 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/d7534d2e-fb32-4895-945b-ea5ccd35fcbb-etcd-ca\") pod \"etcd-operator-b45778765-df5xv\" (UID: \"d7534d2e-fb32-4895-945b-ea5ccd35fcbb\") " pod="openshift-etcd-operator/etcd-operator-b45778765-df5xv" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.726535 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f9cpr\" (UniqueName: \"kubernetes.io/projected/a1f840f6-138f-496d-b64d-fbe6d49e0bdd-kube-api-access-f9cpr\") pod \"ingress-operator-5b745b69d9-2m47l\" (UID: \"a1f840f6-138f-496d-b64d-fbe6d49e0bdd\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2m47l" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.726551 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/36de8e6a-591b-4365-b434-ebd8197abf41-auth-proxy-config\") pod \"machine-config-operator-74547568cd-2cdx8\" (UID: \"36de8e6a-591b-4365-b434-ebd8197abf41\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2cdx8" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.726568 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f552c254-ade6-4fe4-baa7-04e53c67479b-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-6dljc\" (UID: \"f552c254-ade6-4fe4-baa7-04e53c67479b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-6dljc" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.726583 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l7mpv\" (UniqueName: \"kubernetes.io/projected/7fb1a6fb-ea84-468e-a6ca-02a3604f81a7-kube-api-access-l7mpv\") pod \"machine-config-controller-84d6567774-tpldd\" (UID: \"7fb1a6fb-ea84-468e-a6ca-02a3604f81a7\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-tpldd" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.726599 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d7534d2e-fb32-4895-945b-ea5ccd35fcbb-config\") pod \"etcd-operator-b45778765-df5xv\" (UID: \"d7534d2e-fb32-4895-945b-ea5ccd35fcbb\") " pod="openshift-etcd-operator/etcd-operator-b45778765-df5xv" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.726615 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/3ebf456e-6cb3-4cd4-a20d-a189ce1b9333-plugins-dir\") pod \"csi-hostpathplugin-t5q9w\" (UID: \"3ebf456e-6cb3-4cd4-a20d-a189ce1b9333\") " pod="hostpath-provisioner/csi-hostpathplugin-t5q9w" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.726629 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d8vr6\" (UniqueName: \"kubernetes.io/projected/502b6ff7-05d4-41c4-b41a-31b70fb6247c-kube-api-access-d8vr6\") pod \"ingress-canary-hfwsg\" (UID: \"502b6ff7-05d4-41c4-b41a-31b70fb6247c\") " pod="openshift-ingress-canary/ingress-canary-hfwsg" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.726645 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/513ecde9-f189-4a22-8742-1b262523d462-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-9mdwc\" (UID: \"513ecde9-f189-4a22-8742-1b262523d462\") " pod="openshift-authentication/oauth-openshift-558db77b4-9mdwc" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.726661 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/527edab8-f88f-4d82-b42e-b8d52fcaa7d4-installation-pull-secrets\") pod \"image-registry-697d97f7c8-9nkvv\" (UID: \"527edab8-f88f-4d82-b42e-b8d52fcaa7d4\") " pod="openshift-image-registry/image-registry-697d97f7c8-9nkvv" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.726677 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/513ecde9-f189-4a22-8742-1b262523d462-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-9mdwc\" (UID: \"513ecde9-f189-4a22-8742-1b262523d462\") " pod="openshift-authentication/oauth-openshift-558db77b4-9mdwc" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.726693 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/dc0a29a2-27b1-4f7a-afaa-a0ae6cd290e9-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-gp4jg\" (UID: \"dc0a29a2-27b1-4f7a-afaa-a0ae6cd290e9\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-gp4jg" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.726708 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/c8a9e64e-fad9-466c-84b9-2d33a9f20a3b-metrics-tls\") pod \"dns-default-7v5xl\" (UID: \"c8a9e64e-fad9-466c-84b9-2d33a9f20a3b\") " pod="openshift-dns/dns-default-7v5xl" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.726728 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/856a5985-578c-48e4-a535-01496383f3d4-profile-collector-cert\") pod \"olm-operator-6b444d44fb-52th4\" (UID: \"856a5985-578c-48e4-a535-01496383f3d4\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-52th4" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.726743 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/91e66500-e26e-4421-83e2-62b738656d5d-signing-key\") pod \"service-ca-9c57cc56f-drqfg\" (UID: \"91e66500-e26e-4421-83e2-62b738656d5d\") " pod="openshift-service-ca/service-ca-9c57cc56f-drqfg" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.726758 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d7534d2e-fb32-4895-945b-ea5ccd35fcbb-serving-cert\") pod \"etcd-operator-b45778765-df5xv\" (UID: \"d7534d2e-fb32-4895-945b-ea5ccd35fcbb\") " pod="openshift-etcd-operator/etcd-operator-b45778765-df5xv" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.726772 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4t7sg\" (UniqueName: \"kubernetes.io/projected/513ecde9-f189-4a22-8742-1b262523d462-kube-api-access-4t7sg\") pod \"oauth-openshift-558db77b4-9mdwc\" (UID: \"513ecde9-f189-4a22-8742-1b262523d462\") " pod="openshift-authentication/oauth-openshift-558db77b4-9mdwc" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.726787 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/720e4205-b87d-4fc7-95df-3ad7c4849f97-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-n59jp\" (UID: \"720e4205-b87d-4fc7-95df-3ad7c4849f97\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-n59jp" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.726802 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/aec51473-5be4-4a0b-9ccc-7fcd912f9cf4-serving-cert\") pod \"service-ca-operator-777779d784-7mtl6\" (UID: \"aec51473-5be4-4a0b-9ccc-7fcd912f9cf4\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-7mtl6" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.726819 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/513ecde9-f189-4a22-8742-1b262523d462-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-9mdwc\" (UID: \"513ecde9-f189-4a22-8742-1b262523d462\") " pod="openshift-authentication/oauth-openshift-558db77b4-9mdwc" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.726834 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/83fed8a7-bd5f-463d-b70e-3a01700e8315-metrics-certs\") pod \"router-default-5444994796-mkwq2\" (UID: \"83fed8a7-bd5f-463d-b70e-3a01700e8315\") " pod="openshift-ingress/router-default-5444994796-mkwq2" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.726850 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2jjf\" (UniqueName: \"kubernetes.io/projected/f552c254-ade6-4fe4-baa7-04e53c67479b-kube-api-access-s2jjf\") pod \"cluster-image-registry-operator-dc59b4c8b-6dljc\" (UID: \"f552c254-ade6-4fe4-baa7-04e53c67479b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-6dljc" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.726865 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/513ecde9-f189-4a22-8742-1b262523d462-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-9mdwc\" (UID: \"513ecde9-f189-4a22-8742-1b262523d462\") " pod="openshift-authentication/oauth-openshift-558db77b4-9mdwc" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.726880 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/720e4205-b87d-4fc7-95df-3ad7c4849f97-config\") pod \"kube-apiserver-operator-766d6c64bb-n59jp\" (UID: \"720e4205-b87d-4fc7-95df-3ad7c4849f97\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-n59jp" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.726894 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/fea30e22-e331-4518-a777-b6258c053f06-metrics-tls\") pod \"dns-operator-744455d44c-tkx8j\" (UID: \"fea30e22-e331-4518-a777-b6258c053f06\") " pod="openshift-dns-operator/dns-operator-744455d44c-tkx8j" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.726909 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/3ebf456e-6cb3-4cd4-a20d-a189ce1b9333-registration-dir\") pod \"csi-hostpathplugin-t5q9w\" (UID: \"3ebf456e-6cb3-4cd4-a20d-a189ce1b9333\") " pod="hostpath-provisioner/csi-hostpathplugin-t5q9w" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.726926 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/513ecde9-f189-4a22-8742-1b262523d462-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-9mdwc\" (UID: \"513ecde9-f189-4a22-8742-1b262523d462\") " pod="openshift-authentication/oauth-openshift-558db77b4-9mdwc" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.726956 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s8z52\" (UniqueName: \"kubernetes.io/projected/f353fe2c-cc41-447a-b4b8-45eed47a29e2-kube-api-access-s8z52\") pod \"multus-admission-controller-857f4d67dd-fdcsz\" (UID: \"f353fe2c-cc41-447a-b4b8-45eed47a29e2\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-fdcsz" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.726972 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8jxj7\" (UniqueName: \"kubernetes.io/projected/aec51473-5be4-4a0b-9ccc-7fcd912f9cf4-kube-api-access-8jxj7\") pod \"service-ca-operator-777779d784-7mtl6\" (UID: \"aec51473-5be4-4a0b-9ccc-7fcd912f9cf4\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-7mtl6" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.726989 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/3ebf456e-6cb3-4cd4-a20d-a189ce1b9333-mountpoint-dir\") pod \"csi-hostpathplugin-t5q9w\" (UID: \"3ebf456e-6cb3-4cd4-a20d-a189ce1b9333\") " pod="hostpath-provisioner/csi-hostpathplugin-t5q9w" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.727004 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z7blj\" (UniqueName: \"kubernetes.io/projected/42967bdb-847e-4823-b7b4-979844830e87-kube-api-access-z7blj\") pod \"package-server-manager-789f6589d5-p66d5\" (UID: \"42967bdb-847e-4823-b7b4-979844830e87\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-p66d5" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.727021 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/37d3c5dc-0535-4f2e-aa2f-debd99e0f1e3-serving-cert\") pod \"route-controller-manager-6576b87f9c-4298w\" (UID: \"37d3c5dc-0535-4f2e-aa2f-debd99e0f1e3\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4298w" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.727043 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/720e4205-b87d-4fc7-95df-3ad7c4849f97-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-n59jp\" (UID: \"720e4205-b87d-4fc7-95df-3ad7c4849f97\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-n59jp" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.727058 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/f353fe2c-cc41-447a-b4b8-45eed47a29e2-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-fdcsz\" (UID: \"f353fe2c-cc41-447a-b4b8-45eed47a29e2\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-fdcsz" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.727074 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/513ecde9-f189-4a22-8742-1b262523d462-audit-policies\") pod \"oauth-openshift-558db77b4-9mdwc\" (UID: \"513ecde9-f189-4a22-8742-1b262523d462\") " pod="openshift-authentication/oauth-openshift-558db77b4-9mdwc" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.727088 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/006ce277-7c6b-49a5-9e98-0bf7caee30f9-secret-volume\") pod \"collect-profiles-29500380-2gwrp\" (UID: \"006ce277-7c6b-49a5-9e98-0bf7caee30f9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500380-2gwrp" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.727113 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qr7gn\" (UniqueName: \"kubernetes.io/projected/cb1b5d6b-7d42-4b40-99e3-4827a5379133-kube-api-access-qr7gn\") pod \"packageserver-d55dfcdfc-njspk\" (UID: \"cb1b5d6b-7d42-4b40-99e3-4827a5379133\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-njspk" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.727129 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d7534d2e-fb32-4895-945b-ea5ccd35fcbb-etcd-client\") pod \"etcd-operator-b45778765-df5xv\" (UID: \"d7534d2e-fb32-4895-945b-ea5ccd35fcbb\") " pod="openshift-etcd-operator/etcd-operator-b45778765-df5xv" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.727144 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/501c9952-ea7c-4f45-b181-e25abae6986b-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-9flm5\" (UID: \"501c9952-ea7c-4f45-b181-e25abae6986b\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9flm5" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.727160 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/527edab8-f88f-4d82-b42e-b8d52fcaa7d4-trusted-ca\") pod \"image-registry-697d97f7c8-9nkvv\" (UID: \"527edab8-f88f-4d82-b42e-b8d52fcaa7d4\") " pod="openshift-image-registry/image-registry-697d97f7c8-9nkvv" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.727176 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/88f8519e-b576-4bef-96d6-771b78c67c11-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-dlpkk\" (UID: \"88f8519e-b576-4bef-96d6-771b78c67c11\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-dlpkk" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.727191 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/88f8519e-b576-4bef-96d6-771b78c67c11-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-dlpkk\" (UID: \"88f8519e-b576-4bef-96d6-771b78c67c11\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-dlpkk" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.727208 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/dc0a29a2-27b1-4f7a-afaa-a0ae6cd290e9-service-ca-bundle\") pod \"authentication-operator-69f744f599-gp4jg\" (UID: \"dc0a29a2-27b1-4f7a-afaa-a0ae6cd290e9\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-gp4jg" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.727224 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s9xmw\" (UniqueName: \"kubernetes.io/projected/fac965f9-5e91-4c50-a538-48c1301fa072-kube-api-access-s9xmw\") pod \"marketplace-operator-79b997595-wqbtk\" (UID: \"fac965f9-5e91-4c50-a538-48c1301fa072\") " pod="openshift-marketplace/marketplace-operator-79b997595-wqbtk" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.727239 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/856a5985-578c-48e4-a535-01496383f3d4-srv-cert\") pod \"olm-operator-6b444d44fb-52th4\" (UID: \"856a5985-578c-48e4-a535-01496383f3d4\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-52th4" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.727254 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a1f840f6-138f-496d-b64d-fbe6d49e0bdd-metrics-tls\") pod \"ingress-operator-5b745b69d9-2m47l\" (UID: \"a1f840f6-138f-496d-b64d-fbe6d49e0bdd\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2m47l" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.727270 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pq5xc\" (UniqueName: \"kubernetes.io/projected/545e08e9-b2c4-493e-8fc7-8353fcab998d-kube-api-access-pq5xc\") pod \"control-plane-machine-set-operator-78cbb6b69f-qq96b\" (UID: \"545e08e9-b2c4-493e-8fc7-8353fcab998d\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qq96b" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.727461 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/513ecde9-f189-4a22-8742-1b262523d462-audit-dir\") pod \"oauth-openshift-558db77b4-9mdwc\" (UID: \"513ecde9-f189-4a22-8742-1b262523d462\") " pod="openshift-authentication/oauth-openshift-558db77b4-9mdwc" Feb 02 09:09:40 crc kubenswrapper[4764]: E0202 09:09:40.727531 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 09:09:41.227517553 +0000 UTC m=+144.161241641 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.728398 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/37d3c5dc-0535-4f2e-aa2f-debd99e0f1e3-config\") pod \"route-controller-manager-6576b87f9c-4298w\" (UID: \"37d3c5dc-0535-4f2e-aa2f-debd99e0f1e3\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4298w" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.730698 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/36de8e6a-591b-4365-b434-ebd8197abf41-images\") pod \"machine-config-operator-74547568cd-2cdx8\" (UID: \"36de8e6a-591b-4365-b434-ebd8197abf41\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2cdx8" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.733539 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/006ce277-7c6b-49a5-9e98-0bf7caee30f9-config-volume\") pod \"collect-profiles-29500380-2gwrp\" (UID: \"006ce277-7c6b-49a5-9e98-0bf7caee30f9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500380-2gwrp" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.733996 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/513ecde9-f189-4a22-8742-1b262523d462-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-9mdwc\" (UID: \"513ecde9-f189-4a22-8742-1b262523d462\") " pod="openshift-authentication/oauth-openshift-558db77b4-9mdwc" Feb 02 09:09:40 crc kubenswrapper[4764]: W0202 09:09:40.737745 4764 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod08bc3386_ee9a_4a3c_aa52_22d667971076.slice/crio-bc95b1a93c8165caa41e29a2af81f19a14151409f5a33b4fb30afdcee1197f0c WatchSource:0}: Error finding container bc95b1a93c8165caa41e29a2af81f19a14151409f5a33b4fb30afdcee1197f0c: Status 404 returned error can't find the container with id bc95b1a93c8165caa41e29a2af81f19a14151409f5a33b4fb30afdcee1197f0c Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.742640 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/513ecde9-f189-4a22-8742-1b262523d462-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-9mdwc\" (UID: \"513ecde9-f189-4a22-8742-1b262523d462\") " pod="openshift-authentication/oauth-openshift-558db77b4-9mdwc" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.745525 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/d7534d2e-fb32-4895-945b-ea5ccd35fcbb-etcd-ca\") pod \"etcd-operator-b45778765-df5xv\" (UID: \"d7534d2e-fb32-4895-945b-ea5ccd35fcbb\") " pod="openshift-etcd-operator/etcd-operator-b45778765-df5xv" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.745528 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/d7534d2e-fb32-4895-945b-ea5ccd35fcbb-etcd-service-ca\") pod \"etcd-operator-b45778765-df5xv\" (UID: \"d7534d2e-fb32-4895-945b-ea5ccd35fcbb\") " pod="openshift-etcd-operator/etcd-operator-b45778765-df5xv" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.749181 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/36de8e6a-591b-4365-b434-ebd8197abf41-auth-proxy-config\") pod \"machine-config-operator-74547568cd-2cdx8\" (UID: \"36de8e6a-591b-4365-b434-ebd8197abf41\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2cdx8" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.749837 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/527edab8-f88f-4d82-b42e-b8d52fcaa7d4-ca-trust-extracted\") pod \"image-registry-697d97f7c8-9nkvv\" (UID: \"527edab8-f88f-4d82-b42e-b8d52fcaa7d4\") " pod="openshift-image-registry/image-registry-697d97f7c8-9nkvv" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.749895 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d7534d2e-fb32-4895-945b-ea5ccd35fcbb-config\") pod \"etcd-operator-b45778765-df5xv\" (UID: \"d7534d2e-fb32-4895-945b-ea5ccd35fcbb\") " pod="openshift-etcd-operator/etcd-operator-b45778765-df5xv" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.754157 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/cb1b5d6b-7d42-4b40-99e3-4827a5379133-webhook-cert\") pod \"packageserver-d55dfcdfc-njspk\" (UID: \"cb1b5d6b-7d42-4b40-99e3-4827a5379133\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-njspk" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.755897 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/83fed8a7-bd5f-463d-b70e-3a01700e8315-service-ca-bundle\") pod \"router-default-5444994796-mkwq2\" (UID: \"83fed8a7-bd5f-463d-b70e-3a01700e8315\") " pod="openshift-ingress/router-default-5444994796-mkwq2" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.756681 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f552c254-ade6-4fe4-baa7-04e53c67479b-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-6dljc\" (UID: \"f552c254-ade6-4fe4-baa7-04e53c67479b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-6dljc" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.757321 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/83fed8a7-bd5f-463d-b70e-3a01700e8315-stats-auth\") pod \"router-default-5444994796-mkwq2\" (UID: \"83fed8a7-bd5f-463d-b70e-3a01700e8315\") " pod="openshift-ingress/router-default-5444994796-mkwq2" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.761484 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/36de8e6a-591b-4365-b434-ebd8197abf41-proxy-tls\") pod \"machine-config-operator-74547568cd-2cdx8\" (UID: \"36de8e6a-591b-4365-b434-ebd8197abf41\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2cdx8" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.762092 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/83fed8a7-bd5f-463d-b70e-3a01700e8315-metrics-certs\") pod \"router-default-5444994796-mkwq2\" (UID: \"83fed8a7-bd5f-463d-b70e-3a01700e8315\") " pod="openshift-ingress/router-default-5444994796-mkwq2" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.762208 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/513ecde9-f189-4a22-8742-1b262523d462-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-9mdwc\" (UID: \"513ecde9-f189-4a22-8742-1b262523d462\") " pod="openshift-authentication/oauth-openshift-558db77b4-9mdwc" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.762312 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/91e66500-e26e-4421-83e2-62b738656d5d-signing-key\") pod \"service-ca-9c57cc56f-drqfg\" (UID: \"91e66500-e26e-4421-83e2-62b738656d5d\") " pod="openshift-service-ca/service-ca-9c57cc56f-drqfg" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.762652 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/1ea380a7-6b81-4938-a6f0-0be18293846d-profile-collector-cert\") pod \"catalog-operator-68c6474976-mmcq4\" (UID: \"1ea380a7-6b81-4938-a6f0-0be18293846d\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mmcq4" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.763050 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/1ea380a7-6b81-4938-a6f0-0be18293846d-srv-cert\") pod \"catalog-operator-68c6474976-mmcq4\" (UID: \"1ea380a7-6b81-4938-a6f0-0be18293846d\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mmcq4" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.763425 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/91e66500-e26e-4421-83e2-62b738656d5d-signing-cabundle\") pod \"service-ca-9c57cc56f-drqfg\" (UID: \"91e66500-e26e-4421-83e2-62b738656d5d\") " pod="openshift-service-ca/service-ca-9c57cc56f-drqfg" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.765706 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/720e4205-b87d-4fc7-95df-3ad7c4849f97-config\") pod \"kube-apiserver-operator-766d6c64bb-n59jp\" (UID: \"720e4205-b87d-4fc7-95df-3ad7c4849f97\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-n59jp" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.766432 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/dc0a29a2-27b1-4f7a-afaa-a0ae6cd290e9-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-gp4jg\" (UID: \"dc0a29a2-27b1-4f7a-afaa-a0ae6cd290e9\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-gp4jg" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.768257 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/513ecde9-f189-4a22-8742-1b262523d462-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-9mdwc\" (UID: \"513ecde9-f189-4a22-8742-1b262523d462\") " pod="openshift-authentication/oauth-openshift-558db77b4-9mdwc" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.768495 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/513ecde9-f189-4a22-8742-1b262523d462-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-9mdwc\" (UID: \"513ecde9-f189-4a22-8742-1b262523d462\") " pod="openshift-authentication/oauth-openshift-558db77b4-9mdwc" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.771327 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dc0a29a2-27b1-4f7a-afaa-a0ae6cd290e9-serving-cert\") pod \"authentication-operator-69f744f599-gp4jg\" (UID: \"dc0a29a2-27b1-4f7a-afaa-a0ae6cd290e9\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-gp4jg" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.769802 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/cb1b5d6b-7d42-4b40-99e3-4827a5379133-tmpfs\") pod \"packageserver-d55dfcdfc-njspk\" (UID: \"cb1b5d6b-7d42-4b40-99e3-4827a5379133\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-njspk" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.776258 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/527edab8-f88f-4d82-b42e-b8d52fcaa7d4-registry-certificates\") pod \"image-registry-697d97f7c8-9nkvv\" (UID: \"527edab8-f88f-4d82-b42e-b8d52fcaa7d4\") " pod="openshift-image-registry/image-registry-697d97f7c8-9nkvv" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.777281 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/513ecde9-f189-4a22-8742-1b262523d462-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-9mdwc\" (UID: \"513ecde9-f189-4a22-8742-1b262523d462\") " pod="openshift-authentication/oauth-openshift-558db77b4-9mdwc" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.783840 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dc0a29a2-27b1-4f7a-afaa-a0ae6cd290e9-config\") pod \"authentication-operator-69f744f599-gp4jg\" (UID: \"dc0a29a2-27b1-4f7a-afaa-a0ae6cd290e9\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-gp4jg" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.783994 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/37d3c5dc-0535-4f2e-aa2f-debd99e0f1e3-client-ca\") pod \"route-controller-manager-6576b87f9c-4298w\" (UID: \"37d3c5dc-0535-4f2e-aa2f-debd99e0f1e3\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4298w" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.784302 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a1f840f6-138f-496d-b64d-fbe6d49e0bdd-trusted-ca\") pod \"ingress-operator-5b745b69d9-2m47l\" (UID: \"a1f840f6-138f-496d-b64d-fbe6d49e0bdd\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2m47l" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.784462 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/513ecde9-f189-4a22-8742-1b262523d462-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-9mdwc\" (UID: \"513ecde9-f189-4a22-8742-1b262523d462\") " pod="openshift-authentication/oauth-openshift-558db77b4-9mdwc" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.786399 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/545e08e9-b2c4-493e-8fc7-8353fcab998d-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-qq96b\" (UID: \"545e08e9-b2c4-493e-8fc7-8353fcab998d\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qq96b" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.787814 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/88f8519e-b576-4bef-96d6-771b78c67c11-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-dlpkk\" (UID: \"88f8519e-b576-4bef-96d6-771b78c67c11\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-dlpkk" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.788316 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/513ecde9-f189-4a22-8742-1b262523d462-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-9mdwc\" (UID: \"513ecde9-f189-4a22-8742-1b262523d462\") " pod="openshift-authentication/oauth-openshift-558db77b4-9mdwc" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.788050 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/513ecde9-f189-4a22-8742-1b262523d462-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-9mdwc\" (UID: \"513ecde9-f189-4a22-8742-1b262523d462\") " pod="openshift-authentication/oauth-openshift-558db77b4-9mdwc" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.790348 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/f552c254-ade6-4fe4-baa7-04e53c67479b-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-6dljc\" (UID: \"f552c254-ade6-4fe4-baa7-04e53c67479b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-6dljc" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.793203 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/83fed8a7-bd5f-463d-b70e-3a01700e8315-default-certificate\") pod \"router-default-5444994796-mkwq2\" (UID: \"83fed8a7-bd5f-463d-b70e-3a01700e8315\") " pod="openshift-ingress/router-default-5444994796-mkwq2" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.794266 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pq5xc\" (UniqueName: \"kubernetes.io/projected/545e08e9-b2c4-493e-8fc7-8353fcab998d-kube-api-access-pq5xc\") pod \"control-plane-machine-set-operator-78cbb6b69f-qq96b\" (UID: \"545e08e9-b2c4-493e-8fc7-8353fcab998d\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qq96b" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.795239 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/527edab8-f88f-4d82-b42e-b8d52fcaa7d4-trusted-ca\") pod \"image-registry-697d97f7c8-9nkvv\" (UID: \"527edab8-f88f-4d82-b42e-b8d52fcaa7d4\") " pod="openshift-image-registry/image-registry-697d97f7c8-9nkvv" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.796315 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/88f8519e-b576-4bef-96d6-771b78c67c11-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-dlpkk\" (UID: \"88f8519e-b576-4bef-96d6-771b78c67c11\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-dlpkk" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.796900 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/dc0a29a2-27b1-4f7a-afaa-a0ae6cd290e9-service-ca-bundle\") pod \"authentication-operator-69f744f599-gp4jg\" (UID: \"dc0a29a2-27b1-4f7a-afaa-a0ae6cd290e9\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-gp4jg" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.799347 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-7k5qb"] Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.803570 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d7534d2e-fb32-4895-945b-ea5ccd35fcbb-serving-cert\") pod \"etcd-operator-b45778765-df5xv\" (UID: \"d7534d2e-fb32-4895-945b-ea5ccd35fcbb\") " pod="openshift-etcd-operator/etcd-operator-b45778765-df5xv" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.803668 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/527edab8-f88f-4d82-b42e-b8d52fcaa7d4-installation-pull-secrets\") pod \"image-registry-697d97f7c8-9nkvv\" (UID: \"527edab8-f88f-4d82-b42e-b8d52fcaa7d4\") " pod="openshift-image-registry/image-registry-697d97f7c8-9nkvv" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.804120 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/527edab8-f88f-4d82-b42e-b8d52fcaa7d4-registry-tls\") pod \"image-registry-697d97f7c8-9nkvv\" (UID: \"527edab8-f88f-4d82-b42e-b8d52fcaa7d4\") " pod="openshift-image-registry/image-registry-697d97f7c8-9nkvv" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.814268 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/513ecde9-f189-4a22-8742-1b262523d462-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-9mdwc\" (UID: \"513ecde9-f189-4a22-8742-1b262523d462\") " pod="openshift-authentication/oauth-openshift-558db77b4-9mdwc" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.814725 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/513ecde9-f189-4a22-8742-1b262523d462-audit-policies\") pod \"oauth-openshift-558db77b4-9mdwc\" (UID: \"513ecde9-f189-4a22-8742-1b262523d462\") " pod="openshift-authentication/oauth-openshift-558db77b4-9mdwc" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.814876 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/513ecde9-f189-4a22-8742-1b262523d462-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-9mdwc\" (UID: \"513ecde9-f189-4a22-8742-1b262523d462\") " pod="openshift-authentication/oauth-openshift-558db77b4-9mdwc" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.820288 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/856a5985-578c-48e4-a535-01496383f3d4-srv-cert\") pod \"olm-operator-6b444d44fb-52th4\" (UID: \"856a5985-578c-48e4-a535-01496383f3d4\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-52th4" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.820340 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/cb1b5d6b-7d42-4b40-99e3-4827a5379133-apiservice-cert\") pod \"packageserver-d55dfcdfc-njspk\" (UID: \"cb1b5d6b-7d42-4b40-99e3-4827a5379133\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-njspk" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.822752 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/fea30e22-e331-4518-a777-b6258c053f06-metrics-tls\") pod \"dns-operator-744455d44c-tkx8j\" (UID: \"fea30e22-e331-4518-a777-b6258c053f06\") " pod="openshift-dns-operator/dns-operator-744455d44c-tkx8j" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.823491 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/006ce277-7c6b-49a5-9e98-0bf7caee30f9-secret-volume\") pod \"collect-profiles-29500380-2gwrp\" (UID: \"006ce277-7c6b-49a5-9e98-0bf7caee30f9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500380-2gwrp" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.826200 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-4h9v4"] Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.826301 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/856a5985-578c-48e4-a535-01496383f3d4-profile-collector-cert\") pod \"olm-operator-6b444d44fb-52th4\" (UID: \"856a5985-578c-48e4-a535-01496383f3d4\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-52th4" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.836401 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a1f840f6-138f-496d-b64d-fbe6d49e0bdd-metrics-tls\") pod \"ingress-operator-5b745b69d9-2m47l\" (UID: \"a1f840f6-138f-496d-b64d-fbe6d49e0bdd\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2m47l" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.841234 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/720e4205-b87d-4fc7-95df-3ad7c4849f97-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-n59jp\" (UID: \"720e4205-b87d-4fc7-95df-3ad7c4849f97\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-n59jp" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.841313 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7rk5j\" (UniqueName: \"kubernetes.io/projected/3ebf456e-6cb3-4cd4-a20d-a189ce1b9333-kube-api-access-7rk5j\") pod \"csi-hostpathplugin-t5q9w\" (UID: \"3ebf456e-6cb3-4cd4-a20d-a189ce1b9333\") " pod="hostpath-provisioner/csi-hostpathplugin-t5q9w" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.841352 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/42967bdb-847e-4823-b7b4-979844830e87-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-p66d5\" (UID: \"42967bdb-847e-4823-b7b4-979844830e87\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-p66d5" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.841372 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/fac965f9-5e91-4c50-a538-48c1301fa072-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-wqbtk\" (UID: \"fac965f9-5e91-4c50-a538-48c1301fa072\") " pod="openshift-marketplace/marketplace-operator-79b997595-wqbtk" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.841391 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9a4ebc92-5ef5-4088-a509-5399beef3c7d-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-l5jcj\" (UID: \"9a4ebc92-5ef5-4088-a509-5399beef3c7d\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-l5jcj" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.841407 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-99z4n\" (UniqueName: \"kubernetes.io/projected/9a4ebc92-5ef5-4088-a509-5399beef3c7d-kube-api-access-99z4n\") pod \"kube-storage-version-migrator-operator-b67b599dd-l5jcj\" (UID: \"9a4ebc92-5ef5-4088-a509-5399beef3c7d\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-l5jcj" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.841456 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hq2fm\" (UniqueName: \"kubernetes.io/projected/7df533b7-06a3-4a8f-abfa-d7a45272d93d-kube-api-access-hq2fm\") pod \"migrator-59844c95c7-fffd9\" (UID: \"7df533b7-06a3-4a8f-abfa-d7a45272d93d\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-fffd9" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.841489 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/501c9952-ea7c-4f45-b181-e25abae6986b-config\") pod \"kube-controller-manager-operator-78b949d7b-9flm5\" (UID: \"501c9952-ea7c-4f45-b181-e25abae6986b\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9flm5" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.841514 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l7mpv\" (UniqueName: \"kubernetes.io/projected/7fb1a6fb-ea84-468e-a6ca-02a3604f81a7-kube-api-access-l7mpv\") pod \"machine-config-controller-84d6567774-tpldd\" (UID: \"7fb1a6fb-ea84-468e-a6ca-02a3604f81a7\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-tpldd" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.841530 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/3ebf456e-6cb3-4cd4-a20d-a189ce1b9333-plugins-dir\") pod \"csi-hostpathplugin-t5q9w\" (UID: \"3ebf456e-6cb3-4cd4-a20d-a189ce1b9333\") " pod="hostpath-provisioner/csi-hostpathplugin-t5q9w" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.841545 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d8vr6\" (UniqueName: \"kubernetes.io/projected/502b6ff7-05d4-41c4-b41a-31b70fb6247c-kube-api-access-d8vr6\") pod \"ingress-canary-hfwsg\" (UID: \"502b6ff7-05d4-41c4-b41a-31b70fb6247c\") " pod="openshift-ingress-canary/ingress-canary-hfwsg" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.841563 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/c8a9e64e-fad9-466c-84b9-2d33a9f20a3b-metrics-tls\") pod \"dns-default-7v5xl\" (UID: \"c8a9e64e-fad9-466c-84b9-2d33a9f20a3b\") " pod="openshift-dns/dns-default-7v5xl" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.841592 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/aec51473-5be4-4a0b-9ccc-7fcd912f9cf4-serving-cert\") pod \"service-ca-operator-777779d784-7mtl6\" (UID: \"aec51473-5be4-4a0b-9ccc-7fcd912f9cf4\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-7mtl6" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.841616 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/3ebf456e-6cb3-4cd4-a20d-a189ce1b9333-registration-dir\") pod \"csi-hostpathplugin-t5q9w\" (UID: \"3ebf456e-6cb3-4cd4-a20d-a189ce1b9333\") " pod="hostpath-provisioner/csi-hostpathplugin-t5q9w" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.841639 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8jxj7\" (UniqueName: \"kubernetes.io/projected/aec51473-5be4-4a0b-9ccc-7fcd912f9cf4-kube-api-access-8jxj7\") pod \"service-ca-operator-777779d784-7mtl6\" (UID: \"aec51473-5be4-4a0b-9ccc-7fcd912f9cf4\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-7mtl6" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.841660 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/3ebf456e-6cb3-4cd4-a20d-a189ce1b9333-mountpoint-dir\") pod \"csi-hostpathplugin-t5q9w\" (UID: \"3ebf456e-6cb3-4cd4-a20d-a189ce1b9333\") " pod="hostpath-provisioner/csi-hostpathplugin-t5q9w" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.841676 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z7blj\" (UniqueName: \"kubernetes.io/projected/42967bdb-847e-4823-b7b4-979844830e87-kube-api-access-z7blj\") pod \"package-server-manager-789f6589d5-p66d5\" (UID: \"42967bdb-847e-4823-b7b4-979844830e87\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-p66d5" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.841715 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/501c9952-ea7c-4f45-b181-e25abae6986b-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-9flm5\" (UID: \"501c9952-ea7c-4f45-b181-e25abae6986b\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9flm5" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.841740 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s9xmw\" (UniqueName: \"kubernetes.io/projected/fac965f9-5e91-4c50-a538-48c1301fa072-kube-api-access-s9xmw\") pod \"marketplace-operator-79b997595-wqbtk\" (UID: \"fac965f9-5e91-4c50-a538-48c1301fa072\") " pod="openshift-marketplace/marketplace-operator-79b997595-wqbtk" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.841769 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/d17edd1e-7d52-4c32-90a9-7d16723f3875-node-bootstrap-token\") pod \"machine-config-server-b894n\" (UID: \"d17edd1e-7d52-4c32-90a9-7d16723f3875\") " pod="openshift-machine-config-operator/machine-config-server-b894n" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.841791 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/501c9952-ea7c-4f45-b181-e25abae6986b-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-9flm5\" (UID: \"501c9952-ea7c-4f45-b181-e25abae6986b\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9flm5" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.841815 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c8a9e64e-fad9-466c-84b9-2d33a9f20a3b-config-volume\") pod \"dns-default-7v5xl\" (UID: \"c8a9e64e-fad9-466c-84b9-2d33a9f20a3b\") " pod="openshift-dns/dns-default-7v5xl" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.841830 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/7fb1a6fb-ea84-468e-a6ca-02a3604f81a7-proxy-tls\") pod \"machine-config-controller-84d6567774-tpldd\" (UID: \"7fb1a6fb-ea84-468e-a6ca-02a3604f81a7\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-tpldd" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.841843 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/fac965f9-5e91-4c50-a538-48c1301fa072-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-wqbtk\" (UID: \"fac965f9-5e91-4c50-a538-48c1301fa072\") " pod="openshift-marketplace/marketplace-operator-79b997595-wqbtk" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.841856 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/502b6ff7-05d4-41c4-b41a-31b70fb6247c-cert\") pod \"ingress-canary-hfwsg\" (UID: \"502b6ff7-05d4-41c4-b41a-31b70fb6247c\") " pod="openshift-ingress-canary/ingress-canary-hfwsg" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.841870 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/d17edd1e-7d52-4c32-90a9-7d16723f3875-certs\") pod \"machine-config-server-b894n\" (UID: \"d17edd1e-7d52-4c32-90a9-7d16723f3875\") " pod="openshift-machine-config-operator/machine-config-server-b894n" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.841892 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/7fb1a6fb-ea84-468e-a6ca-02a3604f81a7-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-tpldd\" (UID: \"7fb1a6fb-ea84-468e-a6ca-02a3604f81a7\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-tpldd" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.841918 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aec51473-5be4-4a0b-9ccc-7fcd912f9cf4-config\") pod \"service-ca-operator-777779d784-7mtl6\" (UID: \"aec51473-5be4-4a0b-9ccc-7fcd912f9cf4\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-7mtl6" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.841957 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9nkvv\" (UID: \"527edab8-f88f-4d82-b42e-b8d52fcaa7d4\") " pod="openshift-image-registry/image-registry-697d97f7c8-9nkvv" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.841989 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9a4ebc92-5ef5-4088-a509-5399beef3c7d-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-l5jcj\" (UID: \"9a4ebc92-5ef5-4088-a509-5399beef3c7d\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-l5jcj" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.842005 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/3ebf456e-6cb3-4cd4-a20d-a189ce1b9333-csi-data-dir\") pod \"csi-hostpathplugin-t5q9w\" (UID: \"3ebf456e-6cb3-4cd4-a20d-a189ce1b9333\") " pod="hostpath-provisioner/csi-hostpathplugin-t5q9w" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.842025 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9c69v\" (UniqueName: \"kubernetes.io/projected/d17edd1e-7d52-4c32-90a9-7d16723f3875-kube-api-access-9c69v\") pod \"machine-config-server-b894n\" (UID: \"d17edd1e-7d52-4c32-90a9-7d16723f3875\") " pod="openshift-machine-config-operator/machine-config-server-b894n" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.842041 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/3ebf456e-6cb3-4cd4-a20d-a189ce1b9333-socket-dir\") pod \"csi-hostpathplugin-t5q9w\" (UID: \"3ebf456e-6cb3-4cd4-a20d-a189ce1b9333\") " pod="hostpath-provisioner/csi-hostpathplugin-t5q9w" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.842065 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qpvm6\" (UniqueName: \"kubernetes.io/projected/c8a9e64e-fad9-466c-84b9-2d33a9f20a3b-kube-api-access-qpvm6\") pod \"dns-default-7v5xl\" (UID: \"c8a9e64e-fad9-466c-84b9-2d33a9f20a3b\") " pod="openshift-dns/dns-default-7v5xl" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.842844 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/501c9952-ea7c-4f45-b181-e25abae6986b-config\") pod \"kube-controller-manager-operator-78b949d7b-9flm5\" (UID: \"501c9952-ea7c-4f45-b181-e25abae6986b\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9flm5" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.844591 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/3ebf456e-6cb3-4cd4-a20d-a189ce1b9333-plugins-dir\") pod \"csi-hostpathplugin-t5q9w\" (UID: \"3ebf456e-6cb3-4cd4-a20d-a189ce1b9333\") " pod="hostpath-provisioner/csi-hostpathplugin-t5q9w" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.845281 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/3ebf456e-6cb3-4cd4-a20d-a189ce1b9333-mountpoint-dir\") pod \"csi-hostpathplugin-t5q9w\" (UID: \"3ebf456e-6cb3-4cd4-a20d-a189ce1b9333\") " pod="hostpath-provisioner/csi-hostpathplugin-t5q9w" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.846961 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/37d3c5dc-0535-4f2e-aa2f-debd99e0f1e3-serving-cert\") pod \"route-controller-manager-6576b87f9c-4298w\" (UID: \"37d3c5dc-0535-4f2e-aa2f-debd99e0f1e3\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4298w" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.847045 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c8a9e64e-fad9-466c-84b9-2d33a9f20a3b-config-volume\") pod \"dns-default-7v5xl\" (UID: \"c8a9e64e-fad9-466c-84b9-2d33a9f20a3b\") " pod="openshift-dns/dns-default-7v5xl" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.847908 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/3ebf456e-6cb3-4cd4-a20d-a189ce1b9333-registration-dir\") pod \"csi-hostpathplugin-t5q9w\" (UID: \"3ebf456e-6cb3-4cd4-a20d-a189ce1b9333\") " pod="hostpath-provisioner/csi-hostpathplugin-t5q9w" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.848497 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aec51473-5be4-4a0b-9ccc-7fcd912f9cf4-config\") pod \"service-ca-operator-777779d784-7mtl6\" (UID: \"aec51473-5be4-4a0b-9ccc-7fcd912f9cf4\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-7mtl6" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.852199 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/3ebf456e-6cb3-4cd4-a20d-a189ce1b9333-csi-data-dir\") pod \"csi-hostpathplugin-t5q9w\" (UID: \"3ebf456e-6cb3-4cd4-a20d-a189ce1b9333\") " pod="hostpath-provisioner/csi-hostpathplugin-t5q9w" Feb 02 09:09:40 crc kubenswrapper[4764]: E0202 09:09:40.852652 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 09:09:41.352635063 +0000 UTC m=+144.286359211 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9nkvv" (UID: "527edab8-f88f-4d82-b42e-b8d52fcaa7d4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.853385 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9a4ebc92-5ef5-4088-a509-5399beef3c7d-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-l5jcj\" (UID: \"9a4ebc92-5ef5-4088-a509-5399beef3c7d\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-l5jcj" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.854153 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/3ebf456e-6cb3-4cd4-a20d-a189ce1b9333-socket-dir\") pod \"csi-hostpathplugin-t5q9w\" (UID: \"3ebf456e-6cb3-4cd4-a20d-a189ce1b9333\") " pod="hostpath-provisioner/csi-hostpathplugin-t5q9w" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.855220 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/7fb1a6fb-ea84-468e-a6ca-02a3604f81a7-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-tpldd\" (UID: \"7fb1a6fb-ea84-468e-a6ca-02a3604f81a7\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-tpldd" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.858726 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/42967bdb-847e-4823-b7b4-979844830e87-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-p66d5\" (UID: \"42967bdb-847e-4823-b7b4-979844830e87\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-p66d5" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.859177 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4btgj\" (UniqueName: \"kubernetes.io/projected/006ce277-7c6b-49a5-9e98-0bf7caee30f9-kube-api-access-4btgj\") pod \"collect-profiles-29500380-2gwrp\" (UID: \"006ce277-7c6b-49a5-9e98-0bf7caee30f9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500380-2gwrp" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.859215 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/fac965f9-5e91-4c50-a538-48c1301fa072-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-wqbtk\" (UID: \"fac965f9-5e91-4c50-a538-48c1301fa072\") " pod="openshift-marketplace/marketplace-operator-79b997595-wqbtk" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.867257 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/502b6ff7-05d4-41c4-b41a-31b70fb6247c-cert\") pod \"ingress-canary-hfwsg\" (UID: \"502b6ff7-05d4-41c4-b41a-31b70fb6247c\") " pod="openshift-ingress-canary/ingress-canary-hfwsg" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.869209 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/aec51473-5be4-4a0b-9ccc-7fcd912f9cf4-serving-cert\") pod \"service-ca-operator-777779d784-7mtl6\" (UID: \"aec51473-5be4-4a0b-9ccc-7fcd912f9cf4\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-7mtl6" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.869999 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a1f840f6-138f-496d-b64d-fbe6d49e0bdd-bound-sa-token\") pod \"ingress-operator-5b745b69d9-2m47l\" (UID: \"a1f840f6-138f-496d-b64d-fbe6d49e0bdd\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2m47l" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.870182 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qx9cd\" (UniqueName: \"kubernetes.io/projected/1ea380a7-6b81-4938-a6f0-0be18293846d-kube-api-access-qx9cd\") pod \"catalog-operator-68c6474976-mmcq4\" (UID: \"1ea380a7-6b81-4938-a6f0-0be18293846d\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mmcq4" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.870404 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d7534d2e-fb32-4895-945b-ea5ccd35fcbb-etcd-client\") pod \"etcd-operator-b45778765-df5xv\" (UID: \"d7534d2e-fb32-4895-945b-ea5ccd35fcbb\") " pod="openshift-etcd-operator/etcd-operator-b45778765-df5xv" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.870794 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/501c9952-ea7c-4f45-b181-e25abae6986b-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-9flm5\" (UID: \"501c9952-ea7c-4f45-b181-e25abae6986b\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9flm5" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.870920 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/c8a9e64e-fad9-466c-84b9-2d33a9f20a3b-metrics-tls\") pod \"dns-default-7v5xl\" (UID: \"c8a9e64e-fad9-466c-84b9-2d33a9f20a3b\") " pod="openshift-dns/dns-default-7v5xl" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.870768 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9a4ebc92-5ef5-4088-a509-5399beef3c7d-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-l5jcj\" (UID: \"9a4ebc92-5ef5-4088-a509-5399beef3c7d\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-l5jcj" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.871092 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6g6m8\" (UniqueName: \"kubernetes.io/projected/fea30e22-e331-4518-a777-b6258c053f06-kube-api-access-6g6m8\") pod \"dns-operator-744455d44c-tkx8j\" (UID: \"fea30e22-e331-4518-a777-b6258c053f06\") " pod="openshift-dns-operator/dns-operator-744455d44c-tkx8j" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.871300 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f552c254-ade6-4fe4-baa7-04e53c67479b-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-6dljc\" (UID: \"f552c254-ade6-4fe4-baa7-04e53c67479b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-6dljc" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.871582 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/f353fe2c-cc41-447a-b4b8-45eed47a29e2-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-fdcsz\" (UID: \"f353fe2c-cc41-447a-b4b8-45eed47a29e2\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-fdcsz" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.878285 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/7fb1a6fb-ea84-468e-a6ca-02a3604f81a7-proxy-tls\") pod \"machine-config-controller-84d6567774-tpldd\" (UID: \"7fb1a6fb-ea84-468e-a6ca-02a3604f81a7\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-tpldd" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.879141 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/d17edd1e-7d52-4c32-90a9-7d16723f3875-certs\") pod \"machine-config-server-b894n\" (UID: \"d17edd1e-7d52-4c32-90a9-7d16723f3875\") " pod="openshift-machine-config-operator/machine-config-server-b894n" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.890195 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/fac965f9-5e91-4c50-a538-48c1301fa072-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-wqbtk\" (UID: \"fac965f9-5e91-4c50-a538-48c1301fa072\") " pod="openshift-marketplace/marketplace-operator-79b997595-wqbtk" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.890342 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/d17edd1e-7d52-4c32-90a9-7d16723f3875-node-bootstrap-token\") pod \"machine-config-server-b894n\" (UID: \"d17edd1e-7d52-4c32-90a9-7d16723f3875\") " pod="openshift-machine-config-operator/machine-config-server-b894n" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.890466 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-tkx8j" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.892611 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fp6f6\" (UniqueName: \"kubernetes.io/projected/856a5985-578c-48e4-a535-01496383f3d4-kube-api-access-fp6f6\") pod \"olm-operator-6b444d44fb-52th4\" (UID: \"856a5985-578c-48e4-a535-01496383f3d4\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-52th4" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.904658 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5l4gj\" (UniqueName: \"kubernetes.io/projected/36de8e6a-591b-4365-b434-ebd8197abf41-kube-api-access-5l4gj\") pod \"machine-config-operator-74547568cd-2cdx8\" (UID: \"36de8e6a-591b-4365-b434-ebd8197abf41\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2cdx8" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.920846 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fb629\" (UniqueName: \"kubernetes.io/projected/dc0a29a2-27b1-4f7a-afaa-a0ae6cd290e9-kube-api-access-fb629\") pod \"authentication-operator-69f744f599-gp4jg\" (UID: \"dc0a29a2-27b1-4f7a-afaa-a0ae6cd290e9\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-gp4jg" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.941885 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/527edab8-f88f-4d82-b42e-b8d52fcaa7d4-bound-sa-token\") pod \"image-registry-697d97f7c8-9nkvv\" (UID: \"527edab8-f88f-4d82-b42e-b8d52fcaa7d4\") " pod="openshift-image-registry/image-registry-697d97f7c8-9nkvv" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.943042 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 09:09:40 crc kubenswrapper[4764]: E0202 09:09:40.944120 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 09:09:41.444069915 +0000 UTC m=+144.377794003 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.947609 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-966nh"] Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.957976 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-52th4" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.962167 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-srt2d\" (UniqueName: \"kubernetes.io/projected/83fed8a7-bd5f-463d-b70e-3a01700e8315-kube-api-access-srt2d\") pod \"router-default-5444994796-mkwq2\" (UID: \"83fed8a7-bd5f-463d-b70e-3a01700e8315\") " pod="openshift-ingress/router-default-5444994796-mkwq2" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.972086 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29500380-2gwrp" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.979787 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qq96b" Feb 02 09:09:40 crc kubenswrapper[4764]: I0202 09:09:40.986039 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2cdx8" Feb 02 09:09:41 crc kubenswrapper[4764]: I0202 09:09:41.001506 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2jjf\" (UniqueName: \"kubernetes.io/projected/f552c254-ade6-4fe4-baa7-04e53c67479b-kube-api-access-s2jjf\") pod \"cluster-image-registry-operator-dc59b4c8b-6dljc\" (UID: \"f552c254-ade6-4fe4-baa7-04e53c67479b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-6dljc" Feb 02 09:09:41 crc kubenswrapper[4764]: I0202 09:09:41.014135 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mmcq4" Feb 02 09:09:41 crc kubenswrapper[4764]: I0202 09:09:41.017674 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f9cpr\" (UniqueName: \"kubernetes.io/projected/a1f840f6-138f-496d-b64d-fbe6d49e0bdd-kube-api-access-f9cpr\") pod \"ingress-operator-5b745b69d9-2m47l\" (UID: \"a1f840f6-138f-496d-b64d-fbe6d49e0bdd\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2m47l" Feb 02 09:09:41 crc kubenswrapper[4764]: I0202 09:09:41.042712 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-95srz\" (UniqueName: \"kubernetes.io/projected/527edab8-f88f-4d82-b42e-b8d52fcaa7d4-kube-api-access-95srz\") pod \"image-registry-697d97f7c8-9nkvv\" (UID: \"527edab8-f88f-4d82-b42e-b8d52fcaa7d4\") " pod="openshift-image-registry/image-registry-697d97f7c8-9nkvv" Feb 02 09:09:41 crc kubenswrapper[4764]: I0202 09:09:41.044685 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9nkvv\" (UID: \"527edab8-f88f-4d82-b42e-b8d52fcaa7d4\") " pod="openshift-image-registry/image-registry-697d97f7c8-9nkvv" Feb 02 09:09:41 crc kubenswrapper[4764]: E0202 09:09:41.048491 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 09:09:41.545012227 +0000 UTC m=+144.478736315 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9nkvv" (UID: "527edab8-f88f-4d82-b42e-b8d52fcaa7d4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:09:41 crc kubenswrapper[4764]: I0202 09:09:41.071863 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9q5cz\" (UniqueName: \"kubernetes.io/projected/37d3c5dc-0535-4f2e-aa2f-debd99e0f1e3-kube-api-access-9q5cz\") pod \"route-controller-manager-6576b87f9c-4298w\" (UID: \"37d3c5dc-0535-4f2e-aa2f-debd99e0f1e3\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4298w" Feb 02 09:09:41 crc kubenswrapper[4764]: I0202 09:09:41.084011 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4t7sg\" (UniqueName: \"kubernetes.io/projected/513ecde9-f189-4a22-8742-1b262523d462-kube-api-access-4t7sg\") pod \"oauth-openshift-558db77b4-9mdwc\" (UID: \"513ecde9-f189-4a22-8742-1b262523d462\") " pod="openshift-authentication/oauth-openshift-558db77b4-9mdwc" Feb 02 09:09:41 crc kubenswrapper[4764]: I0202 09:09:41.086771 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-cnkss"] Feb 02 09:09:41 crc kubenswrapper[4764]: I0202 09:09:41.101978 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-zh65k"] Feb 02 09:09:41 crc kubenswrapper[4764]: I0202 09:09:41.116182 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/720e4205-b87d-4fc7-95df-3ad7c4849f97-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-n59jp\" (UID: \"720e4205-b87d-4fc7-95df-3ad7c4849f97\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-n59jp" Feb 02 09:09:41 crc kubenswrapper[4764]: I0202 09:09:41.126692 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dl7xr\" (UniqueName: \"kubernetes.io/projected/91e66500-e26e-4421-83e2-62b738656d5d-kube-api-access-dl7xr\") pod \"service-ca-9c57cc56f-drqfg\" (UID: \"91e66500-e26e-4421-83e2-62b738656d5d\") " pod="openshift-service-ca/service-ca-9c57cc56f-drqfg" Feb 02 09:09:41 crc kubenswrapper[4764]: I0202 09:09:41.145284 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 09:09:41 crc kubenswrapper[4764]: E0202 09:09:41.145744 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 09:09:41.645729783 +0000 UTC m=+144.579453871 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:09:41 crc kubenswrapper[4764]: I0202 09:09:41.151621 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k27lp\" (UniqueName: \"kubernetes.io/projected/d7534d2e-fb32-4895-945b-ea5ccd35fcbb-kube-api-access-k27lp\") pod \"etcd-operator-b45778765-df5xv\" (UID: \"d7534d2e-fb32-4895-945b-ea5ccd35fcbb\") " pod="openshift-etcd-operator/etcd-operator-b45778765-df5xv" Feb 02 09:09:41 crc kubenswrapper[4764]: I0202 09:09:41.157171 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-gp4jg" Feb 02 09:09:41 crc kubenswrapper[4764]: I0202 09:09:41.171405 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s8z52\" (UniqueName: \"kubernetes.io/projected/f353fe2c-cc41-447a-b4b8-45eed47a29e2-kube-api-access-s8z52\") pod \"multus-admission-controller-857f4d67dd-fdcsz\" (UID: \"f353fe2c-cc41-447a-b4b8-45eed47a29e2\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-fdcsz" Feb 02 09:09:41 crc kubenswrapper[4764]: I0202 09:09:41.172747 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-9mdwc" Feb 02 09:09:41 crc kubenswrapper[4764]: I0202 09:09:41.195735 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-6dljc" Feb 02 09:09:41 crc kubenswrapper[4764]: I0202 09:09:41.197040 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/88f8519e-b576-4bef-96d6-771b78c67c11-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-dlpkk\" (UID: \"88f8519e-b576-4bef-96d6-771b78c67c11\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-dlpkk" Feb 02 09:09:41 crc kubenswrapper[4764]: I0202 09:09:41.208250 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qr7gn\" (UniqueName: \"kubernetes.io/projected/cb1b5d6b-7d42-4b40-99e3-4827a5379133-kube-api-access-qr7gn\") pod \"packageserver-d55dfcdfc-njspk\" (UID: \"cb1b5d6b-7d42-4b40-99e3-4827a5379133\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-njspk" Feb 02 09:09:41 crc kubenswrapper[4764]: I0202 09:09:41.208594 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-df5xv" Feb 02 09:09:41 crc kubenswrapper[4764]: I0202 09:09:41.210682 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-fdcsz" Feb 02 09:09:41 crc kubenswrapper[4764]: I0202 09:09:41.223156 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-njspk" Feb 02 09:09:41 crc kubenswrapper[4764]: I0202 09:09:41.224312 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qpvm6\" (UniqueName: \"kubernetes.io/projected/c8a9e64e-fad9-466c-84b9-2d33a9f20a3b-kube-api-access-qpvm6\") pod \"dns-default-7v5xl\" (UID: \"c8a9e64e-fad9-466c-84b9-2d33a9f20a3b\") " pod="openshift-dns/dns-default-7v5xl" Feb 02 09:09:41 crc kubenswrapper[4764]: I0202 09:09:41.230351 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4298w" Feb 02 09:09:41 crc kubenswrapper[4764]: I0202 09:09:41.238181 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2m47l" Feb 02 09:09:41 crc kubenswrapper[4764]: I0202 09:09:41.242612 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-mkwq2" Feb 02 09:09:41 crc kubenswrapper[4764]: I0202 09:09:41.246594 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9nkvv\" (UID: \"527edab8-f88f-4d82-b42e-b8d52fcaa7d4\") " pod="openshift-image-registry/image-registry-697d97f7c8-9nkvv" Feb 02 09:09:41 crc kubenswrapper[4764]: E0202 09:09:41.246988 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 09:09:41.746976924 +0000 UTC m=+144.680701012 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9nkvv" (UID: "527edab8-f88f-4d82-b42e-b8d52fcaa7d4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:09:41 crc kubenswrapper[4764]: I0202 09:09:41.251023 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-n59jp" Feb 02 09:09:41 crc kubenswrapper[4764]: I0202 09:09:41.252108 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-99z4n\" (UniqueName: \"kubernetes.io/projected/9a4ebc92-5ef5-4088-a509-5399beef3c7d-kube-api-access-99z4n\") pod \"kube-storage-version-migrator-operator-b67b599dd-l5jcj\" (UID: \"9a4ebc92-5ef5-4088-a509-5399beef3c7d\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-l5jcj" Feb 02 09:09:41 crc kubenswrapper[4764]: I0202 09:09:41.269964 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-dlpkk" Feb 02 09:09:41 crc kubenswrapper[4764]: I0202 09:09:41.272239 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hq2fm\" (UniqueName: \"kubernetes.io/projected/7df533b7-06a3-4a8f-abfa-d7a45272d93d-kube-api-access-hq2fm\") pod \"migrator-59844c95c7-fffd9\" (UID: \"7df533b7-06a3-4a8f-abfa-d7a45272d93d\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-fffd9" Feb 02 09:09:41 crc kubenswrapper[4764]: I0202 09:09:41.295143 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-drqfg" Feb 02 09:09:41 crc kubenswrapper[4764]: I0202 09:09:41.298225 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l7mpv\" (UniqueName: \"kubernetes.io/projected/7fb1a6fb-ea84-468e-a6ca-02a3604f81a7-kube-api-access-l7mpv\") pod \"machine-config-controller-84d6567774-tpldd\" (UID: \"7fb1a6fb-ea84-468e-a6ca-02a3604f81a7\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-tpldd" Feb 02 09:09:41 crc kubenswrapper[4764]: I0202 09:09:41.309884 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8jxj7\" (UniqueName: \"kubernetes.io/projected/aec51473-5be4-4a0b-9ccc-7fcd912f9cf4-kube-api-access-8jxj7\") pod \"service-ca-operator-777779d784-7mtl6\" (UID: \"aec51473-5be4-4a0b-9ccc-7fcd912f9cf4\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-7mtl6" Feb 02 09:09:41 crc kubenswrapper[4764]: I0202 09:09:41.314696 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-tkx8j"] Feb 02 09:09:41 crc kubenswrapper[4764]: I0202 09:09:41.322389 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-fffd9" Feb 02 09:09:41 crc kubenswrapper[4764]: I0202 09:09:41.329712 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-l5jcj" Feb 02 09:09:41 crc kubenswrapper[4764]: I0202 09:09:41.342711 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-7mtl6" Feb 02 09:09:41 crc kubenswrapper[4764]: I0202 09:09:41.347534 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 09:09:41 crc kubenswrapper[4764]: E0202 09:09:41.347868 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 09:09:41.847852254 +0000 UTC m=+144.781576342 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:09:41 crc kubenswrapper[4764]: I0202 09:09:41.348480 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-tpldd" Feb 02 09:09:41 crc kubenswrapper[4764]: I0202 09:09:41.349659 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z7blj\" (UniqueName: \"kubernetes.io/projected/42967bdb-847e-4823-b7b4-979844830e87-kube-api-access-z7blj\") pod \"package-server-manager-789f6589d5-p66d5\" (UID: \"42967bdb-847e-4823-b7b4-979844830e87\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-p66d5" Feb 02 09:09:41 crc kubenswrapper[4764]: I0202 09:09:41.351185 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d8vr6\" (UniqueName: \"kubernetes.io/projected/502b6ff7-05d4-41c4-b41a-31b70fb6247c-kube-api-access-d8vr6\") pod \"ingress-canary-hfwsg\" (UID: \"502b6ff7-05d4-41c4-b41a-31b70fb6247c\") " pod="openshift-ingress-canary/ingress-canary-hfwsg" Feb 02 09:09:41 crc kubenswrapper[4764]: I0202 09:09:41.361047 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/501c9952-ea7c-4f45-b181-e25abae6986b-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-9flm5\" (UID: \"501c9952-ea7c-4f45-b181-e25abae6986b\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9flm5" Feb 02 09:09:41 crc kubenswrapper[4764]: I0202 09:09:41.366292 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-52th4"] Feb 02 09:09:41 crc kubenswrapper[4764]: I0202 09:09:41.386732 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s9xmw\" (UniqueName: \"kubernetes.io/projected/fac965f9-5e91-4c50-a538-48c1301fa072-kube-api-access-s9xmw\") pod \"marketplace-operator-79b997595-wqbtk\" (UID: \"fac965f9-5e91-4c50-a538-48c1301fa072\") " pod="openshift-marketplace/marketplace-operator-79b997595-wqbtk" Feb 02 09:09:41 crc kubenswrapper[4764]: I0202 09:09:41.395373 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-hfwsg" Feb 02 09:09:41 crc kubenswrapper[4764]: I0202 09:09:41.399064 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-7v5xl" Feb 02 09:09:41 crc kubenswrapper[4764]: I0202 09:09:41.401182 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7rk5j\" (UniqueName: \"kubernetes.io/projected/3ebf456e-6cb3-4cd4-a20d-a189ce1b9333-kube-api-access-7rk5j\") pod \"csi-hostpathplugin-t5q9w\" (UID: \"3ebf456e-6cb3-4cd4-a20d-a189ce1b9333\") " pod="hostpath-provisioner/csi-hostpathplugin-t5q9w" Feb 02 09:09:41 crc kubenswrapper[4764]: I0202 09:09:41.425451 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9c69v\" (UniqueName: \"kubernetes.io/projected/d17edd1e-7d52-4c32-90a9-7d16723f3875-kube-api-access-9c69v\") pod \"machine-config-server-b894n\" (UID: \"d17edd1e-7d52-4c32-90a9-7d16723f3875\") " pod="openshift-machine-config-operator/machine-config-server-b894n" Feb 02 09:09:41 crc kubenswrapper[4764]: I0202 09:09:41.435797 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29500380-2gwrp"] Feb 02 09:09:41 crc kubenswrapper[4764]: I0202 09:09:41.453074 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9nkvv\" (UID: \"527edab8-f88f-4d82-b42e-b8d52fcaa7d4\") " pod="openshift-image-registry/image-registry-697d97f7c8-9nkvv" Feb 02 09:09:41 crc kubenswrapper[4764]: E0202 09:09:41.453406 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 09:09:41.953393467 +0000 UTC m=+144.887117555 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9nkvv" (UID: "527edab8-f88f-4d82-b42e-b8d52fcaa7d4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:09:41 crc kubenswrapper[4764]: I0202 09:09:41.496660 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qq96b"] Feb 02 09:09:41 crc kubenswrapper[4764]: I0202 09:09:41.535585 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mmcq4"] Feb 02 09:09:41 crc kubenswrapper[4764]: I0202 09:09:41.550757 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-2cdx8"] Feb 02 09:09:41 crc kubenswrapper[4764]: I0202 09:09:41.553766 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 09:09:41 crc kubenswrapper[4764]: E0202 09:09:41.554067 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 09:09:42.054052491 +0000 UTC m=+144.987776579 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:09:41 crc kubenswrapper[4764]: I0202 09:09:41.600210 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-p66d5" Feb 02 09:09:41 crc kubenswrapper[4764]: I0202 09:09:41.604642 4764 generic.go:334] "Generic (PLEG): container finished" podID="6c0cd275-710b-43d8-a9b9-b46bc379454b" containerID="f8b5b3eef201fe3919b9a3fe28acd1df0f45e29e6e3b8f8acaad71512c8e7a16" exitCode=0 Feb 02 09:09:41 crc kubenswrapper[4764]: I0202 09:09:41.604735 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-h5vrh" event={"ID":"6c0cd275-710b-43d8-a9b9-b46bc379454b","Type":"ContainerDied","Data":"f8b5b3eef201fe3919b9a3fe28acd1df0f45e29e6e3b8f8acaad71512c8e7a16"} Feb 02 09:09:41 crc kubenswrapper[4764]: I0202 09:09:41.604774 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-h5vrh" event={"ID":"6c0cd275-710b-43d8-a9b9-b46bc379454b","Type":"ContainerStarted","Data":"189c2918512f7696d90ff208c61dc73b6e960010a88ec2b8f3a1e38b355083c4"} Feb 02 09:09:41 crc kubenswrapper[4764]: I0202 09:09:41.609485 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9flm5" Feb 02 09:09:41 crc kubenswrapper[4764]: I0202 09:09:41.628371 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-sxb6c" event={"ID":"e4968496-d03e-4450-b85d-240cd75bd900","Type":"ContainerStarted","Data":"d81f3bb521b9dfcf5309d8cd9d8b97f8a4e276b156e26a406b1f79248963bc2b"} Feb 02 09:09:41 crc kubenswrapper[4764]: I0202 09:09:41.628444 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-sxb6c" event={"ID":"e4968496-d03e-4450-b85d-240cd75bd900","Type":"ContainerStarted","Data":"be225897b40b16ddaddf1f37c81f62152ea40d57355be05f48f0a87041236bb6"} Feb 02 09:09:41 crc kubenswrapper[4764]: W0202 09:09:41.640625 4764 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1ea380a7_6b81_4938_a6f0_0be18293846d.slice/crio-a54ce7b7c3222c6bc51a2fc8fa0e8ad8aee95962b9c6c77c0533e5b6160a3b58 WatchSource:0}: Error finding container a54ce7b7c3222c6bc51a2fc8fa0e8ad8aee95962b9c6c77c0533e5b6160a3b58: Status 404 returned error can't find the container with id a54ce7b7c3222c6bc51a2fc8fa0e8ad8aee95962b9c6c77c0533e5b6160a3b58 Feb 02 09:09:41 crc kubenswrapper[4764]: I0202 09:09:41.650553 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-wqbtk" Feb 02 09:09:41 crc kubenswrapper[4764]: I0202 09:09:41.653820 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gzpjj" event={"ID":"8617aa45-3d34-4ab9-9870-67f8ccaeec16","Type":"ContainerStarted","Data":"0ea215b892c7c8ea7b234e8eee97698a58b4ca43e9ef544eeb6893da36876b17"} Feb 02 09:09:41 crc kubenswrapper[4764]: I0202 09:09:41.655485 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9nkvv\" (UID: \"527edab8-f88f-4d82-b42e-b8d52fcaa7d4\") " pod="openshift-image-registry/image-registry-697d97f7c8-9nkvv" Feb 02 09:09:41 crc kubenswrapper[4764]: E0202 09:09:41.656553 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 09:09:42.156536697 +0000 UTC m=+145.090260785 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9nkvv" (UID: "527edab8-f88f-4d82-b42e-b8d52fcaa7d4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:09:41 crc kubenswrapper[4764]: I0202 09:09:41.669297 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-966nh" event={"ID":"70342a10-423d-4657-9fc0-062ab9ab8902","Type":"ContainerStarted","Data":"a3b165019032cbd0b9e2e96ef5294f81bb83582f5c87b5a7ae719e4f9fa104aa"} Feb 02 09:09:41 crc kubenswrapper[4764]: I0202 09:09:41.669357 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-966nh" event={"ID":"70342a10-423d-4657-9fc0-062ab9ab8902","Type":"ContainerStarted","Data":"3a54f4042d11a0f075d8fde8fdfd10430f1e69156df2ed0a877e36c97f22b6ea"} Feb 02 09:09:41 crc kubenswrapper[4764]: I0202 09:09:41.671134 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-966nh" Feb 02 09:09:41 crc kubenswrapper[4764]: I0202 09:09:41.676112 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-t5q9w" Feb 02 09:09:41 crc kubenswrapper[4764]: I0202 09:09:41.678039 4764 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-966nh container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" start-of-body= Feb 02 09:09:41 crc kubenswrapper[4764]: I0202 09:09:41.678135 4764 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-966nh" podUID="70342a10-423d-4657-9fc0-062ab9ab8902" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" Feb 02 09:09:41 crc kubenswrapper[4764]: I0202 09:09:41.690417 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-cnkss" event={"ID":"d874a84d-48e3-4cb2-ac74-4f0db52e9d75","Type":"ContainerStarted","Data":"b1c6c13f4eacde51e30158c53fd5ee984a9d48d0b801c239460e97316334fa43"} Feb 02 09:09:41 crc kubenswrapper[4764]: I0202 09:09:41.691138 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-b894n" Feb 02 09:09:41 crc kubenswrapper[4764]: I0202 09:09:41.703204 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-4h9v4" event={"ID":"1eaf78eb-9719-4737-9f33-a8ab1dcb5dcf","Type":"ContainerStarted","Data":"4a50f8bd73c83a65d401890a6aa949165461db6bde6c797ba9d02dcc223b663e"} Feb 02 09:09:41 crc kubenswrapper[4764]: I0202 09:09:41.703240 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-4h9v4" event={"ID":"1eaf78eb-9719-4737-9f33-a8ab1dcb5dcf","Type":"ContainerStarted","Data":"d853403689c83b975c679276000966673fa9c2ffaafd02b785a6a69e29275311"} Feb 02 09:09:41 crc kubenswrapper[4764]: I0202 09:09:41.718484 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-zh65k" event={"ID":"348075cc-bd79-42f2-bbe2-0e5367092963","Type":"ContainerStarted","Data":"c562aaf58048751b3eff081f460cfd30bd90f74688d3c5e2afd32d5af99315c4"} Feb 02 09:09:41 crc kubenswrapper[4764]: I0202 09:09:41.727799 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-scdc2" event={"ID":"1ceb5fce-77f5-46be-80ff-be6e932bfb35","Type":"ContainerStarted","Data":"405ab824c28becfc4d5526bdea7d23954be05b09f7f22b66e0e73c5fc402098c"} Feb 02 09:09:41 crc kubenswrapper[4764]: I0202 09:09:41.727849 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-scdc2" event={"ID":"1ceb5fce-77f5-46be-80ff-be6e932bfb35","Type":"ContainerStarted","Data":"4ca6c79086216980c1b2a2b5ae95af76a3c03e64421f6e837332ace52541f62e"} Feb 02 09:09:41 crc kubenswrapper[4764]: I0202 09:09:41.728891 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-scdc2" Feb 02 09:09:41 crc kubenswrapper[4764]: I0202 09:09:41.732304 4764 patch_prober.go:28] interesting pod/console-operator-58897d9998-scdc2 container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.21:8443/readyz\": dial tcp 10.217.0.21:8443: connect: connection refused" start-of-body= Feb 02 09:09:41 crc kubenswrapper[4764]: I0202 09:09:41.732345 4764 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-scdc2" podUID="1ceb5fce-77f5-46be-80ff-be6e932bfb35" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.21:8443/readyz\": dial tcp 10.217.0.21:8443: connect: connection refused" Feb 02 09:09:41 crc kubenswrapper[4764]: I0202 09:09:41.742092 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-44gqd" event={"ID":"f02fb0c2-81fc-45ed-8742-f2f8a51a5437","Type":"ContainerStarted","Data":"dbb657586e4fa6936d13e507f11835b2bef018e18daf87bfc3a5f120016f927a"} Feb 02 09:09:41 crc kubenswrapper[4764]: I0202 09:09:41.756464 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 09:09:41 crc kubenswrapper[4764]: E0202 09:09:41.756736 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 09:09:42.256722028 +0000 UTC m=+145.190446106 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:09:41 crc kubenswrapper[4764]: I0202 09:09:41.756823 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9nkvv\" (UID: \"527edab8-f88f-4d82-b42e-b8d52fcaa7d4\") " pod="openshift-image-registry/image-registry-697d97f7c8-9nkvv" Feb 02 09:09:41 crc kubenswrapper[4764]: E0202 09:09:41.758266 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 09:09:42.258246541 +0000 UTC m=+145.191970699 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9nkvv" (UID: "527edab8-f88f-4d82-b42e-b8d52fcaa7d4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:09:41 crc kubenswrapper[4764]: I0202 09:09:41.760410 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-52th4" event={"ID":"856a5985-578c-48e4-a535-01496383f3d4","Type":"ContainerStarted","Data":"6569a095616b0276d9a262bd707572011f0640e00079f88476f6c6644f1ffda3"} Feb 02 09:09:41 crc kubenswrapper[4764]: I0202 09:09:41.763246 4764 generic.go:334] "Generic (PLEG): container finished" podID="40fcb5fe-ee7b-4696-a775-efc703af9b61" containerID="29747dbdd5880f17412105752db8666267c2186458e75ead0b0f1017e11bcc76" exitCode=0 Feb 02 09:09:41 crc kubenswrapper[4764]: I0202 09:09:41.763358 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ngtsq" event={"ID":"40fcb5fe-ee7b-4696-a775-efc703af9b61","Type":"ContainerDied","Data":"29747dbdd5880f17412105752db8666267c2186458e75ead0b0f1017e11bcc76"} Feb 02 09:09:41 crc kubenswrapper[4764]: I0202 09:09:41.763384 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ngtsq" event={"ID":"40fcb5fe-ee7b-4696-a775-efc703af9b61","Type":"ContainerStarted","Data":"1d7b7c3c9003dd05dba750fe87d46fac42b9e5f2ff8003033c9ab32707aef5f0"} Feb 02 09:09:41 crc kubenswrapper[4764]: I0202 09:09:41.781377 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-tkx8j" event={"ID":"fea30e22-e331-4518-a777-b6258c053f06","Type":"ContainerStarted","Data":"35f1648f020d40e22c6cb107e9d934e1be3809944a6ad4b31116b5b9218944e6"} Feb 02 09:09:41 crc kubenswrapper[4764]: I0202 09:09:41.812549 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-7k5qb" event={"ID":"c4b729bd-782f-4393-b162-16c02eb3cb71","Type":"ContainerStarted","Data":"3b02b4c41c96a39e0f243012c369440983dd0f4cf15d01110e3bbaf675737255"} Feb 02 09:09:41 crc kubenswrapper[4764]: I0202 09:09:41.812594 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-7k5qb" event={"ID":"c4b729bd-782f-4393-b162-16c02eb3cb71","Type":"ContainerStarted","Data":"b225e44db69176dca78035fdabd4456d4f60c088bb1368f1bed69ebb28ded19f"} Feb 02 09:09:41 crc kubenswrapper[4764]: I0202 09:09:41.813473 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-7k5qb" Feb 02 09:09:41 crc kubenswrapper[4764]: I0202 09:09:41.814951 4764 patch_prober.go:28] interesting pod/downloads-7954f5f757-7k5qb container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" start-of-body= Feb 02 09:09:41 crc kubenswrapper[4764]: I0202 09:09:41.814987 4764 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-7k5qb" podUID="c4b729bd-782f-4393-b162-16c02eb3cb71" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" Feb 02 09:09:41 crc kubenswrapper[4764]: I0202 09:09:41.849397 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-98wxj" event={"ID":"08bc3386-ee9a-4a3c-aa52-22d667971076","Type":"ContainerStarted","Data":"b36b7d88389773e8854c5675b6f14b039fea62d94388c02b1001ffe4b12ea02c"} Feb 02 09:09:41 crc kubenswrapper[4764]: I0202 09:09:41.849632 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-98wxj" event={"ID":"08bc3386-ee9a-4a3c-aa52-22d667971076","Type":"ContainerStarted","Data":"bc95b1a93c8165caa41e29a2af81f19a14151409f5a33b4fb30afdcee1197f0c"} Feb 02 09:09:41 crc kubenswrapper[4764]: I0202 09:09:41.860415 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 09:09:41 crc kubenswrapper[4764]: E0202 09:09:41.867678 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 09:09:42.367657594 +0000 UTC m=+145.301381672 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:09:41 crc kubenswrapper[4764]: I0202 09:09:41.956153 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-njspk"] Feb 02 09:09:41 crc kubenswrapper[4764]: I0202 09:09:41.957271 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-6dljc"] Feb 02 09:09:41 crc kubenswrapper[4764]: I0202 09:09:41.969866 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9nkvv\" (UID: \"527edab8-f88f-4d82-b42e-b8d52fcaa7d4\") " pod="openshift-image-registry/image-registry-697d97f7c8-9nkvv" Feb 02 09:09:41 crc kubenswrapper[4764]: E0202 09:09:41.970943 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 09:09:42.470917222 +0000 UTC m=+145.404641310 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9nkvv" (UID: "527edab8-f88f-4d82-b42e-b8d52fcaa7d4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:09:41 crc kubenswrapper[4764]: W0202 09:09:41.989168 4764 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf552c254_ade6_4fe4_baa7_04e53c67479b.slice/crio-8768b9bfbb791e22d08ff6845cf0664d57d30e5276ee775ffa2776f7468f37f1 WatchSource:0}: Error finding container 8768b9bfbb791e22d08ff6845cf0664d57d30e5276ee775ffa2776f7468f37f1: Status 404 returned error can't find the container with id 8768b9bfbb791e22d08ff6845cf0664d57d30e5276ee775ffa2776f7468f37f1 Feb 02 09:09:42 crc kubenswrapper[4764]: I0202 09:09:42.014024 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-966nh" podStartSLOduration=124.013992088 podStartE2EDuration="2m4.013992088s" podCreationTimestamp="2026-02-02 09:07:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:09:41.967412613 +0000 UTC m=+144.901136701" watchObservedRunningTime="2026-02-02 09:09:42.013992088 +0000 UTC m=+144.947716166" Feb 02 09:09:42 crc kubenswrapper[4764]: I0202 09:09:42.041129 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gzpjj" podStartSLOduration=124.04110875 podStartE2EDuration="2m4.04110875s" podCreationTimestamp="2026-02-02 09:07:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:09:42.03935693 +0000 UTC m=+144.973081028" watchObservedRunningTime="2026-02-02 09:09:42.04110875 +0000 UTC m=+144.974832868" Feb 02 09:09:42 crc kubenswrapper[4764]: I0202 09:09:42.049582 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-9mdwc"] Feb 02 09:09:42 crc kubenswrapper[4764]: I0202 09:09:42.071986 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 09:09:42 crc kubenswrapper[4764]: E0202 09:09:42.072736 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 09:09:42.572708859 +0000 UTC m=+145.506432987 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:09:42 crc kubenswrapper[4764]: I0202 09:09:42.086989 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-4298w"] Feb 02 09:09:42 crc kubenswrapper[4764]: I0202 09:09:42.104190 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-gp4jg"] Feb 02 09:09:42 crc kubenswrapper[4764]: W0202 09:09:42.159262 4764 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37d3c5dc_0535_4f2e_aa2f_debd99e0f1e3.slice/crio-430f309137b47445cab0755c4c1a2ead26d44204be4c870282523546d0fde4a0 WatchSource:0}: Error finding container 430f309137b47445cab0755c4c1a2ead26d44204be4c870282523546d0fde4a0: Status 404 returned error can't find the container with id 430f309137b47445cab0755c4c1a2ead26d44204be4c870282523546d0fde4a0 Feb 02 09:09:42 crc kubenswrapper[4764]: I0202 09:09:42.176247 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9nkvv\" (UID: \"527edab8-f88f-4d82-b42e-b8d52fcaa7d4\") " pod="openshift-image-registry/image-registry-697d97f7c8-9nkvv" Feb 02 09:09:42 crc kubenswrapper[4764]: E0202 09:09:42.176854 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 09:09:42.676837822 +0000 UTC m=+145.610561910 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9nkvv" (UID: "527edab8-f88f-4d82-b42e-b8d52fcaa7d4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:09:42 crc kubenswrapper[4764]: I0202 09:09:42.252485 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-dlpkk"] Feb 02 09:09:42 crc kubenswrapper[4764]: I0202 09:09:42.269757 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-n59jp"] Feb 02 09:09:42 crc kubenswrapper[4764]: I0202 09:09:42.277401 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-fdcsz"] Feb 02 09:09:42 crc kubenswrapper[4764]: I0202 09:09:42.282662 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 09:09:42 crc kubenswrapper[4764]: E0202 09:09:42.282841 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 09:09:42.782824917 +0000 UTC m=+145.716549005 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:09:42 crc kubenswrapper[4764]: I0202 09:09:42.283369 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9nkvv\" (UID: \"527edab8-f88f-4d82-b42e-b8d52fcaa7d4\") " pod="openshift-image-registry/image-registry-697d97f7c8-9nkvv" Feb 02 09:09:42 crc kubenswrapper[4764]: E0202 09:09:42.283644 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 09:09:42.78363568 +0000 UTC m=+145.717359758 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9nkvv" (UID: "527edab8-f88f-4d82-b42e-b8d52fcaa7d4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:09:42 crc kubenswrapper[4764]: I0202 09:09:42.317804 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-drqfg"] Feb 02 09:09:42 crc kubenswrapper[4764]: I0202 09:09:42.371645 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-df5xv"] Feb 02 09:09:42 crc kubenswrapper[4764]: I0202 09:09:42.377214 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-fffd9"] Feb 02 09:09:42 crc kubenswrapper[4764]: I0202 09:09:42.387659 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 09:09:42 crc kubenswrapper[4764]: E0202 09:09:42.388537 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 09:09:42.888516325 +0000 UTC m=+145.822240413 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:09:42 crc kubenswrapper[4764]: W0202 09:09:42.490889 4764 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7df533b7_06a3_4a8f_abfa_d7a45272d93d.slice/crio-18c149c5102c2b080d12a1805399ec9632285efb2cd532d80beb7ca30f6e5966 WatchSource:0}: Error finding container 18c149c5102c2b080d12a1805399ec9632285efb2cd532d80beb7ca30f6e5966: Status 404 returned error can't find the container with id 18c149c5102c2b080d12a1805399ec9632285efb2cd532d80beb7ca30f6e5966 Feb 02 09:09:42 crc kubenswrapper[4764]: I0202 09:09:42.493130 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9nkvv\" (UID: \"527edab8-f88f-4d82-b42e-b8d52fcaa7d4\") " pod="openshift-image-registry/image-registry-697d97f7c8-9nkvv" Feb 02 09:09:42 crc kubenswrapper[4764]: E0202 09:09:42.493707 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 09:09:42.993672587 +0000 UTC m=+145.927396675 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9nkvv" (UID: "527edab8-f88f-4d82-b42e-b8d52fcaa7d4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:09:42 crc kubenswrapper[4764]: I0202 09:09:42.575239 4764 csr.go:261] certificate signing request csr-h8l4d is approved, waiting to be issued Feb 02 09:09:42 crc kubenswrapper[4764]: I0202 09:09:42.575716 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-sxb6c" podStartSLOduration=124.575697201 podStartE2EDuration="2m4.575697201s" podCreationTimestamp="2026-02-02 09:07:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:09:42.550568696 +0000 UTC m=+145.484292774" watchObservedRunningTime="2026-02-02 09:09:42.575697201 +0000 UTC m=+145.509421289" Feb 02 09:09:42 crc kubenswrapper[4764]: I0202 09:09:42.577867 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-l5jcj"] Feb 02 09:09:42 crc kubenswrapper[4764]: I0202 09:09:42.590548 4764 csr.go:257] certificate signing request csr-h8l4d is issued Feb 02 09:09:42 crc kubenswrapper[4764]: I0202 09:09:42.596561 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 09:09:42 crc kubenswrapper[4764]: E0202 09:09:42.596920 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 09:09:43.096905615 +0000 UTC m=+146.030629703 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:09:42 crc kubenswrapper[4764]: I0202 09:09:42.647375 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-2m47l"] Feb 02 09:09:42 crc kubenswrapper[4764]: I0202 09:09:42.698050 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9nkvv\" (UID: \"527edab8-f88f-4d82-b42e-b8d52fcaa7d4\") " pod="openshift-image-registry/image-registry-697d97f7c8-9nkvv" Feb 02 09:09:42 crc kubenswrapper[4764]: E0202 09:09:42.698414 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 09:09:43.198399962 +0000 UTC m=+146.132124050 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9nkvv" (UID: "527edab8-f88f-4d82-b42e-b8d52fcaa7d4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:09:42 crc kubenswrapper[4764]: W0202 09:09:42.735690 4764 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9a4ebc92_5ef5_4088_a509_5399beef3c7d.slice/crio-2fe719b3282fc1cecbac831d91ded756fbd6fa3cf641cbcf757998ada2ee5ee3 WatchSource:0}: Error finding container 2fe719b3282fc1cecbac831d91ded756fbd6fa3cf641cbcf757998ada2ee5ee3: Status 404 returned error can't find the container with id 2fe719b3282fc1cecbac831d91ded756fbd6fa3cf641cbcf757998ada2ee5ee3 Feb 02 09:09:42 crc kubenswrapper[4764]: I0202 09:09:42.807427 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 09:09:42 crc kubenswrapper[4764]: E0202 09:09:42.807895 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 09:09:43.307874697 +0000 UTC m=+146.241598785 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:09:42 crc kubenswrapper[4764]: I0202 09:09:42.827985 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-t5q9w"] Feb 02 09:09:42 crc kubenswrapper[4764]: I0202 09:09:42.872081 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-hfwsg"] Feb 02 09:09:42 crc kubenswrapper[4764]: I0202 09:09:42.883380 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-tpldd"] Feb 02 09:09:42 crc kubenswrapper[4764]: I0202 09:09:42.891487 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-gp4jg" event={"ID":"dc0a29a2-27b1-4f7a-afaa-a0ae6cd290e9","Type":"ContainerStarted","Data":"ca1c6c234760832614b3eb5c71dbdabd2d8541818382fc717310311afda58e8a"} Feb 02 09:09:42 crc kubenswrapper[4764]: I0202 09:09:42.892710 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-52th4" event={"ID":"856a5985-578c-48e4-a535-01496383f3d4","Type":"ContainerStarted","Data":"61965c3b4b2a7b69ae74f5bf71708db9ed4fd0e2ec302ab0c4fbffdb06ba72a6"} Feb 02 09:09:42 crc kubenswrapper[4764]: I0202 09:09:42.893350 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-52th4" Feb 02 09:09:42 crc kubenswrapper[4764]: I0202 09:09:42.898188 4764 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-52th4 container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.30:8443/healthz\": dial tcp 10.217.0.30:8443: connect: connection refused" start-of-body= Feb 02 09:09:42 crc kubenswrapper[4764]: I0202 09:09:42.898244 4764 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-52th4" podUID="856a5985-578c-48e4-a535-01496383f3d4" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.30:8443/healthz\": dial tcp 10.217.0.30:8443: connect: connection refused" Feb 02 09:09:42 crc kubenswrapper[4764]: I0202 09:09:42.898656 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-drqfg" event={"ID":"91e66500-e26e-4421-83e2-62b738656d5d","Type":"ContainerStarted","Data":"b28f94f1ff3153bf8371b5512ea846ea21ee1b565dda689cc6bbf3fef409bccc"} Feb 02 09:09:42 crc kubenswrapper[4764]: I0202 09:09:42.908790 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9nkvv\" (UID: \"527edab8-f88f-4d82-b42e-b8d52fcaa7d4\") " pod="openshift-image-registry/image-registry-697d97f7c8-9nkvv" Feb 02 09:09:42 crc kubenswrapper[4764]: E0202 09:09:42.910427 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 09:09:43.410415275 +0000 UTC m=+146.344139363 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9nkvv" (UID: "527edab8-f88f-4d82-b42e-b8d52fcaa7d4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:09:42 crc kubenswrapper[4764]: I0202 09:09:42.925470 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2cdx8" event={"ID":"36de8e6a-591b-4365-b434-ebd8197abf41","Type":"ContainerStarted","Data":"f8f714fef7c100950c9c810935df7d20077d1e72f0a296994eb26cee50e852c5"} Feb 02 09:09:42 crc kubenswrapper[4764]: I0202 09:09:42.925672 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9flm5"] Feb 02 09:09:42 crc kubenswrapper[4764]: W0202 09:09:42.953645 4764 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3ebf456e_6cb3_4cd4_a20d_a189ce1b9333.slice/crio-3a8939bf7fcc8e576908e21cfcbb632bc5a39c0f625893286f652596431160d5 WatchSource:0}: Error finding container 3a8939bf7fcc8e576908e21cfcbb632bc5a39c0f625893286f652596431160d5: Status 404 returned error can't find the container with id 3a8939bf7fcc8e576908e21cfcbb632bc5a39c0f625893286f652596431160d5 Feb 02 09:09:42 crc kubenswrapper[4764]: I0202 09:09:42.963587 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-p66d5"] Feb 02 09:09:42 crc kubenswrapper[4764]: I0202 09:09:42.968675 4764 generic.go:334] "Generic (PLEG): container finished" podID="1eaf78eb-9719-4737-9f33-a8ab1dcb5dcf" containerID="4a50f8bd73c83a65d401890a6aa949165461db6bde6c797ba9d02dcc223b663e" exitCode=0 Feb 02 09:09:42 crc kubenswrapper[4764]: I0202 09:09:42.968734 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-4h9v4" event={"ID":"1eaf78eb-9719-4737-9f33-a8ab1dcb5dcf","Type":"ContainerDied","Data":"4a50f8bd73c83a65d401890a6aa949165461db6bde6c797ba9d02dcc223b663e"} Feb 02 09:09:42 crc kubenswrapper[4764]: I0202 09:09:42.982742 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4298w" event={"ID":"37d3c5dc-0535-4f2e-aa2f-debd99e0f1e3","Type":"ContainerStarted","Data":"430f309137b47445cab0755c4c1a2ead26d44204be4c870282523546d0fde4a0"} Feb 02 09:09:42 crc kubenswrapper[4764]: I0202 09:09:42.990584 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-mkwq2" event={"ID":"83fed8a7-bd5f-463d-b70e-3a01700e8315","Type":"ContainerStarted","Data":"1340bafbb76df0f57f6e60cd0bad9a034f48b3f8ad5831f5931d3cb84cd86670"} Feb 02 09:09:42 crc kubenswrapper[4764]: I0202 09:09:42.991461 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-l5jcj" event={"ID":"9a4ebc92-5ef5-4088-a509-5399beef3c7d","Type":"ContainerStarted","Data":"2fe719b3282fc1cecbac831d91ded756fbd6fa3cf641cbcf757998ada2ee5ee3"} Feb 02 09:09:43 crc kubenswrapper[4764]: I0202 09:09:43.001471 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-njspk" event={"ID":"cb1b5d6b-7d42-4b40-99e3-4827a5379133","Type":"ContainerStarted","Data":"2d9031b8bbfc8fbf7eee7ec3dff243e15ccf06d41c087ebeb2657c36b212141a"} Feb 02 09:09:43 crc kubenswrapper[4764]: I0202 09:09:43.009823 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-98wxj" podStartSLOduration=124.009802673 podStartE2EDuration="2m4.009802673s" podCreationTimestamp="2026-02-02 09:07:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:09:43.002106454 +0000 UTC m=+145.935830542" watchObservedRunningTime="2026-02-02 09:09:43.009802673 +0000 UTC m=+145.943526761" Feb 02 09:09:43 crc kubenswrapper[4764]: I0202 09:09:43.012221 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 09:09:43 crc kubenswrapper[4764]: E0202 09:09:43.012592 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 09:09:43.512575732 +0000 UTC m=+146.446299820 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:09:43 crc kubenswrapper[4764]: I0202 09:09:43.021348 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-98wxj" event={"ID":"08bc3386-ee9a-4a3c-aa52-22d667971076","Type":"ContainerStarted","Data":"eb7b656cd3b84a6bf351d781aad741627a203e68ff6e6181ec1ab546d46b533c"} Feb 02 09:09:43 crc kubenswrapper[4764]: I0202 09:09:43.021699 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-7v5xl"] Feb 02 09:09:43 crc kubenswrapper[4764]: I0202 09:09:43.029861 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-wqbtk"] Feb 02 09:09:43 crc kubenswrapper[4764]: I0202 09:09:43.030222 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-fffd9" event={"ID":"7df533b7-06a3-4a8f-abfa-d7a45272d93d","Type":"ContainerStarted","Data":"18c149c5102c2b080d12a1805399ec9632285efb2cd532d80beb7ca30f6e5966"} Feb 02 09:09:43 crc kubenswrapper[4764]: I0202 09:09:43.033073 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-zh65k" event={"ID":"348075cc-bd79-42f2-bbe2-0e5367092963","Type":"ContainerStarted","Data":"da73e8c26f707bcb5b6e55cac1d4e560793c915bd94510c03baa55c55f7cbbd4"} Feb 02 09:09:43 crc kubenswrapper[4764]: I0202 09:09:43.035428 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-dlpkk" event={"ID":"88f8519e-b576-4bef-96d6-771b78c67c11","Type":"ContainerStarted","Data":"ddf43d304869246cfb55e10d870af6bbddcef046f799b1047c495f8c474317dd"} Feb 02 09:09:43 crc kubenswrapper[4764]: I0202 09:09:43.038612 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29500380-2gwrp" event={"ID":"006ce277-7c6b-49a5-9e98-0bf7caee30f9","Type":"ContainerStarted","Data":"3a0f35f6d6c22683f1379ec4e011d9fc24fb03375e1e4b79536567e348c406c7"} Feb 02 09:09:43 crc kubenswrapper[4764]: I0202 09:09:43.038653 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29500380-2gwrp" event={"ID":"006ce277-7c6b-49a5-9e98-0bf7caee30f9","Type":"ContainerStarted","Data":"8137233e318d112629cf934371c3feb215fab8d658a8547450addbf07b690ac7"} Feb 02 09:09:43 crc kubenswrapper[4764]: I0202 09:09:43.040278 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-cnkss" event={"ID":"d874a84d-48e3-4cb2-ac74-4f0db52e9d75","Type":"ContainerStarted","Data":"7f0399b6927158c0f8102427612492ca9041d1130c1c9f64b4196a6117b4f9c2"} Feb 02 09:09:43 crc kubenswrapper[4764]: I0202 09:09:43.041255 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-df5xv" event={"ID":"d7534d2e-fb32-4895-945b-ea5ccd35fcbb","Type":"ContainerStarted","Data":"3af7e1c196c87b0ab201d3eb671f99b9cd6571b4a0a389d558b606de34d154e5"} Feb 02 09:09:43 crc kubenswrapper[4764]: I0202 09:09:43.042024 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-b894n" event={"ID":"d17edd1e-7d52-4c32-90a9-7d16723f3875","Type":"ContainerStarted","Data":"982ee5d7ab7e7196b921b9063b8ba9f65654a75e375bfff1f40a935273c19ac6"} Feb 02 09:09:43 crc kubenswrapper[4764]: I0202 09:09:43.043421 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2m47l" event={"ID":"a1f840f6-138f-496d-b64d-fbe6d49e0bdd","Type":"ContainerStarted","Data":"f70b876e3f9c7ef15007dcd26ddcd5a8c352ff4314ab812e6d18ba2260f20375"} Feb 02 09:09:43 crc kubenswrapper[4764]: I0202 09:09:43.051493 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qq96b" event={"ID":"545e08e9-b2c4-493e-8fc7-8353fcab998d","Type":"ContainerStarted","Data":"f494a507c58e93ae948fd09518d62dbe450d38653fc1ee4dff2026c72ad4a7e7"} Feb 02 09:09:43 crc kubenswrapper[4764]: I0202 09:09:43.055744 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-fdcsz" event={"ID":"f353fe2c-cc41-447a-b4b8-45eed47a29e2","Type":"ContainerStarted","Data":"8801efdfa1d8f0d7b38ddbea6ac9e1ebf108d7a4b626d51dfa28f555c890e157"} Feb 02 09:09:43 crc kubenswrapper[4764]: I0202 09:09:43.069445 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mmcq4" event={"ID":"1ea380a7-6b81-4938-a6f0-0be18293846d","Type":"ContainerStarted","Data":"4d6d470ad0851506f3e03a0c4c017457957060cf8f8db201583a2debadb697b3"} Feb 02 09:09:43 crc kubenswrapper[4764]: I0202 09:09:43.069488 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mmcq4" event={"ID":"1ea380a7-6b81-4938-a6f0-0be18293846d","Type":"ContainerStarted","Data":"a54ce7b7c3222c6bc51a2fc8fa0e8ad8aee95962b9c6c77c0533e5b6160a3b58"} Feb 02 09:09:43 crc kubenswrapper[4764]: I0202 09:09:43.070180 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mmcq4" Feb 02 09:09:43 crc kubenswrapper[4764]: I0202 09:09:43.077961 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-9mdwc" event={"ID":"513ecde9-f189-4a22-8742-1b262523d462","Type":"ContainerStarted","Data":"707c484f25eb263b11a295e4ad9807fda8e4210cb2f90590f5ad4b7631a8fc81"} Feb 02 09:09:43 crc kubenswrapper[4764]: I0202 09:09:43.078866 4764 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-mmcq4 container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.20:8443/healthz\": dial tcp 10.217.0.20:8443: connect: connection refused" start-of-body= Feb 02 09:09:43 crc kubenswrapper[4764]: I0202 09:09:43.078981 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-7mtl6"] Feb 02 09:09:43 crc kubenswrapper[4764]: I0202 09:09:43.078898 4764 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mmcq4" podUID="1ea380a7-6b81-4938-a6f0-0be18293846d" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.20:8443/healthz\": dial tcp 10.217.0.20:8443: connect: connection refused" Feb 02 09:09:43 crc kubenswrapper[4764]: I0202 09:09:43.089170 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-n59jp" event={"ID":"720e4205-b87d-4fc7-95df-3ad7c4849f97","Type":"ContainerStarted","Data":"fcf1c0b423f2abc854d8e4e49c2045d6224787dc225c125181840fd0e80e2f28"} Feb 02 09:09:43 crc kubenswrapper[4764]: I0202 09:09:43.135027 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9nkvv\" (UID: \"527edab8-f88f-4d82-b42e-b8d52fcaa7d4\") " pod="openshift-image-registry/image-registry-697d97f7c8-9nkvv" Feb 02 09:09:43 crc kubenswrapper[4764]: E0202 09:09:43.137376 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 09:09:43.637359973 +0000 UTC m=+146.571084061 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9nkvv" (UID: "527edab8-f88f-4d82-b42e-b8d52fcaa7d4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:09:43 crc kubenswrapper[4764]: I0202 09:09:43.139435 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-7k5qb" podStartSLOduration=125.139417651 podStartE2EDuration="2m5.139417651s" podCreationTimestamp="2026-02-02 09:07:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:09:43.138732832 +0000 UTC m=+146.072456920" watchObservedRunningTime="2026-02-02 09:09:43.139417651 +0000 UTC m=+146.073141759" Feb 02 09:09:43 crc kubenswrapper[4764]: I0202 09:09:43.147253 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-6dljc" event={"ID":"f552c254-ade6-4fe4-baa7-04e53c67479b","Type":"ContainerStarted","Data":"8768b9bfbb791e22d08ff6845cf0664d57d30e5276ee775ffa2776f7468f37f1"} Feb 02 09:09:43 crc kubenswrapper[4764]: I0202 09:09:43.148272 4764 patch_prober.go:28] interesting pod/downloads-7954f5f757-7k5qb container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" start-of-body= Feb 02 09:09:43 crc kubenswrapper[4764]: I0202 09:09:43.148315 4764 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-7k5qb" podUID="c4b729bd-782f-4393-b162-16c02eb3cb71" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" Feb 02 09:09:43 crc kubenswrapper[4764]: I0202 09:09:43.148748 4764 patch_prober.go:28] interesting pod/console-operator-58897d9998-scdc2 container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.21:8443/readyz\": dial tcp 10.217.0.21:8443: connect: connection refused" start-of-body= Feb 02 09:09:43 crc kubenswrapper[4764]: I0202 09:09:43.149094 4764 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-scdc2" podUID="1ceb5fce-77f5-46be-80ff-be6e932bfb35" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.21:8443/readyz\": dial tcp 10.217.0.21:8443: connect: connection refused" Feb 02 09:09:43 crc kubenswrapper[4764]: I0202 09:09:43.148804 4764 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-966nh container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" start-of-body= Feb 02 09:09:43 crc kubenswrapper[4764]: I0202 09:09:43.149147 4764 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-966nh" podUID="70342a10-423d-4657-9fc0-062ab9ab8902" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" Feb 02 09:09:43 crc kubenswrapper[4764]: I0202 09:09:43.178035 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-52th4" podStartSLOduration=124.177999749 podStartE2EDuration="2m4.177999749s" podCreationTimestamp="2026-02-02 09:07:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:09:43.17102385 +0000 UTC m=+146.104747938" watchObservedRunningTime="2026-02-02 09:09:43.177999749 +0000 UTC m=+146.111723837" Feb 02 09:09:43 crc kubenswrapper[4764]: I0202 09:09:43.208074 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-44gqd" podStartSLOduration=125.208052993 podStartE2EDuration="2m5.208052993s" podCreationTimestamp="2026-02-02 09:07:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:09:43.207392154 +0000 UTC m=+146.141116252" watchObservedRunningTime="2026-02-02 09:09:43.208052993 +0000 UTC m=+146.141777081" Feb 02 09:09:43 crc kubenswrapper[4764]: I0202 09:09:43.236778 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 09:09:43 crc kubenswrapper[4764]: E0202 09:09:43.237898 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 09:09:43.737883182 +0000 UTC m=+146.671607270 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:09:43 crc kubenswrapper[4764]: I0202 09:09:43.250579 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-scdc2" podStartSLOduration=125.250556642 podStartE2EDuration="2m5.250556642s" podCreationTimestamp="2026-02-02 09:07:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:09:43.245349504 +0000 UTC m=+146.179073592" watchObservedRunningTime="2026-02-02 09:09:43.250556642 +0000 UTC m=+146.184280730" Feb 02 09:09:43 crc kubenswrapper[4764]: I0202 09:09:43.281601 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mmcq4" podStartSLOduration=124.281582875 podStartE2EDuration="2m4.281582875s" podCreationTimestamp="2026-02-02 09:07:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:09:43.280665069 +0000 UTC m=+146.214389167" watchObservedRunningTime="2026-02-02 09:09:43.281582875 +0000 UTC m=+146.215306963" Feb 02 09:09:43 crc kubenswrapper[4764]: I0202 09:09:43.340777 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9nkvv\" (UID: \"527edab8-f88f-4d82-b42e-b8d52fcaa7d4\") " pod="openshift-image-registry/image-registry-697d97f7c8-9nkvv" Feb 02 09:09:43 crc kubenswrapper[4764]: E0202 09:09:43.341188 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 09:09:43.841171761 +0000 UTC m=+146.774895849 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9nkvv" (UID: "527edab8-f88f-4d82-b42e-b8d52fcaa7d4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:09:43 crc kubenswrapper[4764]: I0202 09:09:43.360226 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-cnkss" podStartSLOduration=125.360207432 podStartE2EDuration="2m5.360207432s" podCreationTimestamp="2026-02-02 09:07:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:09:43.328273184 +0000 UTC m=+146.261997272" watchObservedRunningTime="2026-02-02 09:09:43.360207432 +0000 UTC m=+146.293931550" Feb 02 09:09:43 crc kubenswrapper[4764]: I0202 09:09:43.445545 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 09:09:43 crc kubenswrapper[4764]: E0202 09:09:43.445771 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 09:09:43.945721576 +0000 UTC m=+146.879445664 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:09:43 crc kubenswrapper[4764]: I0202 09:09:43.446115 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9nkvv\" (UID: \"527edab8-f88f-4d82-b42e-b8d52fcaa7d4\") " pod="openshift-image-registry/image-registry-697d97f7c8-9nkvv" Feb 02 09:09:43 crc kubenswrapper[4764]: E0202 09:09:43.446556 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 09:09:43.946541109 +0000 UTC m=+146.880265207 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9nkvv" (UID: "527edab8-f88f-4d82-b42e-b8d52fcaa7d4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:09:43 crc kubenswrapper[4764]: I0202 09:09:43.522880 4764 patch_prober.go:28] interesting pod/machine-config-daemon-4ndm4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 09:09:43 crc kubenswrapper[4764]: I0202 09:09:43.524436 4764 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 09:09:43 crc kubenswrapper[4764]: I0202 09:09:43.548203 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 09:09:43 crc kubenswrapper[4764]: E0202 09:09:43.548658 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 09:09:44.048639884 +0000 UTC m=+146.982363972 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:09:43 crc kubenswrapper[4764]: I0202 09:09:43.548695 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9nkvv\" (UID: \"527edab8-f88f-4d82-b42e-b8d52fcaa7d4\") " pod="openshift-image-registry/image-registry-697d97f7c8-9nkvv" Feb 02 09:09:43 crc kubenswrapper[4764]: E0202 09:09:43.550792 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 09:09:44.050755614 +0000 UTC m=+146.984479702 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9nkvv" (UID: "527edab8-f88f-4d82-b42e-b8d52fcaa7d4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:09:43 crc kubenswrapper[4764]: I0202 09:09:43.592177 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2027-02-02 09:04:42 +0000 UTC, rotation deadline is 2026-11-16 12:11:44.553307197 +0000 UTC Feb 02 09:09:43 crc kubenswrapper[4764]: I0202 09:09:43.592233 4764 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 6891h2m0.961077353s for next certificate rotation Feb 02 09:09:43 crc kubenswrapper[4764]: I0202 09:09:43.650639 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 09:09:43 crc kubenswrapper[4764]: E0202 09:09:43.652229 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 09:09:44.15218809 +0000 UTC m=+147.085912178 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:09:43 crc kubenswrapper[4764]: I0202 09:09:43.752470 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9nkvv\" (UID: \"527edab8-f88f-4d82-b42e-b8d52fcaa7d4\") " pod="openshift-image-registry/image-registry-697d97f7c8-9nkvv" Feb 02 09:09:43 crc kubenswrapper[4764]: E0202 09:09:43.757689 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 09:09:44.257670242 +0000 UTC m=+147.191394330 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9nkvv" (UID: "527edab8-f88f-4d82-b42e-b8d52fcaa7d4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:09:43 crc kubenswrapper[4764]: I0202 09:09:43.855705 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 09:09:43 crc kubenswrapper[4764]: E0202 09:09:43.856215 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 09:09:44.356199475 +0000 UTC m=+147.289923563 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:09:43 crc kubenswrapper[4764]: I0202 09:09:43.959134 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9nkvv\" (UID: \"527edab8-f88f-4d82-b42e-b8d52fcaa7d4\") " pod="openshift-image-registry/image-registry-697d97f7c8-9nkvv" Feb 02 09:09:43 crc kubenswrapper[4764]: E0202 09:09:43.959423 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 09:09:44.459408762 +0000 UTC m=+147.393132850 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9nkvv" (UID: "527edab8-f88f-4d82-b42e-b8d52fcaa7d4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:09:44 crc kubenswrapper[4764]: I0202 09:09:44.060447 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 09:09:44 crc kubenswrapper[4764]: E0202 09:09:44.060793 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 09:09:44.560775806 +0000 UTC m=+147.494499894 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:09:44 crc kubenswrapper[4764]: I0202 09:09:44.163699 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9nkvv\" (UID: \"527edab8-f88f-4d82-b42e-b8d52fcaa7d4\") " pod="openshift-image-registry/image-registry-697d97f7c8-9nkvv" Feb 02 09:09:44 crc kubenswrapper[4764]: E0202 09:09:44.164159 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 09:09:44.664142287 +0000 UTC m=+147.597866375 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9nkvv" (UID: "527edab8-f88f-4d82-b42e-b8d52fcaa7d4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:09:44 crc kubenswrapper[4764]: I0202 09:09:44.265472 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 09:09:44 crc kubenswrapper[4764]: E0202 09:09:44.266137 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 09:09:44.766118849 +0000 UTC m=+147.699842927 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:09:44 crc kubenswrapper[4764]: I0202 09:09:44.291104 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-7mtl6" event={"ID":"aec51473-5be4-4a0b-9ccc-7fcd912f9cf4","Type":"ContainerStarted","Data":"ba1a135a859a36d8427a595d957b66b561d04f2597b788bdb63a32b4bc4aa149"} Feb 02 09:09:44 crc kubenswrapper[4764]: I0202 09:09:44.325482 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-t5q9w" event={"ID":"3ebf456e-6cb3-4cd4-a20d-a189ce1b9333","Type":"ContainerStarted","Data":"3a8939bf7fcc8e576908e21cfcbb632bc5a39c0f625893286f652596431160d5"} Feb 02 09:09:44 crc kubenswrapper[4764]: I0202 09:09:44.361839 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-tpldd" event={"ID":"7fb1a6fb-ea84-468e-a6ca-02a3604f81a7","Type":"ContainerStarted","Data":"c5946bfe07db8cffc8e27d6f0250d48ddb0bf20d9dbfd1b75ddb13c7d26e4754"} Feb 02 09:09:44 crc kubenswrapper[4764]: I0202 09:09:44.369956 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9nkvv\" (UID: \"527edab8-f88f-4d82-b42e-b8d52fcaa7d4\") " pod="openshift-image-registry/image-registry-697d97f7c8-9nkvv" Feb 02 09:09:44 crc kubenswrapper[4764]: E0202 09:09:44.370359 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 09:09:44.870344914 +0000 UTC m=+147.804069002 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9nkvv" (UID: "527edab8-f88f-4d82-b42e-b8d52fcaa7d4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:09:44 crc kubenswrapper[4764]: I0202 09:09:44.406251 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-hfwsg" event={"ID":"502b6ff7-05d4-41c4-b41a-31b70fb6247c","Type":"ContainerStarted","Data":"d537f403fe372996337eacb20a9e74e8575a9888b5d4b177f0babcc12b0ce976"} Feb 02 09:09:44 crc kubenswrapper[4764]: I0202 09:09:44.406298 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-hfwsg" event={"ID":"502b6ff7-05d4-41c4-b41a-31b70fb6247c","Type":"ContainerStarted","Data":"2b87bf4a63492257cf09d9b8f0d7925f78ddb2aa0ca238564097cdecdaddd84a"} Feb 02 09:09:44 crc kubenswrapper[4764]: I0202 09:09:44.452798 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29500380-2gwrp" podStartSLOduration=125.45278114 podStartE2EDuration="2m5.45278114s" podCreationTimestamp="2026-02-02 09:07:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:09:43.357885996 +0000 UTC m=+146.291610084" watchObservedRunningTime="2026-02-02 09:09:44.45278114 +0000 UTC m=+147.386505228" Feb 02 09:09:44 crc kubenswrapper[4764]: I0202 09:09:44.473181 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 09:09:44 crc kubenswrapper[4764]: E0202 09:09:44.473629 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 09:09:44.973605003 +0000 UTC m=+147.907329091 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:09:44 crc kubenswrapper[4764]: I0202 09:09:44.481311 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9nkvv\" (UID: \"527edab8-f88f-4d82-b42e-b8d52fcaa7d4\") " pod="openshift-image-registry/image-registry-697d97f7c8-9nkvv" Feb 02 09:09:44 crc kubenswrapper[4764]: E0202 09:09:44.482725 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 09:09:44.982713022 +0000 UTC m=+147.916437110 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9nkvv" (UID: "527edab8-f88f-4d82-b42e-b8d52fcaa7d4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:09:44 crc kubenswrapper[4764]: I0202 09:09:44.487727 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-p66d5" event={"ID":"42967bdb-847e-4823-b7b4-979844830e87","Type":"ContainerStarted","Data":"50bcdd5348cc4edf541573a85ec7d2bcf9e132672eee7ea75fa0ad13462a34d7"} Feb 02 09:09:44 crc kubenswrapper[4764]: I0202 09:09:44.539419 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4298w" event={"ID":"37d3c5dc-0535-4f2e-aa2f-debd99e0f1e3","Type":"ContainerStarted","Data":"8bb51eb136af1e94516a00d26e9fa4e83fd5eee190f8f5c28e86014b4a44bc77"} Feb 02 09:09:44 crc kubenswrapper[4764]: I0202 09:09:44.541401 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4298w" Feb 02 09:09:44 crc kubenswrapper[4764]: I0202 09:09:44.547463 4764 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-4298w container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.26:8443/healthz\": dial tcp 10.217.0.26:8443: connect: connection refused" start-of-body= Feb 02 09:09:44 crc kubenswrapper[4764]: I0202 09:09:44.547487 4764 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4298w" podUID="37d3c5dc-0535-4f2e-aa2f-debd99e0f1e3" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.26:8443/healthz\": dial tcp 10.217.0.26:8443: connect: connection refused" Feb 02 09:09:44 crc kubenswrapper[4764]: I0202 09:09:44.583245 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 09:09:44 crc kubenswrapper[4764]: E0202 09:09:44.583665 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 09:09:45.083647454 +0000 UTC m=+148.017371542 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:09:44 crc kubenswrapper[4764]: I0202 09:09:44.587519 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-df5xv" event={"ID":"d7534d2e-fb32-4895-945b-ea5ccd35fcbb","Type":"ContainerStarted","Data":"3b0b62374287a35b5fcc345f0c38f7552c857d61f2d005e8e454c390e7999553"} Feb 02 09:09:44 crc kubenswrapper[4764]: I0202 09:09:44.589654 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-hfwsg" podStartSLOduration=6.589637854 podStartE2EDuration="6.589637854s" podCreationTimestamp="2026-02-02 09:09:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:09:44.460947142 +0000 UTC m=+147.394671230" watchObservedRunningTime="2026-02-02 09:09:44.589637854 +0000 UTC m=+147.523361952" Feb 02 09:09:44 crc kubenswrapper[4764]: I0202 09:09:44.590532 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-fffd9" event={"ID":"7df533b7-06a3-4a8f-abfa-d7a45272d93d","Type":"ContainerStarted","Data":"be4a69b2fa3c5eb15487e43664f0a969bad3ad039de7dfe6cd9ef77a631aa74f"} Feb 02 09:09:44 crc kubenswrapper[4764]: I0202 09:09:44.591379 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4298w" podStartSLOduration=125.591373544 podStartE2EDuration="2m5.591373544s" podCreationTimestamp="2026-02-02 09:07:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:09:44.589018196 +0000 UTC m=+147.522742284" watchObservedRunningTime="2026-02-02 09:09:44.591373544 +0000 UTC m=+147.525097632" Feb 02 09:09:44 crc kubenswrapper[4764]: I0202 09:09:44.592076 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-9mdwc" event={"ID":"513ecde9-f189-4a22-8742-1b262523d462","Type":"ContainerStarted","Data":"64648b600bc1ccc6aa5e81e14eb194fd7afbe90c2ec5dcfbbaa066f903a1bb9d"} Feb 02 09:09:44 crc kubenswrapper[4764]: I0202 09:09:44.592670 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-9mdwc" Feb 02 09:09:44 crc kubenswrapper[4764]: I0202 09:09:44.593948 4764 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-9mdwc container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.23:6443/healthz\": dial tcp 10.217.0.23:6443: connect: connection refused" start-of-body= Feb 02 09:09:44 crc kubenswrapper[4764]: I0202 09:09:44.593977 4764 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-9mdwc" podUID="513ecde9-f189-4a22-8742-1b262523d462" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.23:6443/healthz\": dial tcp 10.217.0.23:6443: connect: connection refused" Feb 02 09:09:44 crc kubenswrapper[4764]: I0202 09:09:44.634244 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-l5jcj" event={"ID":"9a4ebc92-5ef5-4088-a509-5399beef3c7d","Type":"ContainerStarted","Data":"0d63da897f2a5fc8bcb646ae583d059946b20bbeee7dd38cdac1aac0787e38cc"} Feb 02 09:09:44 crc kubenswrapper[4764]: I0202 09:09:44.647809 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2cdx8" event={"ID":"36de8e6a-591b-4365-b434-ebd8197abf41","Type":"ContainerStarted","Data":"86004215591bc5538d1e56220b2a5d6c1ad6284625d779f77537884982c95d14"} Feb 02 09:09:44 crc kubenswrapper[4764]: I0202 09:09:44.656196 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2m47l" event={"ID":"a1f840f6-138f-496d-b64d-fbe6d49e0bdd","Type":"ContainerStarted","Data":"02f1463d7ceb9421b417421290078c7962b112826f8e1b057e038787c203d3b1"} Feb 02 09:09:44 crc kubenswrapper[4764]: I0202 09:09:44.663595 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-wqbtk" event={"ID":"fac965f9-5e91-4c50-a538-48c1301fa072","Type":"ContainerStarted","Data":"9b90e50046dcc02abf4663e6e01f3a315df60680b90f7748e87bcd022716f6a1"} Feb 02 09:09:44 crc kubenswrapper[4764]: I0202 09:09:44.664995 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-drqfg" event={"ID":"91e66500-e26e-4421-83e2-62b738656d5d","Type":"ContainerStarted","Data":"a37ed7a3fdc7423170c0135806f63ea2fa95d34e28c9aec33f202abd4fcc23e2"} Feb 02 09:09:44 crc kubenswrapper[4764]: I0202 09:09:44.667636 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ngtsq" event={"ID":"40fcb5fe-ee7b-4696-a775-efc703af9b61","Type":"ContainerStarted","Data":"a3c2ca00994a3c9a918cb5da9c4f4403f96876a5e86010456c2b21ecedffae30"} Feb 02 09:09:44 crc kubenswrapper[4764]: I0202 09:09:44.680510 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-mkwq2" event={"ID":"83fed8a7-bd5f-463d-b70e-3a01700e8315","Type":"ContainerStarted","Data":"4051f2b382355e29c74a945ef0af9b13a7814772b6c38aaed26f6df71b834146"} Feb 02 09:09:44 crc kubenswrapper[4764]: I0202 09:09:44.685056 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9nkvv\" (UID: \"527edab8-f88f-4d82-b42e-b8d52fcaa7d4\") " pod="openshift-image-registry/image-registry-697d97f7c8-9nkvv" Feb 02 09:09:44 crc kubenswrapper[4764]: E0202 09:09:44.686587 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 09:09:45.186576012 +0000 UTC m=+148.120300100 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9nkvv" (UID: "527edab8-f88f-4d82-b42e-b8d52fcaa7d4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:09:44 crc kubenswrapper[4764]: I0202 09:09:44.710655 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-4h9v4" event={"ID":"1eaf78eb-9719-4737-9f33-a8ab1dcb5dcf","Type":"ContainerStarted","Data":"d123ee830ade4caade62bc0079fbd8072a34ae18d280f07c870bcf798b4245e0"} Feb 02 09:09:44 crc kubenswrapper[4764]: I0202 09:09:44.710989 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-4h9v4" Feb 02 09:09:44 crc kubenswrapper[4764]: I0202 09:09:44.720163 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-njspk" event={"ID":"cb1b5d6b-7d42-4b40-99e3-4827a5379133","Type":"ContainerStarted","Data":"c75b828c2bf11e60cd9ab8b5c8c88c94914ced6947d71193746e4e77777c8ef3"} Feb 02 09:09:44 crc kubenswrapper[4764]: I0202 09:09:44.720997 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-njspk" Feb 02 09:09:44 crc kubenswrapper[4764]: I0202 09:09:44.725073 4764 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-njspk container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.22:5443/healthz\": dial tcp 10.217.0.22:5443: connect: connection refused" start-of-body= Feb 02 09:09:44 crc kubenswrapper[4764]: I0202 09:09:44.725112 4764 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-njspk" podUID="cb1b5d6b-7d42-4b40-99e3-4827a5379133" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.22:5443/healthz\": dial tcp 10.217.0.22:5443: connect: connection refused" Feb 02 09:09:44 crc kubenswrapper[4764]: I0202 09:09:44.740986 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-h5vrh" event={"ID":"6c0cd275-710b-43d8-a9b9-b46bc379454b","Type":"ContainerStarted","Data":"3bb444a5f396900c7ff0f06b61a655ad051fa86d942ed6add46c97fbabdcad40"} Feb 02 09:09:44 crc kubenswrapper[4764]: I0202 09:09:44.765118 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-b894n" event={"ID":"d17edd1e-7d52-4c32-90a9-7d16723f3875","Type":"ContainerStarted","Data":"d6b3528fc517e4aa8a7658148ab8a310743f76b544a18892e28c6b451df599e1"} Feb 02 09:09:44 crc kubenswrapper[4764]: I0202 09:09:44.782564 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-7v5xl" event={"ID":"c8a9e64e-fad9-466c-84b9-2d33a9f20a3b","Type":"ContainerStarted","Data":"06c711ab7ba2f9b5c17866827c93f97d1d198d7ac6d36e53e6ca1e26db6dd582"} Feb 02 09:09:44 crc kubenswrapper[4764]: I0202 09:09:44.787508 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 09:09:44 crc kubenswrapper[4764]: E0202 09:09:44.788160 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 09:09:45.288132982 +0000 UTC m=+148.221857070 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:09:44 crc kubenswrapper[4764]: I0202 09:09:44.806524 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-tkx8j" event={"ID":"fea30e22-e331-4518-a777-b6258c053f06","Type":"ContainerStarted","Data":"3ca5bd24aa214fe4c56aab3b0934eada12e80b9ad0be031947821d38ce146fe0"} Feb 02 09:09:44 crc kubenswrapper[4764]: I0202 09:09:44.814246 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-df5xv" podStartSLOduration=126.814223994 podStartE2EDuration="2m6.814223994s" podCreationTimestamp="2026-02-02 09:07:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:09:44.673275144 +0000 UTC m=+147.606999232" watchObservedRunningTime="2026-02-02 09:09:44.814223994 +0000 UTC m=+147.747948102" Feb 02 09:09:44 crc kubenswrapper[4764]: I0202 09:09:44.816199 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-l5jcj" podStartSLOduration=125.81618993 podStartE2EDuration="2m5.81618993s" podCreationTimestamp="2026-02-02 09:07:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:09:44.813736951 +0000 UTC m=+147.747461039" watchObservedRunningTime="2026-02-02 09:09:44.81618993 +0000 UTC m=+147.749914018" Feb 02 09:09:44 crc kubenswrapper[4764]: I0202 09:09:44.829770 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qq96b" event={"ID":"545e08e9-b2c4-493e-8fc7-8353fcab998d","Type":"ContainerStarted","Data":"0224607c1663d77f5d627a133ea370aa1a7577703b0c4085ed8b9e472b02acde"} Feb 02 09:09:44 crc kubenswrapper[4764]: I0202 09:09:44.839174 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9flm5" event={"ID":"501c9952-ea7c-4f45-b181-e25abae6986b","Type":"ContainerStarted","Data":"1a90dcc518223060fc1c949f86eedf55f78f925b04579600edba6f69f099a7bd"} Feb 02 09:09:44 crc kubenswrapper[4764]: I0202 09:09:44.853075 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-6dljc" event={"ID":"f552c254-ade6-4fe4-baa7-04e53c67479b","Type":"ContainerStarted","Data":"0dd3633ee54d0a68233b8af2beaf0c246cff0dc845c61847e0d08566ddc5cc31"} Feb 02 09:09:44 crc kubenswrapper[4764]: I0202 09:09:44.856926 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-zh65k" event={"ID":"348075cc-bd79-42f2-bbe2-0e5367092963","Type":"ContainerStarted","Data":"2783be859f5e9dc53a721b2edd2cdfaf2ff73fa3982203b554f2aefb643111b7"} Feb 02 09:09:44 crc kubenswrapper[4764]: I0202 09:09:44.859727 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-gp4jg" event={"ID":"dc0a29a2-27b1-4f7a-afaa-a0ae6cd290e9","Type":"ContainerStarted","Data":"319becf6713fc7cc0472f5ae519dfd30bfb8f057f887209c204969c6bc269601"} Feb 02 09:09:44 crc kubenswrapper[4764]: I0202 09:09:44.864183 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-dlpkk" event={"ID":"88f8519e-b576-4bef-96d6-771b78c67c11","Type":"ContainerStarted","Data":"1d97cac06a2e43b1b32ef0bbdad2b405f0e8391575ac7f1c8c5448d5b7d29744"} Feb 02 09:09:44 crc kubenswrapper[4764]: I0202 09:09:44.876976 4764 patch_prober.go:28] interesting pod/downloads-7954f5f757-7k5qb container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" start-of-body= Feb 02 09:09:44 crc kubenswrapper[4764]: I0202 09:09:44.877056 4764 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-7k5qb" podUID="c4b729bd-782f-4393-b162-16c02eb3cb71" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" Feb 02 09:09:44 crc kubenswrapper[4764]: I0202 09:09:44.890686 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9nkvv\" (UID: \"527edab8-f88f-4d82-b42e-b8d52fcaa7d4\") " pod="openshift-image-registry/image-registry-697d97f7c8-9nkvv" Feb 02 09:09:44 crc kubenswrapper[4764]: E0202 09:09:44.892790 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 09:09:45.39277863 +0000 UTC m=+148.326502708 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9nkvv" (UID: "527edab8-f88f-4d82-b42e-b8d52fcaa7d4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:09:44 crc kubenswrapper[4764]: I0202 09:09:44.905483 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-966nh" Feb 02 09:09:44 crc kubenswrapper[4764]: I0202 09:09:44.905690 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-52th4" Feb 02 09:09:44 crc kubenswrapper[4764]: I0202 09:09:44.908580 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mmcq4" Feb 02 09:09:44 crc kubenswrapper[4764]: I0202 09:09:44.995592 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-scdc2" Feb 02 09:09:45 crc kubenswrapper[4764]: I0202 09:09:44.998638 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 09:09:45 crc kubenswrapper[4764]: E0202 09:09:45.000676 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 09:09:45.500658049 +0000 UTC m=+148.434382147 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:09:45 crc kubenswrapper[4764]: I0202 09:09:45.029825 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-9mdwc" podStartSLOduration=127.029808579 podStartE2EDuration="2m7.029808579s" podCreationTimestamp="2026-02-02 09:07:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:09:44.939231471 +0000 UTC m=+147.872955559" watchObservedRunningTime="2026-02-02 09:09:45.029808579 +0000 UTC m=+147.963532667" Feb 02 09:09:45 crc kubenswrapper[4764]: I0202 09:09:45.032895 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ngtsq" podStartSLOduration=126.032884356 podStartE2EDuration="2m6.032884356s" podCreationTimestamp="2026-02-02 09:07:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:09:44.996701337 +0000 UTC m=+147.930425415" watchObservedRunningTime="2026-02-02 09:09:45.032884356 +0000 UTC m=+147.966608444" Feb 02 09:09:45 crc kubenswrapper[4764]: I0202 09:09:45.056301 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-drqfg" podStartSLOduration=126.056282662 podStartE2EDuration="2m6.056282662s" podCreationTimestamp="2026-02-02 09:07:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:09:45.046781252 +0000 UTC m=+147.980505350" watchObservedRunningTime="2026-02-02 09:09:45.056282662 +0000 UTC m=+147.990006750" Feb 02 09:09:45 crc kubenswrapper[4764]: I0202 09:09:45.115668 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9nkvv\" (UID: \"527edab8-f88f-4d82-b42e-b8d52fcaa7d4\") " pod="openshift-image-registry/image-registry-697d97f7c8-9nkvv" Feb 02 09:09:45 crc kubenswrapper[4764]: E0202 09:09:45.115958 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 09:09:45.61594562 +0000 UTC m=+148.549669708 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9nkvv" (UID: "527edab8-f88f-4d82-b42e-b8d52fcaa7d4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:09:45 crc kubenswrapper[4764]: I0202 09:09:45.172791 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ngtsq" Feb 02 09:09:45 crc kubenswrapper[4764]: I0202 09:09:45.173151 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ngtsq" Feb 02 09:09:45 crc kubenswrapper[4764]: I0202 09:09:45.197731 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-dlpkk" podStartSLOduration=126.197713376 podStartE2EDuration="2m6.197713376s" podCreationTimestamp="2026-02-02 09:07:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:09:45.197332485 +0000 UTC m=+148.131056583" watchObservedRunningTime="2026-02-02 09:09:45.197713376 +0000 UTC m=+148.131437464" Feb 02 09:09:45 crc kubenswrapper[4764]: I0202 09:09:45.219581 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 09:09:45 crc kubenswrapper[4764]: E0202 09:09:45.219944 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 09:09:45.719913818 +0000 UTC m=+148.653637906 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:09:45 crc kubenswrapper[4764]: I0202 09:09:45.246245 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-mkwq2" Feb 02 09:09:45 crc kubenswrapper[4764]: I0202 09:09:45.250227 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-6dljc" podStartSLOduration=127.25020996 podStartE2EDuration="2m7.25020996s" podCreationTimestamp="2026-02-02 09:07:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:09:45.246795083 +0000 UTC m=+148.180519171" watchObservedRunningTime="2026-02-02 09:09:45.25020996 +0000 UTC m=+148.183934048" Feb 02 09:09:45 crc kubenswrapper[4764]: I0202 09:09:45.268084 4764 patch_prober.go:28] interesting pod/router-default-5444994796-mkwq2 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 02 09:09:45 crc kubenswrapper[4764]: [-]has-synced failed: reason withheld Feb 02 09:09:45 crc kubenswrapper[4764]: [+]process-running ok Feb 02 09:09:45 crc kubenswrapper[4764]: healthz check failed Feb 02 09:09:45 crc kubenswrapper[4764]: I0202 09:09:45.268139 4764 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-mkwq2" podUID="83fed8a7-bd5f-463d-b70e-3a01700e8315" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 02 09:09:45 crc kubenswrapper[4764]: I0202 09:09:45.311465 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-njspk" podStartSLOduration=126.311449462 podStartE2EDuration="2m6.311449462s" podCreationTimestamp="2026-02-02 09:07:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:09:45.311194825 +0000 UTC m=+148.244918913" watchObservedRunningTime="2026-02-02 09:09:45.311449462 +0000 UTC m=+148.245173550" Feb 02 09:09:45 crc kubenswrapper[4764]: I0202 09:09:45.328323 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9nkvv\" (UID: \"527edab8-f88f-4d82-b42e-b8d52fcaa7d4\") " pod="openshift-image-registry/image-registry-697d97f7c8-9nkvv" Feb 02 09:09:45 crc kubenswrapper[4764]: E0202 09:09:45.328660 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 09:09:45.828650642 +0000 UTC m=+148.762374730 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9nkvv" (UID: "527edab8-f88f-4d82-b42e-b8d52fcaa7d4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:09:45 crc kubenswrapper[4764]: I0202 09:09:45.429567 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 09:09:45 crc kubenswrapper[4764]: E0202 09:09:45.429864 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 09:09:45.929849111 +0000 UTC m=+148.863573199 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:09:45 crc kubenswrapper[4764]: I0202 09:09:45.461314 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-mkwq2" podStartSLOduration=127.461299796 podStartE2EDuration="2m7.461299796s" podCreationTimestamp="2026-02-02 09:07:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:09:45.460761871 +0000 UTC m=+148.394485959" watchObservedRunningTime="2026-02-02 09:09:45.461299796 +0000 UTC m=+148.395023884" Feb 02 09:09:45 crc kubenswrapper[4764]: I0202 09:09:45.530895 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9nkvv\" (UID: \"527edab8-f88f-4d82-b42e-b8d52fcaa7d4\") " pod="openshift-image-registry/image-registry-697d97f7c8-9nkvv" Feb 02 09:09:45 crc kubenswrapper[4764]: E0202 09:09:45.531367 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 09:09:46.031350239 +0000 UTC m=+148.965074327 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9nkvv" (UID: "527edab8-f88f-4d82-b42e-b8d52fcaa7d4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:09:45 crc kubenswrapper[4764]: I0202 09:09:45.632587 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 09:09:45 crc kubenswrapper[4764]: E0202 09:09:45.632971 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 09:09:46.13294882 +0000 UTC m=+149.066672908 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:09:45 crc kubenswrapper[4764]: I0202 09:09:45.633366 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9nkvv\" (UID: \"527edab8-f88f-4d82-b42e-b8d52fcaa7d4\") " pod="openshift-image-registry/image-registry-697d97f7c8-9nkvv" Feb 02 09:09:45 crc kubenswrapper[4764]: E0202 09:09:45.633838 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 09:09:46.133827895 +0000 UTC m=+149.067551993 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9nkvv" (UID: "527edab8-f88f-4d82-b42e-b8d52fcaa7d4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:09:45 crc kubenswrapper[4764]: I0202 09:09:45.670748 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ngtsq" Feb 02 09:09:45 crc kubenswrapper[4764]: I0202 09:09:45.735449 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 09:09:45 crc kubenswrapper[4764]: E0202 09:09:45.735967 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 09:09:46.235952861 +0000 UTC m=+149.169676949 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:09:45 crc kubenswrapper[4764]: I0202 09:09:45.783300 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-zh65k" podStartSLOduration=127.783283288 podStartE2EDuration="2m7.783283288s" podCreationTimestamp="2026-02-02 09:07:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:09:45.781489347 +0000 UTC m=+148.715213435" watchObservedRunningTime="2026-02-02 09:09:45.783283288 +0000 UTC m=+148.717007376" Feb 02 09:09:45 crc kubenswrapper[4764]: I0202 09:09:45.836693 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9nkvv\" (UID: \"527edab8-f88f-4d82-b42e-b8d52fcaa7d4\") " pod="openshift-image-registry/image-registry-697d97f7c8-9nkvv" Feb 02 09:09:45 crc kubenswrapper[4764]: E0202 09:09:45.837072 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 09:09:46.337054428 +0000 UTC m=+149.270778516 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9nkvv" (UID: "527edab8-f88f-4d82-b42e-b8d52fcaa7d4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:09:45 crc kubenswrapper[4764]: I0202 09:09:45.873077 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-b894n" podStartSLOduration=7.873063302 podStartE2EDuration="7.873063302s" podCreationTimestamp="2026-02-02 09:09:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:09:45.871353584 +0000 UTC m=+148.805077672" watchObservedRunningTime="2026-02-02 09:09:45.873063302 +0000 UTC m=+148.806787390" Feb 02 09:09:45 crc kubenswrapper[4764]: I0202 09:09:45.888231 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-tpldd" event={"ID":"7fb1a6fb-ea84-468e-a6ca-02a3604f81a7","Type":"ContainerStarted","Data":"2b52e0b019c8dcc55108bf5861f9354b23e00b25a35730f3a38fba79757a0113"} Feb 02 09:09:45 crc kubenswrapper[4764]: I0202 09:09:45.888288 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-tpldd" event={"ID":"7fb1a6fb-ea84-468e-a6ca-02a3604f81a7","Type":"ContainerStarted","Data":"da76034eea72dbf22997bd58d2fbb632870a1acfd0dae86a5c372ea46b51d2a8"} Feb 02 09:09:45 crc kubenswrapper[4764]: I0202 09:09:45.891244 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-p66d5" event={"ID":"42967bdb-847e-4823-b7b4-979844830e87","Type":"ContainerStarted","Data":"b25ec4e1dc5f2b6b833cf6db34cb50797cd44756166691c4155cd0be06a687b8"} Feb 02 09:09:45 crc kubenswrapper[4764]: I0202 09:09:45.891300 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-p66d5" event={"ID":"42967bdb-847e-4823-b7b4-979844830e87","Type":"ContainerStarted","Data":"066de63bd021a447cc54747d5f81afa38b4ed42168f82dc4164c7a2e70233392"} Feb 02 09:09:45 crc kubenswrapper[4764]: I0202 09:09:45.891382 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-p66d5" Feb 02 09:09:45 crc kubenswrapper[4764]: I0202 09:09:45.892810 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-wqbtk" event={"ID":"fac965f9-5e91-4c50-a538-48c1301fa072","Type":"ContainerStarted","Data":"268bfe0e1d8733e258da407a82ee2f78a0dc2d86ecdbc5bafed848d6f8e77c4f"} Feb 02 09:09:45 crc kubenswrapper[4764]: I0202 09:09:45.893134 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-wqbtk" Feb 02 09:09:45 crc kubenswrapper[4764]: I0202 09:09:45.894760 4764 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-wqbtk container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.34:8080/healthz\": dial tcp 10.217.0.34:8080: connect: connection refused" start-of-body= Feb 02 09:09:45 crc kubenswrapper[4764]: I0202 09:09:45.894809 4764 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-wqbtk" podUID="fac965f9-5e91-4c50-a538-48c1301fa072" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.34:8080/healthz\": dial tcp 10.217.0.34:8080: connect: connection refused" Feb 02 09:09:45 crc kubenswrapper[4764]: I0202 09:09:45.897413 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-n59jp" event={"ID":"720e4205-b87d-4fc7-95df-3ad7c4849f97","Type":"ContainerStarted","Data":"761fbfdeb0b0bb6ff96a8b4efe287feeccf64caab47c3cf4eeb4583e3884db98"} Feb 02 09:09:45 crc kubenswrapper[4764]: I0202 09:09:45.913224 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-fffd9" event={"ID":"7df533b7-06a3-4a8f-abfa-d7a45272d93d","Type":"ContainerStarted","Data":"054ec12cf88b728df131d0ec145871fa7b92c72dacd3b6b7ae8c735f8aafc542"} Feb 02 09:09:45 crc kubenswrapper[4764]: I0202 09:09:45.923718 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2cdx8" event={"ID":"36de8e6a-591b-4365-b434-ebd8197abf41","Type":"ContainerStarted","Data":"33251bfd0fd4efb4b465fe2ab0af9246a664c09081b15ecdf6fc6b51f17ab241"} Feb 02 09:09:45 crc kubenswrapper[4764]: I0202 09:09:45.935540 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-7v5xl" event={"ID":"c8a9e64e-fad9-466c-84b9-2d33a9f20a3b","Type":"ContainerStarted","Data":"900950328e14f37be575b3d9db8c4ddcbf268f9d76ecb6804f75b00e29b805e1"} Feb 02 09:09:45 crc kubenswrapper[4764]: I0202 09:09:45.935578 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-7v5xl" event={"ID":"c8a9e64e-fad9-466c-84b9-2d33a9f20a3b","Type":"ContainerStarted","Data":"5b8e8cfabe673ad4ad37eba72f01121909ef8fd5bf79940f34ce595ab39c9148"} Feb 02 09:09:45 crc kubenswrapper[4764]: I0202 09:09:45.935709 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-7v5xl" Feb 02 09:09:45 crc kubenswrapper[4764]: I0202 09:09:45.938025 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 09:09:45 crc kubenswrapper[4764]: E0202 09:09:45.938178 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 09:09:46.438160455 +0000 UTC m=+149.371884543 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:09:45 crc kubenswrapper[4764]: I0202 09:09:45.938270 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9nkvv\" (UID: \"527edab8-f88f-4d82-b42e-b8d52fcaa7d4\") " pod="openshift-image-registry/image-registry-697d97f7c8-9nkvv" Feb 02 09:09:45 crc kubenswrapper[4764]: E0202 09:09:45.938523 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 09:09:46.438501634 +0000 UTC m=+149.372225712 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9nkvv" (UID: "527edab8-f88f-4d82-b42e-b8d52fcaa7d4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:09:45 crc kubenswrapper[4764]: I0202 09:09:45.948524 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2m47l" event={"ID":"a1f840f6-138f-496d-b64d-fbe6d49e0bdd","Type":"ContainerStarted","Data":"8f89f65253d69bb5c5281d051f997a54696ec3909482165de78c54ca5645dba2"} Feb 02 09:09:45 crc kubenswrapper[4764]: I0202 09:09:45.956047 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-7mtl6" event={"ID":"aec51473-5be4-4a0b-9ccc-7fcd912f9cf4","Type":"ContainerStarted","Data":"16b9aa7711a265b6931cb6ee655726b484fbfb5c52076274ad90674956923b5a"} Feb 02 09:09:45 crc kubenswrapper[4764]: I0202 09:09:45.993895 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-h5vrh" event={"ID":"6c0cd275-710b-43d8-a9b9-b46bc379454b","Type":"ContainerStarted","Data":"681a91480a2ea341dc6bc37ac3744a0bbe4e4f322d8ef76dfaf3d96ecee22a0e"} Feb 02 09:09:46 crc kubenswrapper[4764]: I0202 09:09:46.000721 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-t5q9w" event={"ID":"3ebf456e-6cb3-4cd4-a20d-a189ce1b9333","Type":"ContainerStarted","Data":"77ce248cbf5c0258650197181c77c3a7b835ad813a3f4c8765ee6347bbb49558"} Feb 02 09:09:46 crc kubenswrapper[4764]: I0202 09:09:46.006565 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-tkx8j" event={"ID":"fea30e22-e331-4518-a777-b6258c053f06","Type":"ContainerStarted","Data":"af952c1db357e6f55914abffe92a157c0711163430cb2d2a00f3dc2e0efcfc71"} Feb 02 09:09:46 crc kubenswrapper[4764]: I0202 09:09:46.014331 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9flm5" event={"ID":"501c9952-ea7c-4f45-b181-e25abae6986b","Type":"ContainerStarted","Data":"0570aa874f44e163b058d93110034a7def54898bfdeb4aee5e38d66ced757cc9"} Feb 02 09:09:46 crc kubenswrapper[4764]: I0202 09:09:46.021280 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-fdcsz" event={"ID":"f353fe2c-cc41-447a-b4b8-45eed47a29e2","Type":"ContainerStarted","Data":"c7056dfc3b761f32d12246d5dcb40999862e601d5ac9fb1a290a3c359da3d96e"} Feb 02 09:09:46 crc kubenswrapper[4764]: I0202 09:09:46.021348 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-fdcsz" event={"ID":"f353fe2c-cc41-447a-b4b8-45eed47a29e2","Type":"ContainerStarted","Data":"065fdebe369267bb2ff6315ba364f94a9a8c264d5fcb4b74c776307ab4866878"} Feb 02 09:09:46 crc kubenswrapper[4764]: I0202 09:09:46.023017 4764 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-9mdwc container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.23:6443/healthz\": dial tcp 10.217.0.23:6443: connect: connection refused" start-of-body= Feb 02 09:09:46 crc kubenswrapper[4764]: I0202 09:09:46.023056 4764 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-9mdwc" podUID="513ecde9-f189-4a22-8742-1b262523d462" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.23:6443/healthz\": dial tcp 10.217.0.23:6443: connect: connection refused" Feb 02 09:09:46 crc kubenswrapper[4764]: I0202 09:09:46.028311 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qq96b" podStartSLOduration=127.02829941 podStartE2EDuration="2m7.02829941s" podCreationTimestamp="2026-02-02 09:07:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:09:45.940000337 +0000 UTC m=+148.873724425" watchObservedRunningTime="2026-02-02 09:09:46.02829941 +0000 UTC m=+148.962023498" Feb 02 09:09:46 crc kubenswrapper[4764]: I0202 09:09:46.030038 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-4h9v4" podStartSLOduration=128.030033519 podStartE2EDuration="2m8.030033519s" podCreationTimestamp="2026-02-02 09:07:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:09:46.027285451 +0000 UTC m=+148.961009529" watchObservedRunningTime="2026-02-02 09:09:46.030033519 +0000 UTC m=+148.963757607" Feb 02 09:09:46 crc kubenswrapper[4764]: I0202 09:09:46.041747 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 09:09:46 crc kubenswrapper[4764]: I0202 09:09:46.045633 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ngtsq" Feb 02 09:09:46 crc kubenswrapper[4764]: E0202 09:09:46.046392 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 09:09:46.546374334 +0000 UTC m=+149.480098422 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:09:46 crc kubenswrapper[4764]: I0202 09:09:46.081031 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4298w" Feb 02 09:09:46 crc kubenswrapper[4764]: I0202 09:09:46.155274 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9nkvv\" (UID: \"527edab8-f88f-4d82-b42e-b8d52fcaa7d4\") " pod="openshift-image-registry/image-registry-697d97f7c8-9nkvv" Feb 02 09:09:46 crc kubenswrapper[4764]: E0202 09:09:46.161372 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 09:09:46.661353915 +0000 UTC m=+149.595078003 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9nkvv" (UID: "527edab8-f88f-4d82-b42e-b8d52fcaa7d4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:09:46 crc kubenswrapper[4764]: I0202 09:09:46.174493 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-gp4jg" podStartSLOduration=128.174471229 podStartE2EDuration="2m8.174471229s" podCreationTimestamp="2026-02-02 09:07:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:09:46.108970325 +0000 UTC m=+149.042694413" watchObservedRunningTime="2026-02-02 09:09:46.174471229 +0000 UTC m=+149.108195317" Feb 02 09:09:46 crc kubenswrapper[4764]: I0202 09:09:46.202732 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-h5vrh" podStartSLOduration=128.202696852 podStartE2EDuration="2m8.202696852s" podCreationTimestamp="2026-02-02 09:07:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:09:46.179734648 +0000 UTC m=+149.113458736" watchObservedRunningTime="2026-02-02 09:09:46.202696852 +0000 UTC m=+149.136420940" Feb 02 09:09:46 crc kubenswrapper[4764]: I0202 09:09:46.204745 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-ggrpx"] Feb 02 09:09:46 crc kubenswrapper[4764]: I0202 09:09:46.205902 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ggrpx" Feb 02 09:09:46 crc kubenswrapper[4764]: I0202 09:09:46.217656 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Feb 02 09:09:46 crc kubenswrapper[4764]: I0202 09:09:46.248469 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ggrpx"] Feb 02 09:09:46 crc kubenswrapper[4764]: I0202 09:09:46.248878 4764 patch_prober.go:28] interesting pod/router-default-5444994796-mkwq2 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 02 09:09:46 crc kubenswrapper[4764]: [-]has-synced failed: reason withheld Feb 02 09:09:46 crc kubenswrapper[4764]: [+]process-running ok Feb 02 09:09:46 crc kubenswrapper[4764]: healthz check failed Feb 02 09:09:46 crc kubenswrapper[4764]: I0202 09:09:46.249078 4764 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-mkwq2" podUID="83fed8a7-bd5f-463d-b70e-3a01700e8315" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 02 09:09:46 crc kubenswrapper[4764]: I0202 09:09:46.257120 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 09:09:46 crc kubenswrapper[4764]: E0202 09:09:46.257317 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 09:09:46.757290175 +0000 UTC m=+149.691014263 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:09:46 crc kubenswrapper[4764]: I0202 09:09:46.257515 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9nkvv\" (UID: \"527edab8-f88f-4d82-b42e-b8d52fcaa7d4\") " pod="openshift-image-registry/image-registry-697d97f7c8-9nkvv" Feb 02 09:09:46 crc kubenswrapper[4764]: E0202 09:09:46.257840 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 09:09:46.757833061 +0000 UTC m=+149.691557139 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9nkvv" (UID: "527edab8-f88f-4d82-b42e-b8d52fcaa7d4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:09:46 crc kubenswrapper[4764]: I0202 09:09:46.362422 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 09:09:46 crc kubenswrapper[4764]: E0202 09:09:46.362696 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 09:09:46.862644203 +0000 UTC m=+149.796368291 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:09:46 crc kubenswrapper[4764]: I0202 09:09:46.362774 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a40abe98-1419-4f5d-9bed-63f1c89028b1-utilities\") pod \"community-operators-ggrpx\" (UID: \"a40abe98-1419-4f5d-9bed-63f1c89028b1\") " pod="openshift-marketplace/community-operators-ggrpx" Feb 02 09:09:46 crc kubenswrapper[4764]: I0202 09:09:46.362878 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a40abe98-1419-4f5d-9bed-63f1c89028b1-catalog-content\") pod \"community-operators-ggrpx\" (UID: \"a40abe98-1419-4f5d-9bed-63f1c89028b1\") " pod="openshift-marketplace/community-operators-ggrpx" Feb 02 09:09:46 crc kubenswrapper[4764]: I0202 09:09:46.363090 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9nkvv\" (UID: \"527edab8-f88f-4d82-b42e-b8d52fcaa7d4\") " pod="openshift-image-registry/image-registry-697d97f7c8-9nkvv" Feb 02 09:09:46 crc kubenswrapper[4764]: I0202 09:09:46.363190 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-45j59\" (UniqueName: \"kubernetes.io/projected/a40abe98-1419-4f5d-9bed-63f1c89028b1-kube-api-access-45j59\") pod \"community-operators-ggrpx\" (UID: \"a40abe98-1419-4f5d-9bed-63f1c89028b1\") " pod="openshift-marketplace/community-operators-ggrpx" Feb 02 09:09:46 crc kubenswrapper[4764]: E0202 09:09:46.363656 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 09:09:46.863648261 +0000 UTC m=+149.797372349 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9nkvv" (UID: "527edab8-f88f-4d82-b42e-b8d52fcaa7d4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:09:46 crc kubenswrapper[4764]: I0202 09:09:46.426986 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9flm5" podStartSLOduration=127.426968593 podStartE2EDuration="2m7.426968593s" podCreationTimestamp="2026-02-02 09:07:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:09:46.349562221 +0000 UTC m=+149.283286299" watchObservedRunningTime="2026-02-02 09:09:46.426968593 +0000 UTC m=+149.360692681" Feb 02 09:09:46 crc kubenswrapper[4764]: I0202 09:09:46.429848 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-j25pg"] Feb 02 09:09:46 crc kubenswrapper[4764]: I0202 09:09:46.430902 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-j25pg" Feb 02 09:09:46 crc kubenswrapper[4764]: W0202 09:09:46.447044 4764 reflector.go:561] object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g": failed to list *v1.Secret: secrets "certified-operators-dockercfg-4rs5g" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-marketplace": no relationship found between node 'crc' and this object Feb 02 09:09:46 crc kubenswrapper[4764]: E0202 09:09:46.447104 4764 reflector.go:158] "Unhandled Error" err="object-\"openshift-marketplace\"/\"certified-operators-dockercfg-4rs5g\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"certified-operators-dockercfg-4rs5g\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-marketplace\": no relationship found between node 'crc' and this object" logger="UnhandledError" Feb 02 09:09:46 crc kubenswrapper[4764]: I0202 09:09:46.464498 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 09:09:46 crc kubenswrapper[4764]: I0202 09:09:46.464738 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a40abe98-1419-4f5d-9bed-63f1c89028b1-utilities\") pod \"community-operators-ggrpx\" (UID: \"a40abe98-1419-4f5d-9bed-63f1c89028b1\") " pod="openshift-marketplace/community-operators-ggrpx" Feb 02 09:09:46 crc kubenswrapper[4764]: I0202 09:09:46.464778 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a40abe98-1419-4f5d-9bed-63f1c89028b1-catalog-content\") pod \"community-operators-ggrpx\" (UID: \"a40abe98-1419-4f5d-9bed-63f1c89028b1\") " pod="openshift-marketplace/community-operators-ggrpx" Feb 02 09:09:46 crc kubenswrapper[4764]: I0202 09:09:46.464830 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-45j59\" (UniqueName: \"kubernetes.io/projected/a40abe98-1419-4f5d-9bed-63f1c89028b1-kube-api-access-45j59\") pod \"community-operators-ggrpx\" (UID: \"a40abe98-1419-4f5d-9bed-63f1c89028b1\") " pod="openshift-marketplace/community-operators-ggrpx" Feb 02 09:09:46 crc kubenswrapper[4764]: E0202 09:09:46.465203 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 09:09:46.965189721 +0000 UTC m=+149.898913809 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:09:46 crc kubenswrapper[4764]: I0202 09:09:46.465553 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a40abe98-1419-4f5d-9bed-63f1c89028b1-utilities\") pod \"community-operators-ggrpx\" (UID: \"a40abe98-1419-4f5d-9bed-63f1c89028b1\") " pod="openshift-marketplace/community-operators-ggrpx" Feb 02 09:09:46 crc kubenswrapper[4764]: I0202 09:09:46.465762 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a40abe98-1419-4f5d-9bed-63f1c89028b1-catalog-content\") pod \"community-operators-ggrpx\" (UID: \"a40abe98-1419-4f5d-9bed-63f1c89028b1\") " pod="openshift-marketplace/community-operators-ggrpx" Feb 02 09:09:46 crc kubenswrapper[4764]: I0202 09:09:46.520610 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-fdcsz" podStartSLOduration=127.520592397 podStartE2EDuration="2m7.520592397s" podCreationTimestamp="2026-02-02 09:07:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:09:46.484821619 +0000 UTC m=+149.418545707" watchObservedRunningTime="2026-02-02 09:09:46.520592397 +0000 UTC m=+149.454316485" Feb 02 09:09:46 crc kubenswrapper[4764]: I0202 09:09:46.534353 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-j25pg"] Feb 02 09:09:46 crc kubenswrapper[4764]: I0202 09:09:46.552952 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-45j59\" (UniqueName: \"kubernetes.io/projected/a40abe98-1419-4f5d-9bed-63f1c89028b1-kube-api-access-45j59\") pod \"community-operators-ggrpx\" (UID: \"a40abe98-1419-4f5d-9bed-63f1c89028b1\") " pod="openshift-marketplace/community-operators-ggrpx" Feb 02 09:09:46 crc kubenswrapper[4764]: I0202 09:09:46.566008 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9nkvv\" (UID: \"527edab8-f88f-4d82-b42e-b8d52fcaa7d4\") " pod="openshift-image-registry/image-registry-697d97f7c8-9nkvv" Feb 02 09:09:46 crc kubenswrapper[4764]: I0202 09:09:46.566083 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-knbhq\" (UniqueName: \"kubernetes.io/projected/01ede9d5-6b94-4d10-bf6d-3aef751840ff-kube-api-access-knbhq\") pod \"certified-operators-j25pg\" (UID: \"01ede9d5-6b94-4d10-bf6d-3aef751840ff\") " pod="openshift-marketplace/certified-operators-j25pg" Feb 02 09:09:46 crc kubenswrapper[4764]: I0202 09:09:46.566104 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/01ede9d5-6b94-4d10-bf6d-3aef751840ff-utilities\") pod \"certified-operators-j25pg\" (UID: \"01ede9d5-6b94-4d10-bf6d-3aef751840ff\") " pod="openshift-marketplace/certified-operators-j25pg" Feb 02 09:09:46 crc kubenswrapper[4764]: I0202 09:09:46.566140 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/01ede9d5-6b94-4d10-bf6d-3aef751840ff-catalog-content\") pod \"certified-operators-j25pg\" (UID: \"01ede9d5-6b94-4d10-bf6d-3aef751840ff\") " pod="openshift-marketplace/certified-operators-j25pg" Feb 02 09:09:46 crc kubenswrapper[4764]: E0202 09:09:46.566405 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 09:09:47.0663883 +0000 UTC m=+150.000112388 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9nkvv" (UID: "527edab8-f88f-4d82-b42e-b8d52fcaa7d4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:09:46 crc kubenswrapper[4764]: I0202 09:09:46.616038 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2cdx8" podStartSLOduration=127.616021013 podStartE2EDuration="2m7.616021013s" podCreationTimestamp="2026-02-02 09:07:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:09:46.615654412 +0000 UTC m=+149.549378500" watchObservedRunningTime="2026-02-02 09:09:46.616021013 +0000 UTC m=+149.549745101" Feb 02 09:09:46 crc kubenswrapper[4764]: I0202 09:09:46.637257 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-mg5dh"] Feb 02 09:09:46 crc kubenswrapper[4764]: I0202 09:09:46.638130 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mg5dh" Feb 02 09:09:46 crc kubenswrapper[4764]: I0202 09:09:46.667384 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 09:09:46 crc kubenswrapper[4764]: E0202 09:09:46.667714 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 09:09:47.167560349 +0000 UTC m=+150.101284437 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:09:46 crc kubenswrapper[4764]: I0202 09:09:46.668000 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/01ede9d5-6b94-4d10-bf6d-3aef751840ff-catalog-content\") pod \"certified-operators-j25pg\" (UID: \"01ede9d5-6b94-4d10-bf6d-3aef751840ff\") " pod="openshift-marketplace/certified-operators-j25pg" Feb 02 09:09:46 crc kubenswrapper[4764]: I0202 09:09:46.668298 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9nkvv\" (UID: \"527edab8-f88f-4d82-b42e-b8d52fcaa7d4\") " pod="openshift-image-registry/image-registry-697d97f7c8-9nkvv" Feb 02 09:09:46 crc kubenswrapper[4764]: I0202 09:09:46.668417 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/01ede9d5-6b94-4d10-bf6d-3aef751840ff-utilities\") pod \"certified-operators-j25pg\" (UID: \"01ede9d5-6b94-4d10-bf6d-3aef751840ff\") " pod="openshift-marketplace/certified-operators-j25pg" Feb 02 09:09:46 crc kubenswrapper[4764]: I0202 09:09:46.668451 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-knbhq\" (UniqueName: \"kubernetes.io/projected/01ede9d5-6b94-4d10-bf6d-3aef751840ff-kube-api-access-knbhq\") pod \"certified-operators-j25pg\" (UID: \"01ede9d5-6b94-4d10-bf6d-3aef751840ff\") " pod="openshift-marketplace/certified-operators-j25pg" Feb 02 09:09:46 crc kubenswrapper[4764]: I0202 09:09:46.668426 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/01ede9d5-6b94-4d10-bf6d-3aef751840ff-catalog-content\") pod \"certified-operators-j25pg\" (UID: \"01ede9d5-6b94-4d10-bf6d-3aef751840ff\") " pod="openshift-marketplace/certified-operators-j25pg" Feb 02 09:09:46 crc kubenswrapper[4764]: I0202 09:09:46.668628 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/01ede9d5-6b94-4d10-bf6d-3aef751840ff-utilities\") pod \"certified-operators-j25pg\" (UID: \"01ede9d5-6b94-4d10-bf6d-3aef751840ff\") " pod="openshift-marketplace/certified-operators-j25pg" Feb 02 09:09:46 crc kubenswrapper[4764]: E0202 09:09:46.668723 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 09:09:47.168703702 +0000 UTC m=+150.102427790 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9nkvv" (UID: "527edab8-f88f-4d82-b42e-b8d52fcaa7d4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:09:46 crc kubenswrapper[4764]: I0202 09:09:46.695204 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mg5dh"] Feb 02 09:09:46 crc kubenswrapper[4764]: I0202 09:09:46.722837 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-7mtl6" podStartSLOduration=127.722819471 podStartE2EDuration="2m7.722819471s" podCreationTimestamp="2026-02-02 09:07:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:09:46.715580445 +0000 UTC m=+149.649304533" watchObservedRunningTime="2026-02-02 09:09:46.722819471 +0000 UTC m=+149.656543559" Feb 02 09:09:46 crc kubenswrapper[4764]: I0202 09:09:46.738685 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-knbhq\" (UniqueName: \"kubernetes.io/projected/01ede9d5-6b94-4d10-bf6d-3aef751840ff-kube-api-access-knbhq\") pod \"certified-operators-j25pg\" (UID: \"01ede9d5-6b94-4d10-bf6d-3aef751840ff\") " pod="openshift-marketplace/certified-operators-j25pg" Feb 02 09:09:46 crc kubenswrapper[4764]: I0202 09:09:46.771480 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 09:09:46 crc kubenswrapper[4764]: I0202 09:09:46.771849 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nnpm4\" (UniqueName: \"kubernetes.io/projected/35376bf8-2200-49a4-8eb3-f5d7069e73a0-kube-api-access-nnpm4\") pod \"community-operators-mg5dh\" (UID: \"35376bf8-2200-49a4-8eb3-f5d7069e73a0\") " pod="openshift-marketplace/community-operators-mg5dh" Feb 02 09:09:46 crc kubenswrapper[4764]: I0202 09:09:46.771966 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/35376bf8-2200-49a4-8eb3-f5d7069e73a0-catalog-content\") pod \"community-operators-mg5dh\" (UID: \"35376bf8-2200-49a4-8eb3-f5d7069e73a0\") " pod="openshift-marketplace/community-operators-mg5dh" Feb 02 09:09:46 crc kubenswrapper[4764]: I0202 09:09:46.772059 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 09:09:46 crc kubenswrapper[4764]: I0202 09:09:46.772170 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/35376bf8-2200-49a4-8eb3-f5d7069e73a0-utilities\") pod \"community-operators-mg5dh\" (UID: \"35376bf8-2200-49a4-8eb3-f5d7069e73a0\") " pod="openshift-marketplace/community-operators-mg5dh" Feb 02 09:09:46 crc kubenswrapper[4764]: I0202 09:09:46.772258 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 09:09:46 crc kubenswrapper[4764]: I0202 09:09:46.772334 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 09:09:46 crc kubenswrapper[4764]: I0202 09:09:46.772410 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 09:09:46 crc kubenswrapper[4764]: E0202 09:09:46.773684 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 09:09:47.273659487 +0000 UTC m=+150.207383575 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:09:46 crc kubenswrapper[4764]: I0202 09:09:46.777085 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 09:09:46 crc kubenswrapper[4764]: I0202 09:09:46.783481 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 09:09:46 crc kubenswrapper[4764]: I0202 09:09:46.784562 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 09:09:46 crc kubenswrapper[4764]: I0202 09:09:46.785575 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-7p7nt"] Feb 02 09:09:46 crc kubenswrapper[4764]: I0202 09:09:46.786516 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7p7nt" Feb 02 09:09:46 crc kubenswrapper[4764]: I0202 09:09:46.793225 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 09:09:46 crc kubenswrapper[4764]: I0202 09:09:46.830274 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ggrpx" Feb 02 09:09:46 crc kubenswrapper[4764]: I0202 09:09:46.842825 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 09:09:46 crc kubenswrapper[4764]: I0202 09:09:46.853757 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 09:09:46 crc kubenswrapper[4764]: I0202 09:09:46.857017 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 09:09:46 crc kubenswrapper[4764]: I0202 09:09:46.863864 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-tkx8j" podStartSLOduration=128.863848913 podStartE2EDuration="2m8.863848913s" podCreationTimestamp="2026-02-02 09:07:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:09:46.85388055 +0000 UTC m=+149.787604638" watchObservedRunningTime="2026-02-02 09:09:46.863848913 +0000 UTC m=+149.797573001" Feb 02 09:09:46 crc kubenswrapper[4764]: I0202 09:09:46.871329 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7p7nt"] Feb 02 09:09:46 crc kubenswrapper[4764]: I0202 09:09:46.878622 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2e7d8f2e-1913-4b88-942a-995ad472697e-catalog-content\") pod \"certified-operators-7p7nt\" (UID: \"2e7d8f2e-1913-4b88-942a-995ad472697e\") " pod="openshift-marketplace/certified-operators-7p7nt" Feb 02 09:09:46 crc kubenswrapper[4764]: I0202 09:09:46.878662 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nnpm4\" (UniqueName: \"kubernetes.io/projected/35376bf8-2200-49a4-8eb3-f5d7069e73a0-kube-api-access-nnpm4\") pod \"community-operators-mg5dh\" (UID: \"35376bf8-2200-49a4-8eb3-f5d7069e73a0\") " pod="openshift-marketplace/community-operators-mg5dh" Feb 02 09:09:46 crc kubenswrapper[4764]: I0202 09:09:46.878679 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2e7d8f2e-1913-4b88-942a-995ad472697e-utilities\") pod \"certified-operators-7p7nt\" (UID: \"2e7d8f2e-1913-4b88-942a-995ad472697e\") " pod="openshift-marketplace/certified-operators-7p7nt" Feb 02 09:09:46 crc kubenswrapper[4764]: I0202 09:09:46.878701 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/35376bf8-2200-49a4-8eb3-f5d7069e73a0-catalog-content\") pod \"community-operators-mg5dh\" (UID: \"35376bf8-2200-49a4-8eb3-f5d7069e73a0\") " pod="openshift-marketplace/community-operators-mg5dh" Feb 02 09:09:46 crc kubenswrapper[4764]: I0202 09:09:46.878725 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8c22s\" (UniqueName: \"kubernetes.io/projected/2e7d8f2e-1913-4b88-942a-995ad472697e-kube-api-access-8c22s\") pod \"certified-operators-7p7nt\" (UID: \"2e7d8f2e-1913-4b88-942a-995ad472697e\") " pod="openshift-marketplace/certified-operators-7p7nt" Feb 02 09:09:46 crc kubenswrapper[4764]: I0202 09:09:46.878768 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9nkvv\" (UID: \"527edab8-f88f-4d82-b42e-b8d52fcaa7d4\") " pod="openshift-image-registry/image-registry-697d97f7c8-9nkvv" Feb 02 09:09:46 crc kubenswrapper[4764]: I0202 09:09:46.878787 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/35376bf8-2200-49a4-8eb3-f5d7069e73a0-utilities\") pod \"community-operators-mg5dh\" (UID: \"35376bf8-2200-49a4-8eb3-f5d7069e73a0\") " pod="openshift-marketplace/community-operators-mg5dh" Feb 02 09:09:46 crc kubenswrapper[4764]: I0202 09:09:46.879495 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/35376bf8-2200-49a4-8eb3-f5d7069e73a0-utilities\") pod \"community-operators-mg5dh\" (UID: \"35376bf8-2200-49a4-8eb3-f5d7069e73a0\") " pod="openshift-marketplace/community-operators-mg5dh" Feb 02 09:09:46 crc kubenswrapper[4764]: I0202 09:09:46.879990 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/35376bf8-2200-49a4-8eb3-f5d7069e73a0-catalog-content\") pod \"community-operators-mg5dh\" (UID: \"35376bf8-2200-49a4-8eb3-f5d7069e73a0\") " pod="openshift-marketplace/community-operators-mg5dh" Feb 02 09:09:46 crc kubenswrapper[4764]: E0202 09:09:46.880232 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 09:09:47.380221359 +0000 UTC m=+150.313945447 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9nkvv" (UID: "527edab8-f88f-4d82-b42e-b8d52fcaa7d4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:09:46 crc kubenswrapper[4764]: I0202 09:09:46.981222 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 09:09:46 crc kubenswrapper[4764]: E0202 09:09:46.981377 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 09:09:47.481348376 +0000 UTC m=+150.415072464 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:09:46 crc kubenswrapper[4764]: I0202 09:09:46.981414 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8c22s\" (UniqueName: \"kubernetes.io/projected/2e7d8f2e-1913-4b88-942a-995ad472697e-kube-api-access-8c22s\") pod \"certified-operators-7p7nt\" (UID: \"2e7d8f2e-1913-4b88-942a-995ad472697e\") " pod="openshift-marketplace/certified-operators-7p7nt" Feb 02 09:09:46 crc kubenswrapper[4764]: I0202 09:09:46.981505 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9nkvv\" (UID: \"527edab8-f88f-4d82-b42e-b8d52fcaa7d4\") " pod="openshift-image-registry/image-registry-697d97f7c8-9nkvv" Feb 02 09:09:46 crc kubenswrapper[4764]: I0202 09:09:46.981669 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2e7d8f2e-1913-4b88-942a-995ad472697e-catalog-content\") pod \"certified-operators-7p7nt\" (UID: \"2e7d8f2e-1913-4b88-942a-995ad472697e\") " pod="openshift-marketplace/certified-operators-7p7nt" Feb 02 09:09:46 crc kubenswrapper[4764]: I0202 09:09:46.981700 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2e7d8f2e-1913-4b88-942a-995ad472697e-utilities\") pod \"certified-operators-7p7nt\" (UID: \"2e7d8f2e-1913-4b88-942a-995ad472697e\") " pod="openshift-marketplace/certified-operators-7p7nt" Feb 02 09:09:46 crc kubenswrapper[4764]: I0202 09:09:46.982156 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2e7d8f2e-1913-4b88-942a-995ad472697e-utilities\") pod \"certified-operators-7p7nt\" (UID: \"2e7d8f2e-1913-4b88-942a-995ad472697e\") " pod="openshift-marketplace/certified-operators-7p7nt" Feb 02 09:09:46 crc kubenswrapper[4764]: E0202 09:09:46.982209 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 09:09:47.482193741 +0000 UTC m=+150.415917829 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9nkvv" (UID: "527edab8-f88f-4d82-b42e-b8d52fcaa7d4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:09:46 crc kubenswrapper[4764]: I0202 09:09:46.982359 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2e7d8f2e-1913-4b88-942a-995ad472697e-catalog-content\") pod \"certified-operators-7p7nt\" (UID: \"2e7d8f2e-1913-4b88-942a-995ad472697e\") " pod="openshift-marketplace/certified-operators-7p7nt" Feb 02 09:09:47 crc kubenswrapper[4764]: I0202 09:09:47.021578 4764 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-4h9v4 container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.16:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 02 09:09:47 crc kubenswrapper[4764]: I0202 09:09:47.021638 4764 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-4h9v4" podUID="1eaf78eb-9719-4737-9f33-a8ab1dcb5dcf" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.16:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Feb 02 09:09:47 crc kubenswrapper[4764]: I0202 09:09:47.021740 4764 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-njspk container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.22:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 02 09:09:47 crc kubenswrapper[4764]: I0202 09:09:47.021753 4764 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-njspk" podUID="cb1b5d6b-7d42-4b40-99e3-4827a5379133" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.22:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Feb 02 09:09:47 crc kubenswrapper[4764]: I0202 09:09:47.058703 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nnpm4\" (UniqueName: \"kubernetes.io/projected/35376bf8-2200-49a4-8eb3-f5d7069e73a0-kube-api-access-nnpm4\") pod \"community-operators-mg5dh\" (UID: \"35376bf8-2200-49a4-8eb3-f5d7069e73a0\") " pod="openshift-marketplace/community-operators-mg5dh" Feb 02 09:09:47 crc kubenswrapper[4764]: I0202 09:09:47.066279 4764 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-wqbtk container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.34:8080/healthz\": dial tcp 10.217.0.34:8080: connect: connection refused" start-of-body= Feb 02 09:09:47 crc kubenswrapper[4764]: I0202 09:09:47.066348 4764 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-wqbtk" podUID="fac965f9-5e91-4c50-a538-48c1301fa072" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.34:8080/healthz\": dial tcp 10.217.0.34:8080: connect: connection refused" Feb 02 09:09:47 crc kubenswrapper[4764]: I0202 09:09:47.066636 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8c22s\" (UniqueName: \"kubernetes.io/projected/2e7d8f2e-1913-4b88-942a-995ad472697e-kube-api-access-8c22s\") pod \"certified-operators-7p7nt\" (UID: \"2e7d8f2e-1913-4b88-942a-995ad472697e\") " pod="openshift-marketplace/certified-operators-7p7nt" Feb 02 09:09:47 crc kubenswrapper[4764]: I0202 09:09:47.083433 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 09:09:47 crc kubenswrapper[4764]: E0202 09:09:47.083891 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 09:09:47.583872944 +0000 UTC m=+150.517597032 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:09:47 crc kubenswrapper[4764]: I0202 09:09:47.096694 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-fffd9" podStartSLOduration=128.096676258 podStartE2EDuration="2m8.096676258s" podCreationTimestamp="2026-02-02 09:07:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:09:47.023564668 +0000 UTC m=+149.957288756" watchObservedRunningTime="2026-02-02 09:09:47.096676258 +0000 UTC m=+150.030400346" Feb 02 09:09:47 crc kubenswrapper[4764]: I0202 09:09:47.185687 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9nkvv\" (UID: \"527edab8-f88f-4d82-b42e-b8d52fcaa7d4\") " pod="openshift-image-registry/image-registry-697d97f7c8-9nkvv" Feb 02 09:09:47 crc kubenswrapper[4764]: E0202 09:09:47.188719 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 09:09:47.688704607 +0000 UTC m=+150.622428695 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9nkvv" (UID: "527edab8-f88f-4d82-b42e-b8d52fcaa7d4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:09:47 crc kubenswrapper[4764]: I0202 09:09:47.189560 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-tpldd" podStartSLOduration=128.189544121 podStartE2EDuration="2m8.189544121s" podCreationTimestamp="2026-02-02 09:07:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:09:47.097563743 +0000 UTC m=+150.031287831" watchObservedRunningTime="2026-02-02 09:09:47.189544121 +0000 UTC m=+150.123268199" Feb 02 09:09:47 crc kubenswrapper[4764]: I0202 09:09:47.189885 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-7v5xl" podStartSLOduration=9.18988149 podStartE2EDuration="9.18988149s" podCreationTimestamp="2026-02-02 09:09:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:09:47.187459701 +0000 UTC m=+150.121183789" watchObservedRunningTime="2026-02-02 09:09:47.18988149 +0000 UTC m=+150.123605578" Feb 02 09:09:47 crc kubenswrapper[4764]: I0202 09:09:47.246007 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-p66d5" podStartSLOduration=128.245987227 podStartE2EDuration="2m8.245987227s" podCreationTimestamp="2026-02-02 09:07:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:09:47.244497004 +0000 UTC m=+150.178221082" watchObservedRunningTime="2026-02-02 09:09:47.245987227 +0000 UTC m=+150.179711345" Feb 02 09:09:47 crc kubenswrapper[4764]: I0202 09:09:47.247458 4764 patch_prober.go:28] interesting pod/router-default-5444994796-mkwq2 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 02 09:09:47 crc kubenswrapper[4764]: [-]has-synced failed: reason withheld Feb 02 09:09:47 crc kubenswrapper[4764]: [+]process-running ok Feb 02 09:09:47 crc kubenswrapper[4764]: healthz check failed Feb 02 09:09:47 crc kubenswrapper[4764]: I0202 09:09:47.247508 4764 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-mkwq2" podUID="83fed8a7-bd5f-463d-b70e-3a01700e8315" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 02 09:09:47 crc kubenswrapper[4764]: I0202 09:09:47.250057 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mg5dh" Feb 02 09:09:47 crc kubenswrapper[4764]: I0202 09:09:47.280251 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Feb 02 09:09:47 crc kubenswrapper[4764]: I0202 09:09:47.282792 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7p7nt" Feb 02 09:09:47 crc kubenswrapper[4764]: I0202 09:09:47.283618 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-j25pg" Feb 02 09:09:47 crc kubenswrapper[4764]: I0202 09:09:47.287600 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 09:09:47 crc kubenswrapper[4764]: E0202 09:09:47.287814 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 09:09:47.787797166 +0000 UTC m=+150.721521254 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:09:47 crc kubenswrapper[4764]: I0202 09:09:47.287904 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9nkvv\" (UID: \"527edab8-f88f-4d82-b42e-b8d52fcaa7d4\") " pod="openshift-image-registry/image-registry-697d97f7c8-9nkvv" Feb 02 09:09:47 crc kubenswrapper[4764]: E0202 09:09:47.288250 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 09:09:47.788234859 +0000 UTC m=+150.721958947 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9nkvv" (UID: "527edab8-f88f-4d82-b42e-b8d52fcaa7d4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:09:47 crc kubenswrapper[4764]: I0202 09:09:47.314923 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2m47l" podStartSLOduration=129.314908138 podStartE2EDuration="2m9.314908138s" podCreationTimestamp="2026-02-02 09:07:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:09:47.314478445 +0000 UTC m=+150.248202523" watchObservedRunningTime="2026-02-02 09:09:47.314908138 +0000 UTC m=+150.248632226" Feb 02 09:09:47 crc kubenswrapper[4764]: I0202 09:09:47.389337 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 09:09:47 crc kubenswrapper[4764]: E0202 09:09:47.389490 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 09:09:47.889463389 +0000 UTC m=+150.823187477 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:09:47 crc kubenswrapper[4764]: I0202 09:09:47.389593 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9nkvv\" (UID: \"527edab8-f88f-4d82-b42e-b8d52fcaa7d4\") " pod="openshift-image-registry/image-registry-697d97f7c8-9nkvv" Feb 02 09:09:47 crc kubenswrapper[4764]: E0202 09:09:47.389946 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 09:09:47.889914322 +0000 UTC m=+150.823638410 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9nkvv" (UID: "527edab8-f88f-4d82-b42e-b8d52fcaa7d4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:09:47 crc kubenswrapper[4764]: I0202 09:09:47.423514 4764 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-4h9v4 container/openshift-config-operator namespace/openshift-config-operator: Liveness probe status=failure output="Get \"https://10.217.0.16:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 02 09:09:47 crc kubenswrapper[4764]: I0202 09:09:47.423614 4764 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-4h9v4" podUID="1eaf78eb-9719-4737-9f33-a8ab1dcb5dcf" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.16:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Feb 02 09:09:47 crc kubenswrapper[4764]: I0202 09:09:47.466458 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-n59jp" podStartSLOduration=129.466441739 podStartE2EDuration="2m9.466441739s" podCreationTimestamp="2026-02-02 09:07:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:09:47.464476563 +0000 UTC m=+150.398200651" watchObservedRunningTime="2026-02-02 09:09:47.466441739 +0000 UTC m=+150.400165827" Feb 02 09:09:47 crc kubenswrapper[4764]: I0202 09:09:47.491772 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 09:09:47 crc kubenswrapper[4764]: E0202 09:09:47.492173 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 09:09:47.992157921 +0000 UTC m=+150.925882009 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:09:47 crc kubenswrapper[4764]: I0202 09:09:47.592686 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-wqbtk" podStartSLOduration=128.592669861 podStartE2EDuration="2m8.592669861s" podCreationTimestamp="2026-02-02 09:07:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:09:47.531732747 +0000 UTC m=+150.465456835" watchObservedRunningTime="2026-02-02 09:09:47.592669861 +0000 UTC m=+150.526393949" Feb 02 09:09:47 crc kubenswrapper[4764]: I0202 09:09:47.593880 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9nkvv\" (UID: \"527edab8-f88f-4d82-b42e-b8d52fcaa7d4\") " pod="openshift-image-registry/image-registry-697d97f7c8-9nkvv" Feb 02 09:09:47 crc kubenswrapper[4764]: E0202 09:09:47.594175 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 09:09:48.094164863 +0000 UTC m=+151.027888951 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9nkvv" (UID: "527edab8-f88f-4d82-b42e-b8d52fcaa7d4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:09:47 crc kubenswrapper[4764]: I0202 09:09:47.699509 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 09:09:47 crc kubenswrapper[4764]: E0202 09:09:47.699691 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 09:09:48.199676336 +0000 UTC m=+151.133400424 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:09:47 crc kubenswrapper[4764]: I0202 09:09:47.699785 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9nkvv\" (UID: \"527edab8-f88f-4d82-b42e-b8d52fcaa7d4\") " pod="openshift-image-registry/image-registry-697d97f7c8-9nkvv" Feb 02 09:09:47 crc kubenswrapper[4764]: E0202 09:09:47.700098 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 09:09:48.200090927 +0000 UTC m=+151.133815015 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9nkvv" (UID: "527edab8-f88f-4d82-b42e-b8d52fcaa7d4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:09:47 crc kubenswrapper[4764]: I0202 09:09:47.800588 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 09:09:47 crc kubenswrapper[4764]: E0202 09:09:47.800966 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 09:09:48.300947727 +0000 UTC m=+151.234671815 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:09:47 crc kubenswrapper[4764]: I0202 09:09:47.906584 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9nkvv\" (UID: \"527edab8-f88f-4d82-b42e-b8d52fcaa7d4\") " pod="openshift-image-registry/image-registry-697d97f7c8-9nkvv" Feb 02 09:09:47 crc kubenswrapper[4764]: E0202 09:09:47.908776 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 09:09:48.408763675 +0000 UTC m=+151.342487763 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9nkvv" (UID: "527edab8-f88f-4d82-b42e-b8d52fcaa7d4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:09:47 crc kubenswrapper[4764]: I0202 09:09:47.942041 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-njspk" Feb 02 09:09:48 crc kubenswrapper[4764]: I0202 09:09:48.010006 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 09:09:48 crc kubenswrapper[4764]: E0202 09:09:48.010433 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 09:09:48.510402147 +0000 UTC m=+151.444126235 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:09:48 crc kubenswrapper[4764]: I0202 09:09:48.022055 4764 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-4h9v4 container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.16:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 02 09:09:48 crc kubenswrapper[4764]: I0202 09:09:48.022211 4764 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-4h9v4" podUID="1eaf78eb-9719-4737-9f33-a8ab1dcb5dcf" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.16:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Feb 02 09:09:48 crc kubenswrapper[4764]: I0202 09:09:48.069697 4764 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-9mdwc container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.23:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 02 09:09:48 crc kubenswrapper[4764]: I0202 09:09:48.069853 4764 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-9mdwc" podUID="513ecde9-f189-4a22-8742-1b262523d462" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.23:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Feb 02 09:09:48 crc kubenswrapper[4764]: I0202 09:09:48.087304 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-t5q9w" event={"ID":"3ebf456e-6cb3-4cd4-a20d-a189ce1b9333","Type":"ContainerStarted","Data":"d7c2a4ac80ac3deb3e34c4da8f0a4735811bf94ce239b1dec729216072410bdf"} Feb 02 09:09:48 crc kubenswrapper[4764]: I0202 09:09:48.088876 4764 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-wqbtk container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.34:8080/healthz\": dial tcp 10.217.0.34:8080: connect: connection refused" start-of-body= Feb 02 09:09:48 crc kubenswrapper[4764]: I0202 09:09:48.088904 4764 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-wqbtk" podUID="fac965f9-5e91-4c50-a538-48c1301fa072" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.34:8080/healthz\": dial tcp 10.217.0.34:8080: connect: connection refused" Feb 02 09:09:48 crc kubenswrapper[4764]: I0202 09:09:48.111031 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9nkvv\" (UID: \"527edab8-f88f-4d82-b42e-b8d52fcaa7d4\") " pod="openshift-image-registry/image-registry-697d97f7c8-9nkvv" Feb 02 09:09:48 crc kubenswrapper[4764]: E0202 09:09:48.111387 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 09:09:48.61137595 +0000 UTC m=+151.545100038 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9nkvv" (UID: "527edab8-f88f-4d82-b42e-b8d52fcaa7d4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:09:48 crc kubenswrapper[4764]: I0202 09:09:48.212299 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 09:09:48 crc kubenswrapper[4764]: E0202 09:09:48.213557 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 09:09:48.713542567 +0000 UTC m=+151.647266655 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:09:48 crc kubenswrapper[4764]: I0202 09:09:48.294100 4764 patch_prober.go:28] interesting pod/router-default-5444994796-mkwq2 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 02 09:09:48 crc kubenswrapper[4764]: [-]has-synced failed: reason withheld Feb 02 09:09:48 crc kubenswrapper[4764]: [+]process-running ok Feb 02 09:09:48 crc kubenswrapper[4764]: healthz check failed Feb 02 09:09:48 crc kubenswrapper[4764]: I0202 09:09:48.294138 4764 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-mkwq2" podUID="83fed8a7-bd5f-463d-b70e-3a01700e8315" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 02 09:09:48 crc kubenswrapper[4764]: I0202 09:09:48.319515 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9nkvv\" (UID: \"527edab8-f88f-4d82-b42e-b8d52fcaa7d4\") " pod="openshift-image-registry/image-registry-697d97f7c8-9nkvv" Feb 02 09:09:48 crc kubenswrapper[4764]: E0202 09:09:48.320074 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 09:09:48.820062888 +0000 UTC m=+151.753786976 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9nkvv" (UID: "527edab8-f88f-4d82-b42e-b8d52fcaa7d4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:09:48 crc kubenswrapper[4764]: I0202 09:09:48.401384 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-kbk9s"] Feb 02 09:09:48 crc kubenswrapper[4764]: I0202 09:09:48.402365 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kbk9s" Feb 02 09:09:48 crc kubenswrapper[4764]: I0202 09:09:48.422411 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 09:09:48 crc kubenswrapper[4764]: E0202 09:09:48.422770 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 09:09:48.92275737 +0000 UTC m=+151.856481458 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:09:48 crc kubenswrapper[4764]: I0202 09:09:48.423499 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Feb 02 09:09:48 crc kubenswrapper[4764]: I0202 09:09:48.449785 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-kbk9s"] Feb 02 09:09:48 crc kubenswrapper[4764]: I0202 09:09:48.539954 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9nkvv\" (UID: \"527edab8-f88f-4d82-b42e-b8d52fcaa7d4\") " pod="openshift-image-registry/image-registry-697d97f7c8-9nkvv" Feb 02 09:09:48 crc kubenswrapper[4764]: I0202 09:09:48.540008 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6092680d-f31e-4b70-9d81-2d80661f6ac3-utilities\") pod \"redhat-marketplace-kbk9s\" (UID: \"6092680d-f31e-4b70-9d81-2d80661f6ac3\") " pod="openshift-marketplace/redhat-marketplace-kbk9s" Feb 02 09:09:48 crc kubenswrapper[4764]: I0202 09:09:48.540028 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6092680d-f31e-4b70-9d81-2d80661f6ac3-catalog-content\") pod \"redhat-marketplace-kbk9s\" (UID: \"6092680d-f31e-4b70-9d81-2d80661f6ac3\") " pod="openshift-marketplace/redhat-marketplace-kbk9s" Feb 02 09:09:48 crc kubenswrapper[4764]: I0202 09:09:48.540236 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nv9qk\" (UniqueName: \"kubernetes.io/projected/6092680d-f31e-4b70-9d81-2d80661f6ac3-kube-api-access-nv9qk\") pod \"redhat-marketplace-kbk9s\" (UID: \"6092680d-f31e-4b70-9d81-2d80661f6ac3\") " pod="openshift-marketplace/redhat-marketplace-kbk9s" Feb 02 09:09:48 crc kubenswrapper[4764]: E0202 09:09:48.575005 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 09:09:49.040437208 +0000 UTC m=+151.974161296 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9nkvv" (UID: "527edab8-f88f-4d82-b42e-b8d52fcaa7d4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:09:48 crc kubenswrapper[4764]: I0202 09:09:48.643458 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 09:09:48 crc kubenswrapper[4764]: I0202 09:09:48.643718 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6092680d-f31e-4b70-9d81-2d80661f6ac3-utilities\") pod \"redhat-marketplace-kbk9s\" (UID: \"6092680d-f31e-4b70-9d81-2d80661f6ac3\") " pod="openshift-marketplace/redhat-marketplace-kbk9s" Feb 02 09:09:48 crc kubenswrapper[4764]: I0202 09:09:48.643744 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6092680d-f31e-4b70-9d81-2d80661f6ac3-catalog-content\") pod \"redhat-marketplace-kbk9s\" (UID: \"6092680d-f31e-4b70-9d81-2d80661f6ac3\") " pod="openshift-marketplace/redhat-marketplace-kbk9s" Feb 02 09:09:48 crc kubenswrapper[4764]: I0202 09:09:48.643783 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nv9qk\" (UniqueName: \"kubernetes.io/projected/6092680d-f31e-4b70-9d81-2d80661f6ac3-kube-api-access-nv9qk\") pod \"redhat-marketplace-kbk9s\" (UID: \"6092680d-f31e-4b70-9d81-2d80661f6ac3\") " pod="openshift-marketplace/redhat-marketplace-kbk9s" Feb 02 09:09:48 crc kubenswrapper[4764]: E0202 09:09:48.644184 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 09:09:49.14416883 +0000 UTC m=+152.077892908 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:09:48 crc kubenswrapper[4764]: I0202 09:09:48.644516 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6092680d-f31e-4b70-9d81-2d80661f6ac3-utilities\") pod \"redhat-marketplace-kbk9s\" (UID: \"6092680d-f31e-4b70-9d81-2d80661f6ac3\") " pod="openshift-marketplace/redhat-marketplace-kbk9s" Feb 02 09:09:48 crc kubenswrapper[4764]: I0202 09:09:48.644738 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6092680d-f31e-4b70-9d81-2d80661f6ac3-catalog-content\") pod \"redhat-marketplace-kbk9s\" (UID: \"6092680d-f31e-4b70-9d81-2d80661f6ac3\") " pod="openshift-marketplace/redhat-marketplace-kbk9s" Feb 02 09:09:48 crc kubenswrapper[4764]: I0202 09:09:48.745367 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9nkvv\" (UID: \"527edab8-f88f-4d82-b42e-b8d52fcaa7d4\") " pod="openshift-image-registry/image-registry-697d97f7c8-9nkvv" Feb 02 09:09:48 crc kubenswrapper[4764]: E0202 09:09:48.745715 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 09:09:49.245701539 +0000 UTC m=+152.179425627 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9nkvv" (UID: "527edab8-f88f-4d82-b42e-b8d52fcaa7d4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:09:48 crc kubenswrapper[4764]: I0202 09:09:48.814989 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nv9qk\" (UniqueName: \"kubernetes.io/projected/6092680d-f31e-4b70-9d81-2d80661f6ac3-kube-api-access-nv9qk\") pod \"redhat-marketplace-kbk9s\" (UID: \"6092680d-f31e-4b70-9d81-2d80661f6ac3\") " pod="openshift-marketplace/redhat-marketplace-kbk9s" Feb 02 09:09:48 crc kubenswrapper[4764]: I0202 09:09:48.847157 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 09:09:48 crc kubenswrapper[4764]: E0202 09:09:48.847509 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 09:09:49.347493325 +0000 UTC m=+152.281217413 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:09:48 crc kubenswrapper[4764]: I0202 09:09:48.907542 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-hcbn7"] Feb 02 09:09:48 crc kubenswrapper[4764]: I0202 09:09:48.909318 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hcbn7" Feb 02 09:09:48 crc kubenswrapper[4764]: I0202 09:09:48.956671 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9nkvv\" (UID: \"527edab8-f88f-4d82-b42e-b8d52fcaa7d4\") " pod="openshift-image-registry/image-registry-697d97f7c8-9nkvv" Feb 02 09:09:48 crc kubenswrapper[4764]: E0202 09:09:48.956997 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 09:09:49.456982501 +0000 UTC m=+152.390706589 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9nkvv" (UID: "527edab8-f88f-4d82-b42e-b8d52fcaa7d4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:09:49 crc kubenswrapper[4764]: I0202 09:09:49.010052 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hcbn7"] Feb 02 09:09:49 crc kubenswrapper[4764]: I0202 09:09:49.057915 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 09:09:49 crc kubenswrapper[4764]: E0202 09:09:49.058125 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 09:09:49.558097628 +0000 UTC m=+152.491821706 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:09:49 crc kubenswrapper[4764]: I0202 09:09:49.058297 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/88d63f88-8b5e-44de-b8e8-1417bbdb6e8e-catalog-content\") pod \"redhat-marketplace-hcbn7\" (UID: \"88d63f88-8b5e-44de-b8e8-1417bbdb6e8e\") " pod="openshift-marketplace/redhat-marketplace-hcbn7" Feb 02 09:09:49 crc kubenswrapper[4764]: I0202 09:09:49.058420 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ghh25\" (UniqueName: \"kubernetes.io/projected/88d63f88-8b5e-44de-b8e8-1417bbdb6e8e-kube-api-access-ghh25\") pod \"redhat-marketplace-hcbn7\" (UID: \"88d63f88-8b5e-44de-b8e8-1417bbdb6e8e\") " pod="openshift-marketplace/redhat-marketplace-hcbn7" Feb 02 09:09:49 crc kubenswrapper[4764]: I0202 09:09:49.058458 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9nkvv\" (UID: \"527edab8-f88f-4d82-b42e-b8d52fcaa7d4\") " pod="openshift-image-registry/image-registry-697d97f7c8-9nkvv" Feb 02 09:09:49 crc kubenswrapper[4764]: I0202 09:09:49.058484 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/88d63f88-8b5e-44de-b8e8-1417bbdb6e8e-utilities\") pod \"redhat-marketplace-hcbn7\" (UID: \"88d63f88-8b5e-44de-b8e8-1417bbdb6e8e\") " pod="openshift-marketplace/redhat-marketplace-hcbn7" Feb 02 09:09:49 crc kubenswrapper[4764]: E0202 09:09:49.058794 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 09:09:49.558784947 +0000 UTC m=+152.492509025 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9nkvv" (UID: "527edab8-f88f-4d82-b42e-b8d52fcaa7d4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:09:49 crc kubenswrapper[4764]: I0202 09:09:49.066775 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kbk9s" Feb 02 09:09:49 crc kubenswrapper[4764]: I0202 09:09:49.105489 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"78a235d168bbfabe71d2e0bc8c0a3bfb2bb0f3fd730ac9c9bcc08e02d7dfe55f"} Feb 02 09:09:49 crc kubenswrapper[4764]: I0202 09:09:49.105807 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"0f40926e9cbf9be454fb10c7e316e2a88312c8bef90ad38b06b3d2342b82a659"} Feb 02 09:09:49 crc kubenswrapper[4764]: I0202 09:09:49.146883 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-t5q9w" event={"ID":"3ebf456e-6cb3-4cd4-a20d-a189ce1b9333","Type":"ContainerStarted","Data":"f25d404a4a52c32e2ef5c32c9c7c17c4b310ffd6eb965c2bf20a7cbf61caa8a6"} Feb 02 09:09:49 crc kubenswrapper[4764]: I0202 09:09:49.149416 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"8a9a6b938d41f42754566b4b067a75ae1a411f87c2fe8ba9d4032c66f3223c59"} Feb 02 09:09:49 crc kubenswrapper[4764]: I0202 09:09:49.159666 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 09:09:49 crc kubenswrapper[4764]: I0202 09:09:49.159966 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/88d63f88-8b5e-44de-b8e8-1417bbdb6e8e-catalog-content\") pod \"redhat-marketplace-hcbn7\" (UID: \"88d63f88-8b5e-44de-b8e8-1417bbdb6e8e\") " pod="openshift-marketplace/redhat-marketplace-hcbn7" Feb 02 09:09:49 crc kubenswrapper[4764]: I0202 09:09:49.160029 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ghh25\" (UniqueName: \"kubernetes.io/projected/88d63f88-8b5e-44de-b8e8-1417bbdb6e8e-kube-api-access-ghh25\") pod \"redhat-marketplace-hcbn7\" (UID: \"88d63f88-8b5e-44de-b8e8-1417bbdb6e8e\") " pod="openshift-marketplace/redhat-marketplace-hcbn7" Feb 02 09:09:49 crc kubenswrapper[4764]: I0202 09:09:49.160058 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/88d63f88-8b5e-44de-b8e8-1417bbdb6e8e-utilities\") pod \"redhat-marketplace-hcbn7\" (UID: \"88d63f88-8b5e-44de-b8e8-1417bbdb6e8e\") " pod="openshift-marketplace/redhat-marketplace-hcbn7" Feb 02 09:09:49 crc kubenswrapper[4764]: I0202 09:09:49.160548 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/88d63f88-8b5e-44de-b8e8-1417bbdb6e8e-utilities\") pod \"redhat-marketplace-hcbn7\" (UID: \"88d63f88-8b5e-44de-b8e8-1417bbdb6e8e\") " pod="openshift-marketplace/redhat-marketplace-hcbn7" Feb 02 09:09:49 crc kubenswrapper[4764]: E0202 09:09:49.160617 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 09:09:49.660602204 +0000 UTC m=+152.594326292 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:09:49 crc kubenswrapper[4764]: I0202 09:09:49.160836 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/88d63f88-8b5e-44de-b8e8-1417bbdb6e8e-catalog-content\") pod \"redhat-marketplace-hcbn7\" (UID: \"88d63f88-8b5e-44de-b8e8-1417bbdb6e8e\") " pod="openshift-marketplace/redhat-marketplace-hcbn7" Feb 02 09:09:49 crc kubenswrapper[4764]: I0202 09:09:49.168626 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-j25pg"] Feb 02 09:09:49 crc kubenswrapper[4764]: I0202 09:09:49.249410 4764 patch_prober.go:28] interesting pod/router-default-5444994796-mkwq2 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 02 09:09:49 crc kubenswrapper[4764]: [-]has-synced failed: reason withheld Feb 02 09:09:49 crc kubenswrapper[4764]: [+]process-running ok Feb 02 09:09:49 crc kubenswrapper[4764]: healthz check failed Feb 02 09:09:49 crc kubenswrapper[4764]: I0202 09:09:49.249471 4764 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-mkwq2" podUID="83fed8a7-bd5f-463d-b70e-3a01700e8315" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 02 09:09:49 crc kubenswrapper[4764]: I0202 09:09:49.266662 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9nkvv\" (UID: \"527edab8-f88f-4d82-b42e-b8d52fcaa7d4\") " pod="openshift-image-registry/image-registry-697d97f7c8-9nkvv" Feb 02 09:09:49 crc kubenswrapper[4764]: E0202 09:09:49.267642 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 09:09:49.76763021 +0000 UTC m=+152.701354298 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9nkvv" (UID: "527edab8-f88f-4d82-b42e-b8d52fcaa7d4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:09:49 crc kubenswrapper[4764]: I0202 09:09:49.296736 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ghh25\" (UniqueName: \"kubernetes.io/projected/88d63f88-8b5e-44de-b8e8-1417bbdb6e8e-kube-api-access-ghh25\") pod \"redhat-marketplace-hcbn7\" (UID: \"88d63f88-8b5e-44de-b8e8-1417bbdb6e8e\") " pod="openshift-marketplace/redhat-marketplace-hcbn7" Feb 02 09:09:49 crc kubenswrapper[4764]: I0202 09:09:49.370489 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 09:09:49 crc kubenswrapper[4764]: E0202 09:09:49.370622 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 09:09:49.87060708 +0000 UTC m=+152.804331168 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:09:49 crc kubenswrapper[4764]: I0202 09:09:49.370747 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9nkvv\" (UID: \"527edab8-f88f-4d82-b42e-b8d52fcaa7d4\") " pod="openshift-image-registry/image-registry-697d97f7c8-9nkvv" Feb 02 09:09:49 crc kubenswrapper[4764]: E0202 09:09:49.371072 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 09:09:49.871064943 +0000 UTC m=+152.804789031 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9nkvv" (UID: "527edab8-f88f-4d82-b42e-b8d52fcaa7d4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:09:49 crc kubenswrapper[4764]: I0202 09:09:49.374814 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ggrpx"] Feb 02 09:09:49 crc kubenswrapper[4764]: I0202 09:09:49.422780 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-ct55p"] Feb 02 09:09:49 crc kubenswrapper[4764]: I0202 09:09:49.424121 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ct55p" Feb 02 09:09:49 crc kubenswrapper[4764]: I0202 09:09:49.442025 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Feb 02 09:09:49 crc kubenswrapper[4764]: I0202 09:09:49.444150 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-4h9v4" Feb 02 09:09:49 crc kubenswrapper[4764]: I0202 09:09:49.472014 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 09:09:49 crc kubenswrapper[4764]: E0202 09:09:49.472486 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 09:09:49.972468438 +0000 UTC m=+152.906192526 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:09:49 crc kubenswrapper[4764]: I0202 09:09:49.539174 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hcbn7" Feb 02 09:09:49 crc kubenswrapper[4764]: I0202 09:09:49.574713 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9nkvv\" (UID: \"527edab8-f88f-4d82-b42e-b8d52fcaa7d4\") " pod="openshift-image-registry/image-registry-697d97f7c8-9nkvv" Feb 02 09:09:49 crc kubenswrapper[4764]: I0202 09:09:49.574765 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fpp4j\" (UniqueName: \"kubernetes.io/projected/00cb5ef3-3b54-47ac-bfc4-b1d53f04d292-kube-api-access-fpp4j\") pod \"redhat-operators-ct55p\" (UID: \"00cb5ef3-3b54-47ac-bfc4-b1d53f04d292\") " pod="openshift-marketplace/redhat-operators-ct55p" Feb 02 09:09:49 crc kubenswrapper[4764]: I0202 09:09:49.575348 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/00cb5ef3-3b54-47ac-bfc4-b1d53f04d292-utilities\") pod \"redhat-operators-ct55p\" (UID: \"00cb5ef3-3b54-47ac-bfc4-b1d53f04d292\") " pod="openshift-marketplace/redhat-operators-ct55p" Feb 02 09:09:49 crc kubenswrapper[4764]: I0202 09:09:49.575396 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/00cb5ef3-3b54-47ac-bfc4-b1d53f04d292-catalog-content\") pod \"redhat-operators-ct55p\" (UID: \"00cb5ef3-3b54-47ac-bfc4-b1d53f04d292\") " pod="openshift-marketplace/redhat-operators-ct55p" Feb 02 09:09:49 crc kubenswrapper[4764]: E0202 09:09:49.576431 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 09:09:50.076419466 +0000 UTC m=+153.010143554 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9nkvv" (UID: "527edab8-f88f-4d82-b42e-b8d52fcaa7d4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:09:49 crc kubenswrapper[4764]: I0202 09:09:49.678730 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 09:09:49 crc kubenswrapper[4764]: I0202 09:09:49.679287 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/00cb5ef3-3b54-47ac-bfc4-b1d53f04d292-utilities\") pod \"redhat-operators-ct55p\" (UID: \"00cb5ef3-3b54-47ac-bfc4-b1d53f04d292\") " pod="openshift-marketplace/redhat-operators-ct55p" Feb 02 09:09:49 crc kubenswrapper[4764]: I0202 09:09:49.679419 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/00cb5ef3-3b54-47ac-bfc4-b1d53f04d292-catalog-content\") pod \"redhat-operators-ct55p\" (UID: \"00cb5ef3-3b54-47ac-bfc4-b1d53f04d292\") " pod="openshift-marketplace/redhat-operators-ct55p" Feb 02 09:09:49 crc kubenswrapper[4764]: I0202 09:09:49.679565 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fpp4j\" (UniqueName: \"kubernetes.io/projected/00cb5ef3-3b54-47ac-bfc4-b1d53f04d292-kube-api-access-fpp4j\") pod \"redhat-operators-ct55p\" (UID: \"00cb5ef3-3b54-47ac-bfc4-b1d53f04d292\") " pod="openshift-marketplace/redhat-operators-ct55p" Feb 02 09:09:49 crc kubenswrapper[4764]: E0202 09:09:49.679659 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 09:09:50.179622593 +0000 UTC m=+153.113346681 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:09:49 crc kubenswrapper[4764]: I0202 09:09:49.680148 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/00cb5ef3-3b54-47ac-bfc4-b1d53f04d292-utilities\") pod \"redhat-operators-ct55p\" (UID: \"00cb5ef3-3b54-47ac-bfc4-b1d53f04d292\") " pod="openshift-marketplace/redhat-operators-ct55p" Feb 02 09:09:49 crc kubenswrapper[4764]: I0202 09:09:49.680392 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/00cb5ef3-3b54-47ac-bfc4-b1d53f04d292-catalog-content\") pod \"redhat-operators-ct55p\" (UID: \"00cb5ef3-3b54-47ac-bfc4-b1d53f04d292\") " pod="openshift-marketplace/redhat-operators-ct55p" Feb 02 09:09:49 crc kubenswrapper[4764]: I0202 09:09:49.691466 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-ct55p"] Feb 02 09:09:49 crc kubenswrapper[4764]: I0202 09:09:49.740643 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fpp4j\" (UniqueName: \"kubernetes.io/projected/00cb5ef3-3b54-47ac-bfc4-b1d53f04d292-kube-api-access-fpp4j\") pod \"redhat-operators-ct55p\" (UID: \"00cb5ef3-3b54-47ac-bfc4-b1d53f04d292\") " pod="openshift-marketplace/redhat-operators-ct55p" Feb 02 09:09:49 crc kubenswrapper[4764]: I0202 09:09:49.744021 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mg5dh"] Feb 02 09:09:49 crc kubenswrapper[4764]: I0202 09:09:49.761026 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7p7nt"] Feb 02 09:09:49 crc kubenswrapper[4764]: I0202 09:09:49.761375 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ct55p" Feb 02 09:09:49 crc kubenswrapper[4764]: W0202 09:09:49.769966 4764 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod35376bf8_2200_49a4_8eb3_f5d7069e73a0.slice/crio-38d29e8e0d2662f4c04febfb9d1ed4b29d1e3d1ef4a0510c248b2761f1ca4eed WatchSource:0}: Error finding container 38d29e8e0d2662f4c04febfb9d1ed4b29d1e3d1ef4a0510c248b2761f1ca4eed: Status 404 returned error can't find the container with id 38d29e8e0d2662f4c04febfb9d1ed4b29d1e3d1ef4a0510c248b2761f1ca4eed Feb 02 09:09:49 crc kubenswrapper[4764]: I0202 09:09:49.782658 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-w6hd5"] Feb 02 09:09:49 crc kubenswrapper[4764]: I0202 09:09:49.783668 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-w6hd5" Feb 02 09:09:49 crc kubenswrapper[4764]: I0202 09:09:49.784762 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9nkvv\" (UID: \"527edab8-f88f-4d82-b42e-b8d52fcaa7d4\") " pod="openshift-image-registry/image-registry-697d97f7c8-9nkvv" Feb 02 09:09:49 crc kubenswrapper[4764]: E0202 09:09:49.785073 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 09:09:50.285061783 +0000 UTC m=+153.218785871 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9nkvv" (UID: "527edab8-f88f-4d82-b42e-b8d52fcaa7d4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:09:49 crc kubenswrapper[4764]: W0202 09:09:49.814413 4764 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2e7d8f2e_1913_4b88_942a_995ad472697e.slice/crio-d075f0ec2e09cc7ea56f3ed96642ddbc69689f8351a4428d20ee4e10f9e69a40 WatchSource:0}: Error finding container d075f0ec2e09cc7ea56f3ed96642ddbc69689f8351a4428d20ee4e10f9e69a40: Status 404 returned error can't find the container with id d075f0ec2e09cc7ea56f3ed96642ddbc69689f8351a4428d20ee4e10f9e69a40 Feb 02 09:09:49 crc kubenswrapper[4764]: I0202 09:09:49.816319 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-w6hd5"] Feb 02 09:09:49 crc kubenswrapper[4764]: I0202 09:09:49.885462 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 09:09:49 crc kubenswrapper[4764]: I0202 09:09:49.885665 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bnkf7\" (UniqueName: \"kubernetes.io/projected/83d5cb3d-ee7b-44cb-8379-8b52f2c3cadc-kube-api-access-bnkf7\") pod \"redhat-operators-w6hd5\" (UID: \"83d5cb3d-ee7b-44cb-8379-8b52f2c3cadc\") " pod="openshift-marketplace/redhat-operators-w6hd5" Feb 02 09:09:49 crc kubenswrapper[4764]: I0202 09:09:49.885701 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83d5cb3d-ee7b-44cb-8379-8b52f2c3cadc-catalog-content\") pod \"redhat-operators-w6hd5\" (UID: \"83d5cb3d-ee7b-44cb-8379-8b52f2c3cadc\") " pod="openshift-marketplace/redhat-operators-w6hd5" Feb 02 09:09:49 crc kubenswrapper[4764]: I0202 09:09:49.885806 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83d5cb3d-ee7b-44cb-8379-8b52f2c3cadc-utilities\") pod \"redhat-operators-w6hd5\" (UID: \"83d5cb3d-ee7b-44cb-8379-8b52f2c3cadc\") " pod="openshift-marketplace/redhat-operators-w6hd5" Feb 02 09:09:49 crc kubenswrapper[4764]: E0202 09:09:49.885944 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 09:09:50.385892342 +0000 UTC m=+153.319616430 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:09:49 crc kubenswrapper[4764]: I0202 09:09:49.987053 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9nkvv\" (UID: \"527edab8-f88f-4d82-b42e-b8d52fcaa7d4\") " pod="openshift-image-registry/image-registry-697d97f7c8-9nkvv" Feb 02 09:09:49 crc kubenswrapper[4764]: I0202 09:09:49.987106 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83d5cb3d-ee7b-44cb-8379-8b52f2c3cadc-utilities\") pod \"redhat-operators-w6hd5\" (UID: \"83d5cb3d-ee7b-44cb-8379-8b52f2c3cadc\") " pod="openshift-marketplace/redhat-operators-w6hd5" Feb 02 09:09:49 crc kubenswrapper[4764]: I0202 09:09:49.987128 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bnkf7\" (UniqueName: \"kubernetes.io/projected/83d5cb3d-ee7b-44cb-8379-8b52f2c3cadc-kube-api-access-bnkf7\") pod \"redhat-operators-w6hd5\" (UID: \"83d5cb3d-ee7b-44cb-8379-8b52f2c3cadc\") " pod="openshift-marketplace/redhat-operators-w6hd5" Feb 02 09:09:49 crc kubenswrapper[4764]: I0202 09:09:49.987154 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83d5cb3d-ee7b-44cb-8379-8b52f2c3cadc-catalog-content\") pod \"redhat-operators-w6hd5\" (UID: \"83d5cb3d-ee7b-44cb-8379-8b52f2c3cadc\") " pod="openshift-marketplace/redhat-operators-w6hd5" Feb 02 09:09:49 crc kubenswrapper[4764]: I0202 09:09:49.987916 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83d5cb3d-ee7b-44cb-8379-8b52f2c3cadc-catalog-content\") pod \"redhat-operators-w6hd5\" (UID: \"83d5cb3d-ee7b-44cb-8379-8b52f2c3cadc\") " pod="openshift-marketplace/redhat-operators-w6hd5" Feb 02 09:09:49 crc kubenswrapper[4764]: I0202 09:09:49.990484 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83d5cb3d-ee7b-44cb-8379-8b52f2c3cadc-utilities\") pod \"redhat-operators-w6hd5\" (UID: \"83d5cb3d-ee7b-44cb-8379-8b52f2c3cadc\") " pod="openshift-marketplace/redhat-operators-w6hd5" Feb 02 09:09:49 crc kubenswrapper[4764]: E0202 09:09:49.990715 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 09:09:50.490701594 +0000 UTC m=+153.424425682 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9nkvv" (UID: "527edab8-f88f-4d82-b42e-b8d52fcaa7d4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:09:50 crc kubenswrapper[4764]: I0202 09:09:50.067176 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bnkf7\" (UniqueName: \"kubernetes.io/projected/83d5cb3d-ee7b-44cb-8379-8b52f2c3cadc-kube-api-access-bnkf7\") pod \"redhat-operators-w6hd5\" (UID: \"83d5cb3d-ee7b-44cb-8379-8b52f2c3cadc\") " pod="openshift-marketplace/redhat-operators-w6hd5" Feb 02 09:09:50 crc kubenswrapper[4764]: I0202 09:09:50.087847 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 09:09:50 crc kubenswrapper[4764]: E0202 09:09:50.088243 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 09:09:50.588229009 +0000 UTC m=+153.521953097 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:09:50 crc kubenswrapper[4764]: I0202 09:09:50.106221 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-w6hd5" Feb 02 09:09:50 crc kubenswrapper[4764]: I0202 09:09:50.189265 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9nkvv\" (UID: \"527edab8-f88f-4d82-b42e-b8d52fcaa7d4\") " pod="openshift-image-registry/image-registry-697d97f7c8-9nkvv" Feb 02 09:09:50 crc kubenswrapper[4764]: E0202 09:09:50.189551 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 09:09:50.689540442 +0000 UTC m=+153.623264530 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9nkvv" (UID: "527edab8-f88f-4d82-b42e-b8d52fcaa7d4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:09:50 crc kubenswrapper[4764]: I0202 09:09:50.204270 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mg5dh" event={"ID":"35376bf8-2200-49a4-8eb3-f5d7069e73a0","Type":"ContainerStarted","Data":"38d29e8e0d2662f4c04febfb9d1ed4b29d1e3d1ef4a0510c248b2761f1ca4eed"} Feb 02 09:09:50 crc kubenswrapper[4764]: I0202 09:09:50.224147 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7p7nt" event={"ID":"2e7d8f2e-1913-4b88-942a-995ad472697e","Type":"ContainerStarted","Data":"39c8d9f8da4ffed5557047562303f1757f5cb2317304678028c8a785e36bd325"} Feb 02 09:09:50 crc kubenswrapper[4764]: I0202 09:09:50.224206 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7p7nt" event={"ID":"2e7d8f2e-1913-4b88-942a-995ad472697e","Type":"ContainerStarted","Data":"d075f0ec2e09cc7ea56f3ed96642ddbc69689f8351a4428d20ee4e10f9e69a40"} Feb 02 09:09:50 crc kubenswrapper[4764]: I0202 09:09:50.230090 4764 generic.go:334] "Generic (PLEG): container finished" podID="01ede9d5-6b94-4d10-bf6d-3aef751840ff" containerID="15c8dcdbf7dafd0945e7cee7ea86c7e3c74f639942d8cc81473c5ba785830e7d" exitCode=0 Feb 02 09:09:50 crc kubenswrapper[4764]: I0202 09:09:50.230154 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j25pg" event={"ID":"01ede9d5-6b94-4d10-bf6d-3aef751840ff","Type":"ContainerDied","Data":"15c8dcdbf7dafd0945e7cee7ea86c7e3c74f639942d8cc81473c5ba785830e7d"} Feb 02 09:09:50 crc kubenswrapper[4764]: I0202 09:09:50.230180 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j25pg" event={"ID":"01ede9d5-6b94-4d10-bf6d-3aef751840ff","Type":"ContainerStarted","Data":"4afd4af7d19135bfa9e0bb99cecbd3a39b149cfe37e4b26c30eaf267df71b3eb"} Feb 02 09:09:50 crc kubenswrapper[4764]: I0202 09:09:50.233062 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ggrpx" event={"ID":"a40abe98-1419-4f5d-9bed-63f1c89028b1","Type":"ContainerStarted","Data":"290a2505a57e59d6feb7b276cd44e863f761a91d2ed9ce7c3a89f09320ad086d"} Feb 02 09:09:50 crc kubenswrapper[4764]: I0202 09:09:50.233089 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ggrpx" event={"ID":"a40abe98-1419-4f5d-9bed-63f1c89028b1","Type":"ContainerStarted","Data":"bf2b4f22879cad3d2120b2aa1607123e3b5cb77bdf8877486020ee617b44a63c"} Feb 02 09:09:50 crc kubenswrapper[4764]: I0202 09:09:50.249170 4764 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 02 09:09:50 crc kubenswrapper[4764]: I0202 09:09:50.257832 4764 patch_prober.go:28] interesting pod/router-default-5444994796-mkwq2 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 02 09:09:50 crc kubenswrapper[4764]: [-]has-synced failed: reason withheld Feb 02 09:09:50 crc kubenswrapper[4764]: [+]process-running ok Feb 02 09:09:50 crc kubenswrapper[4764]: healthz check failed Feb 02 09:09:50 crc kubenswrapper[4764]: I0202 09:09:50.257891 4764 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-mkwq2" podUID="83fed8a7-bd5f-463d-b70e-3a01700e8315" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 02 09:09:50 crc kubenswrapper[4764]: I0202 09:09:50.273366 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-t5q9w" event={"ID":"3ebf456e-6cb3-4cd4-a20d-a189ce1b9333","Type":"ContainerStarted","Data":"75effeec37cfff4188229e2cb36ff08dfa63303371427a36e860ee965138413c"} Feb 02 09:09:50 crc kubenswrapper[4764]: I0202 09:09:50.290654 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 09:09:50 crc kubenswrapper[4764]: E0202 09:09:50.291086 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 09:09:50.791066241 +0000 UTC m=+153.724790329 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:09:50 crc kubenswrapper[4764]: I0202 09:09:50.298396 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"656a986f598bb7fe92df00564697a2b106875a3d83f2c80358387888e9d7dc09"} Feb 02 09:09:50 crc kubenswrapper[4764]: I0202 09:09:50.299141 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 09:09:50 crc kubenswrapper[4764]: I0202 09:09:50.333232 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-h5vrh" Feb 02 09:09:50 crc kubenswrapper[4764]: I0202 09:09:50.337512 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-h5vrh" Feb 02 09:09:50 crc kubenswrapper[4764]: I0202 09:09:50.338701 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-t5q9w" podStartSLOduration=12.338687876 podStartE2EDuration="12.338687876s" podCreationTimestamp="2026-02-02 09:09:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:09:50.337124211 +0000 UTC m=+153.270848299" watchObservedRunningTime="2026-02-02 09:09:50.338687876 +0000 UTC m=+153.272411964" Feb 02 09:09:50 crc kubenswrapper[4764]: I0202 09:09:50.357211 4764 patch_prober.go:28] interesting pod/apiserver-76f77b778f-h5vrh container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Feb 02 09:09:50 crc kubenswrapper[4764]: [+]log ok Feb 02 09:09:50 crc kubenswrapper[4764]: [+]etcd ok Feb 02 09:09:50 crc kubenswrapper[4764]: [+]poststarthook/start-apiserver-admission-initializer ok Feb 02 09:09:50 crc kubenswrapper[4764]: [+]poststarthook/generic-apiserver-start-informers ok Feb 02 09:09:50 crc kubenswrapper[4764]: [+]poststarthook/max-in-flight-filter ok Feb 02 09:09:50 crc kubenswrapper[4764]: [+]poststarthook/storage-object-count-tracker-hook ok Feb 02 09:09:50 crc kubenswrapper[4764]: [+]poststarthook/image.openshift.io-apiserver-caches ok Feb 02 09:09:50 crc kubenswrapper[4764]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Feb 02 09:09:50 crc kubenswrapper[4764]: [+]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa ok Feb 02 09:09:50 crc kubenswrapper[4764]: [+]poststarthook/project.openshift.io-projectcache ok Feb 02 09:09:50 crc kubenswrapper[4764]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Feb 02 09:09:50 crc kubenswrapper[4764]: [+]poststarthook/openshift.io-startinformers ok Feb 02 09:09:50 crc kubenswrapper[4764]: [+]poststarthook/openshift.io-restmapperupdater ok Feb 02 09:09:50 crc kubenswrapper[4764]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Feb 02 09:09:50 crc kubenswrapper[4764]: livez check failed Feb 02 09:09:50 crc kubenswrapper[4764]: I0202 09:09:50.357262 4764 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-h5vrh" podUID="6c0cd275-710b-43d8-a9b9-b46bc379454b" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 02 09:09:50 crc kubenswrapper[4764]: I0202 09:09:50.387464 4764 patch_prober.go:28] interesting pod/downloads-7954f5f757-7k5qb container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" start-of-body= Feb 02 09:09:50 crc kubenswrapper[4764]: I0202 09:09:50.387525 4764 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-7k5qb" podUID="c4b729bd-782f-4393-b162-16c02eb3cb71" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" Feb 02 09:09:50 crc kubenswrapper[4764]: I0202 09:09:50.387475 4764 patch_prober.go:28] interesting pod/downloads-7954f5f757-7k5qb container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" start-of-body= Feb 02 09:09:50 crc kubenswrapper[4764]: I0202 09:09:50.387598 4764 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-7k5qb" podUID="c4b729bd-782f-4393-b162-16c02eb3cb71" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" Feb 02 09:09:50 crc kubenswrapper[4764]: I0202 09:09:50.391735 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9nkvv\" (UID: \"527edab8-f88f-4d82-b42e-b8d52fcaa7d4\") " pod="openshift-image-registry/image-registry-697d97f7c8-9nkvv" Feb 02 09:09:50 crc kubenswrapper[4764]: E0202 09:09:50.392926 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 09:09:50.892906577 +0000 UTC m=+153.826630725 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9nkvv" (UID: "527edab8-f88f-4d82-b42e-b8d52fcaa7d4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:09:50 crc kubenswrapper[4764]: I0202 09:09:50.462327 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-cnkss" Feb 02 09:09:50 crc kubenswrapper[4764]: I0202 09:09:50.462362 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-cnkss" Feb 02 09:09:50 crc kubenswrapper[4764]: I0202 09:09:50.463057 4764 patch_prober.go:28] interesting pod/console-f9d7485db-cnkss container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.10:8443/health\": dial tcp 10.217.0.10:8443: connect: connection refused" start-of-body= Feb 02 09:09:50 crc kubenswrapper[4764]: I0202 09:09:50.463082 4764 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-cnkss" podUID="d874a84d-48e3-4cb2-ac74-4f0db52e9d75" containerName="console" probeResult="failure" output="Get \"https://10.217.0.10:8443/health\": dial tcp 10.217.0.10:8443: connect: connection refused" Feb 02 09:09:50 crc kubenswrapper[4764]: I0202 09:09:50.492956 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 09:09:50 crc kubenswrapper[4764]: E0202 09:09:50.494136 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 09:09:50.994115717 +0000 UTC m=+153.927839825 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:09:50 crc kubenswrapper[4764]: I0202 09:09:50.525910 4764 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Feb 02 09:09:50 crc kubenswrapper[4764]: I0202 09:09:50.535985 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-kbk9s"] Feb 02 09:09:50 crc kubenswrapper[4764]: I0202 09:09:50.599263 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9nkvv\" (UID: \"527edab8-f88f-4d82-b42e-b8d52fcaa7d4\") " pod="openshift-image-registry/image-registry-697d97f7c8-9nkvv" Feb 02 09:09:50 crc kubenswrapper[4764]: E0202 09:09:50.600369 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 09:09:51.10035517 +0000 UTC m=+154.034079258 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9nkvv" (UID: "527edab8-f88f-4d82-b42e-b8d52fcaa7d4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:09:50 crc kubenswrapper[4764]: I0202 09:09:50.700471 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 09:09:50 crc kubenswrapper[4764]: E0202 09:09:50.700752 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 09:09:51.200734456 +0000 UTC m=+154.134458534 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:09:50 crc kubenswrapper[4764]: I0202 09:09:50.802400 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9nkvv\" (UID: \"527edab8-f88f-4d82-b42e-b8d52fcaa7d4\") " pod="openshift-image-registry/image-registry-697d97f7c8-9nkvv" Feb 02 09:09:50 crc kubenswrapper[4764]: E0202 09:09:50.802784 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 09:09:51.302766199 +0000 UTC m=+154.236490287 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9nkvv" (UID: "527edab8-f88f-4d82-b42e-b8d52fcaa7d4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:09:50 crc kubenswrapper[4764]: I0202 09:09:50.907653 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 09:09:50 crc kubenswrapper[4764]: E0202 09:09:50.908114 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 09:09:51.408082576 +0000 UTC m=+154.341806664 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:09:50 crc kubenswrapper[4764]: I0202 09:09:50.908236 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9nkvv\" (UID: \"527edab8-f88f-4d82-b42e-b8d52fcaa7d4\") " pod="openshift-image-registry/image-registry-697d97f7c8-9nkvv" Feb 02 09:09:50 crc kubenswrapper[4764]: E0202 09:09:50.908585 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 09:09:51.40857042 +0000 UTC m=+154.342294508 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9nkvv" (UID: "527edab8-f88f-4d82-b42e-b8d52fcaa7d4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:09:50 crc kubenswrapper[4764]: I0202 09:09:50.991050 4764 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2026-02-02T09:09:50.525944633Z","Handler":null,"Name":""} Feb 02 09:09:51 crc kubenswrapper[4764]: I0202 09:09:51.008770 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 09:09:51 crc kubenswrapper[4764]: E0202 09:09:51.009151 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 09:09:51.509136981 +0000 UTC m=+154.442861069 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 09:09:51 crc kubenswrapper[4764]: I0202 09:09:51.032794 4764 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Feb 02 09:09:51 crc kubenswrapper[4764]: I0202 09:09:51.032823 4764 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Feb 02 09:09:51 crc kubenswrapper[4764]: I0202 09:09:51.076502 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hcbn7"] Feb 02 09:09:51 crc kubenswrapper[4764]: I0202 09:09:51.114618 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9nkvv\" (UID: \"527edab8-f88f-4d82-b42e-b8d52fcaa7d4\") " pod="openshift-image-registry/image-registry-697d97f7c8-9nkvv" Feb 02 09:09:51 crc kubenswrapper[4764]: I0202 09:09:51.127499 4764 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Feb 02 09:09:51 crc kubenswrapper[4764]: I0202 09:09:51.127527 4764 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9nkvv\" (UID: \"527edab8-f88f-4d82-b42e-b8d52fcaa7d4\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-9nkvv" Feb 02 09:09:51 crc kubenswrapper[4764]: I0202 09:09:51.214255 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-9mdwc" Feb 02 09:09:51 crc kubenswrapper[4764]: I0202 09:09:51.244252 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-mkwq2" Feb 02 09:09:51 crc kubenswrapper[4764]: I0202 09:09:51.254466 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-w6hd5"] Feb 02 09:09:51 crc kubenswrapper[4764]: I0202 09:09:51.255887 4764 patch_prober.go:28] interesting pod/router-default-5444994796-mkwq2 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 02 09:09:51 crc kubenswrapper[4764]: [-]has-synced failed: reason withheld Feb 02 09:09:51 crc kubenswrapper[4764]: [+]process-running ok Feb 02 09:09:51 crc kubenswrapper[4764]: healthz check failed Feb 02 09:09:51 crc kubenswrapper[4764]: I0202 09:09:51.255963 4764 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-mkwq2" podUID="83fed8a7-bd5f-463d-b70e-3a01700e8315" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 02 09:09:51 crc kubenswrapper[4764]: I0202 09:09:51.314000 4764 generic.go:334] "Generic (PLEG): container finished" podID="35376bf8-2200-49a4-8eb3-f5d7069e73a0" containerID="890230c30c2f811cfdc185446a0738979101d42a5f7332d0e7bd8edfb53af670" exitCode=0 Feb 02 09:09:51 crc kubenswrapper[4764]: I0202 09:09:51.314067 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mg5dh" event={"ID":"35376bf8-2200-49a4-8eb3-f5d7069e73a0","Type":"ContainerDied","Data":"890230c30c2f811cfdc185446a0738979101d42a5f7332d0e7bd8edfb53af670"} Feb 02 09:09:51 crc kubenswrapper[4764]: I0202 09:09:51.315793 4764 generic.go:334] "Generic (PLEG): container finished" podID="2e7d8f2e-1913-4b88-942a-995ad472697e" containerID="39c8d9f8da4ffed5557047562303f1757f5cb2317304678028c8a785e36bd325" exitCode=0 Feb 02 09:09:51 crc kubenswrapper[4764]: I0202 09:09:51.315830 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7p7nt" event={"ID":"2e7d8f2e-1913-4b88-942a-995ad472697e","Type":"ContainerDied","Data":"39c8d9f8da4ffed5557047562303f1757f5cb2317304678028c8a785e36bd325"} Feb 02 09:09:51 crc kubenswrapper[4764]: I0202 09:09:51.359658 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"9029a102e315ab3e8804dc49cf4d6be3fa51f8253438e1314a71ff860c71d355"} Feb 02 09:09:51 crc kubenswrapper[4764]: I0202 09:09:51.359703 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"df17c78769823eeb7c477c42f7734b9bef601d43f915f12224ee48ab6b002fab"} Feb 02 09:09:51 crc kubenswrapper[4764]: I0202 09:09:51.373081 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w6hd5" event={"ID":"83d5cb3d-ee7b-44cb-8379-8b52f2c3cadc","Type":"ContainerStarted","Data":"b16395216d7d99de002dce85a619c0d855fc5ca42a1b516108fcc982859314f4"} Feb 02 09:09:51 crc kubenswrapper[4764]: I0202 09:09:51.404977 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-ct55p"] Feb 02 09:09:51 crc kubenswrapper[4764]: I0202 09:09:51.405882 4764 generic.go:334] "Generic (PLEG): container finished" podID="6092680d-f31e-4b70-9d81-2d80661f6ac3" containerID="171e9563f8ad099d9e9c74f1bff0e55abe393140c07ff9c612151163ecb679ee" exitCode=0 Feb 02 09:09:51 crc kubenswrapper[4764]: I0202 09:09:51.405972 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kbk9s" event={"ID":"6092680d-f31e-4b70-9d81-2d80661f6ac3","Type":"ContainerDied","Data":"171e9563f8ad099d9e9c74f1bff0e55abe393140c07ff9c612151163ecb679ee"} Feb 02 09:09:51 crc kubenswrapper[4764]: I0202 09:09:51.405996 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kbk9s" event={"ID":"6092680d-f31e-4b70-9d81-2d80661f6ac3","Type":"ContainerStarted","Data":"b5261920939a0f8c06654f503c8da17cc956f864e29b35fe95aaa77044b2c2d5"} Feb 02 09:09:51 crc kubenswrapper[4764]: I0202 09:09:51.421326 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9nkvv\" (UID: \"527edab8-f88f-4d82-b42e-b8d52fcaa7d4\") " pod="openshift-image-registry/image-registry-697d97f7c8-9nkvv" Feb 02 09:09:51 crc kubenswrapper[4764]: I0202 09:09:51.434647 4764 generic.go:334] "Generic (PLEG): container finished" podID="a40abe98-1419-4f5d-9bed-63f1c89028b1" containerID="290a2505a57e59d6feb7b276cd44e863f761a91d2ed9ce7c3a89f09320ad086d" exitCode=0 Feb 02 09:09:51 crc kubenswrapper[4764]: I0202 09:09:51.434745 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ggrpx" event={"ID":"a40abe98-1419-4f5d-9bed-63f1c89028b1","Type":"ContainerDied","Data":"290a2505a57e59d6feb7b276cd44e863f761a91d2ed9ce7c3a89f09320ad086d"} Feb 02 09:09:51 crc kubenswrapper[4764]: I0202 09:09:51.442899 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hcbn7" event={"ID":"88d63f88-8b5e-44de-b8e8-1417bbdb6e8e","Type":"ContainerStarted","Data":"79bb3443189e6281d7de325470986c4c39a9563b198dfa9d99a1ef8766854151"} Feb 02 09:09:51 crc kubenswrapper[4764]: W0202 09:09:51.483480 4764 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod00cb5ef3_3b54_47ac_bfc4_b1d53f04d292.slice/crio-00838ebc3146464fb7e67e2e6bc4614aea995f37837f9d843754c1c48fecd95b WatchSource:0}: Error finding container 00838ebc3146464fb7e67e2e6bc4614aea995f37837f9d843754c1c48fecd95b: Status 404 returned error can't find the container with id 00838ebc3146464fb7e67e2e6bc4614aea995f37837f9d843754c1c48fecd95b Feb 02 09:09:51 crc kubenswrapper[4764]: I0202 09:09:51.520389 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 09:09:51 crc kubenswrapper[4764]: I0202 09:09:51.659642 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-wqbtk" Feb 02 09:09:51 crc kubenswrapper[4764]: I0202 09:09:51.662373 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Feb 02 09:09:51 crc kubenswrapper[4764]: I0202 09:09:51.718459 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-9nkvv" Feb 02 09:09:51 crc kubenswrapper[4764]: I0202 09:09:51.868502 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Feb 02 09:09:52 crc kubenswrapper[4764]: I0202 09:09:52.236777 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-9nkvv"] Feb 02 09:09:52 crc kubenswrapper[4764]: I0202 09:09:52.245634 4764 patch_prober.go:28] interesting pod/router-default-5444994796-mkwq2 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 02 09:09:52 crc kubenswrapper[4764]: [-]has-synced failed: reason withheld Feb 02 09:09:52 crc kubenswrapper[4764]: [+]process-running ok Feb 02 09:09:52 crc kubenswrapper[4764]: healthz check failed Feb 02 09:09:52 crc kubenswrapper[4764]: I0202 09:09:52.245674 4764 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-mkwq2" podUID="83fed8a7-bd5f-463d-b70e-3a01700e8315" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 02 09:09:52 crc kubenswrapper[4764]: W0202 09:09:52.295532 4764 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod527edab8_f88f_4d82_b42e_b8d52fcaa7d4.slice/crio-0b65b00bc2d8eff15f7d32631379347926baf789e3e9c91110f3de30b1b31c5e WatchSource:0}: Error finding container 0b65b00bc2d8eff15f7d32631379347926baf789e3e9c91110f3de30b1b31c5e: Status 404 returned error can't find the container with id 0b65b00bc2d8eff15f7d32631379347926baf789e3e9c91110f3de30b1b31c5e Feb 02 09:09:52 crc kubenswrapper[4764]: I0202 09:09:52.473149 4764 generic.go:334] "Generic (PLEG): container finished" podID="88d63f88-8b5e-44de-b8e8-1417bbdb6e8e" containerID="93c4bd63dd1eaf39ff8d7650175e8905d0f1a5e49eba7f09bdcb37a3cf07c003" exitCode=0 Feb 02 09:09:52 crc kubenswrapper[4764]: I0202 09:09:52.473806 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hcbn7" event={"ID":"88d63f88-8b5e-44de-b8e8-1417bbdb6e8e","Type":"ContainerDied","Data":"93c4bd63dd1eaf39ff8d7650175e8905d0f1a5e49eba7f09bdcb37a3cf07c003"} Feb 02 09:09:52 crc kubenswrapper[4764]: I0202 09:09:52.506220 4764 generic.go:334] "Generic (PLEG): container finished" podID="00cb5ef3-3b54-47ac-bfc4-b1d53f04d292" containerID="9a274df71070056d0ba90d60479e4818f9ac34548bc19489c14a87da15dc6c8b" exitCode=0 Feb 02 09:09:52 crc kubenswrapper[4764]: I0202 09:09:52.506514 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ct55p" event={"ID":"00cb5ef3-3b54-47ac-bfc4-b1d53f04d292","Type":"ContainerDied","Data":"9a274df71070056d0ba90d60479e4818f9ac34548bc19489c14a87da15dc6c8b"} Feb 02 09:09:52 crc kubenswrapper[4764]: I0202 09:09:52.506558 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ct55p" event={"ID":"00cb5ef3-3b54-47ac-bfc4-b1d53f04d292","Type":"ContainerStarted","Data":"00838ebc3146464fb7e67e2e6bc4614aea995f37837f9d843754c1c48fecd95b"} Feb 02 09:09:52 crc kubenswrapper[4764]: I0202 09:09:52.525061 4764 generic.go:334] "Generic (PLEG): container finished" podID="83d5cb3d-ee7b-44cb-8379-8b52f2c3cadc" containerID="6fa7e7d68b57782853195c602eee842d6e29de75934eb6d982d1ca4fa4dd059c" exitCode=0 Feb 02 09:09:52 crc kubenswrapper[4764]: I0202 09:09:52.525153 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w6hd5" event={"ID":"83d5cb3d-ee7b-44cb-8379-8b52f2c3cadc","Type":"ContainerDied","Data":"6fa7e7d68b57782853195c602eee842d6e29de75934eb6d982d1ca4fa4dd059c"} Feb 02 09:09:52 crc kubenswrapper[4764]: I0202 09:09:52.529633 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-9nkvv" event={"ID":"527edab8-f88f-4d82-b42e-b8d52fcaa7d4","Type":"ContainerStarted","Data":"0b65b00bc2d8eff15f7d32631379347926baf789e3e9c91110f3de30b1b31c5e"} Feb 02 09:09:52 crc kubenswrapper[4764]: I0202 09:09:52.667160 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Feb 02 09:09:52 crc kubenswrapper[4764]: I0202 09:09:52.668609 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 02 09:09:52 crc kubenswrapper[4764]: I0202 09:09:52.670915 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Feb 02 09:09:52 crc kubenswrapper[4764]: I0202 09:09:52.671249 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Feb 02 09:09:52 crc kubenswrapper[4764]: I0202 09:09:52.679466 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Feb 02 09:09:52 crc kubenswrapper[4764]: I0202 09:09:52.749848 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2eaf8a34-8fef-4fc4-be10-8a67f5c91ac5-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"2eaf8a34-8fef-4fc4-be10-8a67f5c91ac5\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 02 09:09:52 crc kubenswrapper[4764]: I0202 09:09:52.750013 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2eaf8a34-8fef-4fc4-be10-8a67f5c91ac5-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"2eaf8a34-8fef-4fc4-be10-8a67f5c91ac5\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 02 09:09:52 crc kubenswrapper[4764]: I0202 09:09:52.851797 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2eaf8a34-8fef-4fc4-be10-8a67f5c91ac5-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"2eaf8a34-8fef-4fc4-be10-8a67f5c91ac5\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 02 09:09:52 crc kubenswrapper[4764]: I0202 09:09:52.851855 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2eaf8a34-8fef-4fc4-be10-8a67f5c91ac5-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"2eaf8a34-8fef-4fc4-be10-8a67f5c91ac5\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 02 09:09:52 crc kubenswrapper[4764]: I0202 09:09:52.852291 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2eaf8a34-8fef-4fc4-be10-8a67f5c91ac5-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"2eaf8a34-8fef-4fc4-be10-8a67f5c91ac5\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 02 09:09:52 crc kubenswrapper[4764]: I0202 09:09:52.887742 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2eaf8a34-8fef-4fc4-be10-8a67f5c91ac5-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"2eaf8a34-8fef-4fc4-be10-8a67f5c91ac5\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 02 09:09:53 crc kubenswrapper[4764]: I0202 09:09:53.005451 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 02 09:09:53 crc kubenswrapper[4764]: I0202 09:09:53.259837 4764 patch_prober.go:28] interesting pod/router-default-5444994796-mkwq2 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 02 09:09:53 crc kubenswrapper[4764]: [-]has-synced failed: reason withheld Feb 02 09:09:53 crc kubenswrapper[4764]: [+]process-running ok Feb 02 09:09:53 crc kubenswrapper[4764]: healthz check failed Feb 02 09:09:53 crc kubenswrapper[4764]: I0202 09:09:53.260179 4764 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-mkwq2" podUID="83fed8a7-bd5f-463d-b70e-3a01700e8315" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 02 09:09:53 crc kubenswrapper[4764]: I0202 09:09:53.542084 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-9nkvv" event={"ID":"527edab8-f88f-4d82-b42e-b8d52fcaa7d4","Type":"ContainerStarted","Data":"37fa793b00f263b65d2c4b4c33beaeacd1c8c3cc4fac5d41b7d06caa9978028f"} Feb 02 09:09:53 crc kubenswrapper[4764]: I0202 09:09:53.543656 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-9nkvv" Feb 02 09:09:53 crc kubenswrapper[4764]: I0202 09:09:53.548918 4764 generic.go:334] "Generic (PLEG): container finished" podID="006ce277-7c6b-49a5-9e98-0bf7caee30f9" containerID="3a0f35f6d6c22683f1379ec4e011d9fc24fb03375e1e4b79536567e348c406c7" exitCode=0 Feb 02 09:09:53 crc kubenswrapper[4764]: I0202 09:09:53.548981 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29500380-2gwrp" event={"ID":"006ce277-7c6b-49a5-9e98-0bf7caee30f9","Type":"ContainerDied","Data":"3a0f35f6d6c22683f1379ec4e011d9fc24fb03375e1e4b79536567e348c406c7"} Feb 02 09:09:53 crc kubenswrapper[4764]: I0202 09:09:53.562344 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Feb 02 09:09:53 crc kubenswrapper[4764]: I0202 09:09:53.576127 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-9nkvv" podStartSLOduration=135.576102241 podStartE2EDuration="2m15.576102241s" podCreationTimestamp="2026-02-02 09:07:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:09:53.568139925 +0000 UTC m=+156.501864023" watchObservedRunningTime="2026-02-02 09:09:53.576102241 +0000 UTC m=+156.509826329" Feb 02 09:09:53 crc kubenswrapper[4764]: W0202 09:09:53.578296 4764 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod2eaf8a34_8fef_4fc4_be10_8a67f5c91ac5.slice/crio-52f9808b0c9af00cbbfb619ec15ac71a76b3205f6c53cc1cdc629016b24ba1c4 WatchSource:0}: Error finding container 52f9808b0c9af00cbbfb619ec15ac71a76b3205f6c53cc1cdc629016b24ba1c4: Status 404 returned error can't find the container with id 52f9808b0c9af00cbbfb619ec15ac71a76b3205f6c53cc1cdc629016b24ba1c4 Feb 02 09:09:54 crc kubenswrapper[4764]: I0202 09:09:54.251266 4764 patch_prober.go:28] interesting pod/router-default-5444994796-mkwq2 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 02 09:09:54 crc kubenswrapper[4764]: [-]has-synced failed: reason withheld Feb 02 09:09:54 crc kubenswrapper[4764]: [+]process-running ok Feb 02 09:09:54 crc kubenswrapper[4764]: healthz check failed Feb 02 09:09:54 crc kubenswrapper[4764]: I0202 09:09:54.251732 4764 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-mkwq2" podUID="83fed8a7-bd5f-463d-b70e-3a01700e8315" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 02 09:09:54 crc kubenswrapper[4764]: I0202 09:09:54.461748 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Feb 02 09:09:54 crc kubenswrapper[4764]: I0202 09:09:54.462416 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 02 09:09:54 crc kubenswrapper[4764]: I0202 09:09:54.471147 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Feb 02 09:09:54 crc kubenswrapper[4764]: I0202 09:09:54.471424 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Feb 02 09:09:54 crc kubenswrapper[4764]: I0202 09:09:54.473361 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Feb 02 09:09:54 crc kubenswrapper[4764]: I0202 09:09:54.502563 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/044e67dd-c049-428c-be8e-7afab5bbe10a-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"044e67dd-c049-428c-be8e-7afab5bbe10a\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 02 09:09:54 crc kubenswrapper[4764]: I0202 09:09:54.502697 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/044e67dd-c049-428c-be8e-7afab5bbe10a-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"044e67dd-c049-428c-be8e-7afab5bbe10a\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 02 09:09:54 crc kubenswrapper[4764]: I0202 09:09:54.577347 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"2eaf8a34-8fef-4fc4-be10-8a67f5c91ac5","Type":"ContainerStarted","Data":"dd5a61449a7674216f394330ff560fd6331ed1a9983571917b1207ca82ee5a49"} Feb 02 09:09:54 crc kubenswrapper[4764]: I0202 09:09:54.577388 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"2eaf8a34-8fef-4fc4-be10-8a67f5c91ac5","Type":"ContainerStarted","Data":"52f9808b0c9af00cbbfb619ec15ac71a76b3205f6c53cc1cdc629016b24ba1c4"} Feb 02 09:09:54 crc kubenswrapper[4764]: I0202 09:09:54.602386 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=2.6023713920000002 podStartE2EDuration="2.602371392s" podCreationTimestamp="2026-02-02 09:09:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:09:54.599317255 +0000 UTC m=+157.533041343" watchObservedRunningTime="2026-02-02 09:09:54.602371392 +0000 UTC m=+157.536095480" Feb 02 09:09:54 crc kubenswrapper[4764]: I0202 09:09:54.604050 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/044e67dd-c049-428c-be8e-7afab5bbe10a-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"044e67dd-c049-428c-be8e-7afab5bbe10a\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 02 09:09:54 crc kubenswrapper[4764]: I0202 09:09:54.604213 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/044e67dd-c049-428c-be8e-7afab5bbe10a-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"044e67dd-c049-428c-be8e-7afab5bbe10a\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 02 09:09:54 crc kubenswrapper[4764]: I0202 09:09:54.604793 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/044e67dd-c049-428c-be8e-7afab5bbe10a-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"044e67dd-c049-428c-be8e-7afab5bbe10a\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 02 09:09:54 crc kubenswrapper[4764]: I0202 09:09:54.625216 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/044e67dd-c049-428c-be8e-7afab5bbe10a-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"044e67dd-c049-428c-be8e-7afab5bbe10a\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 02 09:09:54 crc kubenswrapper[4764]: I0202 09:09:54.818801 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 02 09:09:55 crc kubenswrapper[4764]: I0202 09:09:55.062860 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29500380-2gwrp" Feb 02 09:09:55 crc kubenswrapper[4764]: I0202 09:09:55.114620 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4btgj\" (UniqueName: \"kubernetes.io/projected/006ce277-7c6b-49a5-9e98-0bf7caee30f9-kube-api-access-4btgj\") pod \"006ce277-7c6b-49a5-9e98-0bf7caee30f9\" (UID: \"006ce277-7c6b-49a5-9e98-0bf7caee30f9\") " Feb 02 09:09:55 crc kubenswrapper[4764]: I0202 09:09:55.115049 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/006ce277-7c6b-49a5-9e98-0bf7caee30f9-config-volume\") pod \"006ce277-7c6b-49a5-9e98-0bf7caee30f9\" (UID: \"006ce277-7c6b-49a5-9e98-0bf7caee30f9\") " Feb 02 09:09:55 crc kubenswrapper[4764]: I0202 09:09:55.115093 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/006ce277-7c6b-49a5-9e98-0bf7caee30f9-secret-volume\") pod \"006ce277-7c6b-49a5-9e98-0bf7caee30f9\" (UID: \"006ce277-7c6b-49a5-9e98-0bf7caee30f9\") " Feb 02 09:09:55 crc kubenswrapper[4764]: I0202 09:09:55.117427 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/006ce277-7c6b-49a5-9e98-0bf7caee30f9-config-volume" (OuterVolumeSpecName: "config-volume") pod "006ce277-7c6b-49a5-9e98-0bf7caee30f9" (UID: "006ce277-7c6b-49a5-9e98-0bf7caee30f9"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:09:55 crc kubenswrapper[4764]: I0202 09:09:55.142733 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/006ce277-7c6b-49a5-9e98-0bf7caee30f9-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "006ce277-7c6b-49a5-9e98-0bf7caee30f9" (UID: "006ce277-7c6b-49a5-9e98-0bf7caee30f9"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:09:55 crc kubenswrapper[4764]: I0202 09:09:55.143297 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/006ce277-7c6b-49a5-9e98-0bf7caee30f9-kube-api-access-4btgj" (OuterVolumeSpecName: "kube-api-access-4btgj") pod "006ce277-7c6b-49a5-9e98-0bf7caee30f9" (UID: "006ce277-7c6b-49a5-9e98-0bf7caee30f9"). InnerVolumeSpecName "kube-api-access-4btgj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:09:55 crc kubenswrapper[4764]: I0202 09:09:55.217455 4764 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/006ce277-7c6b-49a5-9e98-0bf7caee30f9-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 02 09:09:55 crc kubenswrapper[4764]: I0202 09:09:55.217483 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4btgj\" (UniqueName: \"kubernetes.io/projected/006ce277-7c6b-49a5-9e98-0bf7caee30f9-kube-api-access-4btgj\") on node \"crc\" DevicePath \"\"" Feb 02 09:09:55 crc kubenswrapper[4764]: I0202 09:09:55.217491 4764 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/006ce277-7c6b-49a5-9e98-0bf7caee30f9-config-volume\") on node \"crc\" DevicePath \"\"" Feb 02 09:09:55 crc kubenswrapper[4764]: I0202 09:09:55.254163 4764 patch_prober.go:28] interesting pod/router-default-5444994796-mkwq2 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 02 09:09:55 crc kubenswrapper[4764]: [-]has-synced failed: reason withheld Feb 02 09:09:55 crc kubenswrapper[4764]: [+]process-running ok Feb 02 09:09:55 crc kubenswrapper[4764]: healthz check failed Feb 02 09:09:55 crc kubenswrapper[4764]: I0202 09:09:55.254244 4764 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-mkwq2" podUID="83fed8a7-bd5f-463d-b70e-3a01700e8315" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 02 09:09:55 crc kubenswrapper[4764]: I0202 09:09:55.342060 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-h5vrh" Feb 02 09:09:55 crc kubenswrapper[4764]: I0202 09:09:55.353006 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-h5vrh" Feb 02 09:09:55 crc kubenswrapper[4764]: I0202 09:09:55.622200 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29500380-2gwrp" Feb 02 09:09:55 crc kubenswrapper[4764]: I0202 09:09:55.624046 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29500380-2gwrp" event={"ID":"006ce277-7c6b-49a5-9e98-0bf7caee30f9","Type":"ContainerDied","Data":"8137233e318d112629cf934371c3feb215fab8d658a8547450addbf07b690ac7"} Feb 02 09:09:55 crc kubenswrapper[4764]: I0202 09:09:55.624114 4764 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8137233e318d112629cf934371c3feb215fab8d658a8547450addbf07b690ac7" Feb 02 09:09:55 crc kubenswrapper[4764]: I0202 09:09:55.640345 4764 generic.go:334] "Generic (PLEG): container finished" podID="2eaf8a34-8fef-4fc4-be10-8a67f5c91ac5" containerID="dd5a61449a7674216f394330ff560fd6331ed1a9983571917b1207ca82ee5a49" exitCode=0 Feb 02 09:09:55 crc kubenswrapper[4764]: I0202 09:09:55.641231 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"2eaf8a34-8fef-4fc4-be10-8a67f5c91ac5","Type":"ContainerDied","Data":"dd5a61449a7674216f394330ff560fd6331ed1a9983571917b1207ca82ee5a49"} Feb 02 09:09:55 crc kubenswrapper[4764]: I0202 09:09:55.748264 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Feb 02 09:09:56 crc kubenswrapper[4764]: I0202 09:09:56.246526 4764 patch_prober.go:28] interesting pod/router-default-5444994796-mkwq2 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 02 09:09:56 crc kubenswrapper[4764]: [-]has-synced failed: reason withheld Feb 02 09:09:56 crc kubenswrapper[4764]: [+]process-running ok Feb 02 09:09:56 crc kubenswrapper[4764]: healthz check failed Feb 02 09:09:56 crc kubenswrapper[4764]: I0202 09:09:56.246584 4764 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-mkwq2" podUID="83fed8a7-bd5f-463d-b70e-3a01700e8315" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 02 09:09:56 crc kubenswrapper[4764]: I0202 09:09:56.402580 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-7v5xl" Feb 02 09:09:56 crc kubenswrapper[4764]: I0202 09:09:56.677399 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"044e67dd-c049-428c-be8e-7afab5bbe10a","Type":"ContainerStarted","Data":"02b465bd4fbf51ac4b5275884920e01dc799be4e963c759f7c9c99fba60be180"} Feb 02 09:09:57 crc kubenswrapper[4764]: I0202 09:09:57.246901 4764 patch_prober.go:28] interesting pod/router-default-5444994796-mkwq2 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 02 09:09:57 crc kubenswrapper[4764]: [-]has-synced failed: reason withheld Feb 02 09:09:57 crc kubenswrapper[4764]: [+]process-running ok Feb 02 09:09:57 crc kubenswrapper[4764]: healthz check failed Feb 02 09:09:57 crc kubenswrapper[4764]: I0202 09:09:57.246975 4764 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-mkwq2" podUID="83fed8a7-bd5f-463d-b70e-3a01700e8315" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 02 09:09:57 crc kubenswrapper[4764]: I0202 09:09:57.341256 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 02 09:09:57 crc kubenswrapper[4764]: I0202 09:09:57.387286 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2eaf8a34-8fef-4fc4-be10-8a67f5c91ac5-kube-api-access\") pod \"2eaf8a34-8fef-4fc4-be10-8a67f5c91ac5\" (UID: \"2eaf8a34-8fef-4fc4-be10-8a67f5c91ac5\") " Feb 02 09:09:57 crc kubenswrapper[4764]: I0202 09:09:57.387329 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2eaf8a34-8fef-4fc4-be10-8a67f5c91ac5-kubelet-dir\") pod \"2eaf8a34-8fef-4fc4-be10-8a67f5c91ac5\" (UID: \"2eaf8a34-8fef-4fc4-be10-8a67f5c91ac5\") " Feb 02 09:09:57 crc kubenswrapper[4764]: I0202 09:09:57.387567 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2eaf8a34-8fef-4fc4-be10-8a67f5c91ac5-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "2eaf8a34-8fef-4fc4-be10-8a67f5c91ac5" (UID: "2eaf8a34-8fef-4fc4-be10-8a67f5c91ac5"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 09:09:57 crc kubenswrapper[4764]: I0202 09:09:57.422231 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2eaf8a34-8fef-4fc4-be10-8a67f5c91ac5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "2eaf8a34-8fef-4fc4-be10-8a67f5c91ac5" (UID: "2eaf8a34-8fef-4fc4-be10-8a67f5c91ac5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:09:57 crc kubenswrapper[4764]: I0202 09:09:57.490156 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2eaf8a34-8fef-4fc4-be10-8a67f5c91ac5-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 02 09:09:57 crc kubenswrapper[4764]: I0202 09:09:57.490223 4764 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2eaf8a34-8fef-4fc4-be10-8a67f5c91ac5-kubelet-dir\") on node \"crc\" DevicePath \"\"" Feb 02 09:09:57 crc kubenswrapper[4764]: I0202 09:09:57.719743 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 02 09:09:57 crc kubenswrapper[4764]: I0202 09:09:57.719738 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"2eaf8a34-8fef-4fc4-be10-8a67f5c91ac5","Type":"ContainerDied","Data":"52f9808b0c9af00cbbfb619ec15ac71a76b3205f6c53cc1cdc629016b24ba1c4"} Feb 02 09:09:57 crc kubenswrapper[4764]: I0202 09:09:57.720350 4764 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="52f9808b0c9af00cbbfb619ec15ac71a76b3205f6c53cc1cdc629016b24ba1c4" Feb 02 09:09:57 crc kubenswrapper[4764]: I0202 09:09:57.736822 4764 generic.go:334] "Generic (PLEG): container finished" podID="044e67dd-c049-428c-be8e-7afab5bbe10a" containerID="cb1ae5c6a0fbb383ed4f645f0bc60e44865759644e4294407b7b6cb7f7e40f51" exitCode=0 Feb 02 09:09:57 crc kubenswrapper[4764]: I0202 09:09:57.736855 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"044e67dd-c049-428c-be8e-7afab5bbe10a","Type":"ContainerDied","Data":"cb1ae5c6a0fbb383ed4f645f0bc60e44865759644e4294407b7b6cb7f7e40f51"} Feb 02 09:09:58 crc kubenswrapper[4764]: I0202 09:09:58.247892 4764 patch_prober.go:28] interesting pod/router-default-5444994796-mkwq2 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 02 09:09:58 crc kubenswrapper[4764]: [-]has-synced failed: reason withheld Feb 02 09:09:58 crc kubenswrapper[4764]: [+]process-running ok Feb 02 09:09:58 crc kubenswrapper[4764]: healthz check failed Feb 02 09:09:58 crc kubenswrapper[4764]: I0202 09:09:58.247981 4764 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-mkwq2" podUID="83fed8a7-bd5f-463d-b70e-3a01700e8315" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 02 09:09:59 crc kubenswrapper[4764]: I0202 09:09:59.243104 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 02 09:09:59 crc kubenswrapper[4764]: I0202 09:09:59.247781 4764 patch_prober.go:28] interesting pod/router-default-5444994796-mkwq2 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 02 09:09:59 crc kubenswrapper[4764]: [-]has-synced failed: reason withheld Feb 02 09:09:59 crc kubenswrapper[4764]: [+]process-running ok Feb 02 09:09:59 crc kubenswrapper[4764]: healthz check failed Feb 02 09:09:59 crc kubenswrapper[4764]: I0202 09:09:59.247830 4764 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-mkwq2" podUID="83fed8a7-bd5f-463d-b70e-3a01700e8315" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 02 09:09:59 crc kubenswrapper[4764]: I0202 09:09:59.323768 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/044e67dd-c049-428c-be8e-7afab5bbe10a-kubelet-dir\") pod \"044e67dd-c049-428c-be8e-7afab5bbe10a\" (UID: \"044e67dd-c049-428c-be8e-7afab5bbe10a\") " Feb 02 09:09:59 crc kubenswrapper[4764]: I0202 09:09:59.323825 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/044e67dd-c049-428c-be8e-7afab5bbe10a-kube-api-access\") pod \"044e67dd-c049-428c-be8e-7afab5bbe10a\" (UID: \"044e67dd-c049-428c-be8e-7afab5bbe10a\") " Feb 02 09:09:59 crc kubenswrapper[4764]: I0202 09:09:59.323911 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/044e67dd-c049-428c-be8e-7afab5bbe10a-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "044e67dd-c049-428c-be8e-7afab5bbe10a" (UID: "044e67dd-c049-428c-be8e-7afab5bbe10a"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 09:09:59 crc kubenswrapper[4764]: I0202 09:09:59.324301 4764 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/044e67dd-c049-428c-be8e-7afab5bbe10a-kubelet-dir\") on node \"crc\" DevicePath \"\"" Feb 02 09:09:59 crc kubenswrapper[4764]: I0202 09:09:59.330549 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/044e67dd-c049-428c-be8e-7afab5bbe10a-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "044e67dd-c049-428c-be8e-7afab5bbe10a" (UID: "044e67dd-c049-428c-be8e-7afab5bbe10a"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:09:59 crc kubenswrapper[4764]: I0202 09:09:59.425438 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/044e67dd-c049-428c-be8e-7afab5bbe10a-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 02 09:09:59 crc kubenswrapper[4764]: I0202 09:09:59.770000 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"044e67dd-c049-428c-be8e-7afab5bbe10a","Type":"ContainerDied","Data":"02b465bd4fbf51ac4b5275884920e01dc799be4e963c759f7c9c99fba60be180"} Feb 02 09:09:59 crc kubenswrapper[4764]: I0202 09:09:59.770041 4764 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="02b465bd4fbf51ac4b5275884920e01dc799be4e963c759f7c9c99fba60be180" Feb 02 09:09:59 crc kubenswrapper[4764]: I0202 09:09:59.770081 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 02 09:10:00 crc kubenswrapper[4764]: I0202 09:10:00.249470 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-mkwq2" Feb 02 09:10:00 crc kubenswrapper[4764]: I0202 09:10:00.252557 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-mkwq2" Feb 02 09:10:00 crc kubenswrapper[4764]: I0202 09:10:00.383063 4764 patch_prober.go:28] interesting pod/downloads-7954f5f757-7k5qb container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" start-of-body= Feb 02 09:10:00 crc kubenswrapper[4764]: I0202 09:10:00.383125 4764 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-7k5qb" podUID="c4b729bd-782f-4393-b162-16c02eb3cb71" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" Feb 02 09:10:00 crc kubenswrapper[4764]: I0202 09:10:00.383647 4764 patch_prober.go:28] interesting pod/downloads-7954f5f757-7k5qb container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" start-of-body= Feb 02 09:10:00 crc kubenswrapper[4764]: I0202 09:10:00.383922 4764 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-7k5qb" podUID="c4b729bd-782f-4393-b162-16c02eb3cb71" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" Feb 02 09:10:00 crc kubenswrapper[4764]: I0202 09:10:00.474803 4764 patch_prober.go:28] interesting pod/console-f9d7485db-cnkss container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.10:8443/health\": dial tcp 10.217.0.10:8443: connect: connection refused" start-of-body= Feb 02 09:10:00 crc kubenswrapper[4764]: I0202 09:10:00.474862 4764 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-cnkss" podUID="d874a84d-48e3-4cb2-ac74-4f0db52e9d75" containerName="console" probeResult="failure" output="Get \"https://10.217.0.10:8443/health\": dial tcp 10.217.0.10:8443: connect: connection refused" Feb 02 09:10:01 crc kubenswrapper[4764]: I0202 09:10:01.984832 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/991faa9a-dd25-4f49-82bd-ce60cefd4af2-metrics-certs\") pod \"network-metrics-daemon-kwtmr\" (UID: \"991faa9a-dd25-4f49-82bd-ce60cefd4af2\") " pod="openshift-multus/network-metrics-daemon-kwtmr" Feb 02 09:10:02 crc kubenswrapper[4764]: I0202 09:10:02.006958 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/991faa9a-dd25-4f49-82bd-ce60cefd4af2-metrics-certs\") pod \"network-metrics-daemon-kwtmr\" (UID: \"991faa9a-dd25-4f49-82bd-ce60cefd4af2\") " pod="openshift-multus/network-metrics-daemon-kwtmr" Feb 02 09:10:02 crc kubenswrapper[4764]: I0202 09:10:02.052128 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kwtmr" Feb 02 09:10:03 crc kubenswrapper[4764]: I0202 09:10:03.345415 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-966nh"] Feb 02 09:10:03 crc kubenswrapper[4764]: I0202 09:10:03.345858 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-966nh" podUID="70342a10-423d-4657-9fc0-062ab9ab8902" containerName="controller-manager" containerID="cri-o://a3b165019032cbd0b9e2e96ef5294f81bb83582f5c87b5a7ae719e4f9fa104aa" gracePeriod=30 Feb 02 09:10:03 crc kubenswrapper[4764]: I0202 09:10:03.395417 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-4298w"] Feb 02 09:10:03 crc kubenswrapper[4764]: I0202 09:10:03.395658 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4298w" podUID="37d3c5dc-0535-4f2e-aa2f-debd99e0f1e3" containerName="route-controller-manager" containerID="cri-o://8bb51eb136af1e94516a00d26e9fa4e83fd5eee190f8f5c28e86014b4a44bc77" gracePeriod=30 Feb 02 09:10:03 crc kubenswrapper[4764]: I0202 09:10:03.888743 4764 generic.go:334] "Generic (PLEG): container finished" podID="70342a10-423d-4657-9fc0-062ab9ab8902" containerID="a3b165019032cbd0b9e2e96ef5294f81bb83582f5c87b5a7ae719e4f9fa104aa" exitCode=0 Feb 02 09:10:03 crc kubenswrapper[4764]: I0202 09:10:03.888873 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-966nh" event={"ID":"70342a10-423d-4657-9fc0-062ab9ab8902","Type":"ContainerDied","Data":"a3b165019032cbd0b9e2e96ef5294f81bb83582f5c87b5a7ae719e4f9fa104aa"} Feb 02 09:10:03 crc kubenswrapper[4764]: I0202 09:10:03.896182 4764 generic.go:334] "Generic (PLEG): container finished" podID="37d3c5dc-0535-4f2e-aa2f-debd99e0f1e3" containerID="8bb51eb136af1e94516a00d26e9fa4e83fd5eee190f8f5c28e86014b4a44bc77" exitCode=0 Feb 02 09:10:03 crc kubenswrapper[4764]: I0202 09:10:03.896290 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4298w" event={"ID":"37d3c5dc-0535-4f2e-aa2f-debd99e0f1e3","Type":"ContainerDied","Data":"8bb51eb136af1e94516a00d26e9fa4e83fd5eee190f8f5c28e86014b4a44bc77"} Feb 02 09:10:09 crc kubenswrapper[4764]: I0202 09:10:09.945634 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-kwtmr"] Feb 02 09:10:09 crc kubenswrapper[4764]: W0202 09:10:09.953875 4764 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod991faa9a_dd25_4f49_82bd_ce60cefd4af2.slice/crio-a07c420123d1d8bea7ef5bc4262766681ab7e9a1033006afbf542f167e91cadb WatchSource:0}: Error finding container a07c420123d1d8bea7ef5bc4262766681ab7e9a1033006afbf542f167e91cadb: Status 404 returned error can't find the container with id a07c420123d1d8bea7ef5bc4262766681ab7e9a1033006afbf542f167e91cadb Feb 02 09:10:10 crc kubenswrapper[4764]: I0202 09:10:10.000163 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4298w" Feb 02 09:10:10 crc kubenswrapper[4764]: I0202 09:10:10.065738 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-8576df8657-g98wg"] Feb 02 09:10:10 crc kubenswrapper[4764]: E0202 09:10:10.066026 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37d3c5dc-0535-4f2e-aa2f-debd99e0f1e3" containerName="route-controller-manager" Feb 02 09:10:10 crc kubenswrapper[4764]: I0202 09:10:10.066039 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="37d3c5dc-0535-4f2e-aa2f-debd99e0f1e3" containerName="route-controller-manager" Feb 02 09:10:10 crc kubenswrapper[4764]: E0202 09:10:10.066053 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="044e67dd-c049-428c-be8e-7afab5bbe10a" containerName="pruner" Feb 02 09:10:10 crc kubenswrapper[4764]: I0202 09:10:10.066071 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="044e67dd-c049-428c-be8e-7afab5bbe10a" containerName="pruner" Feb 02 09:10:10 crc kubenswrapper[4764]: E0202 09:10:10.066080 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="006ce277-7c6b-49a5-9e98-0bf7caee30f9" containerName="collect-profiles" Feb 02 09:10:10 crc kubenswrapper[4764]: I0202 09:10:10.066087 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="006ce277-7c6b-49a5-9e98-0bf7caee30f9" containerName="collect-profiles" Feb 02 09:10:10 crc kubenswrapper[4764]: E0202 09:10:10.066102 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2eaf8a34-8fef-4fc4-be10-8a67f5c91ac5" containerName="pruner" Feb 02 09:10:10 crc kubenswrapper[4764]: I0202 09:10:10.066109 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="2eaf8a34-8fef-4fc4-be10-8a67f5c91ac5" containerName="pruner" Feb 02 09:10:10 crc kubenswrapper[4764]: I0202 09:10:10.066196 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="006ce277-7c6b-49a5-9e98-0bf7caee30f9" containerName="collect-profiles" Feb 02 09:10:10 crc kubenswrapper[4764]: I0202 09:10:10.066209 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="2eaf8a34-8fef-4fc4-be10-8a67f5c91ac5" containerName="pruner" Feb 02 09:10:10 crc kubenswrapper[4764]: I0202 09:10:10.066217 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="044e67dd-c049-428c-be8e-7afab5bbe10a" containerName="pruner" Feb 02 09:10:10 crc kubenswrapper[4764]: I0202 09:10:10.066226 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="37d3c5dc-0535-4f2e-aa2f-debd99e0f1e3" containerName="route-controller-manager" Feb 02 09:10:10 crc kubenswrapper[4764]: I0202 09:10:10.066641 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-8576df8657-g98wg" Feb 02 09:10:10 crc kubenswrapper[4764]: I0202 09:10:10.074066 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-8576df8657-g98wg"] Feb 02 09:10:10 crc kubenswrapper[4764]: I0202 09:10:10.128424 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/37d3c5dc-0535-4f2e-aa2f-debd99e0f1e3-config\") pod \"37d3c5dc-0535-4f2e-aa2f-debd99e0f1e3\" (UID: \"37d3c5dc-0535-4f2e-aa2f-debd99e0f1e3\") " Feb 02 09:10:10 crc kubenswrapper[4764]: I0202 09:10:10.128513 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/37d3c5dc-0535-4f2e-aa2f-debd99e0f1e3-serving-cert\") pod \"37d3c5dc-0535-4f2e-aa2f-debd99e0f1e3\" (UID: \"37d3c5dc-0535-4f2e-aa2f-debd99e0f1e3\") " Feb 02 09:10:10 crc kubenswrapper[4764]: I0202 09:10:10.128586 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/37d3c5dc-0535-4f2e-aa2f-debd99e0f1e3-client-ca\") pod \"37d3c5dc-0535-4f2e-aa2f-debd99e0f1e3\" (UID: \"37d3c5dc-0535-4f2e-aa2f-debd99e0f1e3\") " Feb 02 09:10:10 crc kubenswrapper[4764]: I0202 09:10:10.128685 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9q5cz\" (UniqueName: \"kubernetes.io/projected/37d3c5dc-0535-4f2e-aa2f-debd99e0f1e3-kube-api-access-9q5cz\") pod \"37d3c5dc-0535-4f2e-aa2f-debd99e0f1e3\" (UID: \"37d3c5dc-0535-4f2e-aa2f-debd99e0f1e3\") " Feb 02 09:10:10 crc kubenswrapper[4764]: I0202 09:10:10.128869 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-582mg\" (UniqueName: \"kubernetes.io/projected/bb727d55-ab58-4a5d-97be-2d0798619f69-kube-api-access-582mg\") pod \"route-controller-manager-8576df8657-g98wg\" (UID: \"bb727d55-ab58-4a5d-97be-2d0798619f69\") " pod="openshift-route-controller-manager/route-controller-manager-8576df8657-g98wg" Feb 02 09:10:10 crc kubenswrapper[4764]: I0202 09:10:10.129222 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb727d55-ab58-4a5d-97be-2d0798619f69-config\") pod \"route-controller-manager-8576df8657-g98wg\" (UID: \"bb727d55-ab58-4a5d-97be-2d0798619f69\") " pod="openshift-route-controller-manager/route-controller-manager-8576df8657-g98wg" Feb 02 09:10:10 crc kubenswrapper[4764]: I0202 09:10:10.129325 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bb727d55-ab58-4a5d-97be-2d0798619f69-serving-cert\") pod \"route-controller-manager-8576df8657-g98wg\" (UID: \"bb727d55-ab58-4a5d-97be-2d0798619f69\") " pod="openshift-route-controller-manager/route-controller-manager-8576df8657-g98wg" Feb 02 09:10:10 crc kubenswrapper[4764]: I0202 09:10:10.129403 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/bb727d55-ab58-4a5d-97be-2d0798619f69-client-ca\") pod \"route-controller-manager-8576df8657-g98wg\" (UID: \"bb727d55-ab58-4a5d-97be-2d0798619f69\") " pod="openshift-route-controller-manager/route-controller-manager-8576df8657-g98wg" Feb 02 09:10:10 crc kubenswrapper[4764]: I0202 09:10:10.129431 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/37d3c5dc-0535-4f2e-aa2f-debd99e0f1e3-config" (OuterVolumeSpecName: "config") pod "37d3c5dc-0535-4f2e-aa2f-debd99e0f1e3" (UID: "37d3c5dc-0535-4f2e-aa2f-debd99e0f1e3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:10:10 crc kubenswrapper[4764]: I0202 09:10:10.129501 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/37d3c5dc-0535-4f2e-aa2f-debd99e0f1e3-client-ca" (OuterVolumeSpecName: "client-ca") pod "37d3c5dc-0535-4f2e-aa2f-debd99e0f1e3" (UID: "37d3c5dc-0535-4f2e-aa2f-debd99e0f1e3"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:10:10 crc kubenswrapper[4764]: I0202 09:10:10.134314 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37d3c5dc-0535-4f2e-aa2f-debd99e0f1e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "37d3c5dc-0535-4f2e-aa2f-debd99e0f1e3" (UID: "37d3c5dc-0535-4f2e-aa2f-debd99e0f1e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:10:10 crc kubenswrapper[4764]: I0202 09:10:10.134413 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/37d3c5dc-0535-4f2e-aa2f-debd99e0f1e3-kube-api-access-9q5cz" (OuterVolumeSpecName: "kube-api-access-9q5cz") pod "37d3c5dc-0535-4f2e-aa2f-debd99e0f1e3" (UID: "37d3c5dc-0535-4f2e-aa2f-debd99e0f1e3"). InnerVolumeSpecName "kube-api-access-9q5cz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:10:10 crc kubenswrapper[4764]: I0202 09:10:10.144044 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-966nh" Feb 02 09:10:10 crc kubenswrapper[4764]: I0202 09:10:10.230442 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/70342a10-423d-4657-9fc0-062ab9ab8902-config\") pod \"70342a10-423d-4657-9fc0-062ab9ab8902\" (UID: \"70342a10-423d-4657-9fc0-062ab9ab8902\") " Feb 02 09:10:10 crc kubenswrapper[4764]: I0202 09:10:10.230498 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/70342a10-423d-4657-9fc0-062ab9ab8902-serving-cert\") pod \"70342a10-423d-4657-9fc0-062ab9ab8902\" (UID: \"70342a10-423d-4657-9fc0-062ab9ab8902\") " Feb 02 09:10:10 crc kubenswrapper[4764]: I0202 09:10:10.230574 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5jxh9\" (UniqueName: \"kubernetes.io/projected/70342a10-423d-4657-9fc0-062ab9ab8902-kube-api-access-5jxh9\") pod \"70342a10-423d-4657-9fc0-062ab9ab8902\" (UID: \"70342a10-423d-4657-9fc0-062ab9ab8902\") " Feb 02 09:10:10 crc kubenswrapper[4764]: I0202 09:10:10.230595 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/70342a10-423d-4657-9fc0-062ab9ab8902-proxy-ca-bundles\") pod \"70342a10-423d-4657-9fc0-062ab9ab8902\" (UID: \"70342a10-423d-4657-9fc0-062ab9ab8902\") " Feb 02 09:10:10 crc kubenswrapper[4764]: I0202 09:10:10.230613 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/70342a10-423d-4657-9fc0-062ab9ab8902-client-ca\") pod \"70342a10-423d-4657-9fc0-062ab9ab8902\" (UID: \"70342a10-423d-4657-9fc0-062ab9ab8902\") " Feb 02 09:10:10 crc kubenswrapper[4764]: I0202 09:10:10.230747 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bb727d55-ab58-4a5d-97be-2d0798619f69-serving-cert\") pod \"route-controller-manager-8576df8657-g98wg\" (UID: \"bb727d55-ab58-4a5d-97be-2d0798619f69\") " pod="openshift-route-controller-manager/route-controller-manager-8576df8657-g98wg" Feb 02 09:10:10 crc kubenswrapper[4764]: I0202 09:10:10.230789 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/bb727d55-ab58-4a5d-97be-2d0798619f69-client-ca\") pod \"route-controller-manager-8576df8657-g98wg\" (UID: \"bb727d55-ab58-4a5d-97be-2d0798619f69\") " pod="openshift-route-controller-manager/route-controller-manager-8576df8657-g98wg" Feb 02 09:10:10 crc kubenswrapper[4764]: I0202 09:10:10.230808 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-582mg\" (UniqueName: \"kubernetes.io/projected/bb727d55-ab58-4a5d-97be-2d0798619f69-kube-api-access-582mg\") pod \"route-controller-manager-8576df8657-g98wg\" (UID: \"bb727d55-ab58-4a5d-97be-2d0798619f69\") " pod="openshift-route-controller-manager/route-controller-manager-8576df8657-g98wg" Feb 02 09:10:10 crc kubenswrapper[4764]: I0202 09:10:10.230848 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb727d55-ab58-4a5d-97be-2d0798619f69-config\") pod \"route-controller-manager-8576df8657-g98wg\" (UID: \"bb727d55-ab58-4a5d-97be-2d0798619f69\") " pod="openshift-route-controller-manager/route-controller-manager-8576df8657-g98wg" Feb 02 09:10:10 crc kubenswrapper[4764]: I0202 09:10:10.230894 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9q5cz\" (UniqueName: \"kubernetes.io/projected/37d3c5dc-0535-4f2e-aa2f-debd99e0f1e3-kube-api-access-9q5cz\") on node \"crc\" DevicePath \"\"" Feb 02 09:10:10 crc kubenswrapper[4764]: I0202 09:10:10.230905 4764 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/37d3c5dc-0535-4f2e-aa2f-debd99e0f1e3-config\") on node \"crc\" DevicePath \"\"" Feb 02 09:10:10 crc kubenswrapper[4764]: I0202 09:10:10.230914 4764 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/37d3c5dc-0535-4f2e-aa2f-debd99e0f1e3-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 02 09:10:10 crc kubenswrapper[4764]: I0202 09:10:10.230923 4764 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/37d3c5dc-0535-4f2e-aa2f-debd99e0f1e3-client-ca\") on node \"crc\" DevicePath \"\"" Feb 02 09:10:10 crc kubenswrapper[4764]: I0202 09:10:10.231407 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/70342a10-423d-4657-9fc0-062ab9ab8902-client-ca" (OuterVolumeSpecName: "client-ca") pod "70342a10-423d-4657-9fc0-062ab9ab8902" (UID: "70342a10-423d-4657-9fc0-062ab9ab8902"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:10:10 crc kubenswrapper[4764]: I0202 09:10:10.231941 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/70342a10-423d-4657-9fc0-062ab9ab8902-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "70342a10-423d-4657-9fc0-062ab9ab8902" (UID: "70342a10-423d-4657-9fc0-062ab9ab8902"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:10:10 crc kubenswrapper[4764]: I0202 09:10:10.232063 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/bb727d55-ab58-4a5d-97be-2d0798619f69-client-ca\") pod \"route-controller-manager-8576df8657-g98wg\" (UID: \"bb727d55-ab58-4a5d-97be-2d0798619f69\") " pod="openshift-route-controller-manager/route-controller-manager-8576df8657-g98wg" Feb 02 09:10:10 crc kubenswrapper[4764]: I0202 09:10:10.232496 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb727d55-ab58-4a5d-97be-2d0798619f69-config\") pod \"route-controller-manager-8576df8657-g98wg\" (UID: \"bb727d55-ab58-4a5d-97be-2d0798619f69\") " pod="openshift-route-controller-manager/route-controller-manager-8576df8657-g98wg" Feb 02 09:10:10 crc kubenswrapper[4764]: I0202 09:10:10.232916 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/70342a10-423d-4657-9fc0-062ab9ab8902-config" (OuterVolumeSpecName: "config") pod "70342a10-423d-4657-9fc0-062ab9ab8902" (UID: "70342a10-423d-4657-9fc0-062ab9ab8902"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:10:10 crc kubenswrapper[4764]: I0202 09:10:10.234251 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/70342a10-423d-4657-9fc0-062ab9ab8902-kube-api-access-5jxh9" (OuterVolumeSpecName: "kube-api-access-5jxh9") pod "70342a10-423d-4657-9fc0-062ab9ab8902" (UID: "70342a10-423d-4657-9fc0-062ab9ab8902"). InnerVolumeSpecName "kube-api-access-5jxh9". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:10:10 crc kubenswrapper[4764]: I0202 09:10:10.234599 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70342a10-423d-4657-9fc0-062ab9ab8902-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "70342a10-423d-4657-9fc0-062ab9ab8902" (UID: "70342a10-423d-4657-9fc0-062ab9ab8902"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:10:10 crc kubenswrapper[4764]: I0202 09:10:10.236955 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bb727d55-ab58-4a5d-97be-2d0798619f69-serving-cert\") pod \"route-controller-manager-8576df8657-g98wg\" (UID: \"bb727d55-ab58-4a5d-97be-2d0798619f69\") " pod="openshift-route-controller-manager/route-controller-manager-8576df8657-g98wg" Feb 02 09:10:10 crc kubenswrapper[4764]: I0202 09:10:10.246149 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-582mg\" (UniqueName: \"kubernetes.io/projected/bb727d55-ab58-4a5d-97be-2d0798619f69-kube-api-access-582mg\") pod \"route-controller-manager-8576df8657-g98wg\" (UID: \"bb727d55-ab58-4a5d-97be-2d0798619f69\") " pod="openshift-route-controller-manager/route-controller-manager-8576df8657-g98wg" Feb 02 09:10:10 crc kubenswrapper[4764]: I0202 09:10:10.332515 4764 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/70342a10-423d-4657-9fc0-062ab9ab8902-config\") on node \"crc\" DevicePath \"\"" Feb 02 09:10:10 crc kubenswrapper[4764]: I0202 09:10:10.332543 4764 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/70342a10-423d-4657-9fc0-062ab9ab8902-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 02 09:10:10 crc kubenswrapper[4764]: I0202 09:10:10.332553 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5jxh9\" (UniqueName: \"kubernetes.io/projected/70342a10-423d-4657-9fc0-062ab9ab8902-kube-api-access-5jxh9\") on node \"crc\" DevicePath \"\"" Feb 02 09:10:10 crc kubenswrapper[4764]: I0202 09:10:10.332562 4764 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/70342a10-423d-4657-9fc0-062ab9ab8902-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 02 09:10:10 crc kubenswrapper[4764]: I0202 09:10:10.332572 4764 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/70342a10-423d-4657-9fc0-062ab9ab8902-client-ca\") on node \"crc\" DevicePath \"\"" Feb 02 09:10:10 crc kubenswrapper[4764]: I0202 09:10:10.387251 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-7k5qb" Feb 02 09:10:10 crc kubenswrapper[4764]: I0202 09:10:10.389305 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-8576df8657-g98wg" Feb 02 09:10:10 crc kubenswrapper[4764]: I0202 09:10:10.487691 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-cnkss" Feb 02 09:10:10 crc kubenswrapper[4764]: I0202 09:10:10.492167 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-cnkss" Feb 02 09:10:10 crc kubenswrapper[4764]: I0202 09:10:10.749864 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-8576df8657-g98wg"] Feb 02 09:10:10 crc kubenswrapper[4764]: I0202 09:10:10.955460 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-966nh" event={"ID":"70342a10-423d-4657-9fc0-062ab9ab8902","Type":"ContainerDied","Data":"3a54f4042d11a0f075d8fde8fdfd10430f1e69156df2ed0a877e36c97f22b6ea"} Feb 02 09:10:10 crc kubenswrapper[4764]: I0202 09:10:10.955746 4764 scope.go:117] "RemoveContainer" containerID="a3b165019032cbd0b9e2e96ef5294f81bb83582f5c87b5a7ae719e4f9fa104aa" Feb 02 09:10:10 crc kubenswrapper[4764]: I0202 09:10:10.955786 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-966nh" Feb 02 09:10:10 crc kubenswrapper[4764]: I0202 09:10:10.961043 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4298w" Feb 02 09:10:10 crc kubenswrapper[4764]: I0202 09:10:10.965355 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4298w" event={"ID":"37d3c5dc-0535-4f2e-aa2f-debd99e0f1e3","Type":"ContainerDied","Data":"430f309137b47445cab0755c4c1a2ead26d44204be4c870282523546d0fde4a0"} Feb 02 09:10:10 crc kubenswrapper[4764]: I0202 09:10:10.972218 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-kwtmr" event={"ID":"991faa9a-dd25-4f49-82bd-ce60cefd4af2","Type":"ContainerStarted","Data":"f1655c1d605c71a15f6c8267dd8c50174fd05a7910496adfc3bed83be65e3e6c"} Feb 02 09:10:10 crc kubenswrapper[4764]: I0202 09:10:10.973158 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-kwtmr" event={"ID":"991faa9a-dd25-4f49-82bd-ce60cefd4af2","Type":"ContainerStarted","Data":"12715a34e59c6abc99d98a140bd1962d6513307b5836e4af73c18eb74c3d6968"} Feb 02 09:10:10 crc kubenswrapper[4764]: I0202 09:10:10.973178 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-kwtmr" event={"ID":"991faa9a-dd25-4f49-82bd-ce60cefd4af2","Type":"ContainerStarted","Data":"a07c420123d1d8bea7ef5bc4262766681ab7e9a1033006afbf542f167e91cadb"} Feb 02 09:10:10 crc kubenswrapper[4764]: I0202 09:10:10.991471 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-kwtmr" podStartSLOduration=152.99144687 podStartE2EDuration="2m32.99144687s" podCreationTimestamp="2026-02-02 09:07:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:10:10.989526195 +0000 UTC m=+173.923250273" watchObservedRunningTime="2026-02-02 09:10:10.99144687 +0000 UTC m=+173.925170958" Feb 02 09:10:11 crc kubenswrapper[4764]: I0202 09:10:11.010795 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-966nh"] Feb 02 09:10:11 crc kubenswrapper[4764]: I0202 09:10:11.018598 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-966nh"] Feb 02 09:10:11 crc kubenswrapper[4764]: I0202 09:10:11.023527 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-4298w"] Feb 02 09:10:11 crc kubenswrapper[4764]: I0202 09:10:11.026591 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-4298w"] Feb 02 09:10:11 crc kubenswrapper[4764]: I0202 09:10:11.724346 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-9nkvv" Feb 02 09:10:11 crc kubenswrapper[4764]: I0202 09:10:11.833393 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="37d3c5dc-0535-4f2e-aa2f-debd99e0f1e3" path="/var/lib/kubelet/pods/37d3c5dc-0535-4f2e-aa2f-debd99e0f1e3/volumes" Feb 02 09:10:11 crc kubenswrapper[4764]: I0202 09:10:11.834029 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="70342a10-423d-4657-9fc0-062ab9ab8902" path="/var/lib/kubelet/pods/70342a10-423d-4657-9fc0-062ab9ab8902/volumes" Feb 02 09:10:12 crc kubenswrapper[4764]: I0202 09:10:12.876525 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-795d7df575-bxzxp"] Feb 02 09:10:12 crc kubenswrapper[4764]: E0202 09:10:12.877081 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70342a10-423d-4657-9fc0-062ab9ab8902" containerName="controller-manager" Feb 02 09:10:12 crc kubenswrapper[4764]: I0202 09:10:12.877094 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="70342a10-423d-4657-9fc0-062ab9ab8902" containerName="controller-manager" Feb 02 09:10:12 crc kubenswrapper[4764]: I0202 09:10:12.877204 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="70342a10-423d-4657-9fc0-062ab9ab8902" containerName="controller-manager" Feb 02 09:10:12 crc kubenswrapper[4764]: I0202 09:10:12.877626 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-795d7df575-bxzxp" Feb 02 09:10:12 crc kubenswrapper[4764]: I0202 09:10:12.884822 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Feb 02 09:10:12 crc kubenswrapper[4764]: I0202 09:10:12.885795 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Feb 02 09:10:12 crc kubenswrapper[4764]: I0202 09:10:12.889140 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Feb 02 09:10:12 crc kubenswrapper[4764]: I0202 09:10:12.889336 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Feb 02 09:10:12 crc kubenswrapper[4764]: I0202 09:10:12.889461 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Feb 02 09:10:12 crc kubenswrapper[4764]: I0202 09:10:12.889562 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Feb 02 09:10:12 crc kubenswrapper[4764]: I0202 09:10:12.892620 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Feb 02 09:10:12 crc kubenswrapper[4764]: I0202 09:10:12.896138 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-795d7df575-bxzxp"] Feb 02 09:10:12 crc kubenswrapper[4764]: I0202 09:10:12.970379 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j9ljs\" (UniqueName: \"kubernetes.io/projected/2d251b45-e18e-4837-87ed-8841ecd4f321-kube-api-access-j9ljs\") pod \"controller-manager-795d7df575-bxzxp\" (UID: \"2d251b45-e18e-4837-87ed-8841ecd4f321\") " pod="openshift-controller-manager/controller-manager-795d7df575-bxzxp" Feb 02 09:10:12 crc kubenswrapper[4764]: I0202 09:10:12.970442 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2d251b45-e18e-4837-87ed-8841ecd4f321-proxy-ca-bundles\") pod \"controller-manager-795d7df575-bxzxp\" (UID: \"2d251b45-e18e-4837-87ed-8841ecd4f321\") " pod="openshift-controller-manager/controller-manager-795d7df575-bxzxp" Feb 02 09:10:12 crc kubenswrapper[4764]: I0202 09:10:12.970756 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2d251b45-e18e-4837-87ed-8841ecd4f321-serving-cert\") pod \"controller-manager-795d7df575-bxzxp\" (UID: \"2d251b45-e18e-4837-87ed-8841ecd4f321\") " pod="openshift-controller-manager/controller-manager-795d7df575-bxzxp" Feb 02 09:10:12 crc kubenswrapper[4764]: I0202 09:10:12.970909 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2d251b45-e18e-4837-87ed-8841ecd4f321-client-ca\") pod \"controller-manager-795d7df575-bxzxp\" (UID: \"2d251b45-e18e-4837-87ed-8841ecd4f321\") " pod="openshift-controller-manager/controller-manager-795d7df575-bxzxp" Feb 02 09:10:12 crc kubenswrapper[4764]: I0202 09:10:12.971002 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2d251b45-e18e-4837-87ed-8841ecd4f321-config\") pod \"controller-manager-795d7df575-bxzxp\" (UID: \"2d251b45-e18e-4837-87ed-8841ecd4f321\") " pod="openshift-controller-manager/controller-manager-795d7df575-bxzxp" Feb 02 09:10:13 crc kubenswrapper[4764]: I0202 09:10:13.071952 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2d251b45-e18e-4837-87ed-8841ecd4f321-config\") pod \"controller-manager-795d7df575-bxzxp\" (UID: \"2d251b45-e18e-4837-87ed-8841ecd4f321\") " pod="openshift-controller-manager/controller-manager-795d7df575-bxzxp" Feb 02 09:10:13 crc kubenswrapper[4764]: I0202 09:10:13.073270 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2d251b45-e18e-4837-87ed-8841ecd4f321-config\") pod \"controller-manager-795d7df575-bxzxp\" (UID: \"2d251b45-e18e-4837-87ed-8841ecd4f321\") " pod="openshift-controller-manager/controller-manager-795d7df575-bxzxp" Feb 02 09:10:13 crc kubenswrapper[4764]: I0202 09:10:13.073322 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j9ljs\" (UniqueName: \"kubernetes.io/projected/2d251b45-e18e-4837-87ed-8841ecd4f321-kube-api-access-j9ljs\") pod \"controller-manager-795d7df575-bxzxp\" (UID: \"2d251b45-e18e-4837-87ed-8841ecd4f321\") " pod="openshift-controller-manager/controller-manager-795d7df575-bxzxp" Feb 02 09:10:13 crc kubenswrapper[4764]: I0202 09:10:13.073365 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2d251b45-e18e-4837-87ed-8841ecd4f321-proxy-ca-bundles\") pod \"controller-manager-795d7df575-bxzxp\" (UID: \"2d251b45-e18e-4837-87ed-8841ecd4f321\") " pod="openshift-controller-manager/controller-manager-795d7df575-bxzxp" Feb 02 09:10:13 crc kubenswrapper[4764]: I0202 09:10:13.073419 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2d251b45-e18e-4837-87ed-8841ecd4f321-serving-cert\") pod \"controller-manager-795d7df575-bxzxp\" (UID: \"2d251b45-e18e-4837-87ed-8841ecd4f321\") " pod="openshift-controller-manager/controller-manager-795d7df575-bxzxp" Feb 02 09:10:13 crc kubenswrapper[4764]: I0202 09:10:13.073444 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2d251b45-e18e-4837-87ed-8841ecd4f321-client-ca\") pod \"controller-manager-795d7df575-bxzxp\" (UID: \"2d251b45-e18e-4837-87ed-8841ecd4f321\") " pod="openshift-controller-manager/controller-manager-795d7df575-bxzxp" Feb 02 09:10:13 crc kubenswrapper[4764]: I0202 09:10:13.074163 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2d251b45-e18e-4837-87ed-8841ecd4f321-client-ca\") pod \"controller-manager-795d7df575-bxzxp\" (UID: \"2d251b45-e18e-4837-87ed-8841ecd4f321\") " pod="openshift-controller-manager/controller-manager-795d7df575-bxzxp" Feb 02 09:10:13 crc kubenswrapper[4764]: I0202 09:10:13.074799 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2d251b45-e18e-4837-87ed-8841ecd4f321-proxy-ca-bundles\") pod \"controller-manager-795d7df575-bxzxp\" (UID: \"2d251b45-e18e-4837-87ed-8841ecd4f321\") " pod="openshift-controller-manager/controller-manager-795d7df575-bxzxp" Feb 02 09:10:13 crc kubenswrapper[4764]: I0202 09:10:13.091333 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2d251b45-e18e-4837-87ed-8841ecd4f321-serving-cert\") pod \"controller-manager-795d7df575-bxzxp\" (UID: \"2d251b45-e18e-4837-87ed-8841ecd4f321\") " pod="openshift-controller-manager/controller-manager-795d7df575-bxzxp" Feb 02 09:10:13 crc kubenswrapper[4764]: I0202 09:10:13.093984 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j9ljs\" (UniqueName: \"kubernetes.io/projected/2d251b45-e18e-4837-87ed-8841ecd4f321-kube-api-access-j9ljs\") pod \"controller-manager-795d7df575-bxzxp\" (UID: \"2d251b45-e18e-4837-87ed-8841ecd4f321\") " pod="openshift-controller-manager/controller-manager-795d7df575-bxzxp" Feb 02 09:10:13 crc kubenswrapper[4764]: I0202 09:10:13.253748 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-795d7df575-bxzxp" Feb 02 09:10:13 crc kubenswrapper[4764]: I0202 09:10:13.523330 4764 patch_prober.go:28] interesting pod/machine-config-daemon-4ndm4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 09:10:13 crc kubenswrapper[4764]: I0202 09:10:13.523392 4764 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 09:10:15 crc kubenswrapper[4764]: W0202 09:10:15.707175 4764 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbb727d55_ab58_4a5d_97be_2d0798619f69.slice/crio-d3137b3b6b027564a5e2c76c60757b1b3d196ccfe12dde9fd7a3fefeb3908da5 WatchSource:0}: Error finding container d3137b3b6b027564a5e2c76c60757b1b3d196ccfe12dde9fd7a3fefeb3908da5: Status 404 returned error can't find the container with id d3137b3b6b027564a5e2c76c60757b1b3d196ccfe12dde9fd7a3fefeb3908da5 Feb 02 09:10:15 crc kubenswrapper[4764]: I0202 09:10:15.996422 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-8576df8657-g98wg" event={"ID":"bb727d55-ab58-4a5d-97be-2d0798619f69","Type":"ContainerStarted","Data":"d3137b3b6b027564a5e2c76c60757b1b3d196ccfe12dde9fd7a3fefeb3908da5"} Feb 02 09:10:21 crc kubenswrapper[4764]: I0202 09:10:21.610316 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-p66d5" Feb 02 09:10:21 crc kubenswrapper[4764]: E0202 09:10:21.703903 4764 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Feb 02 09:10:21 crc kubenswrapper[4764]: E0202 09:10:21.704095 4764 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-nnpm4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-mg5dh_openshift-marketplace(35376bf8-2200-49a4-8eb3-f5d7069e73a0): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Feb 02 09:10:21 crc kubenswrapper[4764]: E0202 09:10:21.705287 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-mg5dh" podUID="35376bf8-2200-49a4-8eb3-f5d7069e73a0" Feb 02 09:10:21 crc kubenswrapper[4764]: E0202 09:10:21.728811 4764 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Feb 02 09:10:21 crc kubenswrapper[4764]: E0202 09:10:21.729215 4764 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-45j59,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-ggrpx_openshift-marketplace(a40abe98-1419-4f5d-9bed-63f1c89028b1): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Feb 02 09:10:21 crc kubenswrapper[4764]: E0202 09:10:21.731714 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-ggrpx" podUID="a40abe98-1419-4f5d-9bed-63f1c89028b1" Feb 02 09:10:21 crc kubenswrapper[4764]: E0202 09:10:21.761711 4764 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Feb 02 09:10:21 crc kubenswrapper[4764]: E0202 09:10:21.761899 4764 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-knbhq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-j25pg_openshift-marketplace(01ede9d5-6b94-4d10-bf6d-3aef751840ff): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Feb 02 09:10:21 crc kubenswrapper[4764]: E0202 09:10:21.763208 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-j25pg" podUID="01ede9d5-6b94-4d10-bf6d-3aef751840ff" Feb 02 09:10:23 crc kubenswrapper[4764]: I0202 09:10:23.306803 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-795d7df575-bxzxp"] Feb 02 09:10:23 crc kubenswrapper[4764]: I0202 09:10:23.407274 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-8576df8657-g98wg"] Feb 02 09:10:25 crc kubenswrapper[4764]: E0202 09:10:25.554437 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-mg5dh" podUID="35376bf8-2200-49a4-8eb3-f5d7069e73a0" Feb 02 09:10:25 crc kubenswrapper[4764]: E0202 09:10:25.554500 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-j25pg" podUID="01ede9d5-6b94-4d10-bf6d-3aef751840ff" Feb 02 09:10:25 crc kubenswrapper[4764]: E0202 09:10:25.554567 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-ggrpx" podUID="a40abe98-1419-4f5d-9bed-63f1c89028b1" Feb 02 09:10:25 crc kubenswrapper[4764]: I0202 09:10:25.563050 4764 scope.go:117] "RemoveContainer" containerID="8bb51eb136af1e94516a00d26e9fa4e83fd5eee190f8f5c28e86014b4a44bc77" Feb 02 09:10:25 crc kubenswrapper[4764]: E0202 09:10:25.718639 4764 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Feb 02 09:10:25 crc kubenswrapper[4764]: E0202 09:10:25.719018 4764 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-fpp4j,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-ct55p_openshift-marketplace(00cb5ef3-3b54-47ac-bfc4-b1d53f04d292): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Feb 02 09:10:25 crc kubenswrapper[4764]: E0202 09:10:25.720673 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-ct55p" podUID="00cb5ef3-3b54-47ac-bfc4-b1d53f04d292" Feb 02 09:10:25 crc kubenswrapper[4764]: E0202 09:10:25.729146 4764 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Feb 02 09:10:25 crc kubenswrapper[4764]: E0202 09:10:25.729313 4764 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-bnkf7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-w6hd5_openshift-marketplace(83d5cb3d-ee7b-44cb-8379-8b52f2c3cadc): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Feb 02 09:10:25 crc kubenswrapper[4764]: E0202 09:10:25.731517 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-w6hd5" podUID="83d5cb3d-ee7b-44cb-8379-8b52f2c3cadc" Feb 02 09:10:25 crc kubenswrapper[4764]: I0202 09:10:25.865952 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-795d7df575-bxzxp"] Feb 02 09:10:26 crc kubenswrapper[4764]: I0202 09:10:26.049176 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7p7nt" event={"ID":"2e7d8f2e-1913-4b88-942a-995ad472697e","Type":"ContainerStarted","Data":"e9151643c0db78fb3854d26596227986f108e4c0a6f256d439efc94fdfaeab74"} Feb 02 09:10:26 crc kubenswrapper[4764]: I0202 09:10:26.051986 4764 generic.go:334] "Generic (PLEG): container finished" podID="6092680d-f31e-4b70-9d81-2d80661f6ac3" containerID="02b1aaa97c610ac96c94ef0e43e8d9bdb81abe2098a614069bd721975be3d873" exitCode=0 Feb 02 09:10:26 crc kubenswrapper[4764]: I0202 09:10:26.052055 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kbk9s" event={"ID":"6092680d-f31e-4b70-9d81-2d80661f6ac3","Type":"ContainerDied","Data":"02b1aaa97c610ac96c94ef0e43e8d9bdb81abe2098a614069bd721975be3d873"} Feb 02 09:10:26 crc kubenswrapper[4764]: I0202 09:10:26.061447 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-795d7df575-bxzxp" event={"ID":"2d251b45-e18e-4837-87ed-8841ecd4f321","Type":"ContainerStarted","Data":"31764a95985c2616d8969ec95e3f1c475164dc425ed3902e76d972dade278f35"} Feb 02 09:10:26 crc kubenswrapper[4764]: I0202 09:10:26.061493 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-795d7df575-bxzxp" event={"ID":"2d251b45-e18e-4837-87ed-8841ecd4f321","Type":"ContainerStarted","Data":"2b426941382afe0130ac013462147678b28bd44a19dd201a8753c94a63091d19"} Feb 02 09:10:26 crc kubenswrapper[4764]: I0202 09:10:26.061528 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-795d7df575-bxzxp" podUID="2d251b45-e18e-4837-87ed-8841ecd4f321" containerName="controller-manager" containerID="cri-o://31764a95985c2616d8969ec95e3f1c475164dc425ed3902e76d972dade278f35" gracePeriod=30 Feb 02 09:10:26 crc kubenswrapper[4764]: I0202 09:10:26.061606 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-795d7df575-bxzxp" Feb 02 09:10:26 crc kubenswrapper[4764]: I0202 09:10:26.062957 4764 patch_prober.go:28] interesting pod/controller-manager-795d7df575-bxzxp container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.55:8443/healthz\": dial tcp 10.217.0.55:8443: connect: connection refused" start-of-body= Feb 02 09:10:26 crc kubenswrapper[4764]: I0202 09:10:26.062986 4764 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-795d7df575-bxzxp" podUID="2d251b45-e18e-4837-87ed-8841ecd4f321" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.55:8443/healthz\": dial tcp 10.217.0.55:8443: connect: connection refused" Feb 02 09:10:26 crc kubenswrapper[4764]: I0202 09:10:26.066333 4764 generic.go:334] "Generic (PLEG): container finished" podID="88d63f88-8b5e-44de-b8e8-1417bbdb6e8e" containerID="52ba7ace2c80b4d708fb26af11b1a9aa76411436399c3aaf387655ec8c2e271e" exitCode=0 Feb 02 09:10:26 crc kubenswrapper[4764]: I0202 09:10:26.066577 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hcbn7" event={"ID":"88d63f88-8b5e-44de-b8e8-1417bbdb6e8e","Type":"ContainerDied","Data":"52ba7ace2c80b4d708fb26af11b1a9aa76411436399c3aaf387655ec8c2e271e"} Feb 02 09:10:26 crc kubenswrapper[4764]: I0202 09:10:26.076017 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-8576df8657-g98wg" event={"ID":"bb727d55-ab58-4a5d-97be-2d0798619f69","Type":"ContainerStarted","Data":"62809edbb372ceb2a82131950b20c13492e99b8bd80b58e9eeb8a39ac070648b"} Feb 02 09:10:26 crc kubenswrapper[4764]: I0202 09:10:26.076179 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-8576df8657-g98wg" podUID="bb727d55-ab58-4a5d-97be-2d0798619f69" containerName="route-controller-manager" containerID="cri-o://62809edbb372ceb2a82131950b20c13492e99b8bd80b58e9eeb8a39ac070648b" gracePeriod=30 Feb 02 09:10:26 crc kubenswrapper[4764]: I0202 09:10:26.076552 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-8576df8657-g98wg" Feb 02 09:10:26 crc kubenswrapper[4764]: E0202 09:10:26.077573 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-w6hd5" podUID="83d5cb3d-ee7b-44cb-8379-8b52f2c3cadc" Feb 02 09:10:26 crc kubenswrapper[4764]: E0202 09:10:26.077614 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-ct55p" podUID="00cb5ef3-3b54-47ac-bfc4-b1d53f04d292" Feb 02 09:10:26 crc kubenswrapper[4764]: I0202 09:10:26.096441 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-795d7df575-bxzxp" podStartSLOduration=23.096425558 podStartE2EDuration="23.096425558s" podCreationTimestamp="2026-02-02 09:10:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:10:26.09438109 +0000 UTC m=+189.028105178" watchObservedRunningTime="2026-02-02 09:10:26.096425558 +0000 UTC m=+189.030149646" Feb 02 09:10:26 crc kubenswrapper[4764]: I0202 09:10:26.264216 4764 patch_prober.go:28] interesting pod/route-controller-manager-8576df8657-g98wg container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.54:8443/healthz\": read tcp 10.217.0.2:45420->10.217.0.54:8443: read: connection reset by peer" start-of-body= Feb 02 09:10:26 crc kubenswrapper[4764]: I0202 09:10:26.269895 4764 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-8576df8657-g98wg" podUID="bb727d55-ab58-4a5d-97be-2d0798619f69" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.54:8443/healthz\": read tcp 10.217.0.2:45420->10.217.0.54:8443: read: connection reset by peer" Feb 02 09:10:26 crc kubenswrapper[4764]: I0202 09:10:26.485988 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-controller-manager_controller-manager-795d7df575-bxzxp_2d251b45-e18e-4837-87ed-8841ecd4f321/controller-manager/0.log" Feb 02 09:10:26 crc kubenswrapper[4764]: I0202 09:10:26.486265 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-795d7df575-bxzxp" Feb 02 09:10:26 crc kubenswrapper[4764]: I0202 09:10:26.505508 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-8576df8657-g98wg" podStartSLOduration=23.505485966 podStartE2EDuration="23.505485966s" podCreationTimestamp="2026-02-02 09:10:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:10:26.195452715 +0000 UTC m=+189.129176803" watchObservedRunningTime="2026-02-02 09:10:26.505485966 +0000 UTC m=+189.439210044" Feb 02 09:10:26 crc kubenswrapper[4764]: I0202 09:10:26.524117 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-5489d77c9c-qdmth"] Feb 02 09:10:26 crc kubenswrapper[4764]: E0202 09:10:26.524500 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d251b45-e18e-4837-87ed-8841ecd4f321" containerName="controller-manager" Feb 02 09:10:26 crc kubenswrapper[4764]: I0202 09:10:26.524515 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d251b45-e18e-4837-87ed-8841ecd4f321" containerName="controller-manager" Feb 02 09:10:26 crc kubenswrapper[4764]: I0202 09:10:26.524616 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d251b45-e18e-4837-87ed-8841ecd4f321" containerName="controller-manager" Feb 02 09:10:26 crc kubenswrapper[4764]: I0202 09:10:26.525060 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5489d77c9c-qdmth" Feb 02 09:10:26 crc kubenswrapper[4764]: I0202 09:10:26.530514 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5489d77c9c-qdmth"] Feb 02 09:10:26 crc kubenswrapper[4764]: I0202 09:10:26.628536 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-route-controller-manager_route-controller-manager-8576df8657-g98wg_bb727d55-ab58-4a5d-97be-2d0798619f69/route-controller-manager/0.log" Feb 02 09:10:26 crc kubenswrapper[4764]: I0202 09:10:26.628600 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-8576df8657-g98wg" Feb 02 09:10:26 crc kubenswrapper[4764]: I0202 09:10:26.656837 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j9ljs\" (UniqueName: \"kubernetes.io/projected/2d251b45-e18e-4837-87ed-8841ecd4f321-kube-api-access-j9ljs\") pod \"2d251b45-e18e-4837-87ed-8841ecd4f321\" (UID: \"2d251b45-e18e-4837-87ed-8841ecd4f321\") " Feb 02 09:10:26 crc kubenswrapper[4764]: I0202 09:10:26.656884 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2d251b45-e18e-4837-87ed-8841ecd4f321-proxy-ca-bundles\") pod \"2d251b45-e18e-4837-87ed-8841ecd4f321\" (UID: \"2d251b45-e18e-4837-87ed-8841ecd4f321\") " Feb 02 09:10:26 crc kubenswrapper[4764]: I0202 09:10:26.656909 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2d251b45-e18e-4837-87ed-8841ecd4f321-client-ca\") pod \"2d251b45-e18e-4837-87ed-8841ecd4f321\" (UID: \"2d251b45-e18e-4837-87ed-8841ecd4f321\") " Feb 02 09:10:26 crc kubenswrapper[4764]: I0202 09:10:26.656995 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2d251b45-e18e-4837-87ed-8841ecd4f321-serving-cert\") pod \"2d251b45-e18e-4837-87ed-8841ecd4f321\" (UID: \"2d251b45-e18e-4837-87ed-8841ecd4f321\") " Feb 02 09:10:26 crc kubenswrapper[4764]: I0202 09:10:26.657043 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2d251b45-e18e-4837-87ed-8841ecd4f321-config\") pod \"2d251b45-e18e-4837-87ed-8841ecd4f321\" (UID: \"2d251b45-e18e-4837-87ed-8841ecd4f321\") " Feb 02 09:10:26 crc kubenswrapper[4764]: I0202 09:10:26.657200 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/00732cf0-76cd-47f2-807e-363710f70a41-config\") pod \"controller-manager-5489d77c9c-qdmth\" (UID: \"00732cf0-76cd-47f2-807e-363710f70a41\") " pod="openshift-controller-manager/controller-manager-5489d77c9c-qdmth" Feb 02 09:10:26 crc kubenswrapper[4764]: I0202 09:10:26.657232 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hv6s6\" (UniqueName: \"kubernetes.io/projected/00732cf0-76cd-47f2-807e-363710f70a41-kube-api-access-hv6s6\") pod \"controller-manager-5489d77c9c-qdmth\" (UID: \"00732cf0-76cd-47f2-807e-363710f70a41\") " pod="openshift-controller-manager/controller-manager-5489d77c9c-qdmth" Feb 02 09:10:26 crc kubenswrapper[4764]: I0202 09:10:26.657274 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/00732cf0-76cd-47f2-807e-363710f70a41-client-ca\") pod \"controller-manager-5489d77c9c-qdmth\" (UID: \"00732cf0-76cd-47f2-807e-363710f70a41\") " pod="openshift-controller-manager/controller-manager-5489d77c9c-qdmth" Feb 02 09:10:26 crc kubenswrapper[4764]: I0202 09:10:26.657301 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/00732cf0-76cd-47f2-807e-363710f70a41-proxy-ca-bundles\") pod \"controller-manager-5489d77c9c-qdmth\" (UID: \"00732cf0-76cd-47f2-807e-363710f70a41\") " pod="openshift-controller-manager/controller-manager-5489d77c9c-qdmth" Feb 02 09:10:26 crc kubenswrapper[4764]: I0202 09:10:26.657324 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/00732cf0-76cd-47f2-807e-363710f70a41-serving-cert\") pod \"controller-manager-5489d77c9c-qdmth\" (UID: \"00732cf0-76cd-47f2-807e-363710f70a41\") " pod="openshift-controller-manager/controller-manager-5489d77c9c-qdmth" Feb 02 09:10:26 crc kubenswrapper[4764]: I0202 09:10:26.660861 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2d251b45-e18e-4837-87ed-8841ecd4f321-config" (OuterVolumeSpecName: "config") pod "2d251b45-e18e-4837-87ed-8841ecd4f321" (UID: "2d251b45-e18e-4837-87ed-8841ecd4f321"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:10:26 crc kubenswrapper[4764]: I0202 09:10:26.661431 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2d251b45-e18e-4837-87ed-8841ecd4f321-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "2d251b45-e18e-4837-87ed-8841ecd4f321" (UID: "2d251b45-e18e-4837-87ed-8841ecd4f321"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:10:26 crc kubenswrapper[4764]: I0202 09:10:26.661981 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2d251b45-e18e-4837-87ed-8841ecd4f321-client-ca" (OuterVolumeSpecName: "client-ca") pod "2d251b45-e18e-4837-87ed-8841ecd4f321" (UID: "2d251b45-e18e-4837-87ed-8841ecd4f321"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:10:26 crc kubenswrapper[4764]: I0202 09:10:26.663632 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2d251b45-e18e-4837-87ed-8841ecd4f321-kube-api-access-j9ljs" (OuterVolumeSpecName: "kube-api-access-j9ljs") pod "2d251b45-e18e-4837-87ed-8841ecd4f321" (UID: "2d251b45-e18e-4837-87ed-8841ecd4f321"). InnerVolumeSpecName "kube-api-access-j9ljs". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:10:26 crc kubenswrapper[4764]: I0202 09:10:26.665061 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d251b45-e18e-4837-87ed-8841ecd4f321-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "2d251b45-e18e-4837-87ed-8841ecd4f321" (UID: "2d251b45-e18e-4837-87ed-8841ecd4f321"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:10:26 crc kubenswrapper[4764]: I0202 09:10:26.758282 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb727d55-ab58-4a5d-97be-2d0798619f69-config\") pod \"bb727d55-ab58-4a5d-97be-2d0798619f69\" (UID: \"bb727d55-ab58-4a5d-97be-2d0798619f69\") " Feb 02 09:10:26 crc kubenswrapper[4764]: I0202 09:10:26.758337 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-582mg\" (UniqueName: \"kubernetes.io/projected/bb727d55-ab58-4a5d-97be-2d0798619f69-kube-api-access-582mg\") pod \"bb727d55-ab58-4a5d-97be-2d0798619f69\" (UID: \"bb727d55-ab58-4a5d-97be-2d0798619f69\") " Feb 02 09:10:26 crc kubenswrapper[4764]: I0202 09:10:26.758381 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bb727d55-ab58-4a5d-97be-2d0798619f69-serving-cert\") pod \"bb727d55-ab58-4a5d-97be-2d0798619f69\" (UID: \"bb727d55-ab58-4a5d-97be-2d0798619f69\") " Feb 02 09:10:26 crc kubenswrapper[4764]: I0202 09:10:26.758432 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/bb727d55-ab58-4a5d-97be-2d0798619f69-client-ca\") pod \"bb727d55-ab58-4a5d-97be-2d0798619f69\" (UID: \"bb727d55-ab58-4a5d-97be-2d0798619f69\") " Feb 02 09:10:26 crc kubenswrapper[4764]: I0202 09:10:26.758558 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/00732cf0-76cd-47f2-807e-363710f70a41-config\") pod \"controller-manager-5489d77c9c-qdmth\" (UID: \"00732cf0-76cd-47f2-807e-363710f70a41\") " pod="openshift-controller-manager/controller-manager-5489d77c9c-qdmth" Feb 02 09:10:26 crc kubenswrapper[4764]: I0202 09:10:26.758586 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hv6s6\" (UniqueName: \"kubernetes.io/projected/00732cf0-76cd-47f2-807e-363710f70a41-kube-api-access-hv6s6\") pod \"controller-manager-5489d77c9c-qdmth\" (UID: \"00732cf0-76cd-47f2-807e-363710f70a41\") " pod="openshift-controller-manager/controller-manager-5489d77c9c-qdmth" Feb 02 09:10:26 crc kubenswrapper[4764]: I0202 09:10:26.758629 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/00732cf0-76cd-47f2-807e-363710f70a41-client-ca\") pod \"controller-manager-5489d77c9c-qdmth\" (UID: \"00732cf0-76cd-47f2-807e-363710f70a41\") " pod="openshift-controller-manager/controller-manager-5489d77c9c-qdmth" Feb 02 09:10:26 crc kubenswrapper[4764]: I0202 09:10:26.758655 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/00732cf0-76cd-47f2-807e-363710f70a41-proxy-ca-bundles\") pod \"controller-manager-5489d77c9c-qdmth\" (UID: \"00732cf0-76cd-47f2-807e-363710f70a41\") " pod="openshift-controller-manager/controller-manager-5489d77c9c-qdmth" Feb 02 09:10:26 crc kubenswrapper[4764]: I0202 09:10:26.758675 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/00732cf0-76cd-47f2-807e-363710f70a41-serving-cert\") pod \"controller-manager-5489d77c9c-qdmth\" (UID: \"00732cf0-76cd-47f2-807e-363710f70a41\") " pod="openshift-controller-manager/controller-manager-5489d77c9c-qdmth" Feb 02 09:10:26 crc kubenswrapper[4764]: I0202 09:10:26.758724 4764 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2d251b45-e18e-4837-87ed-8841ecd4f321-config\") on node \"crc\" DevicePath \"\"" Feb 02 09:10:26 crc kubenswrapper[4764]: I0202 09:10:26.758734 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j9ljs\" (UniqueName: \"kubernetes.io/projected/2d251b45-e18e-4837-87ed-8841ecd4f321-kube-api-access-j9ljs\") on node \"crc\" DevicePath \"\"" Feb 02 09:10:26 crc kubenswrapper[4764]: I0202 09:10:26.758744 4764 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2d251b45-e18e-4837-87ed-8841ecd4f321-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 02 09:10:26 crc kubenswrapper[4764]: I0202 09:10:26.758752 4764 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2d251b45-e18e-4837-87ed-8841ecd4f321-client-ca\") on node \"crc\" DevicePath \"\"" Feb 02 09:10:26 crc kubenswrapper[4764]: I0202 09:10:26.758760 4764 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2d251b45-e18e-4837-87ed-8841ecd4f321-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 02 09:10:26 crc kubenswrapper[4764]: I0202 09:10:26.759777 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bb727d55-ab58-4a5d-97be-2d0798619f69-client-ca" (OuterVolumeSpecName: "client-ca") pod "bb727d55-ab58-4a5d-97be-2d0798619f69" (UID: "bb727d55-ab58-4a5d-97be-2d0798619f69"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:10:26 crc kubenswrapper[4764]: I0202 09:10:26.760160 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/00732cf0-76cd-47f2-807e-363710f70a41-client-ca\") pod \"controller-manager-5489d77c9c-qdmth\" (UID: \"00732cf0-76cd-47f2-807e-363710f70a41\") " pod="openshift-controller-manager/controller-manager-5489d77c9c-qdmth" Feb 02 09:10:26 crc kubenswrapper[4764]: I0202 09:10:26.760279 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bb727d55-ab58-4a5d-97be-2d0798619f69-config" (OuterVolumeSpecName: "config") pod "bb727d55-ab58-4a5d-97be-2d0798619f69" (UID: "bb727d55-ab58-4a5d-97be-2d0798619f69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:10:26 crc kubenswrapper[4764]: I0202 09:10:26.760560 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/00732cf0-76cd-47f2-807e-363710f70a41-config\") pod \"controller-manager-5489d77c9c-qdmth\" (UID: \"00732cf0-76cd-47f2-807e-363710f70a41\") " pod="openshift-controller-manager/controller-manager-5489d77c9c-qdmth" Feb 02 09:10:26 crc kubenswrapper[4764]: I0202 09:10:26.760714 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/00732cf0-76cd-47f2-807e-363710f70a41-proxy-ca-bundles\") pod \"controller-manager-5489d77c9c-qdmth\" (UID: \"00732cf0-76cd-47f2-807e-363710f70a41\") " pod="openshift-controller-manager/controller-manager-5489d77c9c-qdmth" Feb 02 09:10:26 crc kubenswrapper[4764]: I0202 09:10:26.763538 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb727d55-ab58-4a5d-97be-2d0798619f69-kube-api-access-582mg" (OuterVolumeSpecName: "kube-api-access-582mg") pod "bb727d55-ab58-4a5d-97be-2d0798619f69" (UID: "bb727d55-ab58-4a5d-97be-2d0798619f69"). InnerVolumeSpecName "kube-api-access-582mg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:10:26 crc kubenswrapper[4764]: I0202 09:10:26.773810 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/00732cf0-76cd-47f2-807e-363710f70a41-serving-cert\") pod \"controller-manager-5489d77c9c-qdmth\" (UID: \"00732cf0-76cd-47f2-807e-363710f70a41\") " pod="openshift-controller-manager/controller-manager-5489d77c9c-qdmth" Feb 02 09:10:26 crc kubenswrapper[4764]: I0202 09:10:26.777012 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb727d55-ab58-4a5d-97be-2d0798619f69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bb727d55-ab58-4a5d-97be-2d0798619f69" (UID: "bb727d55-ab58-4a5d-97be-2d0798619f69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:10:26 crc kubenswrapper[4764]: I0202 09:10:26.782833 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hv6s6\" (UniqueName: \"kubernetes.io/projected/00732cf0-76cd-47f2-807e-363710f70a41-kube-api-access-hv6s6\") pod \"controller-manager-5489d77c9c-qdmth\" (UID: \"00732cf0-76cd-47f2-807e-363710f70a41\") " pod="openshift-controller-manager/controller-manager-5489d77c9c-qdmth" Feb 02 09:10:26 crc kubenswrapper[4764]: I0202 09:10:26.845720 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5489d77c9c-qdmth" Feb 02 09:10:26 crc kubenswrapper[4764]: I0202 09:10:26.848766 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 09:10:26 crc kubenswrapper[4764]: I0202 09:10:26.859779 4764 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/bb727d55-ab58-4a5d-97be-2d0798619f69-client-ca\") on node \"crc\" DevicePath \"\"" Feb 02 09:10:26 crc kubenswrapper[4764]: I0202 09:10:26.859954 4764 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb727d55-ab58-4a5d-97be-2d0798619f69-config\") on node \"crc\" DevicePath \"\"" Feb 02 09:10:26 crc kubenswrapper[4764]: I0202 09:10:26.860034 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-582mg\" (UniqueName: \"kubernetes.io/projected/bb727d55-ab58-4a5d-97be-2d0798619f69-kube-api-access-582mg\") on node \"crc\" DevicePath \"\"" Feb 02 09:10:26 crc kubenswrapper[4764]: I0202 09:10:26.860109 4764 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bb727d55-ab58-4a5d-97be-2d0798619f69-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 02 09:10:27 crc kubenswrapper[4764]: I0202 09:10:27.085749 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-route-controller-manager_route-controller-manager-8576df8657-g98wg_bb727d55-ab58-4a5d-97be-2d0798619f69/route-controller-manager/0.log" Feb 02 09:10:27 crc kubenswrapper[4764]: I0202 09:10:27.086019 4764 generic.go:334] "Generic (PLEG): container finished" podID="bb727d55-ab58-4a5d-97be-2d0798619f69" containerID="62809edbb372ceb2a82131950b20c13492e99b8bd80b58e9eeb8a39ac070648b" exitCode=255 Feb 02 09:10:27 crc kubenswrapper[4764]: I0202 09:10:27.086138 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-8576df8657-g98wg" event={"ID":"bb727d55-ab58-4a5d-97be-2d0798619f69","Type":"ContainerDied","Data":"62809edbb372ceb2a82131950b20c13492e99b8bd80b58e9eeb8a39ac070648b"} Feb 02 09:10:27 crc kubenswrapper[4764]: I0202 09:10:27.086216 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-8576df8657-g98wg" event={"ID":"bb727d55-ab58-4a5d-97be-2d0798619f69","Type":"ContainerDied","Data":"d3137b3b6b027564a5e2c76c60757b1b3d196ccfe12dde9fd7a3fefeb3908da5"} Feb 02 09:10:27 crc kubenswrapper[4764]: I0202 09:10:27.086236 4764 scope.go:117] "RemoveContainer" containerID="62809edbb372ceb2a82131950b20c13492e99b8bd80b58e9eeb8a39ac070648b" Feb 02 09:10:27 crc kubenswrapper[4764]: I0202 09:10:27.086237 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-8576df8657-g98wg" Feb 02 09:10:27 crc kubenswrapper[4764]: I0202 09:10:27.091452 4764 generic.go:334] "Generic (PLEG): container finished" podID="2e7d8f2e-1913-4b88-942a-995ad472697e" containerID="e9151643c0db78fb3854d26596227986f108e4c0a6f256d439efc94fdfaeab74" exitCode=0 Feb 02 09:10:27 crc kubenswrapper[4764]: I0202 09:10:27.091513 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7p7nt" event={"ID":"2e7d8f2e-1913-4b88-942a-995ad472697e","Type":"ContainerDied","Data":"e9151643c0db78fb3854d26596227986f108e4c0a6f256d439efc94fdfaeab74"} Feb 02 09:10:27 crc kubenswrapper[4764]: I0202 09:10:27.097342 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kbk9s" event={"ID":"6092680d-f31e-4b70-9d81-2d80661f6ac3","Type":"ContainerStarted","Data":"cae9539080c975e16c67467c3d702f51ce9280b85f572e147d473e1583321863"} Feb 02 09:10:27 crc kubenswrapper[4764]: I0202 09:10:27.101968 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-controller-manager_controller-manager-795d7df575-bxzxp_2d251b45-e18e-4837-87ed-8841ecd4f321/controller-manager/0.log" Feb 02 09:10:27 crc kubenswrapper[4764]: I0202 09:10:27.102008 4764 generic.go:334] "Generic (PLEG): container finished" podID="2d251b45-e18e-4837-87ed-8841ecd4f321" containerID="31764a95985c2616d8969ec95e3f1c475164dc425ed3902e76d972dade278f35" exitCode=2 Feb 02 09:10:27 crc kubenswrapper[4764]: I0202 09:10:27.102090 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-795d7df575-bxzxp" Feb 02 09:10:27 crc kubenswrapper[4764]: I0202 09:10:27.102585 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-795d7df575-bxzxp" event={"ID":"2d251b45-e18e-4837-87ed-8841ecd4f321","Type":"ContainerDied","Data":"31764a95985c2616d8969ec95e3f1c475164dc425ed3902e76d972dade278f35"} Feb 02 09:10:27 crc kubenswrapper[4764]: I0202 09:10:27.102644 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-795d7df575-bxzxp" event={"ID":"2d251b45-e18e-4837-87ed-8841ecd4f321","Type":"ContainerDied","Data":"2b426941382afe0130ac013462147678b28bd44a19dd201a8753c94a63091d19"} Feb 02 09:10:27 crc kubenswrapper[4764]: I0202 09:10:27.105865 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hcbn7" event={"ID":"88d63f88-8b5e-44de-b8e8-1417bbdb6e8e","Type":"ContainerStarted","Data":"6704e1b75fbbb895cf52f07530ca2bd1a31d27e3da65d3db3726c2b9e43b09dc"} Feb 02 09:10:27 crc kubenswrapper[4764]: I0202 09:10:27.126136 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5489d77c9c-qdmth"] Feb 02 09:10:27 crc kubenswrapper[4764]: I0202 09:10:27.128097 4764 scope.go:117] "RemoveContainer" containerID="62809edbb372ceb2a82131950b20c13492e99b8bd80b58e9eeb8a39ac070648b" Feb 02 09:10:27 crc kubenswrapper[4764]: E0202 09:10:27.130033 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"62809edbb372ceb2a82131950b20c13492e99b8bd80b58e9eeb8a39ac070648b\": container with ID starting with 62809edbb372ceb2a82131950b20c13492e99b8bd80b58e9eeb8a39ac070648b not found: ID does not exist" containerID="62809edbb372ceb2a82131950b20c13492e99b8bd80b58e9eeb8a39ac070648b" Feb 02 09:10:27 crc kubenswrapper[4764]: I0202 09:10:27.130082 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"62809edbb372ceb2a82131950b20c13492e99b8bd80b58e9eeb8a39ac070648b"} err="failed to get container status \"62809edbb372ceb2a82131950b20c13492e99b8bd80b58e9eeb8a39ac070648b\": rpc error: code = NotFound desc = could not find container \"62809edbb372ceb2a82131950b20c13492e99b8bd80b58e9eeb8a39ac070648b\": container with ID starting with 62809edbb372ceb2a82131950b20c13492e99b8bd80b58e9eeb8a39ac070648b not found: ID does not exist" Feb 02 09:10:27 crc kubenswrapper[4764]: I0202 09:10:27.130123 4764 scope.go:117] "RemoveContainer" containerID="31764a95985c2616d8969ec95e3f1c475164dc425ed3902e76d972dade278f35" Feb 02 09:10:27 crc kubenswrapper[4764]: I0202 09:10:27.157301 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-kbk9s" podStartSLOduration=4.017451511 podStartE2EDuration="39.157278442s" podCreationTimestamp="2026-02-02 09:09:48 +0000 UTC" firstStartedPulling="2026-02-02 09:09:51.411998854 +0000 UTC m=+154.345722942" lastFinishedPulling="2026-02-02 09:10:26.551825795 +0000 UTC m=+189.485549873" observedRunningTime="2026-02-02 09:10:27.134069092 +0000 UTC m=+190.067793180" watchObservedRunningTime="2026-02-02 09:10:27.157278442 +0000 UTC m=+190.091002520" Feb 02 09:10:27 crc kubenswrapper[4764]: I0202 09:10:27.161759 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-8576df8657-g98wg"] Feb 02 09:10:27 crc kubenswrapper[4764]: I0202 09:10:27.163764 4764 scope.go:117] "RemoveContainer" containerID="31764a95985c2616d8969ec95e3f1c475164dc425ed3902e76d972dade278f35" Feb 02 09:10:27 crc kubenswrapper[4764]: E0202 09:10:27.164183 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"31764a95985c2616d8969ec95e3f1c475164dc425ed3902e76d972dade278f35\": container with ID starting with 31764a95985c2616d8969ec95e3f1c475164dc425ed3902e76d972dade278f35 not found: ID does not exist" containerID="31764a95985c2616d8969ec95e3f1c475164dc425ed3902e76d972dade278f35" Feb 02 09:10:27 crc kubenswrapper[4764]: I0202 09:10:27.164220 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"31764a95985c2616d8969ec95e3f1c475164dc425ed3902e76d972dade278f35"} err="failed to get container status \"31764a95985c2616d8969ec95e3f1c475164dc425ed3902e76d972dade278f35\": rpc error: code = NotFound desc = could not find container \"31764a95985c2616d8969ec95e3f1c475164dc425ed3902e76d972dade278f35\": container with ID starting with 31764a95985c2616d8969ec95e3f1c475164dc425ed3902e76d972dade278f35 not found: ID does not exist" Feb 02 09:10:27 crc kubenswrapper[4764]: I0202 09:10:27.165550 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-8576df8657-g98wg"] Feb 02 09:10:27 crc kubenswrapper[4764]: I0202 09:10:27.169644 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-hcbn7" podStartSLOduration=5.0654129 podStartE2EDuration="39.169632594s" podCreationTimestamp="2026-02-02 09:09:48 +0000 UTC" firstStartedPulling="2026-02-02 09:09:52.502611266 +0000 UTC m=+155.436335354" lastFinishedPulling="2026-02-02 09:10:26.60683096 +0000 UTC m=+189.540555048" observedRunningTime="2026-02-02 09:10:27.168755079 +0000 UTC m=+190.102479167" watchObservedRunningTime="2026-02-02 09:10:27.169632594 +0000 UTC m=+190.103356682" Feb 02 09:10:27 crc kubenswrapper[4764]: I0202 09:10:27.189875 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-795d7df575-bxzxp"] Feb 02 09:10:27 crc kubenswrapper[4764]: I0202 09:10:27.195732 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-795d7df575-bxzxp"] Feb 02 09:10:27 crc kubenswrapper[4764]: I0202 09:10:27.843347 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2d251b45-e18e-4837-87ed-8841ecd4f321" path="/var/lib/kubelet/pods/2d251b45-e18e-4837-87ed-8841ecd4f321/volumes" Feb 02 09:10:27 crc kubenswrapper[4764]: I0202 09:10:27.844149 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bb727d55-ab58-4a5d-97be-2d0798619f69" path="/var/lib/kubelet/pods/bb727d55-ab58-4a5d-97be-2d0798619f69/volumes" Feb 02 09:10:28 crc kubenswrapper[4764]: I0202 09:10:28.112242 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5489d77c9c-qdmth" event={"ID":"00732cf0-76cd-47f2-807e-363710f70a41","Type":"ContainerStarted","Data":"cf20b2c864a21dceb79ed8dc0780a5c05c72afe9786146b9c208cad350a20671"} Feb 02 09:10:28 crc kubenswrapper[4764]: I0202 09:10:28.112293 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5489d77c9c-qdmth" event={"ID":"00732cf0-76cd-47f2-807e-363710f70a41","Type":"ContainerStarted","Data":"153a2288a39c7d11c8f76ff5292e3a934b8e8c726773223f13c598daa4d1d14c"} Feb 02 09:10:28 crc kubenswrapper[4764]: I0202 09:10:28.112587 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-5489d77c9c-qdmth" Feb 02 09:10:28 crc kubenswrapper[4764]: I0202 09:10:28.116384 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7p7nt" event={"ID":"2e7d8f2e-1913-4b88-942a-995ad472697e","Type":"ContainerStarted","Data":"9d726676c72ec613c07aba44610f0cc43b6f5f4f5d9f256e234dfb03cbb48fae"} Feb 02 09:10:28 crc kubenswrapper[4764]: I0202 09:10:28.118208 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-5489d77c9c-qdmth" Feb 02 09:10:28 crc kubenswrapper[4764]: I0202 09:10:28.136289 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-5489d77c9c-qdmth" podStartSLOduration=5.136269388 podStartE2EDuration="5.136269388s" podCreationTimestamp="2026-02-02 09:10:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:10:28.134009044 +0000 UTC m=+191.067733132" watchObservedRunningTime="2026-02-02 09:10:28.136269388 +0000 UTC m=+191.069993476" Feb 02 09:10:28 crc kubenswrapper[4764]: I0202 09:10:28.155109 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-7p7nt" podStartSLOduration=5.674319849 podStartE2EDuration="42.155092384s" podCreationTimestamp="2026-02-02 09:09:46 +0000 UTC" firstStartedPulling="2026-02-02 09:09:51.333041998 +0000 UTC m=+154.266766086" lastFinishedPulling="2026-02-02 09:10:27.813814533 +0000 UTC m=+190.747538621" observedRunningTime="2026-02-02 09:10:28.150656668 +0000 UTC m=+191.084380756" watchObservedRunningTime="2026-02-02 09:10:28.155092384 +0000 UTC m=+191.088816472" Feb 02 09:10:28 crc kubenswrapper[4764]: I0202 09:10:28.889861 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-85b489c88d-fmzgh"] Feb 02 09:10:28 crc kubenswrapper[4764]: E0202 09:10:28.890449 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb727d55-ab58-4a5d-97be-2d0798619f69" containerName="route-controller-manager" Feb 02 09:10:28 crc kubenswrapper[4764]: I0202 09:10:28.890466 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb727d55-ab58-4a5d-97be-2d0798619f69" containerName="route-controller-manager" Feb 02 09:10:28 crc kubenswrapper[4764]: I0202 09:10:28.890587 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb727d55-ab58-4a5d-97be-2d0798619f69" containerName="route-controller-manager" Feb 02 09:10:28 crc kubenswrapper[4764]: I0202 09:10:28.891055 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-85b489c88d-fmzgh" Feb 02 09:10:28 crc kubenswrapper[4764]: I0202 09:10:28.893290 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Feb 02 09:10:28 crc kubenswrapper[4764]: I0202 09:10:28.893724 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Feb 02 09:10:28 crc kubenswrapper[4764]: I0202 09:10:28.894087 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Feb 02 09:10:28 crc kubenswrapper[4764]: I0202 09:10:28.894166 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Feb 02 09:10:28 crc kubenswrapper[4764]: I0202 09:10:28.897303 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Feb 02 09:10:28 crc kubenswrapper[4764]: I0202 09:10:28.897551 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Feb 02 09:10:28 crc kubenswrapper[4764]: I0202 09:10:28.902187 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-85b489c88d-fmzgh"] Feb 02 09:10:28 crc kubenswrapper[4764]: I0202 09:10:28.987133 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0c3637e8-e75a-455b-9511-71f59a1a26bf-serving-cert\") pod \"route-controller-manager-85b489c88d-fmzgh\" (UID: \"0c3637e8-e75a-455b-9511-71f59a1a26bf\") " pod="openshift-route-controller-manager/route-controller-manager-85b489c88d-fmzgh" Feb 02 09:10:28 crc kubenswrapper[4764]: I0202 09:10:28.987203 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0c3637e8-e75a-455b-9511-71f59a1a26bf-client-ca\") pod \"route-controller-manager-85b489c88d-fmzgh\" (UID: \"0c3637e8-e75a-455b-9511-71f59a1a26bf\") " pod="openshift-route-controller-manager/route-controller-manager-85b489c88d-fmzgh" Feb 02 09:10:28 crc kubenswrapper[4764]: I0202 09:10:28.987265 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0c3637e8-e75a-455b-9511-71f59a1a26bf-config\") pod \"route-controller-manager-85b489c88d-fmzgh\" (UID: \"0c3637e8-e75a-455b-9511-71f59a1a26bf\") " pod="openshift-route-controller-manager/route-controller-manager-85b489c88d-fmzgh" Feb 02 09:10:28 crc kubenswrapper[4764]: I0202 09:10:28.987289 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g8qqh\" (UniqueName: \"kubernetes.io/projected/0c3637e8-e75a-455b-9511-71f59a1a26bf-kube-api-access-g8qqh\") pod \"route-controller-manager-85b489c88d-fmzgh\" (UID: \"0c3637e8-e75a-455b-9511-71f59a1a26bf\") " pod="openshift-route-controller-manager/route-controller-manager-85b489c88d-fmzgh" Feb 02 09:10:29 crc kubenswrapper[4764]: I0202 09:10:29.067638 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-kbk9s" Feb 02 09:10:29 crc kubenswrapper[4764]: I0202 09:10:29.067685 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-kbk9s" Feb 02 09:10:29 crc kubenswrapper[4764]: I0202 09:10:29.088422 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0c3637e8-e75a-455b-9511-71f59a1a26bf-serving-cert\") pod \"route-controller-manager-85b489c88d-fmzgh\" (UID: \"0c3637e8-e75a-455b-9511-71f59a1a26bf\") " pod="openshift-route-controller-manager/route-controller-manager-85b489c88d-fmzgh" Feb 02 09:10:29 crc kubenswrapper[4764]: I0202 09:10:29.088501 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0c3637e8-e75a-455b-9511-71f59a1a26bf-client-ca\") pod \"route-controller-manager-85b489c88d-fmzgh\" (UID: \"0c3637e8-e75a-455b-9511-71f59a1a26bf\") " pod="openshift-route-controller-manager/route-controller-manager-85b489c88d-fmzgh" Feb 02 09:10:29 crc kubenswrapper[4764]: I0202 09:10:29.088554 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0c3637e8-e75a-455b-9511-71f59a1a26bf-config\") pod \"route-controller-manager-85b489c88d-fmzgh\" (UID: \"0c3637e8-e75a-455b-9511-71f59a1a26bf\") " pod="openshift-route-controller-manager/route-controller-manager-85b489c88d-fmzgh" Feb 02 09:10:29 crc kubenswrapper[4764]: I0202 09:10:29.088580 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g8qqh\" (UniqueName: \"kubernetes.io/projected/0c3637e8-e75a-455b-9511-71f59a1a26bf-kube-api-access-g8qqh\") pod \"route-controller-manager-85b489c88d-fmzgh\" (UID: \"0c3637e8-e75a-455b-9511-71f59a1a26bf\") " pod="openshift-route-controller-manager/route-controller-manager-85b489c88d-fmzgh" Feb 02 09:10:29 crc kubenswrapper[4764]: I0202 09:10:29.089451 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0c3637e8-e75a-455b-9511-71f59a1a26bf-client-ca\") pod \"route-controller-manager-85b489c88d-fmzgh\" (UID: \"0c3637e8-e75a-455b-9511-71f59a1a26bf\") " pod="openshift-route-controller-manager/route-controller-manager-85b489c88d-fmzgh" Feb 02 09:10:29 crc kubenswrapper[4764]: I0202 09:10:29.089717 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0c3637e8-e75a-455b-9511-71f59a1a26bf-config\") pod \"route-controller-manager-85b489c88d-fmzgh\" (UID: \"0c3637e8-e75a-455b-9511-71f59a1a26bf\") " pod="openshift-route-controller-manager/route-controller-manager-85b489c88d-fmzgh" Feb 02 09:10:29 crc kubenswrapper[4764]: I0202 09:10:29.097022 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0c3637e8-e75a-455b-9511-71f59a1a26bf-serving-cert\") pod \"route-controller-manager-85b489c88d-fmzgh\" (UID: \"0c3637e8-e75a-455b-9511-71f59a1a26bf\") " pod="openshift-route-controller-manager/route-controller-manager-85b489c88d-fmzgh" Feb 02 09:10:29 crc kubenswrapper[4764]: I0202 09:10:29.112160 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g8qqh\" (UniqueName: \"kubernetes.io/projected/0c3637e8-e75a-455b-9511-71f59a1a26bf-kube-api-access-g8qqh\") pod \"route-controller-manager-85b489c88d-fmzgh\" (UID: \"0c3637e8-e75a-455b-9511-71f59a1a26bf\") " pod="openshift-route-controller-manager/route-controller-manager-85b489c88d-fmzgh" Feb 02 09:10:29 crc kubenswrapper[4764]: I0202 09:10:29.205194 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-85b489c88d-fmzgh" Feb 02 09:10:29 crc kubenswrapper[4764]: I0202 09:10:29.231640 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-kbk9s" Feb 02 09:10:29 crc kubenswrapper[4764]: I0202 09:10:29.540418 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-hcbn7" Feb 02 09:10:29 crc kubenswrapper[4764]: I0202 09:10:29.540744 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-hcbn7" Feb 02 09:10:29 crc kubenswrapper[4764]: I0202 09:10:29.585079 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-hcbn7" Feb 02 09:10:29 crc kubenswrapper[4764]: I0202 09:10:29.679701 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-85b489c88d-fmzgh"] Feb 02 09:10:29 crc kubenswrapper[4764]: W0202 09:10:29.691454 4764 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0c3637e8_e75a_455b_9511_71f59a1a26bf.slice/crio-69e5323eef05db5fcd0f66427b0f1e8a2ec3ee19d72b7b261d99bc565fb52e4d WatchSource:0}: Error finding container 69e5323eef05db5fcd0f66427b0f1e8a2ec3ee19d72b7b261d99bc565fb52e4d: Status 404 returned error can't find the container with id 69e5323eef05db5fcd0f66427b0f1e8a2ec3ee19d72b7b261d99bc565fb52e4d Feb 02 09:10:30 crc kubenswrapper[4764]: I0202 09:10:30.129626 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-85b489c88d-fmzgh" event={"ID":"0c3637e8-e75a-455b-9511-71f59a1a26bf","Type":"ContainerStarted","Data":"1a2815b3d227502b432862d5cb65dca636b0b5698855c49be4bc4f6adf3d2948"} Feb 02 09:10:30 crc kubenswrapper[4764]: I0202 09:10:30.130002 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-85b489c88d-fmzgh" event={"ID":"0c3637e8-e75a-455b-9511-71f59a1a26bf","Type":"ContainerStarted","Data":"69e5323eef05db5fcd0f66427b0f1e8a2ec3ee19d72b7b261d99bc565fb52e4d"} Feb 02 09:10:30 crc kubenswrapper[4764]: I0202 09:10:30.130423 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-85b489c88d-fmzgh" Feb 02 09:10:30 crc kubenswrapper[4764]: I0202 09:10:30.152209 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-85b489c88d-fmzgh" podStartSLOduration=7.152191098 podStartE2EDuration="7.152191098s" podCreationTimestamp="2026-02-02 09:10:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:10:30.150625944 +0000 UTC m=+193.084350032" watchObservedRunningTime="2026-02-02 09:10:30.152191098 +0000 UTC m=+193.085915186" Feb 02 09:10:30 crc kubenswrapper[4764]: I0202 09:10:30.625444 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-85b489c88d-fmzgh" Feb 02 09:10:33 crc kubenswrapper[4764]: I0202 09:10:33.855540 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Feb 02 09:10:33 crc kubenswrapper[4764]: I0202 09:10:33.856620 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 02 09:10:33 crc kubenswrapper[4764]: I0202 09:10:33.861726 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Feb 02 09:10:33 crc kubenswrapper[4764]: I0202 09:10:33.864613 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Feb 02 09:10:33 crc kubenswrapper[4764]: I0202 09:10:33.864867 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Feb 02 09:10:33 crc kubenswrapper[4764]: I0202 09:10:33.968092 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/bebcb21f-dd13-44bb-b81a-e8fea92e3777-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"bebcb21f-dd13-44bb-b81a-e8fea92e3777\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 02 09:10:33 crc kubenswrapper[4764]: I0202 09:10:33.968147 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bebcb21f-dd13-44bb-b81a-e8fea92e3777-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"bebcb21f-dd13-44bb-b81a-e8fea92e3777\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 02 09:10:34 crc kubenswrapper[4764]: I0202 09:10:34.070027 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/bebcb21f-dd13-44bb-b81a-e8fea92e3777-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"bebcb21f-dd13-44bb-b81a-e8fea92e3777\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 02 09:10:34 crc kubenswrapper[4764]: I0202 09:10:34.070090 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bebcb21f-dd13-44bb-b81a-e8fea92e3777-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"bebcb21f-dd13-44bb-b81a-e8fea92e3777\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 02 09:10:34 crc kubenswrapper[4764]: I0202 09:10:34.070126 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/bebcb21f-dd13-44bb-b81a-e8fea92e3777-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"bebcb21f-dd13-44bb-b81a-e8fea92e3777\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 02 09:10:34 crc kubenswrapper[4764]: I0202 09:10:34.087784 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bebcb21f-dd13-44bb-b81a-e8fea92e3777-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"bebcb21f-dd13-44bb-b81a-e8fea92e3777\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 02 09:10:34 crc kubenswrapper[4764]: I0202 09:10:34.175424 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 02 09:10:34 crc kubenswrapper[4764]: I0202 09:10:34.612148 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Feb 02 09:10:34 crc kubenswrapper[4764]: W0202 09:10:34.623623 4764 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-podbebcb21f_dd13_44bb_b81a_e8fea92e3777.slice/crio-29b29aa3afe44a8cce84739e84db34f121e5a3cba660a0afeaf41b5a1745f551 WatchSource:0}: Error finding container 29b29aa3afe44a8cce84739e84db34f121e5a3cba660a0afeaf41b5a1745f551: Status 404 returned error can't find the container with id 29b29aa3afe44a8cce84739e84db34f121e5a3cba660a0afeaf41b5a1745f551 Feb 02 09:10:35 crc kubenswrapper[4764]: I0202 09:10:35.161368 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"bebcb21f-dd13-44bb-b81a-e8fea92e3777","Type":"ContainerStarted","Data":"29b29aa3afe44a8cce84739e84db34f121e5a3cba660a0afeaf41b5a1745f551"} Feb 02 09:10:36 crc kubenswrapper[4764]: I0202 09:10:36.168734 4764 generic.go:334] "Generic (PLEG): container finished" podID="bebcb21f-dd13-44bb-b81a-e8fea92e3777" containerID="87976c447c81e0b69e5fcb543a3d667107735f57c62d1f7f091e1f05fa5fa139" exitCode=0 Feb 02 09:10:36 crc kubenswrapper[4764]: I0202 09:10:36.168814 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"bebcb21f-dd13-44bb-b81a-e8fea92e3777","Type":"ContainerDied","Data":"87976c447c81e0b69e5fcb543a3d667107735f57c62d1f7f091e1f05fa5fa139"} Feb 02 09:10:37 crc kubenswrapper[4764]: I0202 09:10:37.283953 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-7p7nt" Feb 02 09:10:37 crc kubenswrapper[4764]: I0202 09:10:37.284210 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-7p7nt" Feb 02 09:10:37 crc kubenswrapper[4764]: I0202 09:10:37.331595 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-7p7nt" Feb 02 09:10:37 crc kubenswrapper[4764]: I0202 09:10:37.511693 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 02 09:10:37 crc kubenswrapper[4764]: I0202 09:10:37.630880 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bebcb21f-dd13-44bb-b81a-e8fea92e3777-kube-api-access\") pod \"bebcb21f-dd13-44bb-b81a-e8fea92e3777\" (UID: \"bebcb21f-dd13-44bb-b81a-e8fea92e3777\") " Feb 02 09:10:37 crc kubenswrapper[4764]: I0202 09:10:37.630976 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/bebcb21f-dd13-44bb-b81a-e8fea92e3777-kubelet-dir\") pod \"bebcb21f-dd13-44bb-b81a-e8fea92e3777\" (UID: \"bebcb21f-dd13-44bb-b81a-e8fea92e3777\") " Feb 02 09:10:37 crc kubenswrapper[4764]: I0202 09:10:37.631354 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bebcb21f-dd13-44bb-b81a-e8fea92e3777-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "bebcb21f-dd13-44bb-b81a-e8fea92e3777" (UID: "bebcb21f-dd13-44bb-b81a-e8fea92e3777"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 09:10:37 crc kubenswrapper[4764]: I0202 09:10:37.637652 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bebcb21f-dd13-44bb-b81a-e8fea92e3777-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "bebcb21f-dd13-44bb-b81a-e8fea92e3777" (UID: "bebcb21f-dd13-44bb-b81a-e8fea92e3777"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:10:37 crc kubenswrapper[4764]: I0202 09:10:37.732748 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bebcb21f-dd13-44bb-b81a-e8fea92e3777-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 02 09:10:37 crc kubenswrapper[4764]: I0202 09:10:37.732790 4764 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/bebcb21f-dd13-44bb-b81a-e8fea92e3777-kubelet-dir\") on node \"crc\" DevicePath \"\"" Feb 02 09:10:38 crc kubenswrapper[4764]: I0202 09:10:38.195722 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w6hd5" event={"ID":"83d5cb3d-ee7b-44cb-8379-8b52f2c3cadc","Type":"ContainerStarted","Data":"40f47d40a3959e7e99b123bb1f1c5e826a5ecc56dfffc0a9e0191c4b61192845"} Feb 02 09:10:38 crc kubenswrapper[4764]: I0202 09:10:38.199490 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ggrpx" event={"ID":"a40abe98-1419-4f5d-9bed-63f1c89028b1","Type":"ContainerStarted","Data":"242ff04a15506716500f4dde700de56665b7f4570bfc4ebbf3e4eb1dc4263f9f"} Feb 02 09:10:38 crc kubenswrapper[4764]: I0202 09:10:38.202045 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 02 09:10:38 crc kubenswrapper[4764]: I0202 09:10:38.202382 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"bebcb21f-dd13-44bb-b81a-e8fea92e3777","Type":"ContainerDied","Data":"29b29aa3afe44a8cce84739e84db34f121e5a3cba660a0afeaf41b5a1745f551"} Feb 02 09:10:38 crc kubenswrapper[4764]: I0202 09:10:38.202400 4764 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="29b29aa3afe44a8cce84739e84db34f121e5a3cba660a0afeaf41b5a1745f551" Feb 02 09:10:38 crc kubenswrapper[4764]: I0202 09:10:38.245582 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-7p7nt" Feb 02 09:10:39 crc kubenswrapper[4764]: I0202 09:10:39.103058 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-kbk9s" Feb 02 09:10:39 crc kubenswrapper[4764]: I0202 09:10:39.211541 4764 generic.go:334] "Generic (PLEG): container finished" podID="35376bf8-2200-49a4-8eb3-f5d7069e73a0" containerID="dcc388287da6c7fe0be64242fb469e15e627e43062209271becd9a55d5519b52" exitCode=0 Feb 02 09:10:39 crc kubenswrapper[4764]: I0202 09:10:39.211629 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mg5dh" event={"ID":"35376bf8-2200-49a4-8eb3-f5d7069e73a0","Type":"ContainerDied","Data":"dcc388287da6c7fe0be64242fb469e15e627e43062209271becd9a55d5519b52"} Feb 02 09:10:39 crc kubenswrapper[4764]: I0202 09:10:39.213857 4764 generic.go:334] "Generic (PLEG): container finished" podID="83d5cb3d-ee7b-44cb-8379-8b52f2c3cadc" containerID="40f47d40a3959e7e99b123bb1f1c5e826a5ecc56dfffc0a9e0191c4b61192845" exitCode=0 Feb 02 09:10:39 crc kubenswrapper[4764]: I0202 09:10:39.213941 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w6hd5" event={"ID":"83d5cb3d-ee7b-44cb-8379-8b52f2c3cadc","Type":"ContainerDied","Data":"40f47d40a3959e7e99b123bb1f1c5e826a5ecc56dfffc0a9e0191c4b61192845"} Feb 02 09:10:39 crc kubenswrapper[4764]: I0202 09:10:39.215706 4764 generic.go:334] "Generic (PLEG): container finished" podID="a40abe98-1419-4f5d-9bed-63f1c89028b1" containerID="242ff04a15506716500f4dde700de56665b7f4570bfc4ebbf3e4eb1dc4263f9f" exitCode=0 Feb 02 09:10:39 crc kubenswrapper[4764]: I0202 09:10:39.216071 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ggrpx" event={"ID":"a40abe98-1419-4f5d-9bed-63f1c89028b1","Type":"ContainerDied","Data":"242ff04a15506716500f4dde700de56665b7f4570bfc4ebbf3e4eb1dc4263f9f"} Feb 02 09:10:39 crc kubenswrapper[4764]: I0202 09:10:39.588268 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-hcbn7" Feb 02 09:10:40 crc kubenswrapper[4764]: I0202 09:10:40.223179 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mg5dh" event={"ID":"35376bf8-2200-49a4-8eb3-f5d7069e73a0","Type":"ContainerStarted","Data":"becc45a3d22ecefde54fc5e4cbe0ccddfebe7098e8f1a039b21df4f22d1c7fe6"} Feb 02 09:10:40 crc kubenswrapper[4764]: I0202 09:10:40.224888 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w6hd5" event={"ID":"83d5cb3d-ee7b-44cb-8379-8b52f2c3cadc","Type":"ContainerStarted","Data":"c542f29f86ec62e18a334179bc561bb45d7418d55b0033866696690b9c5f1c01"} Feb 02 09:10:40 crc kubenswrapper[4764]: I0202 09:10:40.227761 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ggrpx" event={"ID":"a40abe98-1419-4f5d-9bed-63f1c89028b1","Type":"ContainerStarted","Data":"0cbd7d0e3df1db8208d9af33d979c98caaa3f82973cc8263382a73f719638baa"} Feb 02 09:10:40 crc kubenswrapper[4764]: I0202 09:10:40.262230 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-mg5dh" podStartSLOduration=5.9425226989999995 podStartE2EDuration="54.262209952s" podCreationTimestamp="2026-02-02 09:09:46 +0000 UTC" firstStartedPulling="2026-02-02 09:09:51.333303145 +0000 UTC m=+154.267027233" lastFinishedPulling="2026-02-02 09:10:39.652990398 +0000 UTC m=+202.586714486" observedRunningTime="2026-02-02 09:10:40.243641795 +0000 UTC m=+203.177365883" watchObservedRunningTime="2026-02-02 09:10:40.262209952 +0000 UTC m=+203.195934040" Feb 02 09:10:40 crc kubenswrapper[4764]: I0202 09:10:40.262388 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-w6hd5" podStartSLOduration=4.150772009 podStartE2EDuration="51.262383007s" podCreationTimestamp="2026-02-02 09:09:49 +0000 UTC" firstStartedPulling="2026-02-02 09:09:52.530617493 +0000 UTC m=+155.464341581" lastFinishedPulling="2026-02-02 09:10:39.642228491 +0000 UTC m=+202.575952579" observedRunningTime="2026-02-02 09:10:40.262027766 +0000 UTC m=+203.195751874" watchObservedRunningTime="2026-02-02 09:10:40.262383007 +0000 UTC m=+203.196107095" Feb 02 09:10:40 crc kubenswrapper[4764]: I0202 09:10:40.282627 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-ggrpx" podStartSLOduration=5.985046808 podStartE2EDuration="54.282604243s" podCreationTimestamp="2026-02-02 09:09:46 +0000 UTC" firstStartedPulling="2026-02-02 09:09:51.436008867 +0000 UTC m=+154.369732955" lastFinishedPulling="2026-02-02 09:10:39.733566302 +0000 UTC m=+202.667290390" observedRunningTime="2026-02-02 09:10:40.278685917 +0000 UTC m=+203.212410015" watchObservedRunningTime="2026-02-02 09:10:40.282604243 +0000 UTC m=+203.216328331" Feb 02 09:10:41 crc kubenswrapper[4764]: I0202 09:10:41.850846 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Feb 02 09:10:41 crc kubenswrapper[4764]: E0202 09:10:41.851409 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bebcb21f-dd13-44bb-b81a-e8fea92e3777" containerName="pruner" Feb 02 09:10:41 crc kubenswrapper[4764]: I0202 09:10:41.851425 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="bebcb21f-dd13-44bb-b81a-e8fea92e3777" containerName="pruner" Feb 02 09:10:41 crc kubenswrapper[4764]: I0202 09:10:41.851553 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="bebcb21f-dd13-44bb-b81a-e8fea92e3777" containerName="pruner" Feb 02 09:10:41 crc kubenswrapper[4764]: I0202 09:10:41.851973 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Feb 02 09:10:41 crc kubenswrapper[4764]: I0202 09:10:41.855227 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Feb 02 09:10:41 crc kubenswrapper[4764]: I0202 09:10:41.855686 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Feb 02 09:10:41 crc kubenswrapper[4764]: I0202 09:10:41.860903 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7p7nt"] Feb 02 09:10:41 crc kubenswrapper[4764]: I0202 09:10:41.861149 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-7p7nt" podUID="2e7d8f2e-1913-4b88-942a-995ad472697e" containerName="registry-server" containerID="cri-o://9d726676c72ec613c07aba44610f0cc43b6f5f4f5d9f256e234dfb03cbb48fae" gracePeriod=2 Feb 02 09:10:41 crc kubenswrapper[4764]: I0202 09:10:41.870535 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Feb 02 09:10:41 crc kubenswrapper[4764]: I0202 09:10:41.990263 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/f6df65ba-ae7e-48e4-aa24-c5c00b23d4aa-kubelet-dir\") pod \"installer-9-crc\" (UID: \"f6df65ba-ae7e-48e4-aa24-c5c00b23d4aa\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 02 09:10:41 crc kubenswrapper[4764]: I0202 09:10:41.990382 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f6df65ba-ae7e-48e4-aa24-c5c00b23d4aa-kube-api-access\") pod \"installer-9-crc\" (UID: \"f6df65ba-ae7e-48e4-aa24-c5c00b23d4aa\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 02 09:10:41 crc kubenswrapper[4764]: I0202 09:10:41.990438 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f6df65ba-ae7e-48e4-aa24-c5c00b23d4aa-var-lock\") pod \"installer-9-crc\" (UID: \"f6df65ba-ae7e-48e4-aa24-c5c00b23d4aa\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 02 09:10:42 crc kubenswrapper[4764]: I0202 09:10:42.091628 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f6df65ba-ae7e-48e4-aa24-c5c00b23d4aa-kube-api-access\") pod \"installer-9-crc\" (UID: \"f6df65ba-ae7e-48e4-aa24-c5c00b23d4aa\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 02 09:10:42 crc kubenswrapper[4764]: I0202 09:10:42.091684 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f6df65ba-ae7e-48e4-aa24-c5c00b23d4aa-var-lock\") pod \"installer-9-crc\" (UID: \"f6df65ba-ae7e-48e4-aa24-c5c00b23d4aa\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 02 09:10:42 crc kubenswrapper[4764]: I0202 09:10:42.091725 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/f6df65ba-ae7e-48e4-aa24-c5c00b23d4aa-kubelet-dir\") pod \"installer-9-crc\" (UID: \"f6df65ba-ae7e-48e4-aa24-c5c00b23d4aa\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 02 09:10:42 crc kubenswrapper[4764]: I0202 09:10:42.091798 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/f6df65ba-ae7e-48e4-aa24-c5c00b23d4aa-kubelet-dir\") pod \"installer-9-crc\" (UID: \"f6df65ba-ae7e-48e4-aa24-c5c00b23d4aa\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 02 09:10:42 crc kubenswrapper[4764]: I0202 09:10:42.092037 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f6df65ba-ae7e-48e4-aa24-c5c00b23d4aa-var-lock\") pod \"installer-9-crc\" (UID: \"f6df65ba-ae7e-48e4-aa24-c5c00b23d4aa\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 02 09:10:42 crc kubenswrapper[4764]: I0202 09:10:42.120899 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f6df65ba-ae7e-48e4-aa24-c5c00b23d4aa-kube-api-access\") pod \"installer-9-crc\" (UID: \"f6df65ba-ae7e-48e4-aa24-c5c00b23d4aa\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 02 09:10:42 crc kubenswrapper[4764]: I0202 09:10:42.166241 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Feb 02 09:10:42 crc kubenswrapper[4764]: I0202 09:10:42.260206 4764 generic.go:334] "Generic (PLEG): container finished" podID="2e7d8f2e-1913-4b88-942a-995ad472697e" containerID="9d726676c72ec613c07aba44610f0cc43b6f5f4f5d9f256e234dfb03cbb48fae" exitCode=0 Feb 02 09:10:42 crc kubenswrapper[4764]: I0202 09:10:42.260551 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7p7nt" event={"ID":"2e7d8f2e-1913-4b88-942a-995ad472697e","Type":"ContainerDied","Data":"9d726676c72ec613c07aba44610f0cc43b6f5f4f5d9f256e234dfb03cbb48fae"} Feb 02 09:10:42 crc kubenswrapper[4764]: I0202 09:10:42.261888 4764 generic.go:334] "Generic (PLEG): container finished" podID="00cb5ef3-3b54-47ac-bfc4-b1d53f04d292" containerID="e5c51aeae99f7c7c3a6553c26d425d22a804604a09b6d6bd101fe7a4458b2e00" exitCode=0 Feb 02 09:10:42 crc kubenswrapper[4764]: I0202 09:10:42.262017 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ct55p" event={"ID":"00cb5ef3-3b54-47ac-bfc4-b1d53f04d292","Type":"ContainerDied","Data":"e5c51aeae99f7c7c3a6553c26d425d22a804604a09b6d6bd101fe7a4458b2e00"} Feb 02 09:10:42 crc kubenswrapper[4764]: I0202 09:10:42.272848 4764 generic.go:334] "Generic (PLEG): container finished" podID="01ede9d5-6b94-4d10-bf6d-3aef751840ff" containerID="796d368c28e1658dc2a81024ed86d820b72e2e1f251c1951ede876071b19b223" exitCode=0 Feb 02 09:10:42 crc kubenswrapper[4764]: I0202 09:10:42.272889 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j25pg" event={"ID":"01ede9d5-6b94-4d10-bf6d-3aef751840ff","Type":"ContainerDied","Data":"796d368c28e1658dc2a81024ed86d820b72e2e1f251c1951ede876071b19b223"} Feb 02 09:10:42 crc kubenswrapper[4764]: I0202 09:10:42.474083 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7p7nt" Feb 02 09:10:42 crc kubenswrapper[4764]: I0202 09:10:42.474421 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-hcbn7"] Feb 02 09:10:42 crc kubenswrapper[4764]: I0202 09:10:42.474654 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-hcbn7" podUID="88d63f88-8b5e-44de-b8e8-1417bbdb6e8e" containerName="registry-server" containerID="cri-o://6704e1b75fbbb895cf52f07530ca2bd1a31d27e3da65d3db3726c2b9e43b09dc" gracePeriod=2 Feb 02 09:10:42 crc kubenswrapper[4764]: I0202 09:10:42.604268 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2e7d8f2e-1913-4b88-942a-995ad472697e-utilities\") pod \"2e7d8f2e-1913-4b88-942a-995ad472697e\" (UID: \"2e7d8f2e-1913-4b88-942a-995ad472697e\") " Feb 02 09:10:42 crc kubenswrapper[4764]: I0202 09:10:42.604313 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8c22s\" (UniqueName: \"kubernetes.io/projected/2e7d8f2e-1913-4b88-942a-995ad472697e-kube-api-access-8c22s\") pod \"2e7d8f2e-1913-4b88-942a-995ad472697e\" (UID: \"2e7d8f2e-1913-4b88-942a-995ad472697e\") " Feb 02 09:10:42 crc kubenswrapper[4764]: I0202 09:10:42.604350 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2e7d8f2e-1913-4b88-942a-995ad472697e-catalog-content\") pod \"2e7d8f2e-1913-4b88-942a-995ad472697e\" (UID: \"2e7d8f2e-1913-4b88-942a-995ad472697e\") " Feb 02 09:10:42 crc kubenswrapper[4764]: I0202 09:10:42.605375 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2e7d8f2e-1913-4b88-942a-995ad472697e-utilities" (OuterVolumeSpecName: "utilities") pod "2e7d8f2e-1913-4b88-942a-995ad472697e" (UID: "2e7d8f2e-1913-4b88-942a-995ad472697e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 09:10:42 crc kubenswrapper[4764]: I0202 09:10:42.608247 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2e7d8f2e-1913-4b88-942a-995ad472697e-kube-api-access-8c22s" (OuterVolumeSpecName: "kube-api-access-8c22s") pod "2e7d8f2e-1913-4b88-942a-995ad472697e" (UID: "2e7d8f2e-1913-4b88-942a-995ad472697e"). InnerVolumeSpecName "kube-api-access-8c22s". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:10:42 crc kubenswrapper[4764]: I0202 09:10:42.649926 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2e7d8f2e-1913-4b88-942a-995ad472697e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2e7d8f2e-1913-4b88-942a-995ad472697e" (UID: "2e7d8f2e-1913-4b88-942a-995ad472697e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 09:10:42 crc kubenswrapper[4764]: I0202 09:10:42.705944 4764 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2e7d8f2e-1913-4b88-942a-995ad472697e-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 09:10:42 crc kubenswrapper[4764]: I0202 09:10:42.705991 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8c22s\" (UniqueName: \"kubernetes.io/projected/2e7d8f2e-1913-4b88-942a-995ad472697e-kube-api-access-8c22s\") on node \"crc\" DevicePath \"\"" Feb 02 09:10:42 crc kubenswrapper[4764]: I0202 09:10:42.706004 4764 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2e7d8f2e-1913-4b88-942a-995ad472697e-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 09:10:42 crc kubenswrapper[4764]: I0202 09:10:42.732490 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Feb 02 09:10:42 crc kubenswrapper[4764]: W0202 09:10:42.767683 4764 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-podf6df65ba_ae7e_48e4_aa24_c5c00b23d4aa.slice/crio-1a11c9a20f1e10c2d4157e6f16fcd104fe8d3b63d89d4390370b24693f10aec8 WatchSource:0}: Error finding container 1a11c9a20f1e10c2d4157e6f16fcd104fe8d3b63d89d4390370b24693f10aec8: Status 404 returned error can't find the container with id 1a11c9a20f1e10c2d4157e6f16fcd104fe8d3b63d89d4390370b24693f10aec8 Feb 02 09:10:42 crc kubenswrapper[4764]: I0202 09:10:42.897857 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hcbn7" Feb 02 09:10:43 crc kubenswrapper[4764]: I0202 09:10:43.011451 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ghh25\" (UniqueName: \"kubernetes.io/projected/88d63f88-8b5e-44de-b8e8-1417bbdb6e8e-kube-api-access-ghh25\") pod \"88d63f88-8b5e-44de-b8e8-1417bbdb6e8e\" (UID: \"88d63f88-8b5e-44de-b8e8-1417bbdb6e8e\") " Feb 02 09:10:43 crc kubenswrapper[4764]: I0202 09:10:43.011883 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/88d63f88-8b5e-44de-b8e8-1417bbdb6e8e-catalog-content\") pod \"88d63f88-8b5e-44de-b8e8-1417bbdb6e8e\" (UID: \"88d63f88-8b5e-44de-b8e8-1417bbdb6e8e\") " Feb 02 09:10:43 crc kubenswrapper[4764]: I0202 09:10:43.012048 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/88d63f88-8b5e-44de-b8e8-1417bbdb6e8e-utilities\") pod \"88d63f88-8b5e-44de-b8e8-1417bbdb6e8e\" (UID: \"88d63f88-8b5e-44de-b8e8-1417bbdb6e8e\") " Feb 02 09:10:43 crc kubenswrapper[4764]: I0202 09:10:43.012767 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/88d63f88-8b5e-44de-b8e8-1417bbdb6e8e-utilities" (OuterVolumeSpecName: "utilities") pod "88d63f88-8b5e-44de-b8e8-1417bbdb6e8e" (UID: "88d63f88-8b5e-44de-b8e8-1417bbdb6e8e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 09:10:43 crc kubenswrapper[4764]: I0202 09:10:43.017463 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/88d63f88-8b5e-44de-b8e8-1417bbdb6e8e-kube-api-access-ghh25" (OuterVolumeSpecName: "kube-api-access-ghh25") pod "88d63f88-8b5e-44de-b8e8-1417bbdb6e8e" (UID: "88d63f88-8b5e-44de-b8e8-1417bbdb6e8e"). InnerVolumeSpecName "kube-api-access-ghh25". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:10:43 crc kubenswrapper[4764]: I0202 09:10:43.036047 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/88d63f88-8b5e-44de-b8e8-1417bbdb6e8e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "88d63f88-8b5e-44de-b8e8-1417bbdb6e8e" (UID: "88d63f88-8b5e-44de-b8e8-1417bbdb6e8e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 09:10:43 crc kubenswrapper[4764]: I0202 09:10:43.113918 4764 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/88d63f88-8b5e-44de-b8e8-1417bbdb6e8e-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 09:10:43 crc kubenswrapper[4764]: I0202 09:10:43.113966 4764 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/88d63f88-8b5e-44de-b8e8-1417bbdb6e8e-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 09:10:43 crc kubenswrapper[4764]: I0202 09:10:43.113977 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ghh25\" (UniqueName: \"kubernetes.io/projected/88d63f88-8b5e-44de-b8e8-1417bbdb6e8e-kube-api-access-ghh25\") on node \"crc\" DevicePath \"\"" Feb 02 09:10:43 crc kubenswrapper[4764]: I0202 09:10:43.281742 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7p7nt" event={"ID":"2e7d8f2e-1913-4b88-942a-995ad472697e","Type":"ContainerDied","Data":"d075f0ec2e09cc7ea56f3ed96642ddbc69689f8351a4428d20ee4e10f9e69a40"} Feb 02 09:10:43 crc kubenswrapper[4764]: I0202 09:10:43.281788 4764 scope.go:117] "RemoveContainer" containerID="9d726676c72ec613c07aba44610f0cc43b6f5f4f5d9f256e234dfb03cbb48fae" Feb 02 09:10:43 crc kubenswrapper[4764]: I0202 09:10:43.281898 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7p7nt" Feb 02 09:10:43 crc kubenswrapper[4764]: I0202 09:10:43.290671 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ct55p" event={"ID":"00cb5ef3-3b54-47ac-bfc4-b1d53f04d292","Type":"ContainerStarted","Data":"b0e277ec4569fc9807eb7c3cc0a8905c030a0b41601d8880193f95814472f82e"} Feb 02 09:10:43 crc kubenswrapper[4764]: I0202 09:10:43.301707 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j25pg" event={"ID":"01ede9d5-6b94-4d10-bf6d-3aef751840ff","Type":"ContainerStarted","Data":"18af9361e46e52d7b7a0adc034d3d3d089dd4e8f16f90621792ba3be7ceba8f7"} Feb 02 09:10:43 crc kubenswrapper[4764]: I0202 09:10:43.308256 4764 generic.go:334] "Generic (PLEG): container finished" podID="88d63f88-8b5e-44de-b8e8-1417bbdb6e8e" containerID="6704e1b75fbbb895cf52f07530ca2bd1a31d27e3da65d3db3726c2b9e43b09dc" exitCode=0 Feb 02 09:10:43 crc kubenswrapper[4764]: I0202 09:10:43.308325 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hcbn7" event={"ID":"88d63f88-8b5e-44de-b8e8-1417bbdb6e8e","Type":"ContainerDied","Data":"6704e1b75fbbb895cf52f07530ca2bd1a31d27e3da65d3db3726c2b9e43b09dc"} Feb 02 09:10:43 crc kubenswrapper[4764]: I0202 09:10:43.308357 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hcbn7" event={"ID":"88d63f88-8b5e-44de-b8e8-1417bbdb6e8e","Type":"ContainerDied","Data":"79bb3443189e6281d7de325470986c4c39a9563b198dfa9d99a1ef8766854151"} Feb 02 09:10:43 crc kubenswrapper[4764]: I0202 09:10:43.308430 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hcbn7" Feb 02 09:10:43 crc kubenswrapper[4764]: I0202 09:10:43.311179 4764 scope.go:117] "RemoveContainer" containerID="e9151643c0db78fb3854d26596227986f108e4c0a6f256d439efc94fdfaeab74" Feb 02 09:10:43 crc kubenswrapper[4764]: I0202 09:10:43.312065 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"f6df65ba-ae7e-48e4-aa24-c5c00b23d4aa","Type":"ContainerStarted","Data":"be556bb93f0178d943541fb659f6f1d6dfd72897ea2234499f80f374591e4e66"} Feb 02 09:10:43 crc kubenswrapper[4764]: I0202 09:10:43.312183 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"f6df65ba-ae7e-48e4-aa24-c5c00b23d4aa","Type":"ContainerStarted","Data":"1a11c9a20f1e10c2d4157e6f16fcd104fe8d3b63d89d4390370b24693f10aec8"} Feb 02 09:10:43 crc kubenswrapper[4764]: I0202 09:10:43.321460 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-ct55p" podStartSLOduration=4.01363418 podStartE2EDuration="54.321439478s" podCreationTimestamp="2026-02-02 09:09:49 +0000 UTC" firstStartedPulling="2026-02-02 09:09:52.508736151 +0000 UTC m=+155.442460239" lastFinishedPulling="2026-02-02 09:10:42.816541449 +0000 UTC m=+205.750265537" observedRunningTime="2026-02-02 09:10:43.319969375 +0000 UTC m=+206.253693473" watchObservedRunningTime="2026-02-02 09:10:43.321439478 +0000 UTC m=+206.255163566" Feb 02 09:10:43 crc kubenswrapper[4764]: I0202 09:10:43.344199 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-5489d77c9c-qdmth"] Feb 02 09:10:43 crc kubenswrapper[4764]: I0202 09:10:43.344445 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-5489d77c9c-qdmth" podUID="00732cf0-76cd-47f2-807e-363710f70a41" containerName="controller-manager" containerID="cri-o://cf20b2c864a21dceb79ed8dc0780a5c05c72afe9786146b9c208cad350a20671" gracePeriod=30 Feb 02 09:10:43 crc kubenswrapper[4764]: I0202 09:10:43.350289 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=2.350272568 podStartE2EDuration="2.350272568s" podCreationTimestamp="2026-02-02 09:10:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:10:43.349440523 +0000 UTC m=+206.283164611" watchObservedRunningTime="2026-02-02 09:10:43.350272568 +0000 UTC m=+206.283996656" Feb 02 09:10:43 crc kubenswrapper[4764]: I0202 09:10:43.360762 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-85b489c88d-fmzgh"] Feb 02 09:10:43 crc kubenswrapper[4764]: I0202 09:10:43.361205 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-85b489c88d-fmzgh" podUID="0c3637e8-e75a-455b-9511-71f59a1a26bf" containerName="route-controller-manager" containerID="cri-o://1a2815b3d227502b432862d5cb65dca636b0b5698855c49be4bc4f6adf3d2948" gracePeriod=30 Feb 02 09:10:43 crc kubenswrapper[4764]: I0202 09:10:43.381005 4764 scope.go:117] "RemoveContainer" containerID="39c8d9f8da4ffed5557047562303f1757f5cb2317304678028c8a785e36bd325" Feb 02 09:10:43 crc kubenswrapper[4764]: I0202 09:10:43.384725 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7p7nt"] Feb 02 09:10:43 crc kubenswrapper[4764]: I0202 09:10:43.393773 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-7p7nt"] Feb 02 09:10:43 crc kubenswrapper[4764]: I0202 09:10:43.417161 4764 scope.go:117] "RemoveContainer" containerID="6704e1b75fbbb895cf52f07530ca2bd1a31d27e3da65d3db3726c2b9e43b09dc" Feb 02 09:10:43 crc kubenswrapper[4764]: I0202 09:10:43.449879 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-j25pg" podStartSLOduration=4.724861768 podStartE2EDuration="57.449858323s" podCreationTimestamp="2026-02-02 09:09:46 +0000 UTC" firstStartedPulling="2026-02-02 09:09:50.24887128 +0000 UTC m=+153.182595368" lastFinishedPulling="2026-02-02 09:10:42.973867835 +0000 UTC m=+205.907591923" observedRunningTime="2026-02-02 09:10:43.447505064 +0000 UTC m=+206.381229162" watchObservedRunningTime="2026-02-02 09:10:43.449858323 +0000 UTC m=+206.383582411" Feb 02 09:10:43 crc kubenswrapper[4764]: I0202 09:10:43.455264 4764 scope.go:117] "RemoveContainer" containerID="52ba7ace2c80b4d708fb26af11b1a9aa76411436399c3aaf387655ec8c2e271e" Feb 02 09:10:43 crc kubenswrapper[4764]: I0202 09:10:43.484185 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-hcbn7"] Feb 02 09:10:43 crc kubenswrapper[4764]: I0202 09:10:43.490256 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-hcbn7"] Feb 02 09:10:43 crc kubenswrapper[4764]: I0202 09:10:43.523647 4764 patch_prober.go:28] interesting pod/machine-config-daemon-4ndm4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 09:10:43 crc kubenswrapper[4764]: I0202 09:10:43.523703 4764 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 09:10:43 crc kubenswrapper[4764]: I0202 09:10:43.523756 4764 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" Feb 02 09:10:43 crc kubenswrapper[4764]: I0202 09:10:43.524360 4764 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5fdf06410349b7c99aa166e87fbe80bf68e2ae923635a6a978b7e2cd50383f7b"} pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 02 09:10:43 crc kubenswrapper[4764]: I0202 09:10:43.524410 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" containerName="machine-config-daemon" containerID="cri-o://5fdf06410349b7c99aa166e87fbe80bf68e2ae923635a6a978b7e2cd50383f7b" gracePeriod=600 Feb 02 09:10:43 crc kubenswrapper[4764]: I0202 09:10:43.538904 4764 scope.go:117] "RemoveContainer" containerID="93c4bd63dd1eaf39ff8d7650175e8905d0f1a5e49eba7f09bdcb37a3cf07c003" Feb 02 09:10:43 crc kubenswrapper[4764]: I0202 09:10:43.580922 4764 scope.go:117] "RemoveContainer" containerID="6704e1b75fbbb895cf52f07530ca2bd1a31d27e3da65d3db3726c2b9e43b09dc" Feb 02 09:10:43 crc kubenswrapper[4764]: E0202 09:10:43.581561 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6704e1b75fbbb895cf52f07530ca2bd1a31d27e3da65d3db3726c2b9e43b09dc\": container with ID starting with 6704e1b75fbbb895cf52f07530ca2bd1a31d27e3da65d3db3726c2b9e43b09dc not found: ID does not exist" containerID="6704e1b75fbbb895cf52f07530ca2bd1a31d27e3da65d3db3726c2b9e43b09dc" Feb 02 09:10:43 crc kubenswrapper[4764]: I0202 09:10:43.581615 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6704e1b75fbbb895cf52f07530ca2bd1a31d27e3da65d3db3726c2b9e43b09dc"} err="failed to get container status \"6704e1b75fbbb895cf52f07530ca2bd1a31d27e3da65d3db3726c2b9e43b09dc\": rpc error: code = NotFound desc = could not find container \"6704e1b75fbbb895cf52f07530ca2bd1a31d27e3da65d3db3726c2b9e43b09dc\": container with ID starting with 6704e1b75fbbb895cf52f07530ca2bd1a31d27e3da65d3db3726c2b9e43b09dc not found: ID does not exist" Feb 02 09:10:43 crc kubenswrapper[4764]: I0202 09:10:43.581654 4764 scope.go:117] "RemoveContainer" containerID="52ba7ace2c80b4d708fb26af11b1a9aa76411436399c3aaf387655ec8c2e271e" Feb 02 09:10:43 crc kubenswrapper[4764]: E0202 09:10:43.582023 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"52ba7ace2c80b4d708fb26af11b1a9aa76411436399c3aaf387655ec8c2e271e\": container with ID starting with 52ba7ace2c80b4d708fb26af11b1a9aa76411436399c3aaf387655ec8c2e271e not found: ID does not exist" containerID="52ba7ace2c80b4d708fb26af11b1a9aa76411436399c3aaf387655ec8c2e271e" Feb 02 09:10:43 crc kubenswrapper[4764]: I0202 09:10:43.582054 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"52ba7ace2c80b4d708fb26af11b1a9aa76411436399c3aaf387655ec8c2e271e"} err="failed to get container status \"52ba7ace2c80b4d708fb26af11b1a9aa76411436399c3aaf387655ec8c2e271e\": rpc error: code = NotFound desc = could not find container \"52ba7ace2c80b4d708fb26af11b1a9aa76411436399c3aaf387655ec8c2e271e\": container with ID starting with 52ba7ace2c80b4d708fb26af11b1a9aa76411436399c3aaf387655ec8c2e271e not found: ID does not exist" Feb 02 09:10:43 crc kubenswrapper[4764]: I0202 09:10:43.582077 4764 scope.go:117] "RemoveContainer" containerID="93c4bd63dd1eaf39ff8d7650175e8905d0f1a5e49eba7f09bdcb37a3cf07c003" Feb 02 09:10:43 crc kubenswrapper[4764]: E0202 09:10:43.582435 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"93c4bd63dd1eaf39ff8d7650175e8905d0f1a5e49eba7f09bdcb37a3cf07c003\": container with ID starting with 93c4bd63dd1eaf39ff8d7650175e8905d0f1a5e49eba7f09bdcb37a3cf07c003 not found: ID does not exist" containerID="93c4bd63dd1eaf39ff8d7650175e8905d0f1a5e49eba7f09bdcb37a3cf07c003" Feb 02 09:10:43 crc kubenswrapper[4764]: I0202 09:10:43.582474 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"93c4bd63dd1eaf39ff8d7650175e8905d0f1a5e49eba7f09bdcb37a3cf07c003"} err="failed to get container status \"93c4bd63dd1eaf39ff8d7650175e8905d0f1a5e49eba7f09bdcb37a3cf07c003\": rpc error: code = NotFound desc = could not find container \"93c4bd63dd1eaf39ff8d7650175e8905d0f1a5e49eba7f09bdcb37a3cf07c003\": container with ID starting with 93c4bd63dd1eaf39ff8d7650175e8905d0f1a5e49eba7f09bdcb37a3cf07c003 not found: ID does not exist" Feb 02 09:10:43 crc kubenswrapper[4764]: I0202 09:10:43.832349 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2e7d8f2e-1913-4b88-942a-995ad472697e" path="/var/lib/kubelet/pods/2e7d8f2e-1913-4b88-942a-995ad472697e/volumes" Feb 02 09:10:43 crc kubenswrapper[4764]: I0202 09:10:43.833048 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="88d63f88-8b5e-44de-b8e8-1417bbdb6e8e" path="/var/lib/kubelet/pods/88d63f88-8b5e-44de-b8e8-1417bbdb6e8e/volumes" Feb 02 09:10:44 crc kubenswrapper[4764]: I0202 09:10:44.317634 4764 generic.go:334] "Generic (PLEG): container finished" podID="00732cf0-76cd-47f2-807e-363710f70a41" containerID="cf20b2c864a21dceb79ed8dc0780a5c05c72afe9786146b9c208cad350a20671" exitCode=0 Feb 02 09:10:44 crc kubenswrapper[4764]: I0202 09:10:44.317700 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5489d77c9c-qdmth" event={"ID":"00732cf0-76cd-47f2-807e-363710f70a41","Type":"ContainerDied","Data":"cf20b2c864a21dceb79ed8dc0780a5c05c72afe9786146b9c208cad350a20671"} Feb 02 09:10:44 crc kubenswrapper[4764]: I0202 09:10:44.319642 4764 generic.go:334] "Generic (PLEG): container finished" podID="d192f670-9f9d-4539-9641-e4bed73acdd4" containerID="5fdf06410349b7c99aa166e87fbe80bf68e2ae923635a6a978b7e2cd50383f7b" exitCode=0 Feb 02 09:10:44 crc kubenswrapper[4764]: I0202 09:10:44.319698 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" event={"ID":"d192f670-9f9d-4539-9641-e4bed73acdd4","Type":"ContainerDied","Data":"5fdf06410349b7c99aa166e87fbe80bf68e2ae923635a6a978b7e2cd50383f7b"} Feb 02 09:10:44 crc kubenswrapper[4764]: I0202 09:10:44.321249 4764 generic.go:334] "Generic (PLEG): container finished" podID="0c3637e8-e75a-455b-9511-71f59a1a26bf" containerID="1a2815b3d227502b432862d5cb65dca636b0b5698855c49be4bc4f6adf3d2948" exitCode=0 Feb 02 09:10:44 crc kubenswrapper[4764]: I0202 09:10:44.321306 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-85b489c88d-fmzgh" event={"ID":"0c3637e8-e75a-455b-9511-71f59a1a26bf","Type":"ContainerDied","Data":"1a2815b3d227502b432862d5cb65dca636b0b5698855c49be4bc4f6adf3d2948"} Feb 02 09:10:44 crc kubenswrapper[4764]: I0202 09:10:44.537279 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-85b489c88d-fmzgh" Feb 02 09:10:44 crc kubenswrapper[4764]: I0202 09:10:44.569801 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7dc8d6dc8d-gcxph"] Feb 02 09:10:44 crc kubenswrapper[4764]: E0202 09:10:44.573432 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c3637e8-e75a-455b-9511-71f59a1a26bf" containerName="route-controller-manager" Feb 02 09:10:44 crc kubenswrapper[4764]: I0202 09:10:44.573474 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c3637e8-e75a-455b-9511-71f59a1a26bf" containerName="route-controller-manager" Feb 02 09:10:44 crc kubenswrapper[4764]: E0202 09:10:44.573487 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88d63f88-8b5e-44de-b8e8-1417bbdb6e8e" containerName="extract-utilities" Feb 02 09:10:44 crc kubenswrapper[4764]: I0202 09:10:44.573494 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="88d63f88-8b5e-44de-b8e8-1417bbdb6e8e" containerName="extract-utilities" Feb 02 09:10:44 crc kubenswrapper[4764]: E0202 09:10:44.573517 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88d63f88-8b5e-44de-b8e8-1417bbdb6e8e" containerName="extract-content" Feb 02 09:10:44 crc kubenswrapper[4764]: I0202 09:10:44.573525 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="88d63f88-8b5e-44de-b8e8-1417bbdb6e8e" containerName="extract-content" Feb 02 09:10:44 crc kubenswrapper[4764]: E0202 09:10:44.573539 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88d63f88-8b5e-44de-b8e8-1417bbdb6e8e" containerName="registry-server" Feb 02 09:10:44 crc kubenswrapper[4764]: I0202 09:10:44.573546 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="88d63f88-8b5e-44de-b8e8-1417bbdb6e8e" containerName="registry-server" Feb 02 09:10:44 crc kubenswrapper[4764]: E0202 09:10:44.573555 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e7d8f2e-1913-4b88-942a-995ad472697e" containerName="registry-server" Feb 02 09:10:44 crc kubenswrapper[4764]: I0202 09:10:44.573562 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e7d8f2e-1913-4b88-942a-995ad472697e" containerName="registry-server" Feb 02 09:10:44 crc kubenswrapper[4764]: E0202 09:10:44.573571 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e7d8f2e-1913-4b88-942a-995ad472697e" containerName="extract-content" Feb 02 09:10:44 crc kubenswrapper[4764]: I0202 09:10:44.573578 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e7d8f2e-1913-4b88-942a-995ad472697e" containerName="extract-content" Feb 02 09:10:44 crc kubenswrapper[4764]: E0202 09:10:44.573589 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e7d8f2e-1913-4b88-942a-995ad472697e" containerName="extract-utilities" Feb 02 09:10:44 crc kubenswrapper[4764]: I0202 09:10:44.573595 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e7d8f2e-1913-4b88-942a-995ad472697e" containerName="extract-utilities" Feb 02 09:10:44 crc kubenswrapper[4764]: I0202 09:10:44.573708 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="88d63f88-8b5e-44de-b8e8-1417bbdb6e8e" containerName="registry-server" Feb 02 09:10:44 crc kubenswrapper[4764]: I0202 09:10:44.573722 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c3637e8-e75a-455b-9511-71f59a1a26bf" containerName="route-controller-manager" Feb 02 09:10:44 crc kubenswrapper[4764]: I0202 09:10:44.573733 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e7d8f2e-1913-4b88-942a-995ad472697e" containerName="registry-server" Feb 02 09:10:44 crc kubenswrapper[4764]: I0202 09:10:44.574177 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7dc8d6dc8d-gcxph" Feb 02 09:10:44 crc kubenswrapper[4764]: I0202 09:10:44.585512 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7dc8d6dc8d-gcxph"] Feb 02 09:10:44 crc kubenswrapper[4764]: I0202 09:10:44.633425 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g8qqh\" (UniqueName: \"kubernetes.io/projected/0c3637e8-e75a-455b-9511-71f59a1a26bf-kube-api-access-g8qqh\") pod \"0c3637e8-e75a-455b-9511-71f59a1a26bf\" (UID: \"0c3637e8-e75a-455b-9511-71f59a1a26bf\") " Feb 02 09:10:44 crc kubenswrapper[4764]: I0202 09:10:44.633546 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0c3637e8-e75a-455b-9511-71f59a1a26bf-serving-cert\") pod \"0c3637e8-e75a-455b-9511-71f59a1a26bf\" (UID: \"0c3637e8-e75a-455b-9511-71f59a1a26bf\") " Feb 02 09:10:44 crc kubenswrapper[4764]: I0202 09:10:44.633590 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0c3637e8-e75a-455b-9511-71f59a1a26bf-client-ca\") pod \"0c3637e8-e75a-455b-9511-71f59a1a26bf\" (UID: \"0c3637e8-e75a-455b-9511-71f59a1a26bf\") " Feb 02 09:10:44 crc kubenswrapper[4764]: I0202 09:10:44.633612 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0c3637e8-e75a-455b-9511-71f59a1a26bf-config\") pod \"0c3637e8-e75a-455b-9511-71f59a1a26bf\" (UID: \"0c3637e8-e75a-455b-9511-71f59a1a26bf\") " Feb 02 09:10:44 crc kubenswrapper[4764]: I0202 09:10:44.633738 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/75215b19-865c-4ae9-b362-d036a4240375-config\") pod \"route-controller-manager-7dc8d6dc8d-gcxph\" (UID: \"75215b19-865c-4ae9-b362-d036a4240375\") " pod="openshift-route-controller-manager/route-controller-manager-7dc8d6dc8d-gcxph" Feb 02 09:10:44 crc kubenswrapper[4764]: I0202 09:10:44.633791 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/75215b19-865c-4ae9-b362-d036a4240375-serving-cert\") pod \"route-controller-manager-7dc8d6dc8d-gcxph\" (UID: \"75215b19-865c-4ae9-b362-d036a4240375\") " pod="openshift-route-controller-manager/route-controller-manager-7dc8d6dc8d-gcxph" Feb 02 09:10:44 crc kubenswrapper[4764]: I0202 09:10:44.633822 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-djx2w\" (UniqueName: \"kubernetes.io/projected/75215b19-865c-4ae9-b362-d036a4240375-kube-api-access-djx2w\") pod \"route-controller-manager-7dc8d6dc8d-gcxph\" (UID: \"75215b19-865c-4ae9-b362-d036a4240375\") " pod="openshift-route-controller-manager/route-controller-manager-7dc8d6dc8d-gcxph" Feb 02 09:10:44 crc kubenswrapper[4764]: I0202 09:10:44.633850 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/75215b19-865c-4ae9-b362-d036a4240375-client-ca\") pod \"route-controller-manager-7dc8d6dc8d-gcxph\" (UID: \"75215b19-865c-4ae9-b362-d036a4240375\") " pod="openshift-route-controller-manager/route-controller-manager-7dc8d6dc8d-gcxph" Feb 02 09:10:44 crc kubenswrapper[4764]: I0202 09:10:44.634587 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0c3637e8-e75a-455b-9511-71f59a1a26bf-client-ca" (OuterVolumeSpecName: "client-ca") pod "0c3637e8-e75a-455b-9511-71f59a1a26bf" (UID: "0c3637e8-e75a-455b-9511-71f59a1a26bf"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:10:44 crc kubenswrapper[4764]: I0202 09:10:44.635002 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0c3637e8-e75a-455b-9511-71f59a1a26bf-config" (OuterVolumeSpecName: "config") pod "0c3637e8-e75a-455b-9511-71f59a1a26bf" (UID: "0c3637e8-e75a-455b-9511-71f59a1a26bf"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:10:44 crc kubenswrapper[4764]: I0202 09:10:44.638453 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c3637e8-e75a-455b-9511-71f59a1a26bf-kube-api-access-g8qqh" (OuterVolumeSpecName: "kube-api-access-g8qqh") pod "0c3637e8-e75a-455b-9511-71f59a1a26bf" (UID: "0c3637e8-e75a-455b-9511-71f59a1a26bf"). InnerVolumeSpecName "kube-api-access-g8qqh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:10:44 crc kubenswrapper[4764]: I0202 09:10:44.650957 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c3637e8-e75a-455b-9511-71f59a1a26bf-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0c3637e8-e75a-455b-9511-71f59a1a26bf" (UID: "0c3637e8-e75a-455b-9511-71f59a1a26bf"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:10:44 crc kubenswrapper[4764]: I0202 09:10:44.698719 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5489d77c9c-qdmth" Feb 02 09:10:44 crc kubenswrapper[4764]: I0202 09:10:44.735469 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/75215b19-865c-4ae9-b362-d036a4240375-serving-cert\") pod \"route-controller-manager-7dc8d6dc8d-gcxph\" (UID: \"75215b19-865c-4ae9-b362-d036a4240375\") " pod="openshift-route-controller-manager/route-controller-manager-7dc8d6dc8d-gcxph" Feb 02 09:10:44 crc kubenswrapper[4764]: I0202 09:10:44.735524 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-djx2w\" (UniqueName: \"kubernetes.io/projected/75215b19-865c-4ae9-b362-d036a4240375-kube-api-access-djx2w\") pod \"route-controller-manager-7dc8d6dc8d-gcxph\" (UID: \"75215b19-865c-4ae9-b362-d036a4240375\") " pod="openshift-route-controller-manager/route-controller-manager-7dc8d6dc8d-gcxph" Feb 02 09:10:44 crc kubenswrapper[4764]: I0202 09:10:44.735552 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/75215b19-865c-4ae9-b362-d036a4240375-client-ca\") pod \"route-controller-manager-7dc8d6dc8d-gcxph\" (UID: \"75215b19-865c-4ae9-b362-d036a4240375\") " pod="openshift-route-controller-manager/route-controller-manager-7dc8d6dc8d-gcxph" Feb 02 09:10:44 crc kubenswrapper[4764]: I0202 09:10:44.735586 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/75215b19-865c-4ae9-b362-d036a4240375-config\") pod \"route-controller-manager-7dc8d6dc8d-gcxph\" (UID: \"75215b19-865c-4ae9-b362-d036a4240375\") " pod="openshift-route-controller-manager/route-controller-manager-7dc8d6dc8d-gcxph" Feb 02 09:10:44 crc kubenswrapper[4764]: I0202 09:10:44.735630 4764 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0c3637e8-e75a-455b-9511-71f59a1a26bf-client-ca\") on node \"crc\" DevicePath \"\"" Feb 02 09:10:44 crc kubenswrapper[4764]: I0202 09:10:44.735642 4764 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0c3637e8-e75a-455b-9511-71f59a1a26bf-config\") on node \"crc\" DevicePath \"\"" Feb 02 09:10:44 crc kubenswrapper[4764]: I0202 09:10:44.735650 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g8qqh\" (UniqueName: \"kubernetes.io/projected/0c3637e8-e75a-455b-9511-71f59a1a26bf-kube-api-access-g8qqh\") on node \"crc\" DevicePath \"\"" Feb 02 09:10:44 crc kubenswrapper[4764]: I0202 09:10:44.735660 4764 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0c3637e8-e75a-455b-9511-71f59a1a26bf-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 02 09:10:44 crc kubenswrapper[4764]: I0202 09:10:44.736619 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/75215b19-865c-4ae9-b362-d036a4240375-config\") pod \"route-controller-manager-7dc8d6dc8d-gcxph\" (UID: \"75215b19-865c-4ae9-b362-d036a4240375\") " pod="openshift-route-controller-manager/route-controller-manager-7dc8d6dc8d-gcxph" Feb 02 09:10:44 crc kubenswrapper[4764]: I0202 09:10:44.742506 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/75215b19-865c-4ae9-b362-d036a4240375-client-ca\") pod \"route-controller-manager-7dc8d6dc8d-gcxph\" (UID: \"75215b19-865c-4ae9-b362-d036a4240375\") " pod="openshift-route-controller-manager/route-controller-manager-7dc8d6dc8d-gcxph" Feb 02 09:10:44 crc kubenswrapper[4764]: I0202 09:10:44.743611 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/75215b19-865c-4ae9-b362-d036a4240375-serving-cert\") pod \"route-controller-manager-7dc8d6dc8d-gcxph\" (UID: \"75215b19-865c-4ae9-b362-d036a4240375\") " pod="openshift-route-controller-manager/route-controller-manager-7dc8d6dc8d-gcxph" Feb 02 09:10:44 crc kubenswrapper[4764]: I0202 09:10:44.762363 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-djx2w\" (UniqueName: \"kubernetes.io/projected/75215b19-865c-4ae9-b362-d036a4240375-kube-api-access-djx2w\") pod \"route-controller-manager-7dc8d6dc8d-gcxph\" (UID: \"75215b19-865c-4ae9-b362-d036a4240375\") " pod="openshift-route-controller-manager/route-controller-manager-7dc8d6dc8d-gcxph" Feb 02 09:10:44 crc kubenswrapper[4764]: I0202 09:10:44.837021 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/00732cf0-76cd-47f2-807e-363710f70a41-serving-cert\") pod \"00732cf0-76cd-47f2-807e-363710f70a41\" (UID: \"00732cf0-76cd-47f2-807e-363710f70a41\") " Feb 02 09:10:44 crc kubenswrapper[4764]: I0202 09:10:44.837076 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hv6s6\" (UniqueName: \"kubernetes.io/projected/00732cf0-76cd-47f2-807e-363710f70a41-kube-api-access-hv6s6\") pod \"00732cf0-76cd-47f2-807e-363710f70a41\" (UID: \"00732cf0-76cd-47f2-807e-363710f70a41\") " Feb 02 09:10:44 crc kubenswrapper[4764]: I0202 09:10:44.837116 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/00732cf0-76cd-47f2-807e-363710f70a41-proxy-ca-bundles\") pod \"00732cf0-76cd-47f2-807e-363710f70a41\" (UID: \"00732cf0-76cd-47f2-807e-363710f70a41\") " Feb 02 09:10:44 crc kubenswrapper[4764]: I0202 09:10:44.837139 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/00732cf0-76cd-47f2-807e-363710f70a41-client-ca\") pod \"00732cf0-76cd-47f2-807e-363710f70a41\" (UID: \"00732cf0-76cd-47f2-807e-363710f70a41\") " Feb 02 09:10:44 crc kubenswrapper[4764]: I0202 09:10:44.837197 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/00732cf0-76cd-47f2-807e-363710f70a41-config\") pod \"00732cf0-76cd-47f2-807e-363710f70a41\" (UID: \"00732cf0-76cd-47f2-807e-363710f70a41\") " Feb 02 09:10:44 crc kubenswrapper[4764]: I0202 09:10:44.837833 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/00732cf0-76cd-47f2-807e-363710f70a41-client-ca" (OuterVolumeSpecName: "client-ca") pod "00732cf0-76cd-47f2-807e-363710f70a41" (UID: "00732cf0-76cd-47f2-807e-363710f70a41"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:10:44 crc kubenswrapper[4764]: I0202 09:10:44.837826 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/00732cf0-76cd-47f2-807e-363710f70a41-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "00732cf0-76cd-47f2-807e-363710f70a41" (UID: "00732cf0-76cd-47f2-807e-363710f70a41"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:10:44 crc kubenswrapper[4764]: I0202 09:10:44.837987 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/00732cf0-76cd-47f2-807e-363710f70a41-config" (OuterVolumeSpecName: "config") pod "00732cf0-76cd-47f2-807e-363710f70a41" (UID: "00732cf0-76cd-47f2-807e-363710f70a41"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:10:44 crc kubenswrapper[4764]: I0202 09:10:44.841092 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/00732cf0-76cd-47f2-807e-363710f70a41-kube-api-access-hv6s6" (OuterVolumeSpecName: "kube-api-access-hv6s6") pod "00732cf0-76cd-47f2-807e-363710f70a41" (UID: "00732cf0-76cd-47f2-807e-363710f70a41"). InnerVolumeSpecName "kube-api-access-hv6s6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:10:44 crc kubenswrapper[4764]: I0202 09:10:44.842650 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/00732cf0-76cd-47f2-807e-363710f70a41-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "00732cf0-76cd-47f2-807e-363710f70a41" (UID: "00732cf0-76cd-47f2-807e-363710f70a41"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:10:44 crc kubenswrapper[4764]: I0202 09:10:44.895489 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7dc8d6dc8d-gcxph" Feb 02 09:10:44 crc kubenswrapper[4764]: I0202 09:10:44.938190 4764 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/00732cf0-76cd-47f2-807e-363710f70a41-config\") on node \"crc\" DevicePath \"\"" Feb 02 09:10:44 crc kubenswrapper[4764]: I0202 09:10:44.938221 4764 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/00732cf0-76cd-47f2-807e-363710f70a41-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 02 09:10:44 crc kubenswrapper[4764]: I0202 09:10:44.938232 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hv6s6\" (UniqueName: \"kubernetes.io/projected/00732cf0-76cd-47f2-807e-363710f70a41-kube-api-access-hv6s6\") on node \"crc\" DevicePath \"\"" Feb 02 09:10:44 crc kubenswrapper[4764]: I0202 09:10:44.938241 4764 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/00732cf0-76cd-47f2-807e-363710f70a41-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 02 09:10:44 crc kubenswrapper[4764]: I0202 09:10:44.938249 4764 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/00732cf0-76cd-47f2-807e-363710f70a41-client-ca\") on node \"crc\" DevicePath \"\"" Feb 02 09:10:45 crc kubenswrapper[4764]: I0202 09:10:45.123177 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7dc8d6dc8d-gcxph"] Feb 02 09:10:45 crc kubenswrapper[4764]: W0202 09:10:45.131063 4764 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod75215b19_865c_4ae9_b362_d036a4240375.slice/crio-185fa3affdc61c1d40f4c6b752e86c3d374267008dc049bc57294d17478c2b78 WatchSource:0}: Error finding container 185fa3affdc61c1d40f4c6b752e86c3d374267008dc049bc57294d17478c2b78: Status 404 returned error can't find the container with id 185fa3affdc61c1d40f4c6b752e86c3d374267008dc049bc57294d17478c2b78 Feb 02 09:10:45 crc kubenswrapper[4764]: I0202 09:10:45.328557 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5489d77c9c-qdmth" Feb 02 09:10:45 crc kubenswrapper[4764]: I0202 09:10:45.328563 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5489d77c9c-qdmth" event={"ID":"00732cf0-76cd-47f2-807e-363710f70a41","Type":"ContainerDied","Data":"153a2288a39c7d11c8f76ff5292e3a934b8e8c726773223f13c598daa4d1d14c"} Feb 02 09:10:45 crc kubenswrapper[4764]: I0202 09:10:45.329495 4764 scope.go:117] "RemoveContainer" containerID="cf20b2c864a21dceb79ed8dc0780a5c05c72afe9786146b9c208cad350a20671" Feb 02 09:10:45 crc kubenswrapper[4764]: I0202 09:10:45.330728 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" event={"ID":"d192f670-9f9d-4539-9641-e4bed73acdd4","Type":"ContainerStarted","Data":"c2244e8f664081c8f021b4db06125b0a2f32a0cc2a1d90e1f92743595267e6b2"} Feb 02 09:10:45 crc kubenswrapper[4764]: I0202 09:10:45.332733 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-85b489c88d-fmzgh" event={"ID":"0c3637e8-e75a-455b-9511-71f59a1a26bf","Type":"ContainerDied","Data":"69e5323eef05db5fcd0f66427b0f1e8a2ec3ee19d72b7b261d99bc565fb52e4d"} Feb 02 09:10:45 crc kubenswrapper[4764]: I0202 09:10:45.332796 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-85b489c88d-fmzgh" Feb 02 09:10:45 crc kubenswrapper[4764]: I0202 09:10:45.350369 4764 scope.go:117] "RemoveContainer" containerID="1a2815b3d227502b432862d5cb65dca636b0b5698855c49be4bc4f6adf3d2948" Feb 02 09:10:45 crc kubenswrapper[4764]: I0202 09:10:45.350582 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7dc8d6dc8d-gcxph" event={"ID":"75215b19-865c-4ae9-b362-d036a4240375","Type":"ContainerStarted","Data":"0d3e05b78f8bbf3d3ccd1cafe513ead8ea140452f70231041d6ad2412558df23"} Feb 02 09:10:45 crc kubenswrapper[4764]: I0202 09:10:45.350621 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7dc8d6dc8d-gcxph" event={"ID":"75215b19-865c-4ae9-b362-d036a4240375","Type":"ContainerStarted","Data":"185fa3affdc61c1d40f4c6b752e86c3d374267008dc049bc57294d17478c2b78"} Feb 02 09:10:45 crc kubenswrapper[4764]: I0202 09:10:45.350896 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-7dc8d6dc8d-gcxph" Feb 02 09:10:45 crc kubenswrapper[4764]: I0202 09:10:45.386035 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-85b489c88d-fmzgh"] Feb 02 09:10:45 crc kubenswrapper[4764]: I0202 09:10:45.389668 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-85b489c88d-fmzgh"] Feb 02 09:10:45 crc kubenswrapper[4764]: I0202 09:10:45.395483 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-5489d77c9c-qdmth"] Feb 02 09:10:45 crc kubenswrapper[4764]: I0202 09:10:45.398648 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-5489d77c9c-qdmth"] Feb 02 09:10:45 crc kubenswrapper[4764]: I0202 09:10:45.414345 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-7dc8d6dc8d-gcxph" podStartSLOduration=2.414326566 podStartE2EDuration="2.414326566s" podCreationTimestamp="2026-02-02 09:10:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:10:45.413385808 +0000 UTC m=+208.347109896" watchObservedRunningTime="2026-02-02 09:10:45.414326566 +0000 UTC m=+208.348050654" Feb 02 09:10:45 crc kubenswrapper[4764]: I0202 09:10:45.746061 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-7dc8d6dc8d-gcxph" Feb 02 09:10:45 crc kubenswrapper[4764]: I0202 09:10:45.832147 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="00732cf0-76cd-47f2-807e-363710f70a41" path="/var/lib/kubelet/pods/00732cf0-76cd-47f2-807e-363710f70a41/volumes" Feb 02 09:10:45 crc kubenswrapper[4764]: I0202 09:10:45.832660 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0c3637e8-e75a-455b-9511-71f59a1a26bf" path="/var/lib/kubelet/pods/0c3637e8-e75a-455b-9511-71f59a1a26bf/volumes" Feb 02 09:10:46 crc kubenswrapper[4764]: I0202 09:10:46.831108 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-ggrpx" Feb 02 09:10:46 crc kubenswrapper[4764]: I0202 09:10:46.832045 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-ggrpx" Feb 02 09:10:46 crc kubenswrapper[4764]: I0202 09:10:46.880694 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-ggrpx" Feb 02 09:10:46 crc kubenswrapper[4764]: I0202 09:10:46.905064 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-7b546fdc45-d822b"] Feb 02 09:10:46 crc kubenswrapper[4764]: E0202 09:10:46.905251 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00732cf0-76cd-47f2-807e-363710f70a41" containerName="controller-manager" Feb 02 09:10:46 crc kubenswrapper[4764]: I0202 09:10:46.905263 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="00732cf0-76cd-47f2-807e-363710f70a41" containerName="controller-manager" Feb 02 09:10:46 crc kubenswrapper[4764]: I0202 09:10:46.905360 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="00732cf0-76cd-47f2-807e-363710f70a41" containerName="controller-manager" Feb 02 09:10:46 crc kubenswrapper[4764]: I0202 09:10:46.905670 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7b546fdc45-d822b" Feb 02 09:10:46 crc kubenswrapper[4764]: I0202 09:10:46.908813 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Feb 02 09:10:46 crc kubenswrapper[4764]: I0202 09:10:46.908826 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Feb 02 09:10:46 crc kubenswrapper[4764]: I0202 09:10:46.908826 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Feb 02 09:10:46 crc kubenswrapper[4764]: I0202 09:10:46.910159 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Feb 02 09:10:46 crc kubenswrapper[4764]: I0202 09:10:46.911837 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Feb 02 09:10:46 crc kubenswrapper[4764]: I0202 09:10:46.912666 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Feb 02 09:10:46 crc kubenswrapper[4764]: I0202 09:10:46.920396 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Feb 02 09:10:46 crc kubenswrapper[4764]: I0202 09:10:46.922553 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7b546fdc45-d822b"] Feb 02 09:10:46 crc kubenswrapper[4764]: I0202 09:10:46.961463 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3e6ea794-02fe-4760-afff-2436274c39a7-client-ca\") pod \"controller-manager-7b546fdc45-d822b\" (UID: \"3e6ea794-02fe-4760-afff-2436274c39a7\") " pod="openshift-controller-manager/controller-manager-7b546fdc45-d822b" Feb 02 09:10:46 crc kubenswrapper[4764]: I0202 09:10:46.961528 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3e6ea794-02fe-4760-afff-2436274c39a7-config\") pod \"controller-manager-7b546fdc45-d822b\" (UID: \"3e6ea794-02fe-4760-afff-2436274c39a7\") " pod="openshift-controller-manager/controller-manager-7b546fdc45-d822b" Feb 02 09:10:46 crc kubenswrapper[4764]: I0202 09:10:46.961575 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3e6ea794-02fe-4760-afff-2436274c39a7-proxy-ca-bundles\") pod \"controller-manager-7b546fdc45-d822b\" (UID: \"3e6ea794-02fe-4760-afff-2436274c39a7\") " pod="openshift-controller-manager/controller-manager-7b546fdc45-d822b" Feb 02 09:10:46 crc kubenswrapper[4764]: I0202 09:10:46.961608 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3e6ea794-02fe-4760-afff-2436274c39a7-serving-cert\") pod \"controller-manager-7b546fdc45-d822b\" (UID: \"3e6ea794-02fe-4760-afff-2436274c39a7\") " pod="openshift-controller-manager/controller-manager-7b546fdc45-d822b" Feb 02 09:10:46 crc kubenswrapper[4764]: I0202 09:10:46.961628 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-klscg\" (UniqueName: \"kubernetes.io/projected/3e6ea794-02fe-4760-afff-2436274c39a7-kube-api-access-klscg\") pod \"controller-manager-7b546fdc45-d822b\" (UID: \"3e6ea794-02fe-4760-afff-2436274c39a7\") " pod="openshift-controller-manager/controller-manager-7b546fdc45-d822b" Feb 02 09:10:47 crc kubenswrapper[4764]: I0202 09:10:47.063171 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3e6ea794-02fe-4760-afff-2436274c39a7-config\") pod \"controller-manager-7b546fdc45-d822b\" (UID: \"3e6ea794-02fe-4760-afff-2436274c39a7\") " pod="openshift-controller-manager/controller-manager-7b546fdc45-d822b" Feb 02 09:10:47 crc kubenswrapper[4764]: I0202 09:10:47.063243 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3e6ea794-02fe-4760-afff-2436274c39a7-proxy-ca-bundles\") pod \"controller-manager-7b546fdc45-d822b\" (UID: \"3e6ea794-02fe-4760-afff-2436274c39a7\") " pod="openshift-controller-manager/controller-manager-7b546fdc45-d822b" Feb 02 09:10:47 crc kubenswrapper[4764]: I0202 09:10:47.063281 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3e6ea794-02fe-4760-afff-2436274c39a7-serving-cert\") pod \"controller-manager-7b546fdc45-d822b\" (UID: \"3e6ea794-02fe-4760-afff-2436274c39a7\") " pod="openshift-controller-manager/controller-manager-7b546fdc45-d822b" Feb 02 09:10:47 crc kubenswrapper[4764]: I0202 09:10:47.063310 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-klscg\" (UniqueName: \"kubernetes.io/projected/3e6ea794-02fe-4760-afff-2436274c39a7-kube-api-access-klscg\") pod \"controller-manager-7b546fdc45-d822b\" (UID: \"3e6ea794-02fe-4760-afff-2436274c39a7\") " pod="openshift-controller-manager/controller-manager-7b546fdc45-d822b" Feb 02 09:10:47 crc kubenswrapper[4764]: I0202 09:10:47.063353 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3e6ea794-02fe-4760-afff-2436274c39a7-client-ca\") pod \"controller-manager-7b546fdc45-d822b\" (UID: \"3e6ea794-02fe-4760-afff-2436274c39a7\") " pod="openshift-controller-manager/controller-manager-7b546fdc45-d822b" Feb 02 09:10:47 crc kubenswrapper[4764]: I0202 09:10:47.064157 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3e6ea794-02fe-4760-afff-2436274c39a7-client-ca\") pod \"controller-manager-7b546fdc45-d822b\" (UID: \"3e6ea794-02fe-4760-afff-2436274c39a7\") " pod="openshift-controller-manager/controller-manager-7b546fdc45-d822b" Feb 02 09:10:47 crc kubenswrapper[4764]: I0202 09:10:47.065053 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3e6ea794-02fe-4760-afff-2436274c39a7-config\") pod \"controller-manager-7b546fdc45-d822b\" (UID: \"3e6ea794-02fe-4760-afff-2436274c39a7\") " pod="openshift-controller-manager/controller-manager-7b546fdc45-d822b" Feb 02 09:10:47 crc kubenswrapper[4764]: I0202 09:10:47.066279 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3e6ea794-02fe-4760-afff-2436274c39a7-proxy-ca-bundles\") pod \"controller-manager-7b546fdc45-d822b\" (UID: \"3e6ea794-02fe-4760-afff-2436274c39a7\") " pod="openshift-controller-manager/controller-manager-7b546fdc45-d822b" Feb 02 09:10:47 crc kubenswrapper[4764]: I0202 09:10:47.070585 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3e6ea794-02fe-4760-afff-2436274c39a7-serving-cert\") pod \"controller-manager-7b546fdc45-d822b\" (UID: \"3e6ea794-02fe-4760-afff-2436274c39a7\") " pod="openshift-controller-manager/controller-manager-7b546fdc45-d822b" Feb 02 09:10:47 crc kubenswrapper[4764]: I0202 09:10:47.081764 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-klscg\" (UniqueName: \"kubernetes.io/projected/3e6ea794-02fe-4760-afff-2436274c39a7-kube-api-access-klscg\") pod \"controller-manager-7b546fdc45-d822b\" (UID: \"3e6ea794-02fe-4760-afff-2436274c39a7\") " pod="openshift-controller-manager/controller-manager-7b546fdc45-d822b" Feb 02 09:10:47 crc kubenswrapper[4764]: I0202 09:10:47.227426 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7b546fdc45-d822b" Feb 02 09:10:47 crc kubenswrapper[4764]: I0202 09:10:47.251157 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-mg5dh" Feb 02 09:10:47 crc kubenswrapper[4764]: I0202 09:10:47.252269 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-mg5dh" Feb 02 09:10:47 crc kubenswrapper[4764]: I0202 09:10:47.284655 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-j25pg" Feb 02 09:10:47 crc kubenswrapper[4764]: I0202 09:10:47.286293 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-j25pg" Feb 02 09:10:47 crc kubenswrapper[4764]: I0202 09:10:47.305608 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-mg5dh" Feb 02 09:10:47 crc kubenswrapper[4764]: I0202 09:10:47.325316 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-j25pg" Feb 02 09:10:47 crc kubenswrapper[4764]: I0202 09:10:47.431302 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-mg5dh" Feb 02 09:10:47 crc kubenswrapper[4764]: I0202 09:10:47.433688 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-ggrpx" Feb 02 09:10:47 crc kubenswrapper[4764]: I0202 09:10:47.511748 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7b546fdc45-d822b"] Feb 02 09:10:48 crc kubenswrapper[4764]: I0202 09:10:48.251533 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-mg5dh"] Feb 02 09:10:48 crc kubenswrapper[4764]: I0202 09:10:48.375855 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7b546fdc45-d822b" event={"ID":"3e6ea794-02fe-4760-afff-2436274c39a7","Type":"ContainerStarted","Data":"45777c08cc18a86afef82563aa90ee3237e2441d1d3f35408f1482130f7ea2e4"} Feb 02 09:10:48 crc kubenswrapper[4764]: I0202 09:10:48.425839 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-j25pg" Feb 02 09:10:49 crc kubenswrapper[4764]: I0202 09:10:49.382600 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7b546fdc45-d822b" event={"ID":"3e6ea794-02fe-4760-afff-2436274c39a7","Type":"ContainerStarted","Data":"2996a9b216544be526b093021c3544bfea9543a2b5b033396db91567adececbf"} Feb 02 09:10:49 crc kubenswrapper[4764]: I0202 09:10:49.383285 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-mg5dh" podUID="35376bf8-2200-49a4-8eb3-f5d7069e73a0" containerName="registry-server" containerID="cri-o://becc45a3d22ecefde54fc5e4cbe0ccddfebe7098e8f1a039b21df4f22d1c7fe6" gracePeriod=2 Feb 02 09:10:49 crc kubenswrapper[4764]: I0202 09:10:49.407795 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-7b546fdc45-d822b" podStartSLOduration=6.407778775 podStartE2EDuration="6.407778775s" podCreationTimestamp="2026-02-02 09:10:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:10:49.400380017 +0000 UTC m=+212.334104135" watchObservedRunningTime="2026-02-02 09:10:49.407778775 +0000 UTC m=+212.341502863" Feb 02 09:10:49 crc kubenswrapper[4764]: I0202 09:10:49.762011 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-ct55p" Feb 02 09:10:49 crc kubenswrapper[4764]: I0202 09:10:49.762553 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-ct55p" Feb 02 09:10:49 crc kubenswrapper[4764]: I0202 09:10:49.802781 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-ct55p" Feb 02 09:10:50 crc kubenswrapper[4764]: I0202 09:10:50.106812 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-w6hd5" Feb 02 09:10:50 crc kubenswrapper[4764]: I0202 09:10:50.106860 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-w6hd5" Feb 02 09:10:50 crc kubenswrapper[4764]: I0202 09:10:50.169174 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-w6hd5" Feb 02 09:10:50 crc kubenswrapper[4764]: I0202 09:10:50.394440 4764 generic.go:334] "Generic (PLEG): container finished" podID="35376bf8-2200-49a4-8eb3-f5d7069e73a0" containerID="becc45a3d22ecefde54fc5e4cbe0ccddfebe7098e8f1a039b21df4f22d1c7fe6" exitCode=0 Feb 02 09:10:50 crc kubenswrapper[4764]: I0202 09:10:50.394603 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mg5dh" event={"ID":"35376bf8-2200-49a4-8eb3-f5d7069e73a0","Type":"ContainerDied","Data":"becc45a3d22ecefde54fc5e4cbe0ccddfebe7098e8f1a039b21df4f22d1c7fe6"} Feb 02 09:10:50 crc kubenswrapper[4764]: I0202 09:10:50.395426 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-7b546fdc45-d822b" Feb 02 09:10:50 crc kubenswrapper[4764]: I0202 09:10:50.402852 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-7b546fdc45-d822b" Feb 02 09:10:50 crc kubenswrapper[4764]: I0202 09:10:50.442904 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-ct55p" Feb 02 09:10:50 crc kubenswrapper[4764]: I0202 09:10:50.459158 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-w6hd5" Feb 02 09:10:50 crc kubenswrapper[4764]: I0202 09:10:50.922946 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mg5dh" Feb 02 09:10:51 crc kubenswrapper[4764]: I0202 09:10:51.045415 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/35376bf8-2200-49a4-8eb3-f5d7069e73a0-utilities\") pod \"35376bf8-2200-49a4-8eb3-f5d7069e73a0\" (UID: \"35376bf8-2200-49a4-8eb3-f5d7069e73a0\") " Feb 02 09:10:51 crc kubenswrapper[4764]: I0202 09:10:51.045480 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/35376bf8-2200-49a4-8eb3-f5d7069e73a0-catalog-content\") pod \"35376bf8-2200-49a4-8eb3-f5d7069e73a0\" (UID: \"35376bf8-2200-49a4-8eb3-f5d7069e73a0\") " Feb 02 09:10:51 crc kubenswrapper[4764]: I0202 09:10:51.045658 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nnpm4\" (UniqueName: \"kubernetes.io/projected/35376bf8-2200-49a4-8eb3-f5d7069e73a0-kube-api-access-nnpm4\") pod \"35376bf8-2200-49a4-8eb3-f5d7069e73a0\" (UID: \"35376bf8-2200-49a4-8eb3-f5d7069e73a0\") " Feb 02 09:10:51 crc kubenswrapper[4764]: I0202 09:10:51.046167 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/35376bf8-2200-49a4-8eb3-f5d7069e73a0-utilities" (OuterVolumeSpecName: "utilities") pod "35376bf8-2200-49a4-8eb3-f5d7069e73a0" (UID: "35376bf8-2200-49a4-8eb3-f5d7069e73a0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 09:10:51 crc kubenswrapper[4764]: I0202 09:10:51.050776 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/35376bf8-2200-49a4-8eb3-f5d7069e73a0-kube-api-access-nnpm4" (OuterVolumeSpecName: "kube-api-access-nnpm4") pod "35376bf8-2200-49a4-8eb3-f5d7069e73a0" (UID: "35376bf8-2200-49a4-8eb3-f5d7069e73a0"). InnerVolumeSpecName "kube-api-access-nnpm4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:10:51 crc kubenswrapper[4764]: I0202 09:10:51.097757 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/35376bf8-2200-49a4-8eb3-f5d7069e73a0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "35376bf8-2200-49a4-8eb3-f5d7069e73a0" (UID: "35376bf8-2200-49a4-8eb3-f5d7069e73a0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 09:10:51 crc kubenswrapper[4764]: I0202 09:10:51.147466 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nnpm4\" (UniqueName: \"kubernetes.io/projected/35376bf8-2200-49a4-8eb3-f5d7069e73a0-kube-api-access-nnpm4\") on node \"crc\" DevicePath \"\"" Feb 02 09:10:51 crc kubenswrapper[4764]: I0202 09:10:51.147501 4764 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/35376bf8-2200-49a4-8eb3-f5d7069e73a0-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 09:10:51 crc kubenswrapper[4764]: I0202 09:10:51.147513 4764 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/35376bf8-2200-49a4-8eb3-f5d7069e73a0-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 09:10:51 crc kubenswrapper[4764]: I0202 09:10:51.404353 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mg5dh" event={"ID":"35376bf8-2200-49a4-8eb3-f5d7069e73a0","Type":"ContainerDied","Data":"38d29e8e0d2662f4c04febfb9d1ed4b29d1e3d1ef4a0510c248b2761f1ca4eed"} Feb 02 09:10:51 crc kubenswrapper[4764]: I0202 09:10:51.404424 4764 scope.go:117] "RemoveContainer" containerID="becc45a3d22ecefde54fc5e4cbe0ccddfebe7098e8f1a039b21df4f22d1c7fe6" Feb 02 09:10:51 crc kubenswrapper[4764]: I0202 09:10:51.404537 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mg5dh" Feb 02 09:10:51 crc kubenswrapper[4764]: I0202 09:10:51.422311 4764 scope.go:117] "RemoveContainer" containerID="dcc388287da6c7fe0be64242fb469e15e627e43062209271becd9a55d5519b52" Feb 02 09:10:51 crc kubenswrapper[4764]: I0202 09:10:51.445217 4764 scope.go:117] "RemoveContainer" containerID="890230c30c2f811cfdc185446a0738979101d42a5f7332d0e7bd8edfb53af670" Feb 02 09:10:51 crc kubenswrapper[4764]: I0202 09:10:51.465612 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-mg5dh"] Feb 02 09:10:51 crc kubenswrapper[4764]: I0202 09:10:51.482148 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-mg5dh"] Feb 02 09:10:51 crc kubenswrapper[4764]: I0202 09:10:51.831968 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="35376bf8-2200-49a4-8eb3-f5d7069e73a0" path="/var/lib/kubelet/pods/35376bf8-2200-49a4-8eb3-f5d7069e73a0/volumes" Feb 02 09:10:53 crc kubenswrapper[4764]: I0202 09:10:53.252393 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-w6hd5"] Feb 02 09:10:53 crc kubenswrapper[4764]: I0202 09:10:53.253080 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-w6hd5" podUID="83d5cb3d-ee7b-44cb-8379-8b52f2c3cadc" containerName="registry-server" containerID="cri-o://c542f29f86ec62e18a334179bc561bb45d7418d55b0033866696690b9c5f1c01" gracePeriod=2 Feb 02 09:10:53 crc kubenswrapper[4764]: I0202 09:10:53.420408 4764 generic.go:334] "Generic (PLEG): container finished" podID="83d5cb3d-ee7b-44cb-8379-8b52f2c3cadc" containerID="c542f29f86ec62e18a334179bc561bb45d7418d55b0033866696690b9c5f1c01" exitCode=0 Feb 02 09:10:53 crc kubenswrapper[4764]: I0202 09:10:53.420481 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w6hd5" event={"ID":"83d5cb3d-ee7b-44cb-8379-8b52f2c3cadc","Type":"ContainerDied","Data":"c542f29f86ec62e18a334179bc561bb45d7418d55b0033866696690b9c5f1c01"} Feb 02 09:10:53 crc kubenswrapper[4764]: I0202 09:10:53.705851 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-w6hd5" Feb 02 09:10:53 crc kubenswrapper[4764]: I0202 09:10:53.887245 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83d5cb3d-ee7b-44cb-8379-8b52f2c3cadc-utilities\") pod \"83d5cb3d-ee7b-44cb-8379-8b52f2c3cadc\" (UID: \"83d5cb3d-ee7b-44cb-8379-8b52f2c3cadc\") " Feb 02 09:10:53 crc kubenswrapper[4764]: I0202 09:10:53.887289 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bnkf7\" (UniqueName: \"kubernetes.io/projected/83d5cb3d-ee7b-44cb-8379-8b52f2c3cadc-kube-api-access-bnkf7\") pod \"83d5cb3d-ee7b-44cb-8379-8b52f2c3cadc\" (UID: \"83d5cb3d-ee7b-44cb-8379-8b52f2c3cadc\") " Feb 02 09:10:53 crc kubenswrapper[4764]: I0202 09:10:53.887386 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83d5cb3d-ee7b-44cb-8379-8b52f2c3cadc-catalog-content\") pod \"83d5cb3d-ee7b-44cb-8379-8b52f2c3cadc\" (UID: \"83d5cb3d-ee7b-44cb-8379-8b52f2c3cadc\") " Feb 02 09:10:53 crc kubenswrapper[4764]: I0202 09:10:53.888723 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/83d5cb3d-ee7b-44cb-8379-8b52f2c3cadc-utilities" (OuterVolumeSpecName: "utilities") pod "83d5cb3d-ee7b-44cb-8379-8b52f2c3cadc" (UID: "83d5cb3d-ee7b-44cb-8379-8b52f2c3cadc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 09:10:53 crc kubenswrapper[4764]: I0202 09:10:53.893107 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/83d5cb3d-ee7b-44cb-8379-8b52f2c3cadc-kube-api-access-bnkf7" (OuterVolumeSpecName: "kube-api-access-bnkf7") pod "83d5cb3d-ee7b-44cb-8379-8b52f2c3cadc" (UID: "83d5cb3d-ee7b-44cb-8379-8b52f2c3cadc"). InnerVolumeSpecName "kube-api-access-bnkf7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:10:53 crc kubenswrapper[4764]: I0202 09:10:53.988760 4764 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83d5cb3d-ee7b-44cb-8379-8b52f2c3cadc-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 09:10:53 crc kubenswrapper[4764]: I0202 09:10:53.988795 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bnkf7\" (UniqueName: \"kubernetes.io/projected/83d5cb3d-ee7b-44cb-8379-8b52f2c3cadc-kube-api-access-bnkf7\") on node \"crc\" DevicePath \"\"" Feb 02 09:10:54 crc kubenswrapper[4764]: I0202 09:10:54.010824 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/83d5cb3d-ee7b-44cb-8379-8b52f2c3cadc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "83d5cb3d-ee7b-44cb-8379-8b52f2c3cadc" (UID: "83d5cb3d-ee7b-44cb-8379-8b52f2c3cadc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 09:10:54 crc kubenswrapper[4764]: I0202 09:10:54.090382 4764 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83d5cb3d-ee7b-44cb-8379-8b52f2c3cadc-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 09:10:54 crc kubenswrapper[4764]: I0202 09:10:54.427838 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w6hd5" event={"ID":"83d5cb3d-ee7b-44cb-8379-8b52f2c3cadc","Type":"ContainerDied","Data":"b16395216d7d99de002dce85a619c0d855fc5ca42a1b516108fcc982859314f4"} Feb 02 09:10:54 crc kubenswrapper[4764]: I0202 09:10:54.427884 4764 scope.go:117] "RemoveContainer" containerID="c542f29f86ec62e18a334179bc561bb45d7418d55b0033866696690b9c5f1c01" Feb 02 09:10:54 crc kubenswrapper[4764]: I0202 09:10:54.427998 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-w6hd5" Feb 02 09:10:54 crc kubenswrapper[4764]: I0202 09:10:54.450689 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-w6hd5"] Feb 02 09:10:54 crc kubenswrapper[4764]: I0202 09:10:54.453717 4764 scope.go:117] "RemoveContainer" containerID="40f47d40a3959e7e99b123bb1f1c5e826a5ecc56dfffc0a9e0191c4b61192845" Feb 02 09:10:54 crc kubenswrapper[4764]: I0202 09:10:54.455752 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-w6hd5"] Feb 02 09:10:54 crc kubenswrapper[4764]: I0202 09:10:54.467643 4764 scope.go:117] "RemoveContainer" containerID="6fa7e7d68b57782853195c602eee842d6e29de75934eb6d982d1ca4fa4dd059c" Feb 02 09:10:55 crc kubenswrapper[4764]: I0202 09:10:55.831124 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="83d5cb3d-ee7b-44cb-8379-8b52f2c3cadc" path="/var/lib/kubelet/pods/83d5cb3d-ee7b-44cb-8379-8b52f2c3cadc/volumes" Feb 02 09:10:59 crc kubenswrapper[4764]: I0202 09:10:59.986697 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-9mdwc"] Feb 02 09:11:03 crc kubenswrapper[4764]: I0202 09:11:03.308294 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-7b546fdc45-d822b"] Feb 02 09:11:03 crc kubenswrapper[4764]: I0202 09:11:03.308966 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-7b546fdc45-d822b" podUID="3e6ea794-02fe-4760-afff-2436274c39a7" containerName="controller-manager" containerID="cri-o://2996a9b216544be526b093021c3544bfea9543a2b5b033396db91567adececbf" gracePeriod=30 Feb 02 09:11:03 crc kubenswrapper[4764]: I0202 09:11:03.420518 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7dc8d6dc8d-gcxph"] Feb 02 09:11:03 crc kubenswrapper[4764]: I0202 09:11:03.420716 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-7dc8d6dc8d-gcxph" podUID="75215b19-865c-4ae9-b362-d036a4240375" containerName="route-controller-manager" containerID="cri-o://0d3e05b78f8bbf3d3ccd1cafe513ead8ea140452f70231041d6ad2412558df23" gracePeriod=30 Feb 02 09:11:03 crc kubenswrapper[4764]: I0202 09:11:03.486048 4764 generic.go:334] "Generic (PLEG): container finished" podID="3e6ea794-02fe-4760-afff-2436274c39a7" containerID="2996a9b216544be526b093021c3544bfea9543a2b5b033396db91567adececbf" exitCode=0 Feb 02 09:11:03 crc kubenswrapper[4764]: I0202 09:11:03.486090 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7b546fdc45-d822b" event={"ID":"3e6ea794-02fe-4760-afff-2436274c39a7","Type":"ContainerDied","Data":"2996a9b216544be526b093021c3544bfea9543a2b5b033396db91567adececbf"} Feb 02 09:11:03 crc kubenswrapper[4764]: I0202 09:11:03.957378 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7b546fdc45-d822b" Feb 02 09:11:03 crc kubenswrapper[4764]: I0202 09:11:03.963512 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7dc8d6dc8d-gcxph" Feb 02 09:11:04 crc kubenswrapper[4764]: I0202 09:11:04.123740 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-klscg\" (UniqueName: \"kubernetes.io/projected/3e6ea794-02fe-4760-afff-2436274c39a7-kube-api-access-klscg\") pod \"3e6ea794-02fe-4760-afff-2436274c39a7\" (UID: \"3e6ea794-02fe-4760-afff-2436274c39a7\") " Feb 02 09:11:04 crc kubenswrapper[4764]: I0202 09:11:04.123802 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3e6ea794-02fe-4760-afff-2436274c39a7-proxy-ca-bundles\") pod \"3e6ea794-02fe-4760-afff-2436274c39a7\" (UID: \"3e6ea794-02fe-4760-afff-2436274c39a7\") " Feb 02 09:11:04 crc kubenswrapper[4764]: I0202 09:11:04.123838 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/75215b19-865c-4ae9-b362-d036a4240375-serving-cert\") pod \"75215b19-865c-4ae9-b362-d036a4240375\" (UID: \"75215b19-865c-4ae9-b362-d036a4240375\") " Feb 02 09:11:04 crc kubenswrapper[4764]: I0202 09:11:04.123867 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3e6ea794-02fe-4760-afff-2436274c39a7-client-ca\") pod \"3e6ea794-02fe-4760-afff-2436274c39a7\" (UID: \"3e6ea794-02fe-4760-afff-2436274c39a7\") " Feb 02 09:11:04 crc kubenswrapper[4764]: I0202 09:11:04.123888 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/75215b19-865c-4ae9-b362-d036a4240375-client-ca\") pod \"75215b19-865c-4ae9-b362-d036a4240375\" (UID: \"75215b19-865c-4ae9-b362-d036a4240375\") " Feb 02 09:11:04 crc kubenswrapper[4764]: I0202 09:11:04.123925 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3e6ea794-02fe-4760-afff-2436274c39a7-serving-cert\") pod \"3e6ea794-02fe-4760-afff-2436274c39a7\" (UID: \"3e6ea794-02fe-4760-afff-2436274c39a7\") " Feb 02 09:11:04 crc kubenswrapper[4764]: I0202 09:11:04.123968 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3e6ea794-02fe-4760-afff-2436274c39a7-config\") pod \"3e6ea794-02fe-4760-afff-2436274c39a7\" (UID: \"3e6ea794-02fe-4760-afff-2436274c39a7\") " Feb 02 09:11:04 crc kubenswrapper[4764]: I0202 09:11:04.123989 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-djx2w\" (UniqueName: \"kubernetes.io/projected/75215b19-865c-4ae9-b362-d036a4240375-kube-api-access-djx2w\") pod \"75215b19-865c-4ae9-b362-d036a4240375\" (UID: \"75215b19-865c-4ae9-b362-d036a4240375\") " Feb 02 09:11:04 crc kubenswrapper[4764]: I0202 09:11:04.124011 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/75215b19-865c-4ae9-b362-d036a4240375-config\") pod \"75215b19-865c-4ae9-b362-d036a4240375\" (UID: \"75215b19-865c-4ae9-b362-d036a4240375\") " Feb 02 09:11:04 crc kubenswrapper[4764]: I0202 09:11:04.124562 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3e6ea794-02fe-4760-afff-2436274c39a7-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "3e6ea794-02fe-4760-afff-2436274c39a7" (UID: "3e6ea794-02fe-4760-afff-2436274c39a7"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:11:04 crc kubenswrapper[4764]: I0202 09:11:04.124831 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/75215b19-865c-4ae9-b362-d036a4240375-config" (OuterVolumeSpecName: "config") pod "75215b19-865c-4ae9-b362-d036a4240375" (UID: "75215b19-865c-4ae9-b362-d036a4240375"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:11:04 crc kubenswrapper[4764]: I0202 09:11:04.124846 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/75215b19-865c-4ae9-b362-d036a4240375-client-ca" (OuterVolumeSpecName: "client-ca") pod "75215b19-865c-4ae9-b362-d036a4240375" (UID: "75215b19-865c-4ae9-b362-d036a4240375"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:11:04 crc kubenswrapper[4764]: I0202 09:11:04.125526 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3e6ea794-02fe-4760-afff-2436274c39a7-config" (OuterVolumeSpecName: "config") pod "3e6ea794-02fe-4760-afff-2436274c39a7" (UID: "3e6ea794-02fe-4760-afff-2436274c39a7"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:11:04 crc kubenswrapper[4764]: I0202 09:11:04.125619 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3e6ea794-02fe-4760-afff-2436274c39a7-client-ca" (OuterVolumeSpecName: "client-ca") pod "3e6ea794-02fe-4760-afff-2436274c39a7" (UID: "3e6ea794-02fe-4760-afff-2436274c39a7"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:11:04 crc kubenswrapper[4764]: I0202 09:11:04.129002 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3e6ea794-02fe-4760-afff-2436274c39a7-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "3e6ea794-02fe-4760-afff-2436274c39a7" (UID: "3e6ea794-02fe-4760-afff-2436274c39a7"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:11:04 crc kubenswrapper[4764]: I0202 09:11:04.129823 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/75215b19-865c-4ae9-b362-d036a4240375-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "75215b19-865c-4ae9-b362-d036a4240375" (UID: "75215b19-865c-4ae9-b362-d036a4240375"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:11:04 crc kubenswrapper[4764]: I0202 09:11:04.130615 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3e6ea794-02fe-4760-afff-2436274c39a7-kube-api-access-klscg" (OuterVolumeSpecName: "kube-api-access-klscg") pod "3e6ea794-02fe-4760-afff-2436274c39a7" (UID: "3e6ea794-02fe-4760-afff-2436274c39a7"). InnerVolumeSpecName "kube-api-access-klscg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:11:04 crc kubenswrapper[4764]: I0202 09:11:04.131812 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/75215b19-865c-4ae9-b362-d036a4240375-kube-api-access-djx2w" (OuterVolumeSpecName: "kube-api-access-djx2w") pod "75215b19-865c-4ae9-b362-d036a4240375" (UID: "75215b19-865c-4ae9-b362-d036a4240375"). InnerVolumeSpecName "kube-api-access-djx2w". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:11:04 crc kubenswrapper[4764]: I0202 09:11:04.225749 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-klscg\" (UniqueName: \"kubernetes.io/projected/3e6ea794-02fe-4760-afff-2436274c39a7-kube-api-access-klscg\") on node \"crc\" DevicePath \"\"" Feb 02 09:11:04 crc kubenswrapper[4764]: I0202 09:11:04.225803 4764 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3e6ea794-02fe-4760-afff-2436274c39a7-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 02 09:11:04 crc kubenswrapper[4764]: I0202 09:11:04.225823 4764 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/75215b19-865c-4ae9-b362-d036a4240375-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 02 09:11:04 crc kubenswrapper[4764]: I0202 09:11:04.225841 4764 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3e6ea794-02fe-4760-afff-2436274c39a7-client-ca\") on node \"crc\" DevicePath \"\"" Feb 02 09:11:04 crc kubenswrapper[4764]: I0202 09:11:04.225857 4764 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/75215b19-865c-4ae9-b362-d036a4240375-client-ca\") on node \"crc\" DevicePath \"\"" Feb 02 09:11:04 crc kubenswrapper[4764]: I0202 09:11:04.225874 4764 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3e6ea794-02fe-4760-afff-2436274c39a7-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 02 09:11:04 crc kubenswrapper[4764]: I0202 09:11:04.225890 4764 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3e6ea794-02fe-4760-afff-2436274c39a7-config\") on node \"crc\" DevicePath \"\"" Feb 02 09:11:04 crc kubenswrapper[4764]: I0202 09:11:04.225908 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-djx2w\" (UniqueName: \"kubernetes.io/projected/75215b19-865c-4ae9-b362-d036a4240375-kube-api-access-djx2w\") on node \"crc\" DevicePath \"\"" Feb 02 09:11:04 crc kubenswrapper[4764]: I0202 09:11:04.225924 4764 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/75215b19-865c-4ae9-b362-d036a4240375-config\") on node \"crc\" DevicePath \"\"" Feb 02 09:11:04 crc kubenswrapper[4764]: I0202 09:11:04.494011 4764 generic.go:334] "Generic (PLEG): container finished" podID="75215b19-865c-4ae9-b362-d036a4240375" containerID="0d3e05b78f8bbf3d3ccd1cafe513ead8ea140452f70231041d6ad2412558df23" exitCode=0 Feb 02 09:11:04 crc kubenswrapper[4764]: I0202 09:11:04.494111 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7dc8d6dc8d-gcxph" Feb 02 09:11:04 crc kubenswrapper[4764]: I0202 09:11:04.494308 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7dc8d6dc8d-gcxph" event={"ID":"75215b19-865c-4ae9-b362-d036a4240375","Type":"ContainerDied","Data":"0d3e05b78f8bbf3d3ccd1cafe513ead8ea140452f70231041d6ad2412558df23"} Feb 02 09:11:04 crc kubenswrapper[4764]: I0202 09:11:04.494344 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7dc8d6dc8d-gcxph" event={"ID":"75215b19-865c-4ae9-b362-d036a4240375","Type":"ContainerDied","Data":"185fa3affdc61c1d40f4c6b752e86c3d374267008dc049bc57294d17478c2b78"} Feb 02 09:11:04 crc kubenswrapper[4764]: I0202 09:11:04.494362 4764 scope.go:117] "RemoveContainer" containerID="0d3e05b78f8bbf3d3ccd1cafe513ead8ea140452f70231041d6ad2412558df23" Feb 02 09:11:04 crc kubenswrapper[4764]: I0202 09:11:04.496637 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7b546fdc45-d822b" event={"ID":"3e6ea794-02fe-4760-afff-2436274c39a7","Type":"ContainerDied","Data":"45777c08cc18a86afef82563aa90ee3237e2441d1d3f35408f1482130f7ea2e4"} Feb 02 09:11:04 crc kubenswrapper[4764]: I0202 09:11:04.496906 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7b546fdc45-d822b" Feb 02 09:11:04 crc kubenswrapper[4764]: I0202 09:11:04.518735 4764 scope.go:117] "RemoveContainer" containerID="0d3e05b78f8bbf3d3ccd1cafe513ead8ea140452f70231041d6ad2412558df23" Feb 02 09:11:04 crc kubenswrapper[4764]: E0202 09:11:04.519598 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0d3e05b78f8bbf3d3ccd1cafe513ead8ea140452f70231041d6ad2412558df23\": container with ID starting with 0d3e05b78f8bbf3d3ccd1cafe513ead8ea140452f70231041d6ad2412558df23 not found: ID does not exist" containerID="0d3e05b78f8bbf3d3ccd1cafe513ead8ea140452f70231041d6ad2412558df23" Feb 02 09:11:04 crc kubenswrapper[4764]: I0202 09:11:04.519629 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d3e05b78f8bbf3d3ccd1cafe513ead8ea140452f70231041d6ad2412558df23"} err="failed to get container status \"0d3e05b78f8bbf3d3ccd1cafe513ead8ea140452f70231041d6ad2412558df23\": rpc error: code = NotFound desc = could not find container \"0d3e05b78f8bbf3d3ccd1cafe513ead8ea140452f70231041d6ad2412558df23\": container with ID starting with 0d3e05b78f8bbf3d3ccd1cafe513ead8ea140452f70231041d6ad2412558df23 not found: ID does not exist" Feb 02 09:11:04 crc kubenswrapper[4764]: I0202 09:11:04.519649 4764 scope.go:117] "RemoveContainer" containerID="2996a9b216544be526b093021c3544bfea9543a2b5b033396db91567adececbf" Feb 02 09:11:04 crc kubenswrapper[4764]: I0202 09:11:04.540399 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7dc8d6dc8d-gcxph"] Feb 02 09:11:04 crc kubenswrapper[4764]: I0202 09:11:04.543869 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7dc8d6dc8d-gcxph"] Feb 02 09:11:04 crc kubenswrapper[4764]: I0202 09:11:04.558354 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-7b546fdc45-d822b"] Feb 02 09:11:04 crc kubenswrapper[4764]: I0202 09:11:04.565749 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-7b546fdc45-d822b"] Feb 02 09:11:04 crc kubenswrapper[4764]: I0202 09:11:04.921873 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-789c85bd8b-5xbqt"] Feb 02 09:11:04 crc kubenswrapper[4764]: E0202 09:11:04.922266 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83d5cb3d-ee7b-44cb-8379-8b52f2c3cadc" containerName="extract-content" Feb 02 09:11:04 crc kubenswrapper[4764]: I0202 09:11:04.922286 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="83d5cb3d-ee7b-44cb-8379-8b52f2c3cadc" containerName="extract-content" Feb 02 09:11:04 crc kubenswrapper[4764]: E0202 09:11:04.922304 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35376bf8-2200-49a4-8eb3-f5d7069e73a0" containerName="registry-server" Feb 02 09:11:04 crc kubenswrapper[4764]: I0202 09:11:04.922318 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="35376bf8-2200-49a4-8eb3-f5d7069e73a0" containerName="registry-server" Feb 02 09:11:04 crc kubenswrapper[4764]: E0202 09:11:04.922337 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83d5cb3d-ee7b-44cb-8379-8b52f2c3cadc" containerName="extract-utilities" Feb 02 09:11:04 crc kubenswrapper[4764]: I0202 09:11:04.922352 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="83d5cb3d-ee7b-44cb-8379-8b52f2c3cadc" containerName="extract-utilities" Feb 02 09:11:04 crc kubenswrapper[4764]: E0202 09:11:04.922374 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35376bf8-2200-49a4-8eb3-f5d7069e73a0" containerName="extract-utilities" Feb 02 09:11:04 crc kubenswrapper[4764]: I0202 09:11:04.922386 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="35376bf8-2200-49a4-8eb3-f5d7069e73a0" containerName="extract-utilities" Feb 02 09:11:04 crc kubenswrapper[4764]: E0202 09:11:04.922406 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35376bf8-2200-49a4-8eb3-f5d7069e73a0" containerName="extract-content" Feb 02 09:11:04 crc kubenswrapper[4764]: I0202 09:11:04.922418 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="35376bf8-2200-49a4-8eb3-f5d7069e73a0" containerName="extract-content" Feb 02 09:11:04 crc kubenswrapper[4764]: E0202 09:11:04.922434 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75215b19-865c-4ae9-b362-d036a4240375" containerName="route-controller-manager" Feb 02 09:11:04 crc kubenswrapper[4764]: I0202 09:11:04.922446 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="75215b19-865c-4ae9-b362-d036a4240375" containerName="route-controller-manager" Feb 02 09:11:04 crc kubenswrapper[4764]: E0202 09:11:04.922460 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e6ea794-02fe-4760-afff-2436274c39a7" containerName="controller-manager" Feb 02 09:11:04 crc kubenswrapper[4764]: I0202 09:11:04.922471 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e6ea794-02fe-4760-afff-2436274c39a7" containerName="controller-manager" Feb 02 09:11:04 crc kubenswrapper[4764]: E0202 09:11:04.922492 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83d5cb3d-ee7b-44cb-8379-8b52f2c3cadc" containerName="registry-server" Feb 02 09:11:04 crc kubenswrapper[4764]: I0202 09:11:04.922505 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="83d5cb3d-ee7b-44cb-8379-8b52f2c3cadc" containerName="registry-server" Feb 02 09:11:04 crc kubenswrapper[4764]: I0202 09:11:04.922723 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="75215b19-865c-4ae9-b362-d036a4240375" containerName="route-controller-manager" Feb 02 09:11:04 crc kubenswrapper[4764]: I0202 09:11:04.922739 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e6ea794-02fe-4760-afff-2436274c39a7" containerName="controller-manager" Feb 02 09:11:04 crc kubenswrapper[4764]: I0202 09:11:04.922759 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="35376bf8-2200-49a4-8eb3-f5d7069e73a0" containerName="registry-server" Feb 02 09:11:04 crc kubenswrapper[4764]: I0202 09:11:04.922779 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="83d5cb3d-ee7b-44cb-8379-8b52f2c3cadc" containerName="registry-server" Feb 02 09:11:04 crc kubenswrapper[4764]: I0202 09:11:04.923353 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-789c85bd8b-5xbqt" Feb 02 09:11:04 crc kubenswrapper[4764]: I0202 09:11:04.924816 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7bff5df74-nwqtk"] Feb 02 09:11:04 crc kubenswrapper[4764]: I0202 09:11:04.925649 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7bff5df74-nwqtk" Feb 02 09:11:04 crc kubenswrapper[4764]: I0202 09:11:04.929488 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Feb 02 09:11:04 crc kubenswrapper[4764]: I0202 09:11:04.930783 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Feb 02 09:11:04 crc kubenswrapper[4764]: I0202 09:11:04.932093 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Feb 02 09:11:04 crc kubenswrapper[4764]: I0202 09:11:04.932574 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Feb 02 09:11:04 crc kubenswrapper[4764]: I0202 09:11:04.933001 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Feb 02 09:11:04 crc kubenswrapper[4764]: I0202 09:11:04.933176 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-789c85bd8b-5xbqt"] Feb 02 09:11:04 crc kubenswrapper[4764]: I0202 09:11:04.933281 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Feb 02 09:11:04 crc kubenswrapper[4764]: I0202 09:11:04.935060 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7bff5df74-nwqtk"] Feb 02 09:11:04 crc kubenswrapper[4764]: I0202 09:11:04.940150 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Feb 02 09:11:04 crc kubenswrapper[4764]: I0202 09:11:04.940744 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Feb 02 09:11:04 crc kubenswrapper[4764]: I0202 09:11:04.941218 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Feb 02 09:11:04 crc kubenswrapper[4764]: I0202 09:11:04.958979 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Feb 02 09:11:04 crc kubenswrapper[4764]: I0202 09:11:04.959555 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Feb 02 09:11:04 crc kubenswrapper[4764]: I0202 09:11:04.960532 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Feb 02 09:11:04 crc kubenswrapper[4764]: I0202 09:11:04.963899 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Feb 02 09:11:05 crc kubenswrapper[4764]: I0202 09:11:05.037195 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9gjk9\" (UniqueName: \"kubernetes.io/projected/40917c3d-03a6-4a15-88e5-d42ab4b57407-kube-api-access-9gjk9\") pod \"controller-manager-789c85bd8b-5xbqt\" (UID: \"40917c3d-03a6-4a15-88e5-d42ab4b57407\") " pod="openshift-controller-manager/controller-manager-789c85bd8b-5xbqt" Feb 02 09:11:05 crc kubenswrapper[4764]: I0202 09:11:05.037271 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/67b4c8df-c327-4d8e-bc5c-19fd7cb80444-serving-cert\") pod \"route-controller-manager-7bff5df74-nwqtk\" (UID: \"67b4c8df-c327-4d8e-bc5c-19fd7cb80444\") " pod="openshift-route-controller-manager/route-controller-manager-7bff5df74-nwqtk" Feb 02 09:11:05 crc kubenswrapper[4764]: I0202 09:11:05.037298 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/40917c3d-03a6-4a15-88e5-d42ab4b57407-client-ca\") pod \"controller-manager-789c85bd8b-5xbqt\" (UID: \"40917c3d-03a6-4a15-88e5-d42ab4b57407\") " pod="openshift-controller-manager/controller-manager-789c85bd8b-5xbqt" Feb 02 09:11:05 crc kubenswrapper[4764]: I0202 09:11:05.037315 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/40917c3d-03a6-4a15-88e5-d42ab4b57407-serving-cert\") pod \"controller-manager-789c85bd8b-5xbqt\" (UID: \"40917c3d-03a6-4a15-88e5-d42ab4b57407\") " pod="openshift-controller-manager/controller-manager-789c85bd8b-5xbqt" Feb 02 09:11:05 crc kubenswrapper[4764]: I0202 09:11:05.037338 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/40917c3d-03a6-4a15-88e5-d42ab4b57407-config\") pod \"controller-manager-789c85bd8b-5xbqt\" (UID: \"40917c3d-03a6-4a15-88e5-d42ab4b57407\") " pod="openshift-controller-manager/controller-manager-789c85bd8b-5xbqt" Feb 02 09:11:05 crc kubenswrapper[4764]: I0202 09:11:05.037354 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/67b4c8df-c327-4d8e-bc5c-19fd7cb80444-client-ca\") pod \"route-controller-manager-7bff5df74-nwqtk\" (UID: \"67b4c8df-c327-4d8e-bc5c-19fd7cb80444\") " pod="openshift-route-controller-manager/route-controller-manager-7bff5df74-nwqtk" Feb 02 09:11:05 crc kubenswrapper[4764]: I0202 09:11:05.037369 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67b4c8df-c327-4d8e-bc5c-19fd7cb80444-config\") pod \"route-controller-manager-7bff5df74-nwqtk\" (UID: \"67b4c8df-c327-4d8e-bc5c-19fd7cb80444\") " pod="openshift-route-controller-manager/route-controller-manager-7bff5df74-nwqtk" Feb 02 09:11:05 crc kubenswrapper[4764]: I0202 09:11:05.037414 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/40917c3d-03a6-4a15-88e5-d42ab4b57407-proxy-ca-bundles\") pod \"controller-manager-789c85bd8b-5xbqt\" (UID: \"40917c3d-03a6-4a15-88e5-d42ab4b57407\") " pod="openshift-controller-manager/controller-manager-789c85bd8b-5xbqt" Feb 02 09:11:05 crc kubenswrapper[4764]: I0202 09:11:05.037432 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4mr4g\" (UniqueName: \"kubernetes.io/projected/67b4c8df-c327-4d8e-bc5c-19fd7cb80444-kube-api-access-4mr4g\") pod \"route-controller-manager-7bff5df74-nwqtk\" (UID: \"67b4c8df-c327-4d8e-bc5c-19fd7cb80444\") " pod="openshift-route-controller-manager/route-controller-manager-7bff5df74-nwqtk" Feb 02 09:11:05 crc kubenswrapper[4764]: I0202 09:11:05.138385 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/40917c3d-03a6-4a15-88e5-d42ab4b57407-proxy-ca-bundles\") pod \"controller-manager-789c85bd8b-5xbqt\" (UID: \"40917c3d-03a6-4a15-88e5-d42ab4b57407\") " pod="openshift-controller-manager/controller-manager-789c85bd8b-5xbqt" Feb 02 09:11:05 crc kubenswrapper[4764]: I0202 09:11:05.138617 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4mr4g\" (UniqueName: \"kubernetes.io/projected/67b4c8df-c327-4d8e-bc5c-19fd7cb80444-kube-api-access-4mr4g\") pod \"route-controller-manager-7bff5df74-nwqtk\" (UID: \"67b4c8df-c327-4d8e-bc5c-19fd7cb80444\") " pod="openshift-route-controller-manager/route-controller-manager-7bff5df74-nwqtk" Feb 02 09:11:05 crc kubenswrapper[4764]: I0202 09:11:05.138714 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9gjk9\" (UniqueName: \"kubernetes.io/projected/40917c3d-03a6-4a15-88e5-d42ab4b57407-kube-api-access-9gjk9\") pod \"controller-manager-789c85bd8b-5xbqt\" (UID: \"40917c3d-03a6-4a15-88e5-d42ab4b57407\") " pod="openshift-controller-manager/controller-manager-789c85bd8b-5xbqt" Feb 02 09:11:05 crc kubenswrapper[4764]: I0202 09:11:05.138818 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/67b4c8df-c327-4d8e-bc5c-19fd7cb80444-serving-cert\") pod \"route-controller-manager-7bff5df74-nwqtk\" (UID: \"67b4c8df-c327-4d8e-bc5c-19fd7cb80444\") " pod="openshift-route-controller-manager/route-controller-manager-7bff5df74-nwqtk" Feb 02 09:11:05 crc kubenswrapper[4764]: I0202 09:11:05.138891 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/40917c3d-03a6-4a15-88e5-d42ab4b57407-client-ca\") pod \"controller-manager-789c85bd8b-5xbqt\" (UID: \"40917c3d-03a6-4a15-88e5-d42ab4b57407\") " pod="openshift-controller-manager/controller-manager-789c85bd8b-5xbqt" Feb 02 09:11:05 crc kubenswrapper[4764]: I0202 09:11:05.138990 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/40917c3d-03a6-4a15-88e5-d42ab4b57407-serving-cert\") pod \"controller-manager-789c85bd8b-5xbqt\" (UID: \"40917c3d-03a6-4a15-88e5-d42ab4b57407\") " pod="openshift-controller-manager/controller-manager-789c85bd8b-5xbqt" Feb 02 09:11:05 crc kubenswrapper[4764]: I0202 09:11:05.139069 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/40917c3d-03a6-4a15-88e5-d42ab4b57407-config\") pod \"controller-manager-789c85bd8b-5xbqt\" (UID: \"40917c3d-03a6-4a15-88e5-d42ab4b57407\") " pod="openshift-controller-manager/controller-manager-789c85bd8b-5xbqt" Feb 02 09:11:05 crc kubenswrapper[4764]: I0202 09:11:05.139140 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/67b4c8df-c327-4d8e-bc5c-19fd7cb80444-client-ca\") pod \"route-controller-manager-7bff5df74-nwqtk\" (UID: \"67b4c8df-c327-4d8e-bc5c-19fd7cb80444\") " pod="openshift-route-controller-manager/route-controller-manager-7bff5df74-nwqtk" Feb 02 09:11:05 crc kubenswrapper[4764]: I0202 09:11:05.139217 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67b4c8df-c327-4d8e-bc5c-19fd7cb80444-config\") pod \"route-controller-manager-7bff5df74-nwqtk\" (UID: \"67b4c8df-c327-4d8e-bc5c-19fd7cb80444\") " pod="openshift-route-controller-manager/route-controller-manager-7bff5df74-nwqtk" Feb 02 09:11:05 crc kubenswrapper[4764]: I0202 09:11:05.140296 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67b4c8df-c327-4d8e-bc5c-19fd7cb80444-config\") pod \"route-controller-manager-7bff5df74-nwqtk\" (UID: \"67b4c8df-c327-4d8e-bc5c-19fd7cb80444\") " pod="openshift-route-controller-manager/route-controller-manager-7bff5df74-nwqtk" Feb 02 09:11:05 crc kubenswrapper[4764]: I0202 09:11:05.140564 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/40917c3d-03a6-4a15-88e5-d42ab4b57407-client-ca\") pod \"controller-manager-789c85bd8b-5xbqt\" (UID: \"40917c3d-03a6-4a15-88e5-d42ab4b57407\") " pod="openshift-controller-manager/controller-manager-789c85bd8b-5xbqt" Feb 02 09:11:05 crc kubenswrapper[4764]: I0202 09:11:05.141384 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/40917c3d-03a6-4a15-88e5-d42ab4b57407-config\") pod \"controller-manager-789c85bd8b-5xbqt\" (UID: \"40917c3d-03a6-4a15-88e5-d42ab4b57407\") " pod="openshift-controller-manager/controller-manager-789c85bd8b-5xbqt" Feb 02 09:11:05 crc kubenswrapper[4764]: I0202 09:11:05.142039 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/67b4c8df-c327-4d8e-bc5c-19fd7cb80444-client-ca\") pod \"route-controller-manager-7bff5df74-nwqtk\" (UID: \"67b4c8df-c327-4d8e-bc5c-19fd7cb80444\") " pod="openshift-route-controller-manager/route-controller-manager-7bff5df74-nwqtk" Feb 02 09:11:05 crc kubenswrapper[4764]: I0202 09:11:05.142493 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/40917c3d-03a6-4a15-88e5-d42ab4b57407-proxy-ca-bundles\") pod \"controller-manager-789c85bd8b-5xbqt\" (UID: \"40917c3d-03a6-4a15-88e5-d42ab4b57407\") " pod="openshift-controller-manager/controller-manager-789c85bd8b-5xbqt" Feb 02 09:11:05 crc kubenswrapper[4764]: I0202 09:11:05.146175 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/40917c3d-03a6-4a15-88e5-d42ab4b57407-serving-cert\") pod \"controller-manager-789c85bd8b-5xbqt\" (UID: \"40917c3d-03a6-4a15-88e5-d42ab4b57407\") " pod="openshift-controller-manager/controller-manager-789c85bd8b-5xbqt" Feb 02 09:11:05 crc kubenswrapper[4764]: I0202 09:11:05.146555 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/67b4c8df-c327-4d8e-bc5c-19fd7cb80444-serving-cert\") pod \"route-controller-manager-7bff5df74-nwqtk\" (UID: \"67b4c8df-c327-4d8e-bc5c-19fd7cb80444\") " pod="openshift-route-controller-manager/route-controller-manager-7bff5df74-nwqtk" Feb 02 09:11:05 crc kubenswrapper[4764]: I0202 09:11:05.168900 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4mr4g\" (UniqueName: \"kubernetes.io/projected/67b4c8df-c327-4d8e-bc5c-19fd7cb80444-kube-api-access-4mr4g\") pod \"route-controller-manager-7bff5df74-nwqtk\" (UID: \"67b4c8df-c327-4d8e-bc5c-19fd7cb80444\") " pod="openshift-route-controller-manager/route-controller-manager-7bff5df74-nwqtk" Feb 02 09:11:05 crc kubenswrapper[4764]: I0202 09:11:05.171030 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9gjk9\" (UniqueName: \"kubernetes.io/projected/40917c3d-03a6-4a15-88e5-d42ab4b57407-kube-api-access-9gjk9\") pod \"controller-manager-789c85bd8b-5xbqt\" (UID: \"40917c3d-03a6-4a15-88e5-d42ab4b57407\") " pod="openshift-controller-manager/controller-manager-789c85bd8b-5xbqt" Feb 02 09:11:05 crc kubenswrapper[4764]: I0202 09:11:05.266447 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-789c85bd8b-5xbqt" Feb 02 09:11:05 crc kubenswrapper[4764]: I0202 09:11:05.277886 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7bff5df74-nwqtk" Feb 02 09:11:05 crc kubenswrapper[4764]: I0202 09:11:05.692703 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7bff5df74-nwqtk"] Feb 02 09:11:05 crc kubenswrapper[4764]: I0202 09:11:05.759513 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-789c85bd8b-5xbqt"] Feb 02 09:11:05 crc kubenswrapper[4764]: W0202 09:11:05.769589 4764 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod40917c3d_03a6_4a15_88e5_d42ab4b57407.slice/crio-8c0df35e9a1569e4661c4892a2db3fd9b82fdc0a88ef76ff093a31c8ed56bdf9 WatchSource:0}: Error finding container 8c0df35e9a1569e4661c4892a2db3fd9b82fdc0a88ef76ff093a31c8ed56bdf9: Status 404 returned error can't find the container with id 8c0df35e9a1569e4661c4892a2db3fd9b82fdc0a88ef76ff093a31c8ed56bdf9 Feb 02 09:11:05 crc kubenswrapper[4764]: I0202 09:11:05.832836 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3e6ea794-02fe-4760-afff-2436274c39a7" path="/var/lib/kubelet/pods/3e6ea794-02fe-4760-afff-2436274c39a7/volumes" Feb 02 09:11:05 crc kubenswrapper[4764]: I0202 09:11:05.833579 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="75215b19-865c-4ae9-b362-d036a4240375" path="/var/lib/kubelet/pods/75215b19-865c-4ae9-b362-d036a4240375/volumes" Feb 02 09:11:06 crc kubenswrapper[4764]: I0202 09:11:06.512156 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7bff5df74-nwqtk" event={"ID":"67b4c8df-c327-4d8e-bc5c-19fd7cb80444","Type":"ContainerStarted","Data":"1f4f4169fad02ee28073e6f31d456735d3228ad1c8b35232db88ab5415f9039c"} Feb 02 09:11:06 crc kubenswrapper[4764]: I0202 09:11:06.512215 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7bff5df74-nwqtk" event={"ID":"67b4c8df-c327-4d8e-bc5c-19fd7cb80444","Type":"ContainerStarted","Data":"a56adc20eaef7aa45c421ed07f9a0b72f7f85fb688cb7f52911ffa3959fa637a"} Feb 02 09:11:06 crc kubenswrapper[4764]: I0202 09:11:06.513880 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-7bff5df74-nwqtk" Feb 02 09:11:06 crc kubenswrapper[4764]: I0202 09:11:06.515241 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-789c85bd8b-5xbqt" event={"ID":"40917c3d-03a6-4a15-88e5-d42ab4b57407","Type":"ContainerStarted","Data":"b58c50380aaabe4e5a3973332f29b897f75b47add71868b8a5b22f2255065437"} Feb 02 09:11:06 crc kubenswrapper[4764]: I0202 09:11:06.515266 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-789c85bd8b-5xbqt" event={"ID":"40917c3d-03a6-4a15-88e5-d42ab4b57407","Type":"ContainerStarted","Data":"8c0df35e9a1569e4661c4892a2db3fd9b82fdc0a88ef76ff093a31c8ed56bdf9"} Feb 02 09:11:06 crc kubenswrapper[4764]: I0202 09:11:06.515646 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-789c85bd8b-5xbqt" Feb 02 09:11:06 crc kubenswrapper[4764]: I0202 09:11:06.518921 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-789c85bd8b-5xbqt" Feb 02 09:11:06 crc kubenswrapper[4764]: I0202 09:11:06.553271 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-7bff5df74-nwqtk" Feb 02 09:11:06 crc kubenswrapper[4764]: I0202 09:11:06.566700 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-789c85bd8b-5xbqt" podStartSLOduration=3.566683216 podStartE2EDuration="3.566683216s" podCreationTimestamp="2026-02-02 09:11:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:11:06.563344117 +0000 UTC m=+229.497068205" watchObservedRunningTime="2026-02-02 09:11:06.566683216 +0000 UTC m=+229.500407304" Feb 02 09:11:06 crc kubenswrapper[4764]: I0202 09:11:06.568286 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-7bff5df74-nwqtk" podStartSLOduration=3.568280743 podStartE2EDuration="3.568280743s" podCreationTimestamp="2026-02-02 09:11:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:11:06.542212485 +0000 UTC m=+229.475936573" watchObservedRunningTime="2026-02-02 09:11:06.568280743 +0000 UTC m=+229.502004831" Feb 02 09:11:20 crc kubenswrapper[4764]: I0202 09:11:20.685909 4764 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Feb 02 09:11:20 crc kubenswrapper[4764]: I0202 09:11:20.687212 4764 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Feb 02 09:11:20 crc kubenswrapper[4764]: I0202 09:11:20.687326 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 02 09:11:20 crc kubenswrapper[4764]: I0202 09:11:20.687659 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://9e78b6d08e4d616cefbb73e3fd43b3c6886dcd8aa997c7b4948bf1ff63401293" gracePeriod=15 Feb 02 09:11:20 crc kubenswrapper[4764]: I0202 09:11:20.687680 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://d468715c2da33996561c40043216233fd12e21539d0c58dd1f3b892fd9f91a9d" gracePeriod=15 Feb 02 09:11:20 crc kubenswrapper[4764]: I0202 09:11:20.687709 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://9b184f7c936dc4046f74c0018269cbc2c4b7ceddb650b4e871d3f8a707e0df88" gracePeriod=15 Feb 02 09:11:20 crc kubenswrapper[4764]: I0202 09:11:20.687802 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://b44e17b8ca01f549bc10081dd5f109cb1b9fff364bf5bd4fd90d9f0e3eeb04db" gracePeriod=15 Feb 02 09:11:20 crc kubenswrapper[4764]: I0202 09:11:20.687843 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://269c5074371bd00f8363fb3888c11f1df383081a61497aea6578f08700881d50" gracePeriod=15 Feb 02 09:11:20 crc kubenswrapper[4764]: I0202 09:11:20.692190 4764 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Feb 02 09:11:20 crc kubenswrapper[4764]: E0202 09:11:20.692487 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Feb 02 09:11:20 crc kubenswrapper[4764]: I0202 09:11:20.692502 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Feb 02 09:11:20 crc kubenswrapper[4764]: E0202 09:11:20.692512 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 02 09:11:20 crc kubenswrapper[4764]: I0202 09:11:20.692518 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 02 09:11:20 crc kubenswrapper[4764]: E0202 09:11:20.692532 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Feb 02 09:11:20 crc kubenswrapper[4764]: I0202 09:11:20.692538 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Feb 02 09:11:20 crc kubenswrapper[4764]: E0202 09:11:20.692566 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Feb 02 09:11:20 crc kubenswrapper[4764]: I0202 09:11:20.692573 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Feb 02 09:11:20 crc kubenswrapper[4764]: E0202 09:11:20.692589 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Feb 02 09:11:20 crc kubenswrapper[4764]: I0202 09:11:20.692595 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Feb 02 09:11:20 crc kubenswrapper[4764]: E0202 09:11:20.692602 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Feb 02 09:11:20 crc kubenswrapper[4764]: I0202 09:11:20.692607 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Feb 02 09:11:20 crc kubenswrapper[4764]: I0202 09:11:20.692740 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 02 09:11:20 crc kubenswrapper[4764]: I0202 09:11:20.692750 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Feb 02 09:11:20 crc kubenswrapper[4764]: I0202 09:11:20.692760 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Feb 02 09:11:20 crc kubenswrapper[4764]: I0202 09:11:20.692768 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 02 09:11:20 crc kubenswrapper[4764]: I0202 09:11:20.692777 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Feb 02 09:11:20 crc kubenswrapper[4764]: I0202 09:11:20.692803 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Feb 02 09:11:20 crc kubenswrapper[4764]: E0202 09:11:20.692913 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 02 09:11:20 crc kubenswrapper[4764]: I0202 09:11:20.692920 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 02 09:11:20 crc kubenswrapper[4764]: I0202 09:11:20.785740 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 02 09:11:20 crc kubenswrapper[4764]: I0202 09:11:20.785834 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 02 09:11:20 crc kubenswrapper[4764]: I0202 09:11:20.785894 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 02 09:11:20 crc kubenswrapper[4764]: I0202 09:11:20.785967 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 02 09:11:20 crc kubenswrapper[4764]: I0202 09:11:20.786066 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 02 09:11:20 crc kubenswrapper[4764]: I0202 09:11:20.786124 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 02 09:11:20 crc kubenswrapper[4764]: I0202 09:11:20.786175 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 02 09:11:20 crc kubenswrapper[4764]: I0202 09:11:20.786205 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 02 09:11:20 crc kubenswrapper[4764]: I0202 09:11:20.886827 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 02 09:11:20 crc kubenswrapper[4764]: I0202 09:11:20.887136 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 02 09:11:20 crc kubenswrapper[4764]: I0202 09:11:20.887157 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 02 09:11:20 crc kubenswrapper[4764]: I0202 09:11:20.886973 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 02 09:11:20 crc kubenswrapper[4764]: I0202 09:11:20.887205 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 02 09:11:20 crc kubenswrapper[4764]: I0202 09:11:20.887254 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 02 09:11:20 crc kubenswrapper[4764]: I0202 09:11:20.887256 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 02 09:11:20 crc kubenswrapper[4764]: I0202 09:11:20.887279 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 02 09:11:20 crc kubenswrapper[4764]: I0202 09:11:20.887293 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 02 09:11:20 crc kubenswrapper[4764]: I0202 09:11:20.887253 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 02 09:11:20 crc kubenswrapper[4764]: I0202 09:11:20.887313 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 02 09:11:20 crc kubenswrapper[4764]: I0202 09:11:20.887322 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 02 09:11:20 crc kubenswrapper[4764]: I0202 09:11:20.887329 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 02 09:11:20 crc kubenswrapper[4764]: I0202 09:11:20.887343 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 02 09:11:20 crc kubenswrapper[4764]: I0202 09:11:20.887359 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 02 09:11:20 crc kubenswrapper[4764]: I0202 09:11:20.887408 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 02 09:11:21 crc kubenswrapper[4764]: I0202 09:11:21.597719 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Feb 02 09:11:21 crc kubenswrapper[4764]: I0202 09:11:21.599678 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Feb 02 09:11:21 crc kubenswrapper[4764]: I0202 09:11:21.600565 4764 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="269c5074371bd00f8363fb3888c11f1df383081a61497aea6578f08700881d50" exitCode=0 Feb 02 09:11:21 crc kubenswrapper[4764]: I0202 09:11:21.600603 4764 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="b44e17b8ca01f549bc10081dd5f109cb1b9fff364bf5bd4fd90d9f0e3eeb04db" exitCode=0 Feb 02 09:11:21 crc kubenswrapper[4764]: I0202 09:11:21.600620 4764 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="d468715c2da33996561c40043216233fd12e21539d0c58dd1f3b892fd9f91a9d" exitCode=0 Feb 02 09:11:21 crc kubenswrapper[4764]: I0202 09:11:21.600631 4764 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="9b184f7c936dc4046f74c0018269cbc2c4b7ceddb650b4e871d3f8a707e0df88" exitCode=2 Feb 02 09:11:21 crc kubenswrapper[4764]: I0202 09:11:21.600710 4764 scope.go:117] "RemoveContainer" containerID="b25bc9395831524c53f9e565b4efbe74f21792efafeb1dd507161bc91aa0ecdf" Feb 02 09:11:21 crc kubenswrapper[4764]: I0202 09:11:21.603739 4764 generic.go:334] "Generic (PLEG): container finished" podID="f6df65ba-ae7e-48e4-aa24-c5c00b23d4aa" containerID="be556bb93f0178d943541fb659f6f1d6dfd72897ea2234499f80f374591e4e66" exitCode=0 Feb 02 09:11:21 crc kubenswrapper[4764]: I0202 09:11:21.603802 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"f6df65ba-ae7e-48e4-aa24-c5c00b23d4aa","Type":"ContainerDied","Data":"be556bb93f0178d943541fb659f6f1d6dfd72897ea2234499f80f374591e4e66"} Feb 02 09:11:21 crc kubenswrapper[4764]: I0202 09:11:21.604885 4764 status_manager.go:851] "Failed to get status for pod" podUID="f6df65ba-ae7e-48e4-aa24-c5c00b23d4aa" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.217:6443: connect: connection refused" Feb 02 09:11:21 crc kubenswrapper[4764]: I0202 09:11:21.605491 4764 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.217:6443: connect: connection refused" Feb 02 09:11:22 crc kubenswrapper[4764]: I0202 09:11:22.613014 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Feb 02 09:11:23 crc kubenswrapper[4764]: I0202 09:11:23.069358 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Feb 02 09:11:23 crc kubenswrapper[4764]: I0202 09:11:23.070440 4764 status_manager.go:851] "Failed to get status for pod" podUID="f6df65ba-ae7e-48e4-aa24-c5c00b23d4aa" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.217:6443: connect: connection refused" Feb 02 09:11:23 crc kubenswrapper[4764]: I0202 09:11:23.074921 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Feb 02 09:11:23 crc kubenswrapper[4764]: I0202 09:11:23.075588 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 02 09:11:23 crc kubenswrapper[4764]: I0202 09:11:23.076085 4764 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.217:6443: connect: connection refused" Feb 02 09:11:23 crc kubenswrapper[4764]: I0202 09:11:23.076439 4764 status_manager.go:851] "Failed to get status for pod" podUID="f6df65ba-ae7e-48e4-aa24-c5c00b23d4aa" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.217:6443: connect: connection refused" Feb 02 09:11:23 crc kubenswrapper[4764]: I0202 09:11:23.216848 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/f6df65ba-ae7e-48e4-aa24-c5c00b23d4aa-kubelet-dir\") pod \"f6df65ba-ae7e-48e4-aa24-c5c00b23d4aa\" (UID: \"f6df65ba-ae7e-48e4-aa24-c5c00b23d4aa\") " Feb 02 09:11:23 crc kubenswrapper[4764]: I0202 09:11:23.216926 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f6df65ba-ae7e-48e4-aa24-c5c00b23d4aa-var-lock\") pod \"f6df65ba-ae7e-48e4-aa24-c5c00b23d4aa\" (UID: \"f6df65ba-ae7e-48e4-aa24-c5c00b23d4aa\") " Feb 02 09:11:23 crc kubenswrapper[4764]: I0202 09:11:23.217098 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Feb 02 09:11:23 crc kubenswrapper[4764]: I0202 09:11:23.217161 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f6df65ba-ae7e-48e4-aa24-c5c00b23d4aa-kube-api-access\") pod \"f6df65ba-ae7e-48e4-aa24-c5c00b23d4aa\" (UID: \"f6df65ba-ae7e-48e4-aa24-c5c00b23d4aa\") " Feb 02 09:11:23 crc kubenswrapper[4764]: I0202 09:11:23.217212 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Feb 02 09:11:23 crc kubenswrapper[4764]: I0202 09:11:23.217249 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Feb 02 09:11:23 crc kubenswrapper[4764]: I0202 09:11:23.217093 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f6df65ba-ae7e-48e4-aa24-c5c00b23d4aa-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "f6df65ba-ae7e-48e4-aa24-c5c00b23d4aa" (UID: "f6df65ba-ae7e-48e4-aa24-c5c00b23d4aa"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 09:11:23 crc kubenswrapper[4764]: I0202 09:11:23.217092 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f6df65ba-ae7e-48e4-aa24-c5c00b23d4aa-var-lock" (OuterVolumeSpecName: "var-lock") pod "f6df65ba-ae7e-48e4-aa24-c5c00b23d4aa" (UID: "f6df65ba-ae7e-48e4-aa24-c5c00b23d4aa"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 09:11:23 crc kubenswrapper[4764]: I0202 09:11:23.217139 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 09:11:23 crc kubenswrapper[4764]: I0202 09:11:23.217698 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 09:11:23 crc kubenswrapper[4764]: I0202 09:11:23.217701 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 09:11:23 crc kubenswrapper[4764]: I0202 09:11:23.226172 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f6df65ba-ae7e-48e4-aa24-c5c00b23d4aa-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "f6df65ba-ae7e-48e4-aa24-c5c00b23d4aa" (UID: "f6df65ba-ae7e-48e4-aa24-c5c00b23d4aa"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:11:23 crc kubenswrapper[4764]: I0202 09:11:23.319440 4764 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Feb 02 09:11:23 crc kubenswrapper[4764]: I0202 09:11:23.319517 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f6df65ba-ae7e-48e4-aa24-c5c00b23d4aa-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 02 09:11:23 crc kubenswrapper[4764]: I0202 09:11:23.319540 4764 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Feb 02 09:11:23 crc kubenswrapper[4764]: I0202 09:11:23.319556 4764 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Feb 02 09:11:23 crc kubenswrapper[4764]: I0202 09:11:23.319606 4764 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/f6df65ba-ae7e-48e4-aa24-c5c00b23d4aa-kubelet-dir\") on node \"crc\" DevicePath \"\"" Feb 02 09:11:23 crc kubenswrapper[4764]: I0202 09:11:23.319624 4764 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f6df65ba-ae7e-48e4-aa24-c5c00b23d4aa-var-lock\") on node \"crc\" DevicePath \"\"" Feb 02 09:11:23 crc kubenswrapper[4764]: I0202 09:11:23.623042 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Feb 02 09:11:23 crc kubenswrapper[4764]: I0202 09:11:23.623840 4764 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="9e78b6d08e4d616cefbb73e3fd43b3c6886dcd8aa997c7b4948bf1ff63401293" exitCode=0 Feb 02 09:11:23 crc kubenswrapper[4764]: I0202 09:11:23.623922 4764 scope.go:117] "RemoveContainer" containerID="269c5074371bd00f8363fb3888c11f1df383081a61497aea6578f08700881d50" Feb 02 09:11:23 crc kubenswrapper[4764]: I0202 09:11:23.624027 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 02 09:11:23 crc kubenswrapper[4764]: I0202 09:11:23.625822 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"f6df65ba-ae7e-48e4-aa24-c5c00b23d4aa","Type":"ContainerDied","Data":"1a11c9a20f1e10c2d4157e6f16fcd104fe8d3b63d89d4390370b24693f10aec8"} Feb 02 09:11:23 crc kubenswrapper[4764]: I0202 09:11:23.625847 4764 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1a11c9a20f1e10c2d4157e6f16fcd104fe8d3b63d89d4390370b24693f10aec8" Feb 02 09:11:23 crc kubenswrapper[4764]: I0202 09:11:23.625917 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Feb 02 09:11:23 crc kubenswrapper[4764]: I0202 09:11:23.643874 4764 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.217:6443: connect: connection refused" Feb 02 09:11:23 crc kubenswrapper[4764]: I0202 09:11:23.644360 4764 status_manager.go:851] "Failed to get status for pod" podUID="f6df65ba-ae7e-48e4-aa24-c5c00b23d4aa" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.217:6443: connect: connection refused" Feb 02 09:11:23 crc kubenswrapper[4764]: I0202 09:11:23.650296 4764 scope.go:117] "RemoveContainer" containerID="b44e17b8ca01f549bc10081dd5f109cb1b9fff364bf5bd4fd90d9f0e3eeb04db" Feb 02 09:11:23 crc kubenswrapper[4764]: I0202 09:11:23.653153 4764 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.217:6443: connect: connection refused" Feb 02 09:11:23 crc kubenswrapper[4764]: I0202 09:11:23.653710 4764 status_manager.go:851] "Failed to get status for pod" podUID="f6df65ba-ae7e-48e4-aa24-c5c00b23d4aa" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.217:6443: connect: connection refused" Feb 02 09:11:23 crc kubenswrapper[4764]: I0202 09:11:23.668183 4764 scope.go:117] "RemoveContainer" containerID="d468715c2da33996561c40043216233fd12e21539d0c58dd1f3b892fd9f91a9d" Feb 02 09:11:23 crc kubenswrapper[4764]: I0202 09:11:23.685506 4764 scope.go:117] "RemoveContainer" containerID="9b184f7c936dc4046f74c0018269cbc2c4b7ceddb650b4e871d3f8a707e0df88" Feb 02 09:11:23 crc kubenswrapper[4764]: I0202 09:11:23.703756 4764 scope.go:117] "RemoveContainer" containerID="9e78b6d08e4d616cefbb73e3fd43b3c6886dcd8aa997c7b4948bf1ff63401293" Feb 02 09:11:23 crc kubenswrapper[4764]: I0202 09:11:23.729510 4764 scope.go:117] "RemoveContainer" containerID="dfb06a5ec9262cf39b108a66014d15e325cdbd5441a2d9e7a811dd57265c99a6" Feb 02 09:11:23 crc kubenswrapper[4764]: I0202 09:11:23.758974 4764 scope.go:117] "RemoveContainer" containerID="269c5074371bd00f8363fb3888c11f1df383081a61497aea6578f08700881d50" Feb 02 09:11:23 crc kubenswrapper[4764]: E0202 09:11:23.759884 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"269c5074371bd00f8363fb3888c11f1df383081a61497aea6578f08700881d50\": container with ID starting with 269c5074371bd00f8363fb3888c11f1df383081a61497aea6578f08700881d50 not found: ID does not exist" containerID="269c5074371bd00f8363fb3888c11f1df383081a61497aea6578f08700881d50" Feb 02 09:11:23 crc kubenswrapper[4764]: I0202 09:11:23.760013 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"269c5074371bd00f8363fb3888c11f1df383081a61497aea6578f08700881d50"} err="failed to get container status \"269c5074371bd00f8363fb3888c11f1df383081a61497aea6578f08700881d50\": rpc error: code = NotFound desc = could not find container \"269c5074371bd00f8363fb3888c11f1df383081a61497aea6578f08700881d50\": container with ID starting with 269c5074371bd00f8363fb3888c11f1df383081a61497aea6578f08700881d50 not found: ID does not exist" Feb 02 09:11:23 crc kubenswrapper[4764]: I0202 09:11:23.760055 4764 scope.go:117] "RemoveContainer" containerID="b44e17b8ca01f549bc10081dd5f109cb1b9fff364bf5bd4fd90d9f0e3eeb04db" Feb 02 09:11:23 crc kubenswrapper[4764]: E0202 09:11:23.760487 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b44e17b8ca01f549bc10081dd5f109cb1b9fff364bf5bd4fd90d9f0e3eeb04db\": container with ID starting with b44e17b8ca01f549bc10081dd5f109cb1b9fff364bf5bd4fd90d9f0e3eeb04db not found: ID does not exist" containerID="b44e17b8ca01f549bc10081dd5f109cb1b9fff364bf5bd4fd90d9f0e3eeb04db" Feb 02 09:11:23 crc kubenswrapper[4764]: I0202 09:11:23.760972 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b44e17b8ca01f549bc10081dd5f109cb1b9fff364bf5bd4fd90d9f0e3eeb04db"} err="failed to get container status \"b44e17b8ca01f549bc10081dd5f109cb1b9fff364bf5bd4fd90d9f0e3eeb04db\": rpc error: code = NotFound desc = could not find container \"b44e17b8ca01f549bc10081dd5f109cb1b9fff364bf5bd4fd90d9f0e3eeb04db\": container with ID starting with b44e17b8ca01f549bc10081dd5f109cb1b9fff364bf5bd4fd90d9f0e3eeb04db not found: ID does not exist" Feb 02 09:11:23 crc kubenswrapper[4764]: I0202 09:11:23.761077 4764 scope.go:117] "RemoveContainer" containerID="d468715c2da33996561c40043216233fd12e21539d0c58dd1f3b892fd9f91a9d" Feb 02 09:11:23 crc kubenswrapper[4764]: E0202 09:11:23.761390 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d468715c2da33996561c40043216233fd12e21539d0c58dd1f3b892fd9f91a9d\": container with ID starting with d468715c2da33996561c40043216233fd12e21539d0c58dd1f3b892fd9f91a9d not found: ID does not exist" containerID="d468715c2da33996561c40043216233fd12e21539d0c58dd1f3b892fd9f91a9d" Feb 02 09:11:23 crc kubenswrapper[4764]: I0202 09:11:23.761502 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d468715c2da33996561c40043216233fd12e21539d0c58dd1f3b892fd9f91a9d"} err="failed to get container status \"d468715c2da33996561c40043216233fd12e21539d0c58dd1f3b892fd9f91a9d\": rpc error: code = NotFound desc = could not find container \"d468715c2da33996561c40043216233fd12e21539d0c58dd1f3b892fd9f91a9d\": container with ID starting with d468715c2da33996561c40043216233fd12e21539d0c58dd1f3b892fd9f91a9d not found: ID does not exist" Feb 02 09:11:23 crc kubenswrapper[4764]: I0202 09:11:23.761636 4764 scope.go:117] "RemoveContainer" containerID="9b184f7c936dc4046f74c0018269cbc2c4b7ceddb650b4e871d3f8a707e0df88" Feb 02 09:11:23 crc kubenswrapper[4764]: E0202 09:11:23.762003 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9b184f7c936dc4046f74c0018269cbc2c4b7ceddb650b4e871d3f8a707e0df88\": container with ID starting with 9b184f7c936dc4046f74c0018269cbc2c4b7ceddb650b4e871d3f8a707e0df88 not found: ID does not exist" containerID="9b184f7c936dc4046f74c0018269cbc2c4b7ceddb650b4e871d3f8a707e0df88" Feb 02 09:11:23 crc kubenswrapper[4764]: I0202 09:11:23.762102 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9b184f7c936dc4046f74c0018269cbc2c4b7ceddb650b4e871d3f8a707e0df88"} err="failed to get container status \"9b184f7c936dc4046f74c0018269cbc2c4b7ceddb650b4e871d3f8a707e0df88\": rpc error: code = NotFound desc = could not find container \"9b184f7c936dc4046f74c0018269cbc2c4b7ceddb650b4e871d3f8a707e0df88\": container with ID starting with 9b184f7c936dc4046f74c0018269cbc2c4b7ceddb650b4e871d3f8a707e0df88 not found: ID does not exist" Feb 02 09:11:23 crc kubenswrapper[4764]: I0202 09:11:23.762205 4764 scope.go:117] "RemoveContainer" containerID="9e78b6d08e4d616cefbb73e3fd43b3c6886dcd8aa997c7b4948bf1ff63401293" Feb 02 09:11:23 crc kubenswrapper[4764]: E0202 09:11:23.762554 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9e78b6d08e4d616cefbb73e3fd43b3c6886dcd8aa997c7b4948bf1ff63401293\": container with ID starting with 9e78b6d08e4d616cefbb73e3fd43b3c6886dcd8aa997c7b4948bf1ff63401293 not found: ID does not exist" containerID="9e78b6d08e4d616cefbb73e3fd43b3c6886dcd8aa997c7b4948bf1ff63401293" Feb 02 09:11:23 crc kubenswrapper[4764]: I0202 09:11:23.762641 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e78b6d08e4d616cefbb73e3fd43b3c6886dcd8aa997c7b4948bf1ff63401293"} err="failed to get container status \"9e78b6d08e4d616cefbb73e3fd43b3c6886dcd8aa997c7b4948bf1ff63401293\": rpc error: code = NotFound desc = could not find container \"9e78b6d08e4d616cefbb73e3fd43b3c6886dcd8aa997c7b4948bf1ff63401293\": container with ID starting with 9e78b6d08e4d616cefbb73e3fd43b3c6886dcd8aa997c7b4948bf1ff63401293 not found: ID does not exist" Feb 02 09:11:23 crc kubenswrapper[4764]: I0202 09:11:23.762785 4764 scope.go:117] "RemoveContainer" containerID="dfb06a5ec9262cf39b108a66014d15e325cdbd5441a2d9e7a811dd57265c99a6" Feb 02 09:11:23 crc kubenswrapper[4764]: E0202 09:11:23.763128 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dfb06a5ec9262cf39b108a66014d15e325cdbd5441a2d9e7a811dd57265c99a6\": container with ID starting with dfb06a5ec9262cf39b108a66014d15e325cdbd5441a2d9e7a811dd57265c99a6 not found: ID does not exist" containerID="dfb06a5ec9262cf39b108a66014d15e325cdbd5441a2d9e7a811dd57265c99a6" Feb 02 09:11:23 crc kubenswrapper[4764]: I0202 09:11:23.763208 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dfb06a5ec9262cf39b108a66014d15e325cdbd5441a2d9e7a811dd57265c99a6"} err="failed to get container status \"dfb06a5ec9262cf39b108a66014d15e325cdbd5441a2d9e7a811dd57265c99a6\": rpc error: code = NotFound desc = could not find container \"dfb06a5ec9262cf39b108a66014d15e325cdbd5441a2d9e7a811dd57265c99a6\": container with ID starting with dfb06a5ec9262cf39b108a66014d15e325cdbd5441a2d9e7a811dd57265c99a6 not found: ID does not exist" Feb 02 09:11:23 crc kubenswrapper[4764]: I0202 09:11:23.833135 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Feb 02 09:11:25 crc kubenswrapper[4764]: I0202 09:11:25.016225 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-9mdwc" podUID="513ecde9-f189-4a22-8742-1b262523d462" containerName="oauth-openshift" containerID="cri-o://64648b600bc1ccc6aa5e81e14eb194fd7afbe90c2ec5dcfbbaa066f903a1bb9d" gracePeriod=15 Feb 02 09:11:25 crc kubenswrapper[4764]: I0202 09:11:25.490532 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-9mdwc" Feb 02 09:11:25 crc kubenswrapper[4764]: I0202 09:11:25.491390 4764 status_manager.go:851] "Failed to get status for pod" podUID="f6df65ba-ae7e-48e4-aa24-c5c00b23d4aa" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.217:6443: connect: connection refused" Feb 02 09:11:25 crc kubenswrapper[4764]: I0202 09:11:25.491885 4764 status_manager.go:851] "Failed to get status for pod" podUID="513ecde9-f189-4a22-8742-1b262523d462" pod="openshift-authentication/oauth-openshift-558db77b4-9mdwc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-9mdwc\": dial tcp 38.102.83.217:6443: connect: connection refused" Feb 02 09:11:25 crc kubenswrapper[4764]: I0202 09:11:25.639600 4764 generic.go:334] "Generic (PLEG): container finished" podID="513ecde9-f189-4a22-8742-1b262523d462" containerID="64648b600bc1ccc6aa5e81e14eb194fd7afbe90c2ec5dcfbbaa066f903a1bb9d" exitCode=0 Feb 02 09:11:25 crc kubenswrapper[4764]: I0202 09:11:25.639643 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-9mdwc" event={"ID":"513ecde9-f189-4a22-8742-1b262523d462","Type":"ContainerDied","Data":"64648b600bc1ccc6aa5e81e14eb194fd7afbe90c2ec5dcfbbaa066f903a1bb9d"} Feb 02 09:11:25 crc kubenswrapper[4764]: I0202 09:11:25.639675 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-9mdwc" event={"ID":"513ecde9-f189-4a22-8742-1b262523d462","Type":"ContainerDied","Data":"707c484f25eb263b11a295e4ad9807fda8e4210cb2f90590f5ad4b7631a8fc81"} Feb 02 09:11:25 crc kubenswrapper[4764]: I0202 09:11:25.639680 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-9mdwc" Feb 02 09:11:25 crc kubenswrapper[4764]: I0202 09:11:25.639699 4764 scope.go:117] "RemoveContainer" containerID="64648b600bc1ccc6aa5e81e14eb194fd7afbe90c2ec5dcfbbaa066f903a1bb9d" Feb 02 09:11:25 crc kubenswrapper[4764]: I0202 09:11:25.640747 4764 status_manager.go:851] "Failed to get status for pod" podUID="f6df65ba-ae7e-48e4-aa24-c5c00b23d4aa" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.217:6443: connect: connection refused" Feb 02 09:11:25 crc kubenswrapper[4764]: I0202 09:11:25.641275 4764 status_manager.go:851] "Failed to get status for pod" podUID="513ecde9-f189-4a22-8742-1b262523d462" pod="openshift-authentication/oauth-openshift-558db77b4-9mdwc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-9mdwc\": dial tcp 38.102.83.217:6443: connect: connection refused" Feb 02 09:11:25 crc kubenswrapper[4764]: I0202 09:11:25.648614 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/513ecde9-f189-4a22-8742-1b262523d462-v4-0-config-user-template-provider-selection\") pod \"513ecde9-f189-4a22-8742-1b262523d462\" (UID: \"513ecde9-f189-4a22-8742-1b262523d462\") " Feb 02 09:11:25 crc kubenswrapper[4764]: I0202 09:11:25.648653 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/513ecde9-f189-4a22-8742-1b262523d462-v4-0-config-user-template-login\") pod \"513ecde9-f189-4a22-8742-1b262523d462\" (UID: \"513ecde9-f189-4a22-8742-1b262523d462\") " Feb 02 09:11:25 crc kubenswrapper[4764]: I0202 09:11:25.648722 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/513ecde9-f189-4a22-8742-1b262523d462-v4-0-config-user-idp-0-file-data\") pod \"513ecde9-f189-4a22-8742-1b262523d462\" (UID: \"513ecde9-f189-4a22-8742-1b262523d462\") " Feb 02 09:11:25 crc kubenswrapper[4764]: I0202 09:11:25.648776 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/513ecde9-f189-4a22-8742-1b262523d462-v4-0-config-user-template-error\") pod \"513ecde9-f189-4a22-8742-1b262523d462\" (UID: \"513ecde9-f189-4a22-8742-1b262523d462\") " Feb 02 09:11:25 crc kubenswrapper[4764]: I0202 09:11:25.648794 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/513ecde9-f189-4a22-8742-1b262523d462-audit-dir\") pod \"513ecde9-f189-4a22-8742-1b262523d462\" (UID: \"513ecde9-f189-4a22-8742-1b262523d462\") " Feb 02 09:11:25 crc kubenswrapper[4764]: I0202 09:11:25.648829 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/513ecde9-f189-4a22-8742-1b262523d462-v4-0-config-system-service-ca\") pod \"513ecde9-f189-4a22-8742-1b262523d462\" (UID: \"513ecde9-f189-4a22-8742-1b262523d462\") " Feb 02 09:11:25 crc kubenswrapper[4764]: I0202 09:11:25.648849 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/513ecde9-f189-4a22-8742-1b262523d462-v4-0-config-system-router-certs\") pod \"513ecde9-f189-4a22-8742-1b262523d462\" (UID: \"513ecde9-f189-4a22-8742-1b262523d462\") " Feb 02 09:11:25 crc kubenswrapper[4764]: I0202 09:11:25.648883 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/513ecde9-f189-4a22-8742-1b262523d462-v4-0-config-system-session\") pod \"513ecde9-f189-4a22-8742-1b262523d462\" (UID: \"513ecde9-f189-4a22-8742-1b262523d462\") " Feb 02 09:11:25 crc kubenswrapper[4764]: I0202 09:11:25.648899 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/513ecde9-f189-4a22-8742-1b262523d462-v4-0-config-system-ocp-branding-template\") pod \"513ecde9-f189-4a22-8742-1b262523d462\" (UID: \"513ecde9-f189-4a22-8742-1b262523d462\") " Feb 02 09:11:25 crc kubenswrapper[4764]: I0202 09:11:25.648918 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/513ecde9-f189-4a22-8742-1b262523d462-audit-policies\") pod \"513ecde9-f189-4a22-8742-1b262523d462\" (UID: \"513ecde9-f189-4a22-8742-1b262523d462\") " Feb 02 09:11:25 crc kubenswrapper[4764]: I0202 09:11:25.648966 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/513ecde9-f189-4a22-8742-1b262523d462-v4-0-config-system-serving-cert\") pod \"513ecde9-f189-4a22-8742-1b262523d462\" (UID: \"513ecde9-f189-4a22-8742-1b262523d462\") " Feb 02 09:11:25 crc kubenswrapper[4764]: I0202 09:11:25.648992 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4t7sg\" (UniqueName: \"kubernetes.io/projected/513ecde9-f189-4a22-8742-1b262523d462-kube-api-access-4t7sg\") pod \"513ecde9-f189-4a22-8742-1b262523d462\" (UID: \"513ecde9-f189-4a22-8742-1b262523d462\") " Feb 02 09:11:25 crc kubenswrapper[4764]: I0202 09:11:25.649030 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/513ecde9-f189-4a22-8742-1b262523d462-v4-0-config-system-cliconfig\") pod \"513ecde9-f189-4a22-8742-1b262523d462\" (UID: \"513ecde9-f189-4a22-8742-1b262523d462\") " Feb 02 09:11:25 crc kubenswrapper[4764]: I0202 09:11:25.649055 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/513ecde9-f189-4a22-8742-1b262523d462-v4-0-config-system-trusted-ca-bundle\") pod \"513ecde9-f189-4a22-8742-1b262523d462\" (UID: \"513ecde9-f189-4a22-8742-1b262523d462\") " Feb 02 09:11:25 crc kubenswrapper[4764]: I0202 09:11:25.649675 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/513ecde9-f189-4a22-8742-1b262523d462-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "513ecde9-f189-4a22-8742-1b262523d462" (UID: "513ecde9-f189-4a22-8742-1b262523d462"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:11:25 crc kubenswrapper[4764]: I0202 09:11:25.649914 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/513ecde9-f189-4a22-8742-1b262523d462-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "513ecde9-f189-4a22-8742-1b262523d462" (UID: "513ecde9-f189-4a22-8742-1b262523d462"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:11:25 crc kubenswrapper[4764]: I0202 09:11:25.649974 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/513ecde9-f189-4a22-8742-1b262523d462-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "513ecde9-f189-4a22-8742-1b262523d462" (UID: "513ecde9-f189-4a22-8742-1b262523d462"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 09:11:25 crc kubenswrapper[4764]: I0202 09:11:25.650415 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/513ecde9-f189-4a22-8742-1b262523d462-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "513ecde9-f189-4a22-8742-1b262523d462" (UID: "513ecde9-f189-4a22-8742-1b262523d462"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:11:25 crc kubenswrapper[4764]: I0202 09:11:25.650811 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/513ecde9-f189-4a22-8742-1b262523d462-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "513ecde9-f189-4a22-8742-1b262523d462" (UID: "513ecde9-f189-4a22-8742-1b262523d462"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:11:25 crc kubenswrapper[4764]: I0202 09:11:25.654147 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/513ecde9-f189-4a22-8742-1b262523d462-kube-api-access-4t7sg" (OuterVolumeSpecName: "kube-api-access-4t7sg") pod "513ecde9-f189-4a22-8742-1b262523d462" (UID: "513ecde9-f189-4a22-8742-1b262523d462"). InnerVolumeSpecName "kube-api-access-4t7sg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:11:25 crc kubenswrapper[4764]: I0202 09:11:25.656032 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/513ecde9-f189-4a22-8742-1b262523d462-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "513ecde9-f189-4a22-8742-1b262523d462" (UID: "513ecde9-f189-4a22-8742-1b262523d462"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:11:25 crc kubenswrapper[4764]: I0202 09:11:25.656262 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/513ecde9-f189-4a22-8742-1b262523d462-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "513ecde9-f189-4a22-8742-1b262523d462" (UID: "513ecde9-f189-4a22-8742-1b262523d462"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:11:25 crc kubenswrapper[4764]: I0202 09:11:25.656469 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/513ecde9-f189-4a22-8742-1b262523d462-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "513ecde9-f189-4a22-8742-1b262523d462" (UID: "513ecde9-f189-4a22-8742-1b262523d462"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:11:25 crc kubenswrapper[4764]: I0202 09:11:25.661516 4764 scope.go:117] "RemoveContainer" containerID="64648b600bc1ccc6aa5e81e14eb194fd7afbe90c2ec5dcfbbaa066f903a1bb9d" Feb 02 09:11:25 crc kubenswrapper[4764]: E0202 09:11:25.662141 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"64648b600bc1ccc6aa5e81e14eb194fd7afbe90c2ec5dcfbbaa066f903a1bb9d\": container with ID starting with 64648b600bc1ccc6aa5e81e14eb194fd7afbe90c2ec5dcfbbaa066f903a1bb9d not found: ID does not exist" containerID="64648b600bc1ccc6aa5e81e14eb194fd7afbe90c2ec5dcfbbaa066f903a1bb9d" Feb 02 09:11:25 crc kubenswrapper[4764]: I0202 09:11:25.662228 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"64648b600bc1ccc6aa5e81e14eb194fd7afbe90c2ec5dcfbbaa066f903a1bb9d"} err="failed to get container status \"64648b600bc1ccc6aa5e81e14eb194fd7afbe90c2ec5dcfbbaa066f903a1bb9d\": rpc error: code = NotFound desc = could not find container \"64648b600bc1ccc6aa5e81e14eb194fd7afbe90c2ec5dcfbbaa066f903a1bb9d\": container with ID starting with 64648b600bc1ccc6aa5e81e14eb194fd7afbe90c2ec5dcfbbaa066f903a1bb9d not found: ID does not exist" Feb 02 09:11:25 crc kubenswrapper[4764]: I0202 09:11:25.663358 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/513ecde9-f189-4a22-8742-1b262523d462-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "513ecde9-f189-4a22-8742-1b262523d462" (UID: "513ecde9-f189-4a22-8742-1b262523d462"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:11:25 crc kubenswrapper[4764]: I0202 09:11:25.663726 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/513ecde9-f189-4a22-8742-1b262523d462-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "513ecde9-f189-4a22-8742-1b262523d462" (UID: "513ecde9-f189-4a22-8742-1b262523d462"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:11:25 crc kubenswrapper[4764]: I0202 09:11:25.663759 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/513ecde9-f189-4a22-8742-1b262523d462-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "513ecde9-f189-4a22-8742-1b262523d462" (UID: "513ecde9-f189-4a22-8742-1b262523d462"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:11:25 crc kubenswrapper[4764]: I0202 09:11:25.664015 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/513ecde9-f189-4a22-8742-1b262523d462-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "513ecde9-f189-4a22-8742-1b262523d462" (UID: "513ecde9-f189-4a22-8742-1b262523d462"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:11:25 crc kubenswrapper[4764]: I0202 09:11:25.664155 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/513ecde9-f189-4a22-8742-1b262523d462-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "513ecde9-f189-4a22-8742-1b262523d462" (UID: "513ecde9-f189-4a22-8742-1b262523d462"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:11:25 crc kubenswrapper[4764]: E0202 09:11:25.717809 4764 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.217:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 02 09:11:25 crc kubenswrapper[4764]: I0202 09:11:25.718551 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 02 09:11:25 crc kubenswrapper[4764]: E0202 09:11:25.742532 4764 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.217:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.189062fb77ba95b8 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-02 09:11:25.741635 +0000 UTC m=+248.675359098,LastTimestamp:2026-02-02 09:11:25.741635 +0000 UTC m=+248.675359098,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 02 09:11:25 crc kubenswrapper[4764]: I0202 09:11:25.750554 4764 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/513ecde9-f189-4a22-8742-1b262523d462-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Feb 02 09:11:25 crc kubenswrapper[4764]: I0202 09:11:25.750589 4764 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/513ecde9-f189-4a22-8742-1b262523d462-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 09:11:25 crc kubenswrapper[4764]: I0202 09:11:25.750605 4764 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/513ecde9-f189-4a22-8742-1b262523d462-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Feb 02 09:11:25 crc kubenswrapper[4764]: I0202 09:11:25.750619 4764 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/513ecde9-f189-4a22-8742-1b262523d462-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Feb 02 09:11:25 crc kubenswrapper[4764]: I0202 09:11:25.750633 4764 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/513ecde9-f189-4a22-8742-1b262523d462-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Feb 02 09:11:25 crc kubenswrapper[4764]: I0202 09:11:25.750645 4764 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/513ecde9-f189-4a22-8742-1b262523d462-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Feb 02 09:11:25 crc kubenswrapper[4764]: I0202 09:11:25.750657 4764 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/513ecde9-f189-4a22-8742-1b262523d462-audit-dir\") on node \"crc\" DevicePath \"\"" Feb 02 09:11:25 crc kubenswrapper[4764]: I0202 09:11:25.750669 4764 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/513ecde9-f189-4a22-8742-1b262523d462-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Feb 02 09:11:25 crc kubenswrapper[4764]: I0202 09:11:25.750682 4764 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/513ecde9-f189-4a22-8742-1b262523d462-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Feb 02 09:11:25 crc kubenswrapper[4764]: I0202 09:11:25.750736 4764 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/513ecde9-f189-4a22-8742-1b262523d462-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Feb 02 09:11:25 crc kubenswrapper[4764]: I0202 09:11:25.750750 4764 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/513ecde9-f189-4a22-8742-1b262523d462-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Feb 02 09:11:25 crc kubenswrapper[4764]: I0202 09:11:25.750762 4764 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/513ecde9-f189-4a22-8742-1b262523d462-audit-policies\") on node \"crc\" DevicePath \"\"" Feb 02 09:11:25 crc kubenswrapper[4764]: I0202 09:11:25.750775 4764 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/513ecde9-f189-4a22-8742-1b262523d462-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 02 09:11:25 crc kubenswrapper[4764]: I0202 09:11:25.750817 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4t7sg\" (UniqueName: \"kubernetes.io/projected/513ecde9-f189-4a22-8742-1b262523d462-kube-api-access-4t7sg\") on node \"crc\" DevicePath \"\"" Feb 02 09:11:25 crc kubenswrapper[4764]: I0202 09:11:25.944446 4764 status_manager.go:851] "Failed to get status for pod" podUID="f6df65ba-ae7e-48e4-aa24-c5c00b23d4aa" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.217:6443: connect: connection refused" Feb 02 09:11:25 crc kubenswrapper[4764]: I0202 09:11:25.945102 4764 status_manager.go:851] "Failed to get status for pod" podUID="513ecde9-f189-4a22-8742-1b262523d462" pod="openshift-authentication/oauth-openshift-558db77b4-9mdwc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-9mdwc\": dial tcp 38.102.83.217:6443: connect: connection refused" Feb 02 09:11:26 crc kubenswrapper[4764]: I0202 09:11:26.647087 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"80d86a89e0b9d62e35dea277f2821831c8e2e121c28458b00dc0dcfe4d734219"} Feb 02 09:11:26 crc kubenswrapper[4764]: I0202 09:11:26.647150 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"fd8fc58d8c721c976412ff5a59051e5078ccbc4ffb4760674e981c3b3efc06ac"} Feb 02 09:11:26 crc kubenswrapper[4764]: E0202 09:11:26.648004 4764 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.217:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 02 09:11:26 crc kubenswrapper[4764]: I0202 09:11:26.648083 4764 status_manager.go:851] "Failed to get status for pod" podUID="f6df65ba-ae7e-48e4-aa24-c5c00b23d4aa" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.217:6443: connect: connection refused" Feb 02 09:11:26 crc kubenswrapper[4764]: I0202 09:11:26.648580 4764 status_manager.go:851] "Failed to get status for pod" podUID="513ecde9-f189-4a22-8742-1b262523d462" pod="openshift-authentication/oauth-openshift-558db77b4-9mdwc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-9mdwc\": dial tcp 38.102.83.217:6443: connect: connection refused" Feb 02 09:11:27 crc kubenswrapper[4764]: E0202 09:11:27.733548 4764 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.217:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.189062fb77ba95b8 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-02 09:11:25.741635 +0000 UTC m=+248.675359098,LastTimestamp:2026-02-02 09:11:25.741635 +0000 UTC m=+248.675359098,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 02 09:11:27 crc kubenswrapper[4764]: I0202 09:11:27.830031 4764 status_manager.go:851] "Failed to get status for pod" podUID="f6df65ba-ae7e-48e4-aa24-c5c00b23d4aa" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.217:6443: connect: connection refused" Feb 02 09:11:27 crc kubenswrapper[4764]: I0202 09:11:27.831980 4764 status_manager.go:851] "Failed to get status for pod" podUID="513ecde9-f189-4a22-8742-1b262523d462" pod="openshift-authentication/oauth-openshift-558db77b4-9mdwc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-9mdwc\": dial tcp 38.102.83.217:6443: connect: connection refused" Feb 02 09:11:28 crc kubenswrapper[4764]: E0202 09:11:28.480923 4764 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.217:6443: connect: connection refused" Feb 02 09:11:28 crc kubenswrapper[4764]: E0202 09:11:28.481647 4764 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.217:6443: connect: connection refused" Feb 02 09:11:28 crc kubenswrapper[4764]: E0202 09:11:28.482002 4764 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.217:6443: connect: connection refused" Feb 02 09:11:28 crc kubenswrapper[4764]: E0202 09:11:28.482378 4764 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.217:6443: connect: connection refused" Feb 02 09:11:28 crc kubenswrapper[4764]: E0202 09:11:28.482701 4764 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.217:6443: connect: connection refused" Feb 02 09:11:28 crc kubenswrapper[4764]: I0202 09:11:28.482743 4764 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Feb 02 09:11:28 crc kubenswrapper[4764]: E0202 09:11:28.483077 4764 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.217:6443: connect: connection refused" interval="200ms" Feb 02 09:11:28 crc kubenswrapper[4764]: E0202 09:11:28.684596 4764 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.217:6443: connect: connection refused" interval="400ms" Feb 02 09:11:29 crc kubenswrapper[4764]: E0202 09:11:29.085943 4764 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.217:6443: connect: connection refused" interval="800ms" Feb 02 09:11:29 crc kubenswrapper[4764]: E0202 09:11:29.887337 4764 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.217:6443: connect: connection refused" interval="1.6s" Feb 02 09:11:31 crc kubenswrapper[4764]: E0202 09:11:31.488392 4764 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.217:6443: connect: connection refused" interval="3.2s" Feb 02 09:11:32 crc kubenswrapper[4764]: I0202 09:11:32.825187 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 02 09:11:32 crc kubenswrapper[4764]: I0202 09:11:32.826581 4764 status_manager.go:851] "Failed to get status for pod" podUID="f6df65ba-ae7e-48e4-aa24-c5c00b23d4aa" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.217:6443: connect: connection refused" Feb 02 09:11:32 crc kubenswrapper[4764]: I0202 09:11:32.827357 4764 status_manager.go:851] "Failed to get status for pod" podUID="513ecde9-f189-4a22-8742-1b262523d462" pod="openshift-authentication/oauth-openshift-558db77b4-9mdwc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-9mdwc\": dial tcp 38.102.83.217:6443: connect: connection refused" Feb 02 09:11:32 crc kubenswrapper[4764]: I0202 09:11:32.854127 4764 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="87d61aea-46f6-4ee6-a7a4-767a6a67fea4" Feb 02 09:11:32 crc kubenswrapper[4764]: I0202 09:11:32.854276 4764 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="87d61aea-46f6-4ee6-a7a4-767a6a67fea4" Feb 02 09:11:32 crc kubenswrapper[4764]: E0202 09:11:32.854735 4764 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.217:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 02 09:11:32 crc kubenswrapper[4764]: I0202 09:11:32.855352 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 02 09:11:32 crc kubenswrapper[4764]: W0202 09:11:32.888678 4764 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71bb4a3aecc4ba5b26c4b7318770ce13.slice/crio-2429e955b9b7ca11c7c7e0b3c582df661163861f43bc477d8bd4aa6274b166b4 WatchSource:0}: Error finding container 2429e955b9b7ca11c7c7e0b3c582df661163861f43bc477d8bd4aa6274b166b4: Status 404 returned error can't find the container with id 2429e955b9b7ca11c7c7e0b3c582df661163861f43bc477d8bd4aa6274b166b4 Feb 02 09:11:33 crc kubenswrapper[4764]: I0202 09:11:33.691308 4764 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="9910cbb70cd89ec0b262f6859fb8688da57b38d165d2c590a77390b2fd36fa55" exitCode=0 Feb 02 09:11:33 crc kubenswrapper[4764]: I0202 09:11:33.691350 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"9910cbb70cd89ec0b262f6859fb8688da57b38d165d2c590a77390b2fd36fa55"} Feb 02 09:11:33 crc kubenswrapper[4764]: I0202 09:11:33.691377 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"2429e955b9b7ca11c7c7e0b3c582df661163861f43bc477d8bd4aa6274b166b4"} Feb 02 09:11:33 crc kubenswrapper[4764]: I0202 09:11:33.691725 4764 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="87d61aea-46f6-4ee6-a7a4-767a6a67fea4" Feb 02 09:11:33 crc kubenswrapper[4764]: I0202 09:11:33.691772 4764 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="87d61aea-46f6-4ee6-a7a4-767a6a67fea4" Feb 02 09:11:33 crc kubenswrapper[4764]: I0202 09:11:33.692193 4764 status_manager.go:851] "Failed to get status for pod" podUID="f6df65ba-ae7e-48e4-aa24-c5c00b23d4aa" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.217:6443: connect: connection refused" Feb 02 09:11:33 crc kubenswrapper[4764]: E0202 09:11:33.692443 4764 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.217:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 02 09:11:33 crc kubenswrapper[4764]: I0202 09:11:33.692603 4764 status_manager.go:851] "Failed to get status for pod" podUID="513ecde9-f189-4a22-8742-1b262523d462" pod="openshift-authentication/oauth-openshift-558db77b4-9mdwc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-9mdwc\": dial tcp 38.102.83.217:6443: connect: connection refused" Feb 02 09:11:34 crc kubenswrapper[4764]: I0202 09:11:34.704730 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"e4b5c90b987de32a3f99393d242539703efb3df92f5a955cf46984e156771d45"} Feb 02 09:11:34 crc kubenswrapper[4764]: I0202 09:11:34.705373 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"85c3b058c171b717d057555c3948c7cede9306783d262387b8707bbfacea320c"} Feb 02 09:11:34 crc kubenswrapper[4764]: I0202 09:11:34.705387 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"ba98c3d3129fcdd3bd2a4ea404bb6595e379cc2c72be461bee05e49a2320cfda"} Feb 02 09:11:34 crc kubenswrapper[4764]: I0202 09:11:34.708576 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Feb 02 09:11:34 crc kubenswrapper[4764]: I0202 09:11:34.708647 4764 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="60258f84a1d8dcfe48b7e9688ed50f077e201971f20f3f6663fbf1ba58a4611a" exitCode=1 Feb 02 09:11:34 crc kubenswrapper[4764]: I0202 09:11:34.708690 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"60258f84a1d8dcfe48b7e9688ed50f077e201971f20f3f6663fbf1ba58a4611a"} Feb 02 09:11:34 crc kubenswrapper[4764]: I0202 09:11:34.709377 4764 scope.go:117] "RemoveContainer" containerID="60258f84a1d8dcfe48b7e9688ed50f077e201971f20f3f6663fbf1ba58a4611a" Feb 02 09:11:35 crc kubenswrapper[4764]: I0202 09:11:35.717197 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"0afbd41e299d23a4b52a102c850c792285cdd776fab834f7f5acd657f9e6a36c"} Feb 02 09:11:35 crc kubenswrapper[4764]: I0202 09:11:35.717562 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"b5963642554395f966199942d39ed1c20984082d8d2f0d35fb262058e00bf44a"} Feb 02 09:11:35 crc kubenswrapper[4764]: I0202 09:11:35.717655 4764 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="87d61aea-46f6-4ee6-a7a4-767a6a67fea4" Feb 02 09:11:35 crc kubenswrapper[4764]: I0202 09:11:35.717698 4764 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="87d61aea-46f6-4ee6-a7a4-767a6a67fea4" Feb 02 09:11:35 crc kubenswrapper[4764]: I0202 09:11:35.720911 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Feb 02 09:11:35 crc kubenswrapper[4764]: I0202 09:11:35.721014 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"c616c013433ef912ef5ff270bd79e057373450d24844a0aab5fdb2bc3f6884dc"} Feb 02 09:11:37 crc kubenswrapper[4764]: I0202 09:11:37.856394 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 02 09:11:37 crc kubenswrapper[4764]: I0202 09:11:37.856715 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 02 09:11:37 crc kubenswrapper[4764]: I0202 09:11:37.863496 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 02 09:11:40 crc kubenswrapper[4764]: I0202 09:11:40.325273 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 02 09:11:40 crc kubenswrapper[4764]: I0202 09:11:40.726918 4764 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 02 09:11:40 crc kubenswrapper[4764]: I0202 09:11:40.758209 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 02 09:11:40 crc kubenswrapper[4764]: I0202 09:11:40.758298 4764 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="87d61aea-46f6-4ee6-a7a4-767a6a67fea4" Feb 02 09:11:40 crc kubenswrapper[4764]: I0202 09:11:40.758516 4764 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="87d61aea-46f6-4ee6-a7a4-767a6a67fea4" Feb 02 09:11:40 crc kubenswrapper[4764]: I0202 09:11:40.763731 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 02 09:11:40 crc kubenswrapper[4764]: I0202 09:11:40.766139 4764 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="f91932b8-06c5-4835-8616-15e087747360" Feb 02 09:11:41 crc kubenswrapper[4764]: I0202 09:11:41.764655 4764 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="87d61aea-46f6-4ee6-a7a4-767a6a67fea4" Feb 02 09:11:41 crc kubenswrapper[4764]: I0202 09:11:41.765099 4764 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="87d61aea-46f6-4ee6-a7a4-767a6a67fea4" Feb 02 09:11:42 crc kubenswrapper[4764]: I0202 09:11:42.770226 4764 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="87d61aea-46f6-4ee6-a7a4-767a6a67fea4" Feb 02 09:11:42 crc kubenswrapper[4764]: I0202 09:11:42.770256 4764 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="87d61aea-46f6-4ee6-a7a4-767a6a67fea4" Feb 02 09:11:44 crc kubenswrapper[4764]: I0202 09:11:44.134187 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 02 09:11:44 crc kubenswrapper[4764]: I0202 09:11:44.134588 4764 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Feb 02 09:11:44 crc kubenswrapper[4764]: I0202 09:11:44.134659 4764 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Feb 02 09:11:47 crc kubenswrapper[4764]: I0202 09:11:47.846100 4764 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="f91932b8-06c5-4835-8616-15e087747360" Feb 02 09:11:51 crc kubenswrapper[4764]: I0202 09:11:51.044470 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Feb 02 09:11:51 crc kubenswrapper[4764]: I0202 09:11:51.069218 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Feb 02 09:11:51 crc kubenswrapper[4764]: I0202 09:11:51.561920 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Feb 02 09:11:51 crc kubenswrapper[4764]: I0202 09:11:51.914739 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Feb 02 09:11:52 crc kubenswrapper[4764]: I0202 09:11:52.133142 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Feb 02 09:11:52 crc kubenswrapper[4764]: I0202 09:11:52.326891 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Feb 02 09:11:52 crc kubenswrapper[4764]: I0202 09:11:52.429578 4764 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Feb 02 09:11:52 crc kubenswrapper[4764]: I0202 09:11:52.434829 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-9mdwc","openshift-kube-apiserver/kube-apiserver-crc"] Feb 02 09:11:52 crc kubenswrapper[4764]: I0202 09:11:52.434888 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Feb 02 09:11:52 crc kubenswrapper[4764]: I0202 09:11:52.444200 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 02 09:11:52 crc kubenswrapper[4764]: I0202 09:11:52.465310 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=12.465289577 podStartE2EDuration="12.465289577s" podCreationTimestamp="2026-02-02 09:11:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:11:52.463854605 +0000 UTC m=+275.397578733" watchObservedRunningTime="2026-02-02 09:11:52.465289577 +0000 UTC m=+275.399013665" Feb 02 09:11:52 crc kubenswrapper[4764]: I0202 09:11:52.934025 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Feb 02 09:11:53 crc kubenswrapper[4764]: I0202 09:11:53.375602 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Feb 02 09:11:53 crc kubenswrapper[4764]: I0202 09:11:53.403384 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Feb 02 09:11:53 crc kubenswrapper[4764]: I0202 09:11:53.535429 4764 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Feb 02 09:11:53 crc kubenswrapper[4764]: I0202 09:11:53.645781 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Feb 02 09:11:53 crc kubenswrapper[4764]: I0202 09:11:53.775675 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Feb 02 09:11:53 crc kubenswrapper[4764]: I0202 09:11:53.836409 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="513ecde9-f189-4a22-8742-1b262523d462" path="/var/lib/kubelet/pods/513ecde9-f189-4a22-8742-1b262523d462/volumes" Feb 02 09:11:53 crc kubenswrapper[4764]: I0202 09:11:53.865087 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Feb 02 09:11:53 crc kubenswrapper[4764]: I0202 09:11:53.981046 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Feb 02 09:11:54 crc kubenswrapper[4764]: I0202 09:11:54.097851 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Feb 02 09:11:54 crc kubenswrapper[4764]: I0202 09:11:54.134779 4764 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Feb 02 09:11:54 crc kubenswrapper[4764]: I0202 09:11:54.135119 4764 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Feb 02 09:11:54 crc kubenswrapper[4764]: I0202 09:11:54.166812 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Feb 02 09:11:54 crc kubenswrapper[4764]: I0202 09:11:54.268796 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Feb 02 09:11:54 crc kubenswrapper[4764]: I0202 09:11:54.599121 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Feb 02 09:11:54 crc kubenswrapper[4764]: I0202 09:11:54.673210 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Feb 02 09:11:54 crc kubenswrapper[4764]: I0202 09:11:54.704814 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Feb 02 09:11:54 crc kubenswrapper[4764]: I0202 09:11:54.801528 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Feb 02 09:11:54 crc kubenswrapper[4764]: I0202 09:11:54.822159 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Feb 02 09:11:54 crc kubenswrapper[4764]: I0202 09:11:54.853872 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Feb 02 09:11:54 crc kubenswrapper[4764]: I0202 09:11:54.878304 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Feb 02 09:11:55 crc kubenswrapper[4764]: I0202 09:11:55.039185 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Feb 02 09:11:55 crc kubenswrapper[4764]: I0202 09:11:55.201039 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Feb 02 09:11:55 crc kubenswrapper[4764]: I0202 09:11:55.204689 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Feb 02 09:11:55 crc kubenswrapper[4764]: I0202 09:11:55.280565 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Feb 02 09:11:55 crc kubenswrapper[4764]: I0202 09:11:55.321552 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Feb 02 09:11:55 crc kubenswrapper[4764]: I0202 09:11:55.387863 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Feb 02 09:11:55 crc kubenswrapper[4764]: I0202 09:11:55.394360 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Feb 02 09:11:55 crc kubenswrapper[4764]: I0202 09:11:55.639119 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Feb 02 09:11:55 crc kubenswrapper[4764]: I0202 09:11:55.734862 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Feb 02 09:11:55 crc kubenswrapper[4764]: I0202 09:11:55.748364 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Feb 02 09:11:55 crc kubenswrapper[4764]: I0202 09:11:55.773878 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Feb 02 09:11:55 crc kubenswrapper[4764]: I0202 09:11:55.923402 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Feb 02 09:11:55 crc kubenswrapper[4764]: I0202 09:11:55.940454 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Feb 02 09:11:56 crc kubenswrapper[4764]: I0202 09:11:56.091473 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Feb 02 09:11:56 crc kubenswrapper[4764]: I0202 09:11:56.099460 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Feb 02 09:11:56 crc kubenswrapper[4764]: I0202 09:11:56.225249 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Feb 02 09:11:56 crc kubenswrapper[4764]: I0202 09:11:56.297067 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Feb 02 09:11:56 crc kubenswrapper[4764]: I0202 09:11:56.307605 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Feb 02 09:11:56 crc kubenswrapper[4764]: I0202 09:11:56.332133 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Feb 02 09:11:56 crc kubenswrapper[4764]: I0202 09:11:56.353318 4764 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Feb 02 09:11:56 crc kubenswrapper[4764]: I0202 09:11:56.365580 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Feb 02 09:11:56 crc kubenswrapper[4764]: I0202 09:11:56.371573 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Feb 02 09:11:56 crc kubenswrapper[4764]: I0202 09:11:56.458013 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Feb 02 09:11:56 crc kubenswrapper[4764]: I0202 09:11:56.458187 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Feb 02 09:11:56 crc kubenswrapper[4764]: I0202 09:11:56.461234 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Feb 02 09:11:56 crc kubenswrapper[4764]: I0202 09:11:56.463117 4764 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Feb 02 09:11:56 crc kubenswrapper[4764]: I0202 09:11:56.531913 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Feb 02 09:11:56 crc kubenswrapper[4764]: I0202 09:11:56.551086 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Feb 02 09:11:56 crc kubenswrapper[4764]: I0202 09:11:56.698209 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Feb 02 09:11:56 crc kubenswrapper[4764]: I0202 09:11:56.726120 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Feb 02 09:11:56 crc kubenswrapper[4764]: I0202 09:11:56.781799 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Feb 02 09:11:56 crc kubenswrapper[4764]: I0202 09:11:56.838601 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Feb 02 09:11:56 crc kubenswrapper[4764]: I0202 09:11:56.857159 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Feb 02 09:11:56 crc kubenswrapper[4764]: I0202 09:11:56.885940 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Feb 02 09:11:57 crc kubenswrapper[4764]: I0202 09:11:57.063499 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Feb 02 09:11:57 crc kubenswrapper[4764]: I0202 09:11:57.100065 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Feb 02 09:11:57 crc kubenswrapper[4764]: I0202 09:11:57.134000 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Feb 02 09:11:57 crc kubenswrapper[4764]: I0202 09:11:57.139172 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Feb 02 09:11:57 crc kubenswrapper[4764]: I0202 09:11:57.151474 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Feb 02 09:11:57 crc kubenswrapper[4764]: I0202 09:11:57.162123 4764 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Feb 02 09:11:57 crc kubenswrapper[4764]: I0202 09:11:57.217168 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Feb 02 09:11:57 crc kubenswrapper[4764]: I0202 09:11:57.272281 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Feb 02 09:11:57 crc kubenswrapper[4764]: I0202 09:11:57.438753 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Feb 02 09:11:57 crc kubenswrapper[4764]: I0202 09:11:57.468280 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Feb 02 09:11:57 crc kubenswrapper[4764]: I0202 09:11:57.532177 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Feb 02 09:11:57 crc kubenswrapper[4764]: I0202 09:11:57.566116 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Feb 02 09:11:57 crc kubenswrapper[4764]: I0202 09:11:57.637454 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Feb 02 09:11:57 crc kubenswrapper[4764]: I0202 09:11:57.638453 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Feb 02 09:11:57 crc kubenswrapper[4764]: I0202 09:11:57.697456 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Feb 02 09:11:57 crc kubenswrapper[4764]: I0202 09:11:57.731306 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Feb 02 09:11:57 crc kubenswrapper[4764]: I0202 09:11:57.734004 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Feb 02 09:11:57 crc kubenswrapper[4764]: I0202 09:11:57.746322 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Feb 02 09:11:57 crc kubenswrapper[4764]: I0202 09:11:57.759720 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Feb 02 09:11:57 crc kubenswrapper[4764]: I0202 09:11:57.871154 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Feb 02 09:11:57 crc kubenswrapper[4764]: I0202 09:11:57.919979 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Feb 02 09:11:57 crc kubenswrapper[4764]: I0202 09:11:57.924684 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Feb 02 09:11:58 crc kubenswrapper[4764]: I0202 09:11:58.053219 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Feb 02 09:11:58 crc kubenswrapper[4764]: I0202 09:11:58.096409 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Feb 02 09:11:58 crc kubenswrapper[4764]: I0202 09:11:58.122196 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Feb 02 09:11:58 crc kubenswrapper[4764]: I0202 09:11:58.139380 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Feb 02 09:11:58 crc kubenswrapper[4764]: I0202 09:11:58.240734 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Feb 02 09:11:58 crc kubenswrapper[4764]: I0202 09:11:58.249273 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Feb 02 09:11:58 crc kubenswrapper[4764]: I0202 09:11:58.277458 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Feb 02 09:11:58 crc kubenswrapper[4764]: I0202 09:11:58.280432 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Feb 02 09:11:58 crc kubenswrapper[4764]: I0202 09:11:58.378214 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Feb 02 09:11:58 crc kubenswrapper[4764]: I0202 09:11:58.388970 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Feb 02 09:11:58 crc kubenswrapper[4764]: I0202 09:11:58.422551 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Feb 02 09:11:58 crc kubenswrapper[4764]: I0202 09:11:58.425976 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Feb 02 09:11:58 crc kubenswrapper[4764]: I0202 09:11:58.447866 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Feb 02 09:11:58 crc kubenswrapper[4764]: I0202 09:11:58.537603 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Feb 02 09:11:58 crc kubenswrapper[4764]: I0202 09:11:58.579671 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Feb 02 09:11:58 crc kubenswrapper[4764]: I0202 09:11:58.697895 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Feb 02 09:11:58 crc kubenswrapper[4764]: I0202 09:11:58.699813 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Feb 02 09:11:58 crc kubenswrapper[4764]: I0202 09:11:58.755104 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Feb 02 09:11:58 crc kubenswrapper[4764]: I0202 09:11:58.838861 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Feb 02 09:11:58 crc kubenswrapper[4764]: I0202 09:11:58.850761 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Feb 02 09:11:58 crc kubenswrapper[4764]: I0202 09:11:58.913286 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Feb 02 09:11:59 crc kubenswrapper[4764]: I0202 09:11:59.004093 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Feb 02 09:11:59 crc kubenswrapper[4764]: I0202 09:11:59.125493 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Feb 02 09:11:59 crc kubenswrapper[4764]: I0202 09:11:59.152455 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Feb 02 09:11:59 crc kubenswrapper[4764]: I0202 09:11:59.199575 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Feb 02 09:11:59 crc kubenswrapper[4764]: I0202 09:11:59.339271 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Feb 02 09:11:59 crc kubenswrapper[4764]: I0202 09:11:59.459757 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Feb 02 09:11:59 crc kubenswrapper[4764]: I0202 09:11:59.582028 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Feb 02 09:11:59 crc kubenswrapper[4764]: I0202 09:11:59.624720 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Feb 02 09:11:59 crc kubenswrapper[4764]: I0202 09:11:59.674290 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Feb 02 09:11:59 crc kubenswrapper[4764]: I0202 09:11:59.735319 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Feb 02 09:11:59 crc kubenswrapper[4764]: I0202 09:11:59.796736 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Feb 02 09:11:59 crc kubenswrapper[4764]: I0202 09:11:59.913480 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Feb 02 09:11:59 crc kubenswrapper[4764]: I0202 09:11:59.931487 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Feb 02 09:11:59 crc kubenswrapper[4764]: I0202 09:11:59.939303 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Feb 02 09:11:59 crc kubenswrapper[4764]: I0202 09:11:59.962025 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Feb 02 09:11:59 crc kubenswrapper[4764]: I0202 09:11:59.976548 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Feb 02 09:12:00 crc kubenswrapper[4764]: I0202 09:12:00.084991 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Feb 02 09:12:00 crc kubenswrapper[4764]: I0202 09:12:00.115469 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Feb 02 09:12:00 crc kubenswrapper[4764]: I0202 09:12:00.134120 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Feb 02 09:12:00 crc kubenswrapper[4764]: I0202 09:12:00.143146 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Feb 02 09:12:00 crc kubenswrapper[4764]: I0202 09:12:00.313461 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Feb 02 09:12:00 crc kubenswrapper[4764]: I0202 09:12:00.315793 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Feb 02 09:12:00 crc kubenswrapper[4764]: I0202 09:12:00.369045 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Feb 02 09:12:00 crc kubenswrapper[4764]: I0202 09:12:00.397547 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Feb 02 09:12:00 crc kubenswrapper[4764]: I0202 09:12:00.426620 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Feb 02 09:12:00 crc kubenswrapper[4764]: I0202 09:12:00.447061 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Feb 02 09:12:00 crc kubenswrapper[4764]: I0202 09:12:00.473777 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Feb 02 09:12:00 crc kubenswrapper[4764]: I0202 09:12:00.482911 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Feb 02 09:12:00 crc kubenswrapper[4764]: I0202 09:12:00.548649 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Feb 02 09:12:00 crc kubenswrapper[4764]: I0202 09:12:00.709289 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Feb 02 09:12:00 crc kubenswrapper[4764]: I0202 09:12:00.754958 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Feb 02 09:12:00 crc kubenswrapper[4764]: I0202 09:12:00.815611 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Feb 02 09:12:00 crc kubenswrapper[4764]: I0202 09:12:00.901706 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Feb 02 09:12:00 crc kubenswrapper[4764]: I0202 09:12:00.947786 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Feb 02 09:12:00 crc kubenswrapper[4764]: I0202 09:12:00.995712 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Feb 02 09:12:01 crc kubenswrapper[4764]: I0202 09:12:01.058017 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Feb 02 09:12:01 crc kubenswrapper[4764]: I0202 09:12:01.059379 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Feb 02 09:12:01 crc kubenswrapper[4764]: I0202 09:12:01.134550 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Feb 02 09:12:01 crc kubenswrapper[4764]: I0202 09:12:01.156708 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Feb 02 09:12:01 crc kubenswrapper[4764]: I0202 09:12:01.200045 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Feb 02 09:12:01 crc kubenswrapper[4764]: I0202 09:12:01.207105 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Feb 02 09:12:01 crc kubenswrapper[4764]: I0202 09:12:01.340874 4764 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Feb 02 09:12:01 crc kubenswrapper[4764]: I0202 09:12:01.374900 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Feb 02 09:12:01 crc kubenswrapper[4764]: I0202 09:12:01.383911 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Feb 02 09:12:01 crc kubenswrapper[4764]: I0202 09:12:01.492921 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Feb 02 09:12:01 crc kubenswrapper[4764]: I0202 09:12:01.570120 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Feb 02 09:12:01 crc kubenswrapper[4764]: I0202 09:12:01.624478 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Feb 02 09:12:01 crc kubenswrapper[4764]: I0202 09:12:01.624481 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Feb 02 09:12:01 crc kubenswrapper[4764]: I0202 09:12:01.650682 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Feb 02 09:12:01 crc kubenswrapper[4764]: I0202 09:12:01.727632 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Feb 02 09:12:01 crc kubenswrapper[4764]: I0202 09:12:01.859814 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Feb 02 09:12:01 crc kubenswrapper[4764]: I0202 09:12:01.960402 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Feb 02 09:12:01 crc kubenswrapper[4764]: I0202 09:12:01.988301 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Feb 02 09:12:02 crc kubenswrapper[4764]: I0202 09:12:02.004078 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Feb 02 09:12:02 crc kubenswrapper[4764]: I0202 09:12:02.057247 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Feb 02 09:12:02 crc kubenswrapper[4764]: I0202 09:12:02.125867 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Feb 02 09:12:02 crc kubenswrapper[4764]: I0202 09:12:02.131756 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Feb 02 09:12:02 crc kubenswrapper[4764]: I0202 09:12:02.150263 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Feb 02 09:12:02 crc kubenswrapper[4764]: I0202 09:12:02.172976 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Feb 02 09:12:02 crc kubenswrapper[4764]: I0202 09:12:02.255589 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Feb 02 09:12:02 crc kubenswrapper[4764]: I0202 09:12:02.328083 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Feb 02 09:12:02 crc kubenswrapper[4764]: I0202 09:12:02.548147 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Feb 02 09:12:02 crc kubenswrapper[4764]: I0202 09:12:02.580627 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Feb 02 09:12:02 crc kubenswrapper[4764]: I0202 09:12:02.585287 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Feb 02 09:12:02 crc kubenswrapper[4764]: I0202 09:12:02.607719 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Feb 02 09:12:02 crc kubenswrapper[4764]: I0202 09:12:02.717657 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Feb 02 09:12:02 crc kubenswrapper[4764]: I0202 09:12:02.850171 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Feb 02 09:12:02 crc kubenswrapper[4764]: I0202 09:12:02.897621 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Feb 02 09:12:02 crc kubenswrapper[4764]: I0202 09:12:02.915152 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Feb 02 09:12:02 crc kubenswrapper[4764]: I0202 09:12:02.930584 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Feb 02 09:12:02 crc kubenswrapper[4764]: I0202 09:12:02.933739 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Feb 02 09:12:02 crc kubenswrapper[4764]: I0202 09:12:02.947911 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Feb 02 09:12:02 crc kubenswrapper[4764]: I0202 09:12:02.975771 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Feb 02 09:12:03 crc kubenswrapper[4764]: I0202 09:12:03.055487 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Feb 02 09:12:03 crc kubenswrapper[4764]: I0202 09:12:03.089370 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Feb 02 09:12:03 crc kubenswrapper[4764]: I0202 09:12:03.093812 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Feb 02 09:12:03 crc kubenswrapper[4764]: I0202 09:12:03.095478 4764 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Feb 02 09:12:03 crc kubenswrapper[4764]: I0202 09:12:03.095684 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://80d86a89e0b9d62e35dea277f2821831c8e2e121c28458b00dc0dcfe4d734219" gracePeriod=5 Feb 02 09:12:03 crc kubenswrapper[4764]: I0202 09:12:03.098227 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Feb 02 09:12:03 crc kubenswrapper[4764]: I0202 09:12:03.120682 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Feb 02 09:12:03 crc kubenswrapper[4764]: I0202 09:12:03.180269 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Feb 02 09:12:03 crc kubenswrapper[4764]: I0202 09:12:03.216900 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Feb 02 09:12:03 crc kubenswrapper[4764]: I0202 09:12:03.220678 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Feb 02 09:12:03 crc kubenswrapper[4764]: I0202 09:12:03.287343 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-756fdfd6b5-w529b"] Feb 02 09:12:03 crc kubenswrapper[4764]: E0202 09:12:03.287994 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6df65ba-ae7e-48e4-aa24-c5c00b23d4aa" containerName="installer" Feb 02 09:12:03 crc kubenswrapper[4764]: I0202 09:12:03.288130 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6df65ba-ae7e-48e4-aa24-c5c00b23d4aa" containerName="installer" Feb 02 09:12:03 crc kubenswrapper[4764]: E0202 09:12:03.288256 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="513ecde9-f189-4a22-8742-1b262523d462" containerName="oauth-openshift" Feb 02 09:12:03 crc kubenswrapper[4764]: I0202 09:12:03.288387 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="513ecde9-f189-4a22-8742-1b262523d462" containerName="oauth-openshift" Feb 02 09:12:03 crc kubenswrapper[4764]: E0202 09:12:03.288519 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Feb 02 09:12:03 crc kubenswrapper[4764]: I0202 09:12:03.288641 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Feb 02 09:12:03 crc kubenswrapper[4764]: I0202 09:12:03.288913 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="513ecde9-f189-4a22-8742-1b262523d462" containerName="oauth-openshift" Feb 02 09:12:03 crc kubenswrapper[4764]: I0202 09:12:03.289076 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6df65ba-ae7e-48e4-aa24-c5c00b23d4aa" containerName="installer" Feb 02 09:12:03 crc kubenswrapper[4764]: I0202 09:12:03.289184 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Feb 02 09:12:03 crc kubenswrapper[4764]: I0202 09:12:03.289841 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-756fdfd6b5-w529b" Feb 02 09:12:03 crc kubenswrapper[4764]: I0202 09:12:03.295049 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Feb 02 09:12:03 crc kubenswrapper[4764]: I0202 09:12:03.295846 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Feb 02 09:12:03 crc kubenswrapper[4764]: I0202 09:12:03.296005 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Feb 02 09:12:03 crc kubenswrapper[4764]: I0202 09:12:03.296245 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Feb 02 09:12:03 crc kubenswrapper[4764]: I0202 09:12:03.298282 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Feb 02 09:12:03 crc kubenswrapper[4764]: I0202 09:12:03.298557 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Feb 02 09:12:03 crc kubenswrapper[4764]: I0202 09:12:03.298998 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Feb 02 09:12:03 crc kubenswrapper[4764]: I0202 09:12:03.299274 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Feb 02 09:12:03 crc kubenswrapper[4764]: I0202 09:12:03.299616 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Feb 02 09:12:03 crc kubenswrapper[4764]: I0202 09:12:03.299622 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Feb 02 09:12:03 crc kubenswrapper[4764]: I0202 09:12:03.299792 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Feb 02 09:12:03 crc kubenswrapper[4764]: I0202 09:12:03.301647 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-756fdfd6b5-w529b"] Feb 02 09:12:03 crc kubenswrapper[4764]: I0202 09:12:03.303212 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Feb 02 09:12:03 crc kubenswrapper[4764]: I0202 09:12:03.303703 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Feb 02 09:12:03 crc kubenswrapper[4764]: I0202 09:12:03.316618 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Feb 02 09:12:03 crc kubenswrapper[4764]: I0202 09:12:03.324858 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Feb 02 09:12:03 crc kubenswrapper[4764]: I0202 09:12:03.335833 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Feb 02 09:12:03 crc kubenswrapper[4764]: I0202 09:12:03.335130 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Feb 02 09:12:03 crc kubenswrapper[4764]: I0202 09:12:03.358890 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Feb 02 09:12:03 crc kubenswrapper[4764]: I0202 09:12:03.372016 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/dc6ed97a-fb02-4bb8-a363-af06ea7b5e03-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-756fdfd6b5-w529b\" (UID: \"dc6ed97a-fb02-4bb8-a363-af06ea7b5e03\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-w529b" Feb 02 09:12:03 crc kubenswrapper[4764]: I0202 09:12:03.372293 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/dc6ed97a-fb02-4bb8-a363-af06ea7b5e03-v4-0-config-system-cliconfig\") pod \"oauth-openshift-756fdfd6b5-w529b\" (UID: \"dc6ed97a-fb02-4bb8-a363-af06ea7b5e03\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-w529b" Feb 02 09:12:03 crc kubenswrapper[4764]: I0202 09:12:03.372449 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6blxx\" (UniqueName: \"kubernetes.io/projected/dc6ed97a-fb02-4bb8-a363-af06ea7b5e03-kube-api-access-6blxx\") pod \"oauth-openshift-756fdfd6b5-w529b\" (UID: \"dc6ed97a-fb02-4bb8-a363-af06ea7b5e03\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-w529b" Feb 02 09:12:03 crc kubenswrapper[4764]: I0202 09:12:03.372581 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/dc6ed97a-fb02-4bb8-a363-af06ea7b5e03-audit-dir\") pod \"oauth-openshift-756fdfd6b5-w529b\" (UID: \"dc6ed97a-fb02-4bb8-a363-af06ea7b5e03\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-w529b" Feb 02 09:12:03 crc kubenswrapper[4764]: I0202 09:12:03.372725 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/dc6ed97a-fb02-4bb8-a363-af06ea7b5e03-v4-0-config-user-template-error\") pod \"oauth-openshift-756fdfd6b5-w529b\" (UID: \"dc6ed97a-fb02-4bb8-a363-af06ea7b5e03\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-w529b" Feb 02 09:12:03 crc kubenswrapper[4764]: I0202 09:12:03.372847 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/dc6ed97a-fb02-4bb8-a363-af06ea7b5e03-v4-0-config-system-serving-cert\") pod \"oauth-openshift-756fdfd6b5-w529b\" (UID: \"dc6ed97a-fb02-4bb8-a363-af06ea7b5e03\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-w529b" Feb 02 09:12:03 crc kubenswrapper[4764]: I0202 09:12:03.373004 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/dc6ed97a-fb02-4bb8-a363-af06ea7b5e03-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-756fdfd6b5-w529b\" (UID: \"dc6ed97a-fb02-4bb8-a363-af06ea7b5e03\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-w529b" Feb 02 09:12:03 crc kubenswrapper[4764]: I0202 09:12:03.373132 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/dc6ed97a-fb02-4bb8-a363-af06ea7b5e03-audit-policies\") pod \"oauth-openshift-756fdfd6b5-w529b\" (UID: \"dc6ed97a-fb02-4bb8-a363-af06ea7b5e03\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-w529b" Feb 02 09:12:03 crc kubenswrapper[4764]: I0202 09:12:03.373320 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/dc6ed97a-fb02-4bb8-a363-af06ea7b5e03-v4-0-config-system-session\") pod \"oauth-openshift-756fdfd6b5-w529b\" (UID: \"dc6ed97a-fb02-4bb8-a363-af06ea7b5e03\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-w529b" Feb 02 09:12:03 crc kubenswrapper[4764]: I0202 09:12:03.373443 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/dc6ed97a-fb02-4bb8-a363-af06ea7b5e03-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-756fdfd6b5-w529b\" (UID: \"dc6ed97a-fb02-4bb8-a363-af06ea7b5e03\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-w529b" Feb 02 09:12:03 crc kubenswrapper[4764]: I0202 09:12:03.373584 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/dc6ed97a-fb02-4bb8-a363-af06ea7b5e03-v4-0-config-system-router-certs\") pod \"oauth-openshift-756fdfd6b5-w529b\" (UID: \"dc6ed97a-fb02-4bb8-a363-af06ea7b5e03\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-w529b" Feb 02 09:12:03 crc kubenswrapper[4764]: I0202 09:12:03.373720 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/dc6ed97a-fb02-4bb8-a363-af06ea7b5e03-v4-0-config-user-template-login\") pod \"oauth-openshift-756fdfd6b5-w529b\" (UID: \"dc6ed97a-fb02-4bb8-a363-af06ea7b5e03\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-w529b" Feb 02 09:12:03 crc kubenswrapper[4764]: I0202 09:12:03.373850 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/dc6ed97a-fb02-4bb8-a363-af06ea7b5e03-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-756fdfd6b5-w529b\" (UID: \"dc6ed97a-fb02-4bb8-a363-af06ea7b5e03\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-w529b" Feb 02 09:12:03 crc kubenswrapper[4764]: I0202 09:12:03.374227 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/dc6ed97a-fb02-4bb8-a363-af06ea7b5e03-v4-0-config-system-service-ca\") pod \"oauth-openshift-756fdfd6b5-w529b\" (UID: \"dc6ed97a-fb02-4bb8-a363-af06ea7b5e03\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-w529b" Feb 02 09:12:03 crc kubenswrapper[4764]: I0202 09:12:03.475723 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/dc6ed97a-fb02-4bb8-a363-af06ea7b5e03-v4-0-config-system-router-certs\") pod \"oauth-openshift-756fdfd6b5-w529b\" (UID: \"dc6ed97a-fb02-4bb8-a363-af06ea7b5e03\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-w529b" Feb 02 09:12:03 crc kubenswrapper[4764]: I0202 09:12:03.475808 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/dc6ed97a-fb02-4bb8-a363-af06ea7b5e03-v4-0-config-user-template-login\") pod \"oauth-openshift-756fdfd6b5-w529b\" (UID: \"dc6ed97a-fb02-4bb8-a363-af06ea7b5e03\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-w529b" Feb 02 09:12:03 crc kubenswrapper[4764]: I0202 09:12:03.475856 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/dc6ed97a-fb02-4bb8-a363-af06ea7b5e03-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-756fdfd6b5-w529b\" (UID: \"dc6ed97a-fb02-4bb8-a363-af06ea7b5e03\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-w529b" Feb 02 09:12:03 crc kubenswrapper[4764]: I0202 09:12:03.475895 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/dc6ed97a-fb02-4bb8-a363-af06ea7b5e03-v4-0-config-system-service-ca\") pod \"oauth-openshift-756fdfd6b5-w529b\" (UID: \"dc6ed97a-fb02-4bb8-a363-af06ea7b5e03\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-w529b" Feb 02 09:12:03 crc kubenswrapper[4764]: I0202 09:12:03.475960 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/dc6ed97a-fb02-4bb8-a363-af06ea7b5e03-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-756fdfd6b5-w529b\" (UID: \"dc6ed97a-fb02-4bb8-a363-af06ea7b5e03\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-w529b" Feb 02 09:12:03 crc kubenswrapper[4764]: I0202 09:12:03.475997 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/dc6ed97a-fb02-4bb8-a363-af06ea7b5e03-v4-0-config-system-cliconfig\") pod \"oauth-openshift-756fdfd6b5-w529b\" (UID: \"dc6ed97a-fb02-4bb8-a363-af06ea7b5e03\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-w529b" Feb 02 09:12:03 crc kubenswrapper[4764]: I0202 09:12:03.476033 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6blxx\" (UniqueName: \"kubernetes.io/projected/dc6ed97a-fb02-4bb8-a363-af06ea7b5e03-kube-api-access-6blxx\") pod \"oauth-openshift-756fdfd6b5-w529b\" (UID: \"dc6ed97a-fb02-4bb8-a363-af06ea7b5e03\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-w529b" Feb 02 09:12:03 crc kubenswrapper[4764]: I0202 09:12:03.476067 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/dc6ed97a-fb02-4bb8-a363-af06ea7b5e03-audit-dir\") pod \"oauth-openshift-756fdfd6b5-w529b\" (UID: \"dc6ed97a-fb02-4bb8-a363-af06ea7b5e03\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-w529b" Feb 02 09:12:03 crc kubenswrapper[4764]: I0202 09:12:03.476124 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/dc6ed97a-fb02-4bb8-a363-af06ea7b5e03-v4-0-config-user-template-error\") pod \"oauth-openshift-756fdfd6b5-w529b\" (UID: \"dc6ed97a-fb02-4bb8-a363-af06ea7b5e03\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-w529b" Feb 02 09:12:03 crc kubenswrapper[4764]: I0202 09:12:03.476172 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/dc6ed97a-fb02-4bb8-a363-af06ea7b5e03-v4-0-config-system-serving-cert\") pod \"oauth-openshift-756fdfd6b5-w529b\" (UID: \"dc6ed97a-fb02-4bb8-a363-af06ea7b5e03\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-w529b" Feb 02 09:12:03 crc kubenswrapper[4764]: I0202 09:12:03.476214 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/dc6ed97a-fb02-4bb8-a363-af06ea7b5e03-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-756fdfd6b5-w529b\" (UID: \"dc6ed97a-fb02-4bb8-a363-af06ea7b5e03\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-w529b" Feb 02 09:12:03 crc kubenswrapper[4764]: I0202 09:12:03.476248 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/dc6ed97a-fb02-4bb8-a363-af06ea7b5e03-audit-policies\") pod \"oauth-openshift-756fdfd6b5-w529b\" (UID: \"dc6ed97a-fb02-4bb8-a363-af06ea7b5e03\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-w529b" Feb 02 09:12:03 crc kubenswrapper[4764]: I0202 09:12:03.476283 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/dc6ed97a-fb02-4bb8-a363-af06ea7b5e03-v4-0-config-system-session\") pod \"oauth-openshift-756fdfd6b5-w529b\" (UID: \"dc6ed97a-fb02-4bb8-a363-af06ea7b5e03\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-w529b" Feb 02 09:12:03 crc kubenswrapper[4764]: I0202 09:12:03.476316 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/dc6ed97a-fb02-4bb8-a363-af06ea7b5e03-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-756fdfd6b5-w529b\" (UID: \"dc6ed97a-fb02-4bb8-a363-af06ea7b5e03\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-w529b" Feb 02 09:12:03 crc kubenswrapper[4764]: I0202 09:12:03.478534 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/dc6ed97a-fb02-4bb8-a363-af06ea7b5e03-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-756fdfd6b5-w529b\" (UID: \"dc6ed97a-fb02-4bb8-a363-af06ea7b5e03\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-w529b" Feb 02 09:12:03 crc kubenswrapper[4764]: I0202 09:12:03.479546 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/dc6ed97a-fb02-4bb8-a363-af06ea7b5e03-v4-0-config-system-cliconfig\") pod \"oauth-openshift-756fdfd6b5-w529b\" (UID: \"dc6ed97a-fb02-4bb8-a363-af06ea7b5e03\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-w529b" Feb 02 09:12:03 crc kubenswrapper[4764]: I0202 09:12:03.480560 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/dc6ed97a-fb02-4bb8-a363-af06ea7b5e03-audit-policies\") pod \"oauth-openshift-756fdfd6b5-w529b\" (UID: \"dc6ed97a-fb02-4bb8-a363-af06ea7b5e03\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-w529b" Feb 02 09:12:03 crc kubenswrapper[4764]: I0202 09:12:03.482043 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/dc6ed97a-fb02-4bb8-a363-af06ea7b5e03-v4-0-config-system-service-ca\") pod \"oauth-openshift-756fdfd6b5-w529b\" (UID: \"dc6ed97a-fb02-4bb8-a363-af06ea7b5e03\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-w529b" Feb 02 09:12:03 crc kubenswrapper[4764]: I0202 09:12:03.482134 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/dc6ed97a-fb02-4bb8-a363-af06ea7b5e03-audit-dir\") pod \"oauth-openshift-756fdfd6b5-w529b\" (UID: \"dc6ed97a-fb02-4bb8-a363-af06ea7b5e03\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-w529b" Feb 02 09:12:03 crc kubenswrapper[4764]: I0202 09:12:03.485688 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/dc6ed97a-fb02-4bb8-a363-af06ea7b5e03-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-756fdfd6b5-w529b\" (UID: \"dc6ed97a-fb02-4bb8-a363-af06ea7b5e03\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-w529b" Feb 02 09:12:03 crc kubenswrapper[4764]: I0202 09:12:03.485992 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/dc6ed97a-fb02-4bb8-a363-af06ea7b5e03-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-756fdfd6b5-w529b\" (UID: \"dc6ed97a-fb02-4bb8-a363-af06ea7b5e03\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-w529b" Feb 02 09:12:03 crc kubenswrapper[4764]: I0202 09:12:03.486450 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/dc6ed97a-fb02-4bb8-a363-af06ea7b5e03-v4-0-config-system-router-certs\") pod \"oauth-openshift-756fdfd6b5-w529b\" (UID: \"dc6ed97a-fb02-4bb8-a363-af06ea7b5e03\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-w529b" Feb 02 09:12:03 crc kubenswrapper[4764]: I0202 09:12:03.488868 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/dc6ed97a-fb02-4bb8-a363-af06ea7b5e03-v4-0-config-user-template-login\") pod \"oauth-openshift-756fdfd6b5-w529b\" (UID: \"dc6ed97a-fb02-4bb8-a363-af06ea7b5e03\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-w529b" Feb 02 09:12:03 crc kubenswrapper[4764]: I0202 09:12:03.490503 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/dc6ed97a-fb02-4bb8-a363-af06ea7b5e03-v4-0-config-system-serving-cert\") pod \"oauth-openshift-756fdfd6b5-w529b\" (UID: \"dc6ed97a-fb02-4bb8-a363-af06ea7b5e03\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-w529b" Feb 02 09:12:03 crc kubenswrapper[4764]: I0202 09:12:03.491693 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/dc6ed97a-fb02-4bb8-a363-af06ea7b5e03-v4-0-config-system-session\") pod \"oauth-openshift-756fdfd6b5-w529b\" (UID: \"dc6ed97a-fb02-4bb8-a363-af06ea7b5e03\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-w529b" Feb 02 09:12:03 crc kubenswrapper[4764]: I0202 09:12:03.493900 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/dc6ed97a-fb02-4bb8-a363-af06ea7b5e03-v4-0-config-user-template-error\") pod \"oauth-openshift-756fdfd6b5-w529b\" (UID: \"dc6ed97a-fb02-4bb8-a363-af06ea7b5e03\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-w529b" Feb 02 09:12:03 crc kubenswrapper[4764]: I0202 09:12:03.497402 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/dc6ed97a-fb02-4bb8-a363-af06ea7b5e03-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-756fdfd6b5-w529b\" (UID: \"dc6ed97a-fb02-4bb8-a363-af06ea7b5e03\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-w529b" Feb 02 09:12:03 crc kubenswrapper[4764]: I0202 09:12:03.499731 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6blxx\" (UniqueName: \"kubernetes.io/projected/dc6ed97a-fb02-4bb8-a363-af06ea7b5e03-kube-api-access-6blxx\") pod \"oauth-openshift-756fdfd6b5-w529b\" (UID: \"dc6ed97a-fb02-4bb8-a363-af06ea7b5e03\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-w529b" Feb 02 09:12:03 crc kubenswrapper[4764]: I0202 09:12:03.554025 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Feb 02 09:12:03 crc kubenswrapper[4764]: I0202 09:12:03.608637 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Feb 02 09:12:03 crc kubenswrapper[4764]: I0202 09:12:03.609033 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-756fdfd6b5-w529b" Feb 02 09:12:03 crc kubenswrapper[4764]: I0202 09:12:03.622282 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Feb 02 09:12:03 crc kubenswrapper[4764]: I0202 09:12:03.663715 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Feb 02 09:12:03 crc kubenswrapper[4764]: I0202 09:12:03.749367 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Feb 02 09:12:03 crc kubenswrapper[4764]: I0202 09:12:03.909020 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Feb 02 09:12:03 crc kubenswrapper[4764]: I0202 09:12:03.911900 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Feb 02 09:12:03 crc kubenswrapper[4764]: I0202 09:12:03.919650 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Feb 02 09:12:04 crc kubenswrapper[4764]: I0202 09:12:04.050250 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Feb 02 09:12:04 crc kubenswrapper[4764]: I0202 09:12:04.054282 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Feb 02 09:12:04 crc kubenswrapper[4764]: I0202 09:12:04.073350 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-756fdfd6b5-w529b"] Feb 02 09:12:04 crc kubenswrapper[4764]: I0202 09:12:04.073508 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Feb 02 09:12:04 crc kubenswrapper[4764]: I0202 09:12:04.134409 4764 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Feb 02 09:12:04 crc kubenswrapper[4764]: I0202 09:12:04.134538 4764 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Feb 02 09:12:04 crc kubenswrapper[4764]: I0202 09:12:04.134629 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 02 09:12:04 crc kubenswrapper[4764]: I0202 09:12:04.135894 4764 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="kube-controller-manager" containerStatusID={"Type":"cri-o","ID":"c616c013433ef912ef5ff270bd79e057373450d24844a0aab5fdb2bc3f6884dc"} pod="openshift-kube-controller-manager/kube-controller-manager-crc" containerMessage="Container kube-controller-manager failed startup probe, will be restarted" Feb 02 09:12:04 crc kubenswrapper[4764]: I0202 09:12:04.136162 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" containerID="cri-o://c616c013433ef912ef5ff270bd79e057373450d24844a0aab5fdb2bc3f6884dc" gracePeriod=30 Feb 02 09:12:04 crc kubenswrapper[4764]: I0202 09:12:04.136650 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Feb 02 09:12:04 crc kubenswrapper[4764]: I0202 09:12:04.157805 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Feb 02 09:12:04 crc kubenswrapper[4764]: I0202 09:12:04.168912 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Feb 02 09:12:04 crc kubenswrapper[4764]: I0202 09:12:04.257578 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Feb 02 09:12:04 crc kubenswrapper[4764]: I0202 09:12:04.284685 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Feb 02 09:12:04 crc kubenswrapper[4764]: I0202 09:12:04.309912 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Feb 02 09:12:04 crc kubenswrapper[4764]: I0202 09:12:04.325337 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Feb 02 09:12:04 crc kubenswrapper[4764]: I0202 09:12:04.327551 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Feb 02 09:12:04 crc kubenswrapper[4764]: I0202 09:12:04.343200 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Feb 02 09:12:04 crc kubenswrapper[4764]: I0202 09:12:04.415303 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Feb 02 09:12:04 crc kubenswrapper[4764]: I0202 09:12:04.542782 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Feb 02 09:12:04 crc kubenswrapper[4764]: I0202 09:12:04.618143 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Feb 02 09:12:04 crc kubenswrapper[4764]: I0202 09:12:04.748299 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Feb 02 09:12:04 crc kubenswrapper[4764]: I0202 09:12:04.915833 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Feb 02 09:12:04 crc kubenswrapper[4764]: I0202 09:12:04.937286 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-756fdfd6b5-w529b" event={"ID":"dc6ed97a-fb02-4bb8-a363-af06ea7b5e03","Type":"ContainerStarted","Data":"1dd7eec1c3b407f939154dbe4c07c7a3a5a6daddd9a186d291a44c9a614089d3"} Feb 02 09:12:04 crc kubenswrapper[4764]: I0202 09:12:04.937330 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-756fdfd6b5-w529b" event={"ID":"dc6ed97a-fb02-4bb8-a363-af06ea7b5e03","Type":"ContainerStarted","Data":"afccde5524cdb83a4a2b88d285d627f83f1f481d564accd3b38e3feb8cd59f34"} Feb 02 09:12:04 crc kubenswrapper[4764]: I0202 09:12:04.938915 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-756fdfd6b5-w529b" Feb 02 09:12:04 crc kubenswrapper[4764]: I0202 09:12:04.947182 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-756fdfd6b5-w529b" Feb 02 09:12:04 crc kubenswrapper[4764]: I0202 09:12:04.969450 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-756fdfd6b5-w529b" podStartSLOduration=64.969434188 podStartE2EDuration="1m4.969434188s" podCreationTimestamp="2026-02-02 09:11:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:12:04.965256906 +0000 UTC m=+287.898980994" watchObservedRunningTime="2026-02-02 09:12:04.969434188 +0000 UTC m=+287.903158276" Feb 02 09:12:05 crc kubenswrapper[4764]: I0202 09:12:05.020837 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Feb 02 09:12:05 crc kubenswrapper[4764]: I0202 09:12:05.028198 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Feb 02 09:12:05 crc kubenswrapper[4764]: I0202 09:12:05.094688 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Feb 02 09:12:05 crc kubenswrapper[4764]: I0202 09:12:05.099853 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Feb 02 09:12:05 crc kubenswrapper[4764]: I0202 09:12:05.159342 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Feb 02 09:12:05 crc kubenswrapper[4764]: I0202 09:12:05.164365 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Feb 02 09:12:05 crc kubenswrapper[4764]: I0202 09:12:05.226738 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Feb 02 09:12:05 crc kubenswrapper[4764]: I0202 09:12:05.458119 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Feb 02 09:12:05 crc kubenswrapper[4764]: I0202 09:12:05.724693 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Feb 02 09:12:05 crc kubenswrapper[4764]: I0202 09:12:05.881773 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Feb 02 09:12:06 crc kubenswrapper[4764]: I0202 09:12:06.098598 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Feb 02 09:12:06 crc kubenswrapper[4764]: I0202 09:12:06.212541 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Feb 02 09:12:06 crc kubenswrapper[4764]: I0202 09:12:06.257222 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Feb 02 09:12:06 crc kubenswrapper[4764]: I0202 09:12:06.298734 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Feb 02 09:12:06 crc kubenswrapper[4764]: I0202 09:12:06.401631 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Feb 02 09:12:06 crc kubenswrapper[4764]: I0202 09:12:06.504308 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Feb 02 09:12:06 crc kubenswrapper[4764]: I0202 09:12:06.701980 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Feb 02 09:12:06 crc kubenswrapper[4764]: I0202 09:12:06.771437 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Feb 02 09:12:07 crc kubenswrapper[4764]: I0202 09:12:07.294786 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Feb 02 09:12:07 crc kubenswrapper[4764]: I0202 09:12:07.310320 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Feb 02 09:12:07 crc kubenswrapper[4764]: I0202 09:12:07.407342 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Feb 02 09:12:07 crc kubenswrapper[4764]: I0202 09:12:07.619601 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Feb 02 09:12:07 crc kubenswrapper[4764]: I0202 09:12:07.683160 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Feb 02 09:12:07 crc kubenswrapper[4764]: I0202 09:12:07.908472 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Feb 02 09:12:08 crc kubenswrapper[4764]: I0202 09:12:08.226020 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Feb 02 09:12:08 crc kubenswrapper[4764]: I0202 09:12:08.226468 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 02 09:12:08 crc kubenswrapper[4764]: I0202 09:12:08.343734 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 02 09:12:08 crc kubenswrapper[4764]: I0202 09:12:08.343783 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 02 09:12:08 crc kubenswrapper[4764]: I0202 09:12:08.343836 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 02 09:12:08 crc kubenswrapper[4764]: I0202 09:12:08.343861 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 02 09:12:08 crc kubenswrapper[4764]: I0202 09:12:08.343852 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 09:12:08 crc kubenswrapper[4764]: I0202 09:12:08.343899 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 02 09:12:08 crc kubenswrapper[4764]: I0202 09:12:08.343915 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 09:12:08 crc kubenswrapper[4764]: I0202 09:12:08.343960 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 09:12:08 crc kubenswrapper[4764]: I0202 09:12:08.343995 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 09:12:08 crc kubenswrapper[4764]: I0202 09:12:08.344201 4764 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Feb 02 09:12:08 crc kubenswrapper[4764]: I0202 09:12:08.344219 4764 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Feb 02 09:12:08 crc kubenswrapper[4764]: I0202 09:12:08.344229 4764 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Feb 02 09:12:08 crc kubenswrapper[4764]: I0202 09:12:08.344239 4764 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Feb 02 09:12:08 crc kubenswrapper[4764]: I0202 09:12:08.354757 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 09:12:08 crc kubenswrapper[4764]: I0202 09:12:08.445623 4764 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Feb 02 09:12:08 crc kubenswrapper[4764]: I0202 09:12:08.977919 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Feb 02 09:12:08 crc kubenswrapper[4764]: I0202 09:12:08.978038 4764 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="80d86a89e0b9d62e35dea277f2821831c8e2e121c28458b00dc0dcfe4d734219" exitCode=137 Feb 02 09:12:08 crc kubenswrapper[4764]: I0202 09:12:08.978112 4764 scope.go:117] "RemoveContainer" containerID="80d86a89e0b9d62e35dea277f2821831c8e2e121c28458b00dc0dcfe4d734219" Feb 02 09:12:08 crc kubenswrapper[4764]: I0202 09:12:08.978158 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 02 09:12:09 crc kubenswrapper[4764]: I0202 09:12:09.002654 4764 scope.go:117] "RemoveContainer" containerID="80d86a89e0b9d62e35dea277f2821831c8e2e121c28458b00dc0dcfe4d734219" Feb 02 09:12:09 crc kubenswrapper[4764]: E0202 09:12:09.006179 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"80d86a89e0b9d62e35dea277f2821831c8e2e121c28458b00dc0dcfe4d734219\": container with ID starting with 80d86a89e0b9d62e35dea277f2821831c8e2e121c28458b00dc0dcfe4d734219 not found: ID does not exist" containerID="80d86a89e0b9d62e35dea277f2821831c8e2e121c28458b00dc0dcfe4d734219" Feb 02 09:12:09 crc kubenswrapper[4764]: I0202 09:12:09.006256 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80d86a89e0b9d62e35dea277f2821831c8e2e121c28458b00dc0dcfe4d734219"} err="failed to get container status \"80d86a89e0b9d62e35dea277f2821831c8e2e121c28458b00dc0dcfe4d734219\": rpc error: code = NotFound desc = could not find container \"80d86a89e0b9d62e35dea277f2821831c8e2e121c28458b00dc0dcfe4d734219\": container with ID starting with 80d86a89e0b9d62e35dea277f2821831c8e2e121c28458b00dc0dcfe4d734219 not found: ID does not exist" Feb 02 09:12:09 crc kubenswrapper[4764]: I0202 09:12:09.385387 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Feb 02 09:12:09 crc kubenswrapper[4764]: I0202 09:12:09.465889 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Feb 02 09:12:09 crc kubenswrapper[4764]: I0202 09:12:09.836624 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Feb 02 09:12:17 crc kubenswrapper[4764]: I0202 09:12:17.641746 4764 cert_rotation.go:91] certificate rotation detected, shutting down client connections to start using new credentials Feb 02 09:12:35 crc kubenswrapper[4764]: I0202 09:12:35.134417 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/1.log" Feb 02 09:12:35 crc kubenswrapper[4764]: I0202 09:12:35.138223 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Feb 02 09:12:35 crc kubenswrapper[4764]: I0202 09:12:35.138296 4764 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="c616c013433ef912ef5ff270bd79e057373450d24844a0aab5fdb2bc3f6884dc" exitCode=137 Feb 02 09:12:35 crc kubenswrapper[4764]: I0202 09:12:35.138338 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"c616c013433ef912ef5ff270bd79e057373450d24844a0aab5fdb2bc3f6884dc"} Feb 02 09:12:35 crc kubenswrapper[4764]: I0202 09:12:35.138374 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"bb455a599374e5f3ea01a4eb5135b484624eb763171f8fcaf814fb72517f6364"} Feb 02 09:12:35 crc kubenswrapper[4764]: I0202 09:12:35.138394 4764 scope.go:117] "RemoveContainer" containerID="60258f84a1d8dcfe48b7e9688ed50f077e201971f20f3f6663fbf1ba58a4611a" Feb 02 09:12:36 crc kubenswrapper[4764]: I0202 09:12:36.147162 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/1.log" Feb 02 09:12:40 crc kubenswrapper[4764]: I0202 09:12:40.324331 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 02 09:12:44 crc kubenswrapper[4764]: I0202 09:12:44.134674 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 02 09:12:44 crc kubenswrapper[4764]: I0202 09:12:44.138854 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 02 09:12:44 crc kubenswrapper[4764]: I0202 09:12:44.192756 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 02 09:13:08 crc kubenswrapper[4764]: I0202 09:13:08.579170 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-j25pg"] Feb 02 09:13:08 crc kubenswrapper[4764]: I0202 09:13:08.580191 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-j25pg" podUID="01ede9d5-6b94-4d10-bf6d-3aef751840ff" containerName="registry-server" containerID="cri-o://18af9361e46e52d7b7a0adc034d3d3d089dd4e8f16f90621792ba3be7ceba8f7" gracePeriod=30 Feb 02 09:13:08 crc kubenswrapper[4764]: I0202 09:13:08.589279 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-ggrpx"] Feb 02 09:13:08 crc kubenswrapper[4764]: I0202 09:13:08.589757 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-ggrpx" podUID="a40abe98-1419-4f5d-9bed-63f1c89028b1" containerName="registry-server" containerID="cri-o://0cbd7d0e3df1db8208d9af33d979c98caaa3f82973cc8263382a73f719638baa" gracePeriod=30 Feb 02 09:13:08 crc kubenswrapper[4764]: I0202 09:13:08.608216 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-wqbtk"] Feb 02 09:13:08 crc kubenswrapper[4764]: I0202 09:13:08.608437 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-wqbtk" podUID="fac965f9-5e91-4c50-a538-48c1301fa072" containerName="marketplace-operator" containerID="cri-o://268bfe0e1d8733e258da407a82ee2f78a0dc2d86ecdbc5bafed848d6f8e77c4f" gracePeriod=30 Feb 02 09:13:08 crc kubenswrapper[4764]: I0202 09:13:08.621253 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-kbk9s"] Feb 02 09:13:08 crc kubenswrapper[4764]: I0202 09:13:08.621574 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-kbk9s" podUID="6092680d-f31e-4b70-9d81-2d80661f6ac3" containerName="registry-server" containerID="cri-o://cae9539080c975e16c67467c3d702f51ce9280b85f572e147d473e1583321863" gracePeriod=30 Feb 02 09:13:08 crc kubenswrapper[4764]: I0202 09:13:08.629630 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-ct55p"] Feb 02 09:13:08 crc kubenswrapper[4764]: I0202 09:13:08.629889 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-ct55p" podUID="00cb5ef3-3b54-47ac-bfc4-b1d53f04d292" containerName="registry-server" containerID="cri-o://b0e277ec4569fc9807eb7c3cc0a8905c030a0b41601d8880193f95814472f82e" gracePeriod=30 Feb 02 09:13:08 crc kubenswrapper[4764]: I0202 09:13:08.642221 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-f4d5w"] Feb 02 09:13:08 crc kubenswrapper[4764]: I0202 09:13:08.643025 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-f4d5w" Feb 02 09:13:08 crc kubenswrapper[4764]: I0202 09:13:08.649401 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-f4d5w"] Feb 02 09:13:08 crc kubenswrapper[4764]: I0202 09:13:08.783245 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/a03d1d75-d8a0-407c-ab02-d151d5ba00b6-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-f4d5w\" (UID: \"a03d1d75-d8a0-407c-ab02-d151d5ba00b6\") " pod="openshift-marketplace/marketplace-operator-79b997595-f4d5w" Feb 02 09:13:08 crc kubenswrapper[4764]: I0202 09:13:08.783311 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a03d1d75-d8a0-407c-ab02-d151d5ba00b6-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-f4d5w\" (UID: \"a03d1d75-d8a0-407c-ab02-d151d5ba00b6\") " pod="openshift-marketplace/marketplace-operator-79b997595-f4d5w" Feb 02 09:13:08 crc kubenswrapper[4764]: I0202 09:13:08.783348 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2bmjk\" (UniqueName: \"kubernetes.io/projected/a03d1d75-d8a0-407c-ab02-d151d5ba00b6-kube-api-access-2bmjk\") pod \"marketplace-operator-79b997595-f4d5w\" (UID: \"a03d1d75-d8a0-407c-ab02-d151d5ba00b6\") " pod="openshift-marketplace/marketplace-operator-79b997595-f4d5w" Feb 02 09:13:08 crc kubenswrapper[4764]: I0202 09:13:08.884827 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/a03d1d75-d8a0-407c-ab02-d151d5ba00b6-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-f4d5w\" (UID: \"a03d1d75-d8a0-407c-ab02-d151d5ba00b6\") " pod="openshift-marketplace/marketplace-operator-79b997595-f4d5w" Feb 02 09:13:08 crc kubenswrapper[4764]: I0202 09:13:08.884868 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a03d1d75-d8a0-407c-ab02-d151d5ba00b6-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-f4d5w\" (UID: \"a03d1d75-d8a0-407c-ab02-d151d5ba00b6\") " pod="openshift-marketplace/marketplace-operator-79b997595-f4d5w" Feb 02 09:13:08 crc kubenswrapper[4764]: I0202 09:13:08.884897 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2bmjk\" (UniqueName: \"kubernetes.io/projected/a03d1d75-d8a0-407c-ab02-d151d5ba00b6-kube-api-access-2bmjk\") pod \"marketplace-operator-79b997595-f4d5w\" (UID: \"a03d1d75-d8a0-407c-ab02-d151d5ba00b6\") " pod="openshift-marketplace/marketplace-operator-79b997595-f4d5w" Feb 02 09:13:08 crc kubenswrapper[4764]: I0202 09:13:08.891836 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a03d1d75-d8a0-407c-ab02-d151d5ba00b6-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-f4d5w\" (UID: \"a03d1d75-d8a0-407c-ab02-d151d5ba00b6\") " pod="openshift-marketplace/marketplace-operator-79b997595-f4d5w" Feb 02 09:13:08 crc kubenswrapper[4764]: I0202 09:13:08.894799 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/a03d1d75-d8a0-407c-ab02-d151d5ba00b6-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-f4d5w\" (UID: \"a03d1d75-d8a0-407c-ab02-d151d5ba00b6\") " pod="openshift-marketplace/marketplace-operator-79b997595-f4d5w" Feb 02 09:13:08 crc kubenswrapper[4764]: I0202 09:13:08.905626 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2bmjk\" (UniqueName: \"kubernetes.io/projected/a03d1d75-d8a0-407c-ab02-d151d5ba00b6-kube-api-access-2bmjk\") pod \"marketplace-operator-79b997595-f4d5w\" (UID: \"a03d1d75-d8a0-407c-ab02-d151d5ba00b6\") " pod="openshift-marketplace/marketplace-operator-79b997595-f4d5w" Feb 02 09:13:08 crc kubenswrapper[4764]: I0202 09:13:08.965309 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-f4d5w" Feb 02 09:13:09 crc kubenswrapper[4764]: I0202 09:13:09.040179 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ggrpx" Feb 02 09:13:09 crc kubenswrapper[4764]: E0202 09:13:09.068611 4764 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of cae9539080c975e16c67467c3d702f51ce9280b85f572e147d473e1583321863 is running failed: container process not found" containerID="cae9539080c975e16c67467c3d702f51ce9280b85f572e147d473e1583321863" cmd=["grpc_health_probe","-addr=:50051"] Feb 02 09:13:09 crc kubenswrapper[4764]: E0202 09:13:09.068946 4764 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of cae9539080c975e16c67467c3d702f51ce9280b85f572e147d473e1583321863 is running failed: container process not found" containerID="cae9539080c975e16c67467c3d702f51ce9280b85f572e147d473e1583321863" cmd=["grpc_health_probe","-addr=:50051"] Feb 02 09:13:09 crc kubenswrapper[4764]: E0202 09:13:09.069225 4764 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of cae9539080c975e16c67467c3d702f51ce9280b85f572e147d473e1583321863 is running failed: container process not found" containerID="cae9539080c975e16c67467c3d702f51ce9280b85f572e147d473e1583321863" cmd=["grpc_health_probe","-addr=:50051"] Feb 02 09:13:09 crc kubenswrapper[4764]: E0202 09:13:09.069259 4764 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of cae9539080c975e16c67467c3d702f51ce9280b85f572e147d473e1583321863 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/redhat-marketplace-kbk9s" podUID="6092680d-f31e-4b70-9d81-2d80661f6ac3" containerName="registry-server" Feb 02 09:13:09 crc kubenswrapper[4764]: I0202 09:13:09.194443 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a40abe98-1419-4f5d-9bed-63f1c89028b1-catalog-content\") pod \"a40abe98-1419-4f5d-9bed-63f1c89028b1\" (UID: \"a40abe98-1419-4f5d-9bed-63f1c89028b1\") " Feb 02 09:13:09 crc kubenswrapper[4764]: I0202 09:13:09.194492 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a40abe98-1419-4f5d-9bed-63f1c89028b1-utilities\") pod \"a40abe98-1419-4f5d-9bed-63f1c89028b1\" (UID: \"a40abe98-1419-4f5d-9bed-63f1c89028b1\") " Feb 02 09:13:09 crc kubenswrapper[4764]: I0202 09:13:09.194568 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-45j59\" (UniqueName: \"kubernetes.io/projected/a40abe98-1419-4f5d-9bed-63f1c89028b1-kube-api-access-45j59\") pod \"a40abe98-1419-4f5d-9bed-63f1c89028b1\" (UID: \"a40abe98-1419-4f5d-9bed-63f1c89028b1\") " Feb 02 09:13:09 crc kubenswrapper[4764]: I0202 09:13:09.195434 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a40abe98-1419-4f5d-9bed-63f1c89028b1-utilities" (OuterVolumeSpecName: "utilities") pod "a40abe98-1419-4f5d-9bed-63f1c89028b1" (UID: "a40abe98-1419-4f5d-9bed-63f1c89028b1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 09:13:09 crc kubenswrapper[4764]: I0202 09:13:09.197085 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ct55p" Feb 02 09:13:09 crc kubenswrapper[4764]: I0202 09:13:09.199153 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a40abe98-1419-4f5d-9bed-63f1c89028b1-kube-api-access-45j59" (OuterVolumeSpecName: "kube-api-access-45j59") pod "a40abe98-1419-4f5d-9bed-63f1c89028b1" (UID: "a40abe98-1419-4f5d-9bed-63f1c89028b1"). InnerVolumeSpecName "kube-api-access-45j59". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:13:09 crc kubenswrapper[4764]: I0202 09:13:09.243448 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-j25pg" Feb 02 09:13:09 crc kubenswrapper[4764]: I0202 09:13:09.250137 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-wqbtk" Feb 02 09:13:09 crc kubenswrapper[4764]: I0202 09:13:09.290416 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a40abe98-1419-4f5d-9bed-63f1c89028b1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a40abe98-1419-4f5d-9bed-63f1c89028b1" (UID: "a40abe98-1419-4f5d-9bed-63f1c89028b1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 09:13:09 crc kubenswrapper[4764]: I0202 09:13:09.300295 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fpp4j\" (UniqueName: \"kubernetes.io/projected/00cb5ef3-3b54-47ac-bfc4-b1d53f04d292-kube-api-access-fpp4j\") pod \"00cb5ef3-3b54-47ac-bfc4-b1d53f04d292\" (UID: \"00cb5ef3-3b54-47ac-bfc4-b1d53f04d292\") " Feb 02 09:13:09 crc kubenswrapper[4764]: I0202 09:13:09.300457 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/00cb5ef3-3b54-47ac-bfc4-b1d53f04d292-utilities\") pod \"00cb5ef3-3b54-47ac-bfc4-b1d53f04d292\" (UID: \"00cb5ef3-3b54-47ac-bfc4-b1d53f04d292\") " Feb 02 09:13:09 crc kubenswrapper[4764]: I0202 09:13:09.300483 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/00cb5ef3-3b54-47ac-bfc4-b1d53f04d292-catalog-content\") pod \"00cb5ef3-3b54-47ac-bfc4-b1d53f04d292\" (UID: \"00cb5ef3-3b54-47ac-bfc4-b1d53f04d292\") " Feb 02 09:13:09 crc kubenswrapper[4764]: I0202 09:13:09.300667 4764 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a40abe98-1419-4f5d-9bed-63f1c89028b1-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 09:13:09 crc kubenswrapper[4764]: I0202 09:13:09.300680 4764 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a40abe98-1419-4f5d-9bed-63f1c89028b1-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 09:13:09 crc kubenswrapper[4764]: I0202 09:13:09.300688 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-45j59\" (UniqueName: \"kubernetes.io/projected/a40abe98-1419-4f5d-9bed-63f1c89028b1-kube-api-access-45j59\") on node \"crc\" DevicePath \"\"" Feb 02 09:13:09 crc kubenswrapper[4764]: I0202 09:13:09.301489 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/00cb5ef3-3b54-47ac-bfc4-b1d53f04d292-utilities" (OuterVolumeSpecName: "utilities") pod "00cb5ef3-3b54-47ac-bfc4-b1d53f04d292" (UID: "00cb5ef3-3b54-47ac-bfc4-b1d53f04d292"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 09:13:09 crc kubenswrapper[4764]: I0202 09:13:09.308070 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/00cb5ef3-3b54-47ac-bfc4-b1d53f04d292-kube-api-access-fpp4j" (OuterVolumeSpecName: "kube-api-access-fpp4j") pod "00cb5ef3-3b54-47ac-bfc4-b1d53f04d292" (UID: "00cb5ef3-3b54-47ac-bfc4-b1d53f04d292"). InnerVolumeSpecName "kube-api-access-fpp4j". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:13:09 crc kubenswrapper[4764]: I0202 09:13:09.331246 4764 generic.go:334] "Generic (PLEG): container finished" podID="01ede9d5-6b94-4d10-bf6d-3aef751840ff" containerID="18af9361e46e52d7b7a0adc034d3d3d089dd4e8f16f90621792ba3be7ceba8f7" exitCode=0 Feb 02 09:13:09 crc kubenswrapper[4764]: I0202 09:13:09.331327 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j25pg" event={"ID":"01ede9d5-6b94-4d10-bf6d-3aef751840ff","Type":"ContainerDied","Data":"18af9361e46e52d7b7a0adc034d3d3d089dd4e8f16f90621792ba3be7ceba8f7"} Feb 02 09:13:09 crc kubenswrapper[4764]: I0202 09:13:09.331350 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j25pg" event={"ID":"01ede9d5-6b94-4d10-bf6d-3aef751840ff","Type":"ContainerDied","Data":"4afd4af7d19135bfa9e0bb99cecbd3a39b149cfe37e4b26c30eaf267df71b3eb"} Feb 02 09:13:09 crc kubenswrapper[4764]: I0202 09:13:09.331383 4764 scope.go:117] "RemoveContainer" containerID="18af9361e46e52d7b7a0adc034d3d3d089dd4e8f16f90621792ba3be7ceba8f7" Feb 02 09:13:09 crc kubenswrapper[4764]: I0202 09:13:09.331495 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-j25pg" Feb 02 09:13:09 crc kubenswrapper[4764]: I0202 09:13:09.339788 4764 generic.go:334] "Generic (PLEG): container finished" podID="a40abe98-1419-4f5d-9bed-63f1c89028b1" containerID="0cbd7d0e3df1db8208d9af33d979c98caaa3f82973cc8263382a73f719638baa" exitCode=0 Feb 02 09:13:09 crc kubenswrapper[4764]: I0202 09:13:09.339836 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ggrpx" Feb 02 09:13:09 crc kubenswrapper[4764]: I0202 09:13:09.339851 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ggrpx" event={"ID":"a40abe98-1419-4f5d-9bed-63f1c89028b1","Type":"ContainerDied","Data":"0cbd7d0e3df1db8208d9af33d979c98caaa3f82973cc8263382a73f719638baa"} Feb 02 09:13:09 crc kubenswrapper[4764]: I0202 09:13:09.339880 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ggrpx" event={"ID":"a40abe98-1419-4f5d-9bed-63f1c89028b1","Type":"ContainerDied","Data":"bf2b4f22879cad3d2120b2aa1607123e3b5cb77bdf8877486020ee617b44a63c"} Feb 02 09:13:09 crc kubenswrapper[4764]: I0202 09:13:09.343194 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kbk9s" Feb 02 09:13:09 crc kubenswrapper[4764]: I0202 09:13:09.344214 4764 generic.go:334] "Generic (PLEG): container finished" podID="00cb5ef3-3b54-47ac-bfc4-b1d53f04d292" containerID="b0e277ec4569fc9807eb7c3cc0a8905c030a0b41601d8880193f95814472f82e" exitCode=0 Feb 02 09:13:09 crc kubenswrapper[4764]: I0202 09:13:09.344264 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ct55p" event={"ID":"00cb5ef3-3b54-47ac-bfc4-b1d53f04d292","Type":"ContainerDied","Data":"b0e277ec4569fc9807eb7c3cc0a8905c030a0b41601d8880193f95814472f82e"} Feb 02 09:13:09 crc kubenswrapper[4764]: I0202 09:13:09.344427 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ct55p" event={"ID":"00cb5ef3-3b54-47ac-bfc4-b1d53f04d292","Type":"ContainerDied","Data":"00838ebc3146464fb7e67e2e6bc4614aea995f37837f9d843754c1c48fecd95b"} Feb 02 09:13:09 crc kubenswrapper[4764]: I0202 09:13:09.344621 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ct55p" Feb 02 09:13:09 crc kubenswrapper[4764]: I0202 09:13:09.353330 4764 generic.go:334] "Generic (PLEG): container finished" podID="6092680d-f31e-4b70-9d81-2d80661f6ac3" containerID="cae9539080c975e16c67467c3d702f51ce9280b85f572e147d473e1583321863" exitCode=0 Feb 02 09:13:09 crc kubenswrapper[4764]: I0202 09:13:09.353404 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kbk9s" event={"ID":"6092680d-f31e-4b70-9d81-2d80661f6ac3","Type":"ContainerDied","Data":"cae9539080c975e16c67467c3d702f51ce9280b85f572e147d473e1583321863"} Feb 02 09:13:09 crc kubenswrapper[4764]: I0202 09:13:09.353635 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kbk9s" Feb 02 09:13:09 crc kubenswrapper[4764]: I0202 09:13:09.354906 4764 generic.go:334] "Generic (PLEG): container finished" podID="fac965f9-5e91-4c50-a538-48c1301fa072" containerID="268bfe0e1d8733e258da407a82ee2f78a0dc2d86ecdbc5bafed848d6f8e77c4f" exitCode=0 Feb 02 09:13:09 crc kubenswrapper[4764]: I0202 09:13:09.355010 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-wqbtk" event={"ID":"fac965f9-5e91-4c50-a538-48c1301fa072","Type":"ContainerDied","Data":"268bfe0e1d8733e258da407a82ee2f78a0dc2d86ecdbc5bafed848d6f8e77c4f"} Feb 02 09:13:09 crc kubenswrapper[4764]: I0202 09:13:09.355031 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-wqbtk" event={"ID":"fac965f9-5e91-4c50-a538-48c1301fa072","Type":"ContainerDied","Data":"9b90e50046dcc02abf4663e6e01f3a315df60680b90f7748e87bcd022716f6a1"} Feb 02 09:13:09 crc kubenswrapper[4764]: I0202 09:13:09.355071 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-wqbtk" Feb 02 09:13:09 crc kubenswrapper[4764]: I0202 09:13:09.367737 4764 scope.go:117] "RemoveContainer" containerID="796d368c28e1658dc2a81024ed86d820b72e2e1f251c1951ede876071b19b223" Feb 02 09:13:09 crc kubenswrapper[4764]: I0202 09:13:09.400601 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-f4d5w"] Feb 02 09:13:09 crc kubenswrapper[4764]: I0202 09:13:09.402031 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-ggrpx"] Feb 02 09:13:09 crc kubenswrapper[4764]: I0202 09:13:09.402216 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/fac965f9-5e91-4c50-a538-48c1301fa072-marketplace-trusted-ca\") pod \"fac965f9-5e91-4c50-a538-48c1301fa072\" (UID: \"fac965f9-5e91-4c50-a538-48c1301fa072\") " Feb 02 09:13:09 crc kubenswrapper[4764]: I0202 09:13:09.402333 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/fac965f9-5e91-4c50-a538-48c1301fa072-marketplace-operator-metrics\") pod \"fac965f9-5e91-4c50-a538-48c1301fa072\" (UID: \"fac965f9-5e91-4c50-a538-48c1301fa072\") " Feb 02 09:13:09 crc kubenswrapper[4764]: I0202 09:13:09.402371 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s9xmw\" (UniqueName: \"kubernetes.io/projected/fac965f9-5e91-4c50-a538-48c1301fa072-kube-api-access-s9xmw\") pod \"fac965f9-5e91-4c50-a538-48c1301fa072\" (UID: \"fac965f9-5e91-4c50-a538-48c1301fa072\") " Feb 02 09:13:09 crc kubenswrapper[4764]: I0202 09:13:09.402389 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/01ede9d5-6b94-4d10-bf6d-3aef751840ff-catalog-content\") pod \"01ede9d5-6b94-4d10-bf6d-3aef751840ff\" (UID: \"01ede9d5-6b94-4d10-bf6d-3aef751840ff\") " Feb 02 09:13:09 crc kubenswrapper[4764]: I0202 09:13:09.402415 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/01ede9d5-6b94-4d10-bf6d-3aef751840ff-utilities\") pod \"01ede9d5-6b94-4d10-bf6d-3aef751840ff\" (UID: \"01ede9d5-6b94-4d10-bf6d-3aef751840ff\") " Feb 02 09:13:09 crc kubenswrapper[4764]: I0202 09:13:09.402431 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-knbhq\" (UniqueName: \"kubernetes.io/projected/01ede9d5-6b94-4d10-bf6d-3aef751840ff-kube-api-access-knbhq\") pod \"01ede9d5-6b94-4d10-bf6d-3aef751840ff\" (UID: \"01ede9d5-6b94-4d10-bf6d-3aef751840ff\") " Feb 02 09:13:09 crc kubenswrapper[4764]: I0202 09:13:09.403288 4764 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/00cb5ef3-3b54-47ac-bfc4-b1d53f04d292-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 09:13:09 crc kubenswrapper[4764]: I0202 09:13:09.403410 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fpp4j\" (UniqueName: \"kubernetes.io/projected/00cb5ef3-3b54-47ac-bfc4-b1d53f04d292-kube-api-access-fpp4j\") on node \"crc\" DevicePath \"\"" Feb 02 09:13:09 crc kubenswrapper[4764]: I0202 09:13:09.405911 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/01ede9d5-6b94-4d10-bf6d-3aef751840ff-utilities" (OuterVolumeSpecName: "utilities") pod "01ede9d5-6b94-4d10-bf6d-3aef751840ff" (UID: "01ede9d5-6b94-4d10-bf6d-3aef751840ff"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 09:13:09 crc kubenswrapper[4764]: I0202 09:13:09.406733 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fac965f9-5e91-4c50-a538-48c1301fa072-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "fac965f9-5e91-4c50-a538-48c1301fa072" (UID: "fac965f9-5e91-4c50-a538-48c1301fa072"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:13:09 crc kubenswrapper[4764]: I0202 09:13:09.407627 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fac965f9-5e91-4c50-a538-48c1301fa072-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "fac965f9-5e91-4c50-a538-48c1301fa072" (UID: "fac965f9-5e91-4c50-a538-48c1301fa072"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:13:09 crc kubenswrapper[4764]: I0202 09:13:09.408307 4764 scope.go:117] "RemoveContainer" containerID="15c8dcdbf7dafd0945e7cee7ea86c7e3c74f639942d8cc81473c5ba785830e7d" Feb 02 09:13:09 crc kubenswrapper[4764]: I0202 09:13:09.411012 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-ggrpx"] Feb 02 09:13:09 crc kubenswrapper[4764]: I0202 09:13:09.419955 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ede9d5-6b94-4d10-bf6d-3aef751840ff-kube-api-access-knbhq" (OuterVolumeSpecName: "kube-api-access-knbhq") pod "01ede9d5-6b94-4d10-bf6d-3aef751840ff" (UID: "01ede9d5-6b94-4d10-bf6d-3aef751840ff"). InnerVolumeSpecName "kube-api-access-knbhq". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:13:09 crc kubenswrapper[4764]: W0202 09:13:09.421144 4764 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda03d1d75_d8a0_407c_ab02_d151d5ba00b6.slice/crio-902a00976ebf5d11d8688225b60c652bad924f19af9c0f2e5388310709340268 WatchSource:0}: Error finding container 902a00976ebf5d11d8688225b60c652bad924f19af9c0f2e5388310709340268: Status 404 returned error can't find the container with id 902a00976ebf5d11d8688225b60c652bad924f19af9c0f2e5388310709340268 Feb 02 09:13:09 crc kubenswrapper[4764]: I0202 09:13:09.422896 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fac965f9-5e91-4c50-a538-48c1301fa072-kube-api-access-s9xmw" (OuterVolumeSpecName: "kube-api-access-s9xmw") pod "fac965f9-5e91-4c50-a538-48c1301fa072" (UID: "fac965f9-5e91-4c50-a538-48c1301fa072"). InnerVolumeSpecName "kube-api-access-s9xmw". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:13:09 crc kubenswrapper[4764]: I0202 09:13:09.434651 4764 scope.go:117] "RemoveContainer" containerID="18af9361e46e52d7b7a0adc034d3d3d089dd4e8f16f90621792ba3be7ceba8f7" Feb 02 09:13:09 crc kubenswrapper[4764]: E0202 09:13:09.435123 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"18af9361e46e52d7b7a0adc034d3d3d089dd4e8f16f90621792ba3be7ceba8f7\": container with ID starting with 18af9361e46e52d7b7a0adc034d3d3d089dd4e8f16f90621792ba3be7ceba8f7 not found: ID does not exist" containerID="18af9361e46e52d7b7a0adc034d3d3d089dd4e8f16f90621792ba3be7ceba8f7" Feb 02 09:13:09 crc kubenswrapper[4764]: I0202 09:13:09.435188 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"18af9361e46e52d7b7a0adc034d3d3d089dd4e8f16f90621792ba3be7ceba8f7"} err="failed to get container status \"18af9361e46e52d7b7a0adc034d3d3d089dd4e8f16f90621792ba3be7ceba8f7\": rpc error: code = NotFound desc = could not find container \"18af9361e46e52d7b7a0adc034d3d3d089dd4e8f16f90621792ba3be7ceba8f7\": container with ID starting with 18af9361e46e52d7b7a0adc034d3d3d089dd4e8f16f90621792ba3be7ceba8f7 not found: ID does not exist" Feb 02 09:13:09 crc kubenswrapper[4764]: I0202 09:13:09.435236 4764 scope.go:117] "RemoveContainer" containerID="796d368c28e1658dc2a81024ed86d820b72e2e1f251c1951ede876071b19b223" Feb 02 09:13:09 crc kubenswrapper[4764]: E0202 09:13:09.436187 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"796d368c28e1658dc2a81024ed86d820b72e2e1f251c1951ede876071b19b223\": container with ID starting with 796d368c28e1658dc2a81024ed86d820b72e2e1f251c1951ede876071b19b223 not found: ID does not exist" containerID="796d368c28e1658dc2a81024ed86d820b72e2e1f251c1951ede876071b19b223" Feb 02 09:13:09 crc kubenswrapper[4764]: I0202 09:13:09.436221 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"796d368c28e1658dc2a81024ed86d820b72e2e1f251c1951ede876071b19b223"} err="failed to get container status \"796d368c28e1658dc2a81024ed86d820b72e2e1f251c1951ede876071b19b223\": rpc error: code = NotFound desc = could not find container \"796d368c28e1658dc2a81024ed86d820b72e2e1f251c1951ede876071b19b223\": container with ID starting with 796d368c28e1658dc2a81024ed86d820b72e2e1f251c1951ede876071b19b223 not found: ID does not exist" Feb 02 09:13:09 crc kubenswrapper[4764]: I0202 09:13:09.436249 4764 scope.go:117] "RemoveContainer" containerID="15c8dcdbf7dafd0945e7cee7ea86c7e3c74f639942d8cc81473c5ba785830e7d" Feb 02 09:13:09 crc kubenswrapper[4764]: E0202 09:13:09.437580 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"15c8dcdbf7dafd0945e7cee7ea86c7e3c74f639942d8cc81473c5ba785830e7d\": container with ID starting with 15c8dcdbf7dafd0945e7cee7ea86c7e3c74f639942d8cc81473c5ba785830e7d not found: ID does not exist" containerID="15c8dcdbf7dafd0945e7cee7ea86c7e3c74f639942d8cc81473c5ba785830e7d" Feb 02 09:13:09 crc kubenswrapper[4764]: I0202 09:13:09.437610 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"15c8dcdbf7dafd0945e7cee7ea86c7e3c74f639942d8cc81473c5ba785830e7d"} err="failed to get container status \"15c8dcdbf7dafd0945e7cee7ea86c7e3c74f639942d8cc81473c5ba785830e7d\": rpc error: code = NotFound desc = could not find container \"15c8dcdbf7dafd0945e7cee7ea86c7e3c74f639942d8cc81473c5ba785830e7d\": container with ID starting with 15c8dcdbf7dafd0945e7cee7ea86c7e3c74f639942d8cc81473c5ba785830e7d not found: ID does not exist" Feb 02 09:13:09 crc kubenswrapper[4764]: I0202 09:13:09.437624 4764 scope.go:117] "RemoveContainer" containerID="0cbd7d0e3df1db8208d9af33d979c98caaa3f82973cc8263382a73f719638baa" Feb 02 09:13:09 crc kubenswrapper[4764]: I0202 09:13:09.455102 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/01ede9d5-6b94-4d10-bf6d-3aef751840ff-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "01ede9d5-6b94-4d10-bf6d-3aef751840ff" (UID: "01ede9d5-6b94-4d10-bf6d-3aef751840ff"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 09:13:09 crc kubenswrapper[4764]: I0202 09:13:09.460396 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/00cb5ef3-3b54-47ac-bfc4-b1d53f04d292-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "00cb5ef3-3b54-47ac-bfc4-b1d53f04d292" (UID: "00cb5ef3-3b54-47ac-bfc4-b1d53f04d292"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 09:13:09 crc kubenswrapper[4764]: I0202 09:13:09.472576 4764 scope.go:117] "RemoveContainer" containerID="242ff04a15506716500f4dde700de56665b7f4570bfc4ebbf3e4eb1dc4263f9f" Feb 02 09:13:09 crc kubenswrapper[4764]: I0202 09:13:09.498334 4764 scope.go:117] "RemoveContainer" containerID="290a2505a57e59d6feb7b276cd44e863f761a91d2ed9ce7c3a89f09320ad086d" Feb 02 09:13:09 crc kubenswrapper[4764]: I0202 09:13:09.504202 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6092680d-f31e-4b70-9d81-2d80661f6ac3-utilities\") pod \"6092680d-f31e-4b70-9d81-2d80661f6ac3\" (UID: \"6092680d-f31e-4b70-9d81-2d80661f6ac3\") " Feb 02 09:13:09 crc kubenswrapper[4764]: I0202 09:13:09.504260 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6092680d-f31e-4b70-9d81-2d80661f6ac3-catalog-content\") pod \"6092680d-f31e-4b70-9d81-2d80661f6ac3\" (UID: \"6092680d-f31e-4b70-9d81-2d80661f6ac3\") " Feb 02 09:13:09 crc kubenswrapper[4764]: I0202 09:13:09.504327 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nv9qk\" (UniqueName: \"kubernetes.io/projected/6092680d-f31e-4b70-9d81-2d80661f6ac3-kube-api-access-nv9qk\") pod \"6092680d-f31e-4b70-9d81-2d80661f6ac3\" (UID: \"6092680d-f31e-4b70-9d81-2d80661f6ac3\") " Feb 02 09:13:09 crc kubenswrapper[4764]: I0202 09:13:09.504518 4764 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/fac965f9-5e91-4c50-a538-48c1301fa072-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Feb 02 09:13:09 crc kubenswrapper[4764]: I0202 09:13:09.504532 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s9xmw\" (UniqueName: \"kubernetes.io/projected/fac965f9-5e91-4c50-a538-48c1301fa072-kube-api-access-s9xmw\") on node \"crc\" DevicePath \"\"" Feb 02 09:13:09 crc kubenswrapper[4764]: I0202 09:13:09.504542 4764 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/01ede9d5-6b94-4d10-bf6d-3aef751840ff-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 09:13:09 crc kubenswrapper[4764]: I0202 09:13:09.504550 4764 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/01ede9d5-6b94-4d10-bf6d-3aef751840ff-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 09:13:09 crc kubenswrapper[4764]: I0202 09:13:09.504559 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-knbhq\" (UniqueName: \"kubernetes.io/projected/01ede9d5-6b94-4d10-bf6d-3aef751840ff-kube-api-access-knbhq\") on node \"crc\" DevicePath \"\"" Feb 02 09:13:09 crc kubenswrapper[4764]: I0202 09:13:09.504567 4764 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/00cb5ef3-3b54-47ac-bfc4-b1d53f04d292-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 09:13:09 crc kubenswrapper[4764]: I0202 09:13:09.504578 4764 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/fac965f9-5e91-4c50-a538-48c1301fa072-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 02 09:13:09 crc kubenswrapper[4764]: I0202 09:13:09.505431 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6092680d-f31e-4b70-9d81-2d80661f6ac3-utilities" (OuterVolumeSpecName: "utilities") pod "6092680d-f31e-4b70-9d81-2d80661f6ac3" (UID: "6092680d-f31e-4b70-9d81-2d80661f6ac3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 09:13:09 crc kubenswrapper[4764]: I0202 09:13:09.509819 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6092680d-f31e-4b70-9d81-2d80661f6ac3-kube-api-access-nv9qk" (OuterVolumeSpecName: "kube-api-access-nv9qk") pod "6092680d-f31e-4b70-9d81-2d80661f6ac3" (UID: "6092680d-f31e-4b70-9d81-2d80661f6ac3"). InnerVolumeSpecName "kube-api-access-nv9qk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:13:09 crc kubenswrapper[4764]: I0202 09:13:09.524862 4764 scope.go:117] "RemoveContainer" containerID="0cbd7d0e3df1db8208d9af33d979c98caaa3f82973cc8263382a73f719638baa" Feb 02 09:13:09 crc kubenswrapper[4764]: E0202 09:13:09.525660 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0cbd7d0e3df1db8208d9af33d979c98caaa3f82973cc8263382a73f719638baa\": container with ID starting with 0cbd7d0e3df1db8208d9af33d979c98caaa3f82973cc8263382a73f719638baa not found: ID does not exist" containerID="0cbd7d0e3df1db8208d9af33d979c98caaa3f82973cc8263382a73f719638baa" Feb 02 09:13:09 crc kubenswrapper[4764]: I0202 09:13:09.525697 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0cbd7d0e3df1db8208d9af33d979c98caaa3f82973cc8263382a73f719638baa"} err="failed to get container status \"0cbd7d0e3df1db8208d9af33d979c98caaa3f82973cc8263382a73f719638baa\": rpc error: code = NotFound desc = could not find container \"0cbd7d0e3df1db8208d9af33d979c98caaa3f82973cc8263382a73f719638baa\": container with ID starting with 0cbd7d0e3df1db8208d9af33d979c98caaa3f82973cc8263382a73f719638baa not found: ID does not exist" Feb 02 09:13:09 crc kubenswrapper[4764]: I0202 09:13:09.525717 4764 scope.go:117] "RemoveContainer" containerID="242ff04a15506716500f4dde700de56665b7f4570bfc4ebbf3e4eb1dc4263f9f" Feb 02 09:13:09 crc kubenswrapper[4764]: E0202 09:13:09.526108 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"242ff04a15506716500f4dde700de56665b7f4570bfc4ebbf3e4eb1dc4263f9f\": container with ID starting with 242ff04a15506716500f4dde700de56665b7f4570bfc4ebbf3e4eb1dc4263f9f not found: ID does not exist" containerID="242ff04a15506716500f4dde700de56665b7f4570bfc4ebbf3e4eb1dc4263f9f" Feb 02 09:13:09 crc kubenswrapper[4764]: I0202 09:13:09.526129 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"242ff04a15506716500f4dde700de56665b7f4570bfc4ebbf3e4eb1dc4263f9f"} err="failed to get container status \"242ff04a15506716500f4dde700de56665b7f4570bfc4ebbf3e4eb1dc4263f9f\": rpc error: code = NotFound desc = could not find container \"242ff04a15506716500f4dde700de56665b7f4570bfc4ebbf3e4eb1dc4263f9f\": container with ID starting with 242ff04a15506716500f4dde700de56665b7f4570bfc4ebbf3e4eb1dc4263f9f not found: ID does not exist" Feb 02 09:13:09 crc kubenswrapper[4764]: I0202 09:13:09.526141 4764 scope.go:117] "RemoveContainer" containerID="290a2505a57e59d6feb7b276cd44e863f761a91d2ed9ce7c3a89f09320ad086d" Feb 02 09:13:09 crc kubenswrapper[4764]: E0202 09:13:09.527063 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"290a2505a57e59d6feb7b276cd44e863f761a91d2ed9ce7c3a89f09320ad086d\": container with ID starting with 290a2505a57e59d6feb7b276cd44e863f761a91d2ed9ce7c3a89f09320ad086d not found: ID does not exist" containerID="290a2505a57e59d6feb7b276cd44e863f761a91d2ed9ce7c3a89f09320ad086d" Feb 02 09:13:09 crc kubenswrapper[4764]: I0202 09:13:09.527086 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"290a2505a57e59d6feb7b276cd44e863f761a91d2ed9ce7c3a89f09320ad086d"} err="failed to get container status \"290a2505a57e59d6feb7b276cd44e863f761a91d2ed9ce7c3a89f09320ad086d\": rpc error: code = NotFound desc = could not find container \"290a2505a57e59d6feb7b276cd44e863f761a91d2ed9ce7c3a89f09320ad086d\": container with ID starting with 290a2505a57e59d6feb7b276cd44e863f761a91d2ed9ce7c3a89f09320ad086d not found: ID does not exist" Feb 02 09:13:09 crc kubenswrapper[4764]: I0202 09:13:09.527100 4764 scope.go:117] "RemoveContainer" containerID="b0e277ec4569fc9807eb7c3cc0a8905c030a0b41601d8880193f95814472f82e" Feb 02 09:13:09 crc kubenswrapper[4764]: I0202 09:13:09.535204 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6092680d-f31e-4b70-9d81-2d80661f6ac3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6092680d-f31e-4b70-9d81-2d80661f6ac3" (UID: "6092680d-f31e-4b70-9d81-2d80661f6ac3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 09:13:09 crc kubenswrapper[4764]: I0202 09:13:09.544376 4764 scope.go:117] "RemoveContainer" containerID="e5c51aeae99f7c7c3a6553c26d425d22a804604a09b6d6bd101fe7a4458b2e00" Feb 02 09:13:09 crc kubenswrapper[4764]: I0202 09:13:09.559947 4764 scope.go:117] "RemoveContainer" containerID="9a274df71070056d0ba90d60479e4818f9ac34548bc19489c14a87da15dc6c8b" Feb 02 09:13:09 crc kubenswrapper[4764]: I0202 09:13:09.572118 4764 scope.go:117] "RemoveContainer" containerID="b0e277ec4569fc9807eb7c3cc0a8905c030a0b41601d8880193f95814472f82e" Feb 02 09:13:09 crc kubenswrapper[4764]: E0202 09:13:09.572643 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b0e277ec4569fc9807eb7c3cc0a8905c030a0b41601d8880193f95814472f82e\": container with ID starting with b0e277ec4569fc9807eb7c3cc0a8905c030a0b41601d8880193f95814472f82e not found: ID does not exist" containerID="b0e277ec4569fc9807eb7c3cc0a8905c030a0b41601d8880193f95814472f82e" Feb 02 09:13:09 crc kubenswrapper[4764]: I0202 09:13:09.572670 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b0e277ec4569fc9807eb7c3cc0a8905c030a0b41601d8880193f95814472f82e"} err="failed to get container status \"b0e277ec4569fc9807eb7c3cc0a8905c030a0b41601d8880193f95814472f82e\": rpc error: code = NotFound desc = could not find container \"b0e277ec4569fc9807eb7c3cc0a8905c030a0b41601d8880193f95814472f82e\": container with ID starting with b0e277ec4569fc9807eb7c3cc0a8905c030a0b41601d8880193f95814472f82e not found: ID does not exist" Feb 02 09:13:09 crc kubenswrapper[4764]: I0202 09:13:09.572713 4764 scope.go:117] "RemoveContainer" containerID="e5c51aeae99f7c7c3a6553c26d425d22a804604a09b6d6bd101fe7a4458b2e00" Feb 02 09:13:09 crc kubenswrapper[4764]: E0202 09:13:09.573161 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e5c51aeae99f7c7c3a6553c26d425d22a804604a09b6d6bd101fe7a4458b2e00\": container with ID starting with e5c51aeae99f7c7c3a6553c26d425d22a804604a09b6d6bd101fe7a4458b2e00 not found: ID does not exist" containerID="e5c51aeae99f7c7c3a6553c26d425d22a804604a09b6d6bd101fe7a4458b2e00" Feb 02 09:13:09 crc kubenswrapper[4764]: I0202 09:13:09.573192 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e5c51aeae99f7c7c3a6553c26d425d22a804604a09b6d6bd101fe7a4458b2e00"} err="failed to get container status \"e5c51aeae99f7c7c3a6553c26d425d22a804604a09b6d6bd101fe7a4458b2e00\": rpc error: code = NotFound desc = could not find container \"e5c51aeae99f7c7c3a6553c26d425d22a804604a09b6d6bd101fe7a4458b2e00\": container with ID starting with e5c51aeae99f7c7c3a6553c26d425d22a804604a09b6d6bd101fe7a4458b2e00 not found: ID does not exist" Feb 02 09:13:09 crc kubenswrapper[4764]: I0202 09:13:09.573215 4764 scope.go:117] "RemoveContainer" containerID="9a274df71070056d0ba90d60479e4818f9ac34548bc19489c14a87da15dc6c8b" Feb 02 09:13:09 crc kubenswrapper[4764]: E0202 09:13:09.573571 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9a274df71070056d0ba90d60479e4818f9ac34548bc19489c14a87da15dc6c8b\": container with ID starting with 9a274df71070056d0ba90d60479e4818f9ac34548bc19489c14a87da15dc6c8b not found: ID does not exist" containerID="9a274df71070056d0ba90d60479e4818f9ac34548bc19489c14a87da15dc6c8b" Feb 02 09:13:09 crc kubenswrapper[4764]: I0202 09:13:09.573600 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9a274df71070056d0ba90d60479e4818f9ac34548bc19489c14a87da15dc6c8b"} err="failed to get container status \"9a274df71070056d0ba90d60479e4818f9ac34548bc19489c14a87da15dc6c8b\": rpc error: code = NotFound desc = could not find container \"9a274df71070056d0ba90d60479e4818f9ac34548bc19489c14a87da15dc6c8b\": container with ID starting with 9a274df71070056d0ba90d60479e4818f9ac34548bc19489c14a87da15dc6c8b not found: ID does not exist" Feb 02 09:13:09 crc kubenswrapper[4764]: I0202 09:13:09.573617 4764 scope.go:117] "RemoveContainer" containerID="cae9539080c975e16c67467c3d702f51ce9280b85f572e147d473e1583321863" Feb 02 09:13:09 crc kubenswrapper[4764]: I0202 09:13:09.589788 4764 scope.go:117] "RemoveContainer" containerID="02b1aaa97c610ac96c94ef0e43e8d9bdb81abe2098a614069bd721975be3d873" Feb 02 09:13:09 crc kubenswrapper[4764]: I0202 09:13:09.602061 4764 scope.go:117] "RemoveContainer" containerID="171e9563f8ad099d9e9c74f1bff0e55abe393140c07ff9c612151163ecb679ee" Feb 02 09:13:09 crc kubenswrapper[4764]: I0202 09:13:09.605273 4764 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6092680d-f31e-4b70-9d81-2d80661f6ac3-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 09:13:09 crc kubenswrapper[4764]: I0202 09:13:09.605298 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nv9qk\" (UniqueName: \"kubernetes.io/projected/6092680d-f31e-4b70-9d81-2d80661f6ac3-kube-api-access-nv9qk\") on node \"crc\" DevicePath \"\"" Feb 02 09:13:09 crc kubenswrapper[4764]: I0202 09:13:09.605311 4764 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6092680d-f31e-4b70-9d81-2d80661f6ac3-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 09:13:09 crc kubenswrapper[4764]: I0202 09:13:09.618500 4764 scope.go:117] "RemoveContainer" containerID="268bfe0e1d8733e258da407a82ee2f78a0dc2d86ecdbc5bafed848d6f8e77c4f" Feb 02 09:13:09 crc kubenswrapper[4764]: I0202 09:13:09.631439 4764 scope.go:117] "RemoveContainer" containerID="268bfe0e1d8733e258da407a82ee2f78a0dc2d86ecdbc5bafed848d6f8e77c4f" Feb 02 09:13:09 crc kubenswrapper[4764]: E0202 09:13:09.631736 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"268bfe0e1d8733e258da407a82ee2f78a0dc2d86ecdbc5bafed848d6f8e77c4f\": container with ID starting with 268bfe0e1d8733e258da407a82ee2f78a0dc2d86ecdbc5bafed848d6f8e77c4f not found: ID does not exist" containerID="268bfe0e1d8733e258da407a82ee2f78a0dc2d86ecdbc5bafed848d6f8e77c4f" Feb 02 09:13:09 crc kubenswrapper[4764]: I0202 09:13:09.631765 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"268bfe0e1d8733e258da407a82ee2f78a0dc2d86ecdbc5bafed848d6f8e77c4f"} err="failed to get container status \"268bfe0e1d8733e258da407a82ee2f78a0dc2d86ecdbc5bafed848d6f8e77c4f\": rpc error: code = NotFound desc = could not find container \"268bfe0e1d8733e258da407a82ee2f78a0dc2d86ecdbc5bafed848d6f8e77c4f\": container with ID starting with 268bfe0e1d8733e258da407a82ee2f78a0dc2d86ecdbc5bafed848d6f8e77c4f not found: ID does not exist" Feb 02 09:13:09 crc kubenswrapper[4764]: I0202 09:13:09.665103 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-j25pg"] Feb 02 09:13:09 crc kubenswrapper[4764]: I0202 09:13:09.673242 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-j25pg"] Feb 02 09:13:09 crc kubenswrapper[4764]: I0202 09:13:09.685954 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-ct55p"] Feb 02 09:13:09 crc kubenswrapper[4764]: I0202 09:13:09.688987 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-ct55p"] Feb 02 09:13:09 crc kubenswrapper[4764]: I0202 09:13:09.698515 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-wqbtk"] Feb 02 09:13:09 crc kubenswrapper[4764]: I0202 09:13:09.700992 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-wqbtk"] Feb 02 09:13:09 crc kubenswrapper[4764]: I0202 09:13:09.705528 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-kbk9s"] Feb 02 09:13:09 crc kubenswrapper[4764]: I0202 09:13:09.708711 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-kbk9s"] Feb 02 09:13:09 crc kubenswrapper[4764]: I0202 09:13:09.832827 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="00cb5ef3-3b54-47ac-bfc4-b1d53f04d292" path="/var/lib/kubelet/pods/00cb5ef3-3b54-47ac-bfc4-b1d53f04d292/volumes" Feb 02 09:13:09 crc kubenswrapper[4764]: I0202 09:13:09.833632 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ede9d5-6b94-4d10-bf6d-3aef751840ff" path="/var/lib/kubelet/pods/01ede9d5-6b94-4d10-bf6d-3aef751840ff/volumes" Feb 02 09:13:09 crc kubenswrapper[4764]: I0202 09:13:09.834200 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6092680d-f31e-4b70-9d81-2d80661f6ac3" path="/var/lib/kubelet/pods/6092680d-f31e-4b70-9d81-2d80661f6ac3/volumes" Feb 02 09:13:09 crc kubenswrapper[4764]: I0202 09:13:09.835177 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a40abe98-1419-4f5d-9bed-63f1c89028b1" path="/var/lib/kubelet/pods/a40abe98-1419-4f5d-9bed-63f1c89028b1/volumes" Feb 02 09:13:09 crc kubenswrapper[4764]: I0202 09:13:09.835737 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fac965f9-5e91-4c50-a538-48c1301fa072" path="/var/lib/kubelet/pods/fac965f9-5e91-4c50-a538-48c1301fa072/volumes" Feb 02 09:13:10 crc kubenswrapper[4764]: I0202 09:13:10.363680 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-f4d5w" event={"ID":"a03d1d75-d8a0-407c-ab02-d151d5ba00b6","Type":"ContainerStarted","Data":"75867dd834098d7d824672b974c56255f4475366356e78d5625a3fa87ebd2053"} Feb 02 09:13:10 crc kubenswrapper[4764]: I0202 09:13:10.363862 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-f4d5w" Feb 02 09:13:10 crc kubenswrapper[4764]: I0202 09:13:10.363966 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-f4d5w" event={"ID":"a03d1d75-d8a0-407c-ab02-d151d5ba00b6","Type":"ContainerStarted","Data":"902a00976ebf5d11d8688225b60c652bad924f19af9c0f2e5388310709340268"} Feb 02 09:13:10 crc kubenswrapper[4764]: I0202 09:13:10.371550 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-f4d5w" Feb 02 09:13:10 crc kubenswrapper[4764]: I0202 09:13:10.392410 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-f4d5w" podStartSLOduration=2.392393742 podStartE2EDuration="2.392393742s" podCreationTimestamp="2026-02-02 09:13:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:13:10.389163425 +0000 UTC m=+353.322887543" watchObservedRunningTime="2026-02-02 09:13:10.392393742 +0000 UTC m=+353.326117830" Feb 02 09:13:10 crc kubenswrapper[4764]: I0202 09:13:10.405642 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-vp4vx"] Feb 02 09:13:10 crc kubenswrapper[4764]: E0202 09:13:10.405904 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00cb5ef3-3b54-47ac-bfc4-b1d53f04d292" containerName="extract-utilities" Feb 02 09:13:10 crc kubenswrapper[4764]: I0202 09:13:10.405943 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="00cb5ef3-3b54-47ac-bfc4-b1d53f04d292" containerName="extract-utilities" Feb 02 09:13:10 crc kubenswrapper[4764]: E0202 09:13:10.405955 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6092680d-f31e-4b70-9d81-2d80661f6ac3" containerName="registry-server" Feb 02 09:13:10 crc kubenswrapper[4764]: I0202 09:13:10.405966 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="6092680d-f31e-4b70-9d81-2d80661f6ac3" containerName="registry-server" Feb 02 09:13:10 crc kubenswrapper[4764]: E0202 09:13:10.405979 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a40abe98-1419-4f5d-9bed-63f1c89028b1" containerName="extract-utilities" Feb 02 09:13:10 crc kubenswrapper[4764]: I0202 09:13:10.405988 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="a40abe98-1419-4f5d-9bed-63f1c89028b1" containerName="extract-utilities" Feb 02 09:13:10 crc kubenswrapper[4764]: E0202 09:13:10.406000 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6092680d-f31e-4b70-9d81-2d80661f6ac3" containerName="extract-content" Feb 02 09:13:10 crc kubenswrapper[4764]: I0202 09:13:10.406007 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="6092680d-f31e-4b70-9d81-2d80661f6ac3" containerName="extract-content" Feb 02 09:13:10 crc kubenswrapper[4764]: E0202 09:13:10.406016 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01ede9d5-6b94-4d10-bf6d-3aef751840ff" containerName="extract-content" Feb 02 09:13:10 crc kubenswrapper[4764]: I0202 09:13:10.406023 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="01ede9d5-6b94-4d10-bf6d-3aef751840ff" containerName="extract-content" Feb 02 09:13:10 crc kubenswrapper[4764]: E0202 09:13:10.406034 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a40abe98-1419-4f5d-9bed-63f1c89028b1" containerName="extract-content" Feb 02 09:13:10 crc kubenswrapper[4764]: I0202 09:13:10.406040 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="a40abe98-1419-4f5d-9bed-63f1c89028b1" containerName="extract-content" Feb 02 09:13:10 crc kubenswrapper[4764]: E0202 09:13:10.406050 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01ede9d5-6b94-4d10-bf6d-3aef751840ff" containerName="registry-server" Feb 02 09:13:10 crc kubenswrapper[4764]: I0202 09:13:10.406056 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="01ede9d5-6b94-4d10-bf6d-3aef751840ff" containerName="registry-server" Feb 02 09:13:10 crc kubenswrapper[4764]: E0202 09:13:10.406066 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00cb5ef3-3b54-47ac-bfc4-b1d53f04d292" containerName="extract-content" Feb 02 09:13:10 crc kubenswrapper[4764]: I0202 09:13:10.406072 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="00cb5ef3-3b54-47ac-bfc4-b1d53f04d292" containerName="extract-content" Feb 02 09:13:10 crc kubenswrapper[4764]: E0202 09:13:10.406082 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6092680d-f31e-4b70-9d81-2d80661f6ac3" containerName="extract-utilities" Feb 02 09:13:10 crc kubenswrapper[4764]: I0202 09:13:10.406089 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="6092680d-f31e-4b70-9d81-2d80661f6ac3" containerName="extract-utilities" Feb 02 09:13:10 crc kubenswrapper[4764]: E0202 09:13:10.406101 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00cb5ef3-3b54-47ac-bfc4-b1d53f04d292" containerName="registry-server" Feb 02 09:13:10 crc kubenswrapper[4764]: I0202 09:13:10.406109 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="00cb5ef3-3b54-47ac-bfc4-b1d53f04d292" containerName="registry-server" Feb 02 09:13:10 crc kubenswrapper[4764]: E0202 09:13:10.406122 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fac965f9-5e91-4c50-a538-48c1301fa072" containerName="marketplace-operator" Feb 02 09:13:10 crc kubenswrapper[4764]: I0202 09:13:10.406129 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="fac965f9-5e91-4c50-a538-48c1301fa072" containerName="marketplace-operator" Feb 02 09:13:10 crc kubenswrapper[4764]: E0202 09:13:10.406139 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a40abe98-1419-4f5d-9bed-63f1c89028b1" containerName="registry-server" Feb 02 09:13:10 crc kubenswrapper[4764]: I0202 09:13:10.406145 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="a40abe98-1419-4f5d-9bed-63f1c89028b1" containerName="registry-server" Feb 02 09:13:10 crc kubenswrapper[4764]: E0202 09:13:10.406153 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01ede9d5-6b94-4d10-bf6d-3aef751840ff" containerName="extract-utilities" Feb 02 09:13:10 crc kubenswrapper[4764]: I0202 09:13:10.406158 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="01ede9d5-6b94-4d10-bf6d-3aef751840ff" containerName="extract-utilities" Feb 02 09:13:10 crc kubenswrapper[4764]: I0202 09:13:10.406239 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="fac965f9-5e91-4c50-a538-48c1301fa072" containerName="marketplace-operator" Feb 02 09:13:10 crc kubenswrapper[4764]: I0202 09:13:10.406247 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="6092680d-f31e-4b70-9d81-2d80661f6ac3" containerName="registry-server" Feb 02 09:13:10 crc kubenswrapper[4764]: I0202 09:13:10.406261 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="a40abe98-1419-4f5d-9bed-63f1c89028b1" containerName="registry-server" Feb 02 09:13:10 crc kubenswrapper[4764]: I0202 09:13:10.406270 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="01ede9d5-6b94-4d10-bf6d-3aef751840ff" containerName="registry-server" Feb 02 09:13:10 crc kubenswrapper[4764]: I0202 09:13:10.406281 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="00cb5ef3-3b54-47ac-bfc4-b1d53f04d292" containerName="registry-server" Feb 02 09:13:10 crc kubenswrapper[4764]: I0202 09:13:10.407119 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vp4vx" Feb 02 09:13:10 crc kubenswrapper[4764]: I0202 09:13:10.420516 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Feb 02 09:13:10 crc kubenswrapper[4764]: I0202 09:13:10.420527 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-vp4vx"] Feb 02 09:13:10 crc kubenswrapper[4764]: I0202 09:13:10.521230 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7e8b7ce6-dc1a-474e-83e0-659a275c7aaa-catalog-content\") pod \"redhat-marketplace-vp4vx\" (UID: \"7e8b7ce6-dc1a-474e-83e0-659a275c7aaa\") " pod="openshift-marketplace/redhat-marketplace-vp4vx" Feb 02 09:13:10 crc kubenswrapper[4764]: I0202 09:13:10.521305 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4hwbl\" (UniqueName: \"kubernetes.io/projected/7e8b7ce6-dc1a-474e-83e0-659a275c7aaa-kube-api-access-4hwbl\") pod \"redhat-marketplace-vp4vx\" (UID: \"7e8b7ce6-dc1a-474e-83e0-659a275c7aaa\") " pod="openshift-marketplace/redhat-marketplace-vp4vx" Feb 02 09:13:10 crc kubenswrapper[4764]: I0202 09:13:10.521762 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7e8b7ce6-dc1a-474e-83e0-659a275c7aaa-utilities\") pod \"redhat-marketplace-vp4vx\" (UID: \"7e8b7ce6-dc1a-474e-83e0-659a275c7aaa\") " pod="openshift-marketplace/redhat-marketplace-vp4vx" Feb 02 09:13:10 crc kubenswrapper[4764]: I0202 09:13:10.623326 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7e8b7ce6-dc1a-474e-83e0-659a275c7aaa-catalog-content\") pod \"redhat-marketplace-vp4vx\" (UID: \"7e8b7ce6-dc1a-474e-83e0-659a275c7aaa\") " pod="openshift-marketplace/redhat-marketplace-vp4vx" Feb 02 09:13:10 crc kubenswrapper[4764]: I0202 09:13:10.623372 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4hwbl\" (UniqueName: \"kubernetes.io/projected/7e8b7ce6-dc1a-474e-83e0-659a275c7aaa-kube-api-access-4hwbl\") pod \"redhat-marketplace-vp4vx\" (UID: \"7e8b7ce6-dc1a-474e-83e0-659a275c7aaa\") " pod="openshift-marketplace/redhat-marketplace-vp4vx" Feb 02 09:13:10 crc kubenswrapper[4764]: I0202 09:13:10.623407 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7e8b7ce6-dc1a-474e-83e0-659a275c7aaa-utilities\") pod \"redhat-marketplace-vp4vx\" (UID: \"7e8b7ce6-dc1a-474e-83e0-659a275c7aaa\") " pod="openshift-marketplace/redhat-marketplace-vp4vx" Feb 02 09:13:10 crc kubenswrapper[4764]: I0202 09:13:10.623793 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7e8b7ce6-dc1a-474e-83e0-659a275c7aaa-utilities\") pod \"redhat-marketplace-vp4vx\" (UID: \"7e8b7ce6-dc1a-474e-83e0-659a275c7aaa\") " pod="openshift-marketplace/redhat-marketplace-vp4vx" Feb 02 09:13:10 crc kubenswrapper[4764]: I0202 09:13:10.624139 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7e8b7ce6-dc1a-474e-83e0-659a275c7aaa-catalog-content\") pod \"redhat-marketplace-vp4vx\" (UID: \"7e8b7ce6-dc1a-474e-83e0-659a275c7aaa\") " pod="openshift-marketplace/redhat-marketplace-vp4vx" Feb 02 09:13:10 crc kubenswrapper[4764]: I0202 09:13:10.652906 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4hwbl\" (UniqueName: \"kubernetes.io/projected/7e8b7ce6-dc1a-474e-83e0-659a275c7aaa-kube-api-access-4hwbl\") pod \"redhat-marketplace-vp4vx\" (UID: \"7e8b7ce6-dc1a-474e-83e0-659a275c7aaa\") " pod="openshift-marketplace/redhat-marketplace-vp4vx" Feb 02 09:13:10 crc kubenswrapper[4764]: I0202 09:13:10.778992 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vp4vx" Feb 02 09:13:11 crc kubenswrapper[4764]: I0202 09:13:11.514731 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-rc4qs"] Feb 02 09:13:11 crc kubenswrapper[4764]: I0202 09:13:11.516619 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rc4qs" Feb 02 09:13:11 crc kubenswrapper[4764]: I0202 09:13:11.519695 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Feb 02 09:13:11 crc kubenswrapper[4764]: I0202 09:13:11.526632 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rc4qs"] Feb 02 09:13:11 crc kubenswrapper[4764]: I0202 09:13:11.639809 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9443ada1-5f42-40df-b8e2-2c072dd7f264-catalog-content\") pod \"redhat-operators-rc4qs\" (UID: \"9443ada1-5f42-40df-b8e2-2c072dd7f264\") " pod="openshift-marketplace/redhat-operators-rc4qs" Feb 02 09:13:11 crc kubenswrapper[4764]: I0202 09:13:11.639858 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9443ada1-5f42-40df-b8e2-2c072dd7f264-utilities\") pod \"redhat-operators-rc4qs\" (UID: \"9443ada1-5f42-40df-b8e2-2c072dd7f264\") " pod="openshift-marketplace/redhat-operators-rc4qs" Feb 02 09:13:11 crc kubenswrapper[4764]: I0202 09:13:11.639877 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-59tw8\" (UniqueName: \"kubernetes.io/projected/9443ada1-5f42-40df-b8e2-2c072dd7f264-kube-api-access-59tw8\") pod \"redhat-operators-rc4qs\" (UID: \"9443ada1-5f42-40df-b8e2-2c072dd7f264\") " pod="openshift-marketplace/redhat-operators-rc4qs" Feb 02 09:13:11 crc kubenswrapper[4764]: I0202 09:13:11.741414 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9443ada1-5f42-40df-b8e2-2c072dd7f264-utilities\") pod \"redhat-operators-rc4qs\" (UID: \"9443ada1-5f42-40df-b8e2-2c072dd7f264\") " pod="openshift-marketplace/redhat-operators-rc4qs" Feb 02 09:13:11 crc kubenswrapper[4764]: I0202 09:13:11.741484 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-59tw8\" (UniqueName: \"kubernetes.io/projected/9443ada1-5f42-40df-b8e2-2c072dd7f264-kube-api-access-59tw8\") pod \"redhat-operators-rc4qs\" (UID: \"9443ada1-5f42-40df-b8e2-2c072dd7f264\") " pod="openshift-marketplace/redhat-operators-rc4qs" Feb 02 09:13:11 crc kubenswrapper[4764]: I0202 09:13:11.741635 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9443ada1-5f42-40df-b8e2-2c072dd7f264-catalog-content\") pod \"redhat-operators-rc4qs\" (UID: \"9443ada1-5f42-40df-b8e2-2c072dd7f264\") " pod="openshift-marketplace/redhat-operators-rc4qs" Feb 02 09:13:11 crc kubenswrapper[4764]: I0202 09:13:11.741977 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9443ada1-5f42-40df-b8e2-2c072dd7f264-utilities\") pod \"redhat-operators-rc4qs\" (UID: \"9443ada1-5f42-40df-b8e2-2c072dd7f264\") " pod="openshift-marketplace/redhat-operators-rc4qs" Feb 02 09:13:11 crc kubenswrapper[4764]: I0202 09:13:11.742281 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9443ada1-5f42-40df-b8e2-2c072dd7f264-catalog-content\") pod \"redhat-operators-rc4qs\" (UID: \"9443ada1-5f42-40df-b8e2-2c072dd7f264\") " pod="openshift-marketplace/redhat-operators-rc4qs" Feb 02 09:13:11 crc kubenswrapper[4764]: I0202 09:13:11.778240 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-59tw8\" (UniqueName: \"kubernetes.io/projected/9443ada1-5f42-40df-b8e2-2c072dd7f264-kube-api-access-59tw8\") pod \"redhat-operators-rc4qs\" (UID: \"9443ada1-5f42-40df-b8e2-2c072dd7f264\") " pod="openshift-marketplace/redhat-operators-rc4qs" Feb 02 09:13:11 crc kubenswrapper[4764]: I0202 09:13:11.830806 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-vp4vx"] Feb 02 09:13:11 crc kubenswrapper[4764]: I0202 09:13:11.844437 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rc4qs" Feb 02 09:13:12 crc kubenswrapper[4764]: I0202 09:13:12.013086 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rc4qs"] Feb 02 09:13:12 crc kubenswrapper[4764]: W0202 09:13:12.051409 4764 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9443ada1_5f42_40df_b8e2_2c072dd7f264.slice/crio-fd79fae9223aef427ad812f8dfe3e82ebf2ed03ff4436f507b5886844777f2d1 WatchSource:0}: Error finding container fd79fae9223aef427ad812f8dfe3e82ebf2ed03ff4436f507b5886844777f2d1: Status 404 returned error can't find the container with id fd79fae9223aef427ad812f8dfe3e82ebf2ed03ff4436f507b5886844777f2d1 Feb 02 09:13:12 crc kubenswrapper[4764]: I0202 09:13:12.466090 4764 generic.go:334] "Generic (PLEG): container finished" podID="9443ada1-5f42-40df-b8e2-2c072dd7f264" containerID="59f2c23d72bc11c5b628e277c20ad3ee7e34578a7c9cdb6068e5ed7f9f2da2a6" exitCode=0 Feb 02 09:13:12 crc kubenswrapper[4764]: I0202 09:13:12.466157 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rc4qs" event={"ID":"9443ada1-5f42-40df-b8e2-2c072dd7f264","Type":"ContainerDied","Data":"59f2c23d72bc11c5b628e277c20ad3ee7e34578a7c9cdb6068e5ed7f9f2da2a6"} Feb 02 09:13:12 crc kubenswrapper[4764]: I0202 09:13:12.466183 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rc4qs" event={"ID":"9443ada1-5f42-40df-b8e2-2c072dd7f264","Type":"ContainerStarted","Data":"fd79fae9223aef427ad812f8dfe3e82ebf2ed03ff4436f507b5886844777f2d1"} Feb 02 09:13:12 crc kubenswrapper[4764]: I0202 09:13:12.471559 4764 generic.go:334] "Generic (PLEG): container finished" podID="7e8b7ce6-dc1a-474e-83e0-659a275c7aaa" containerID="bcc1fde8869b8c80d9af527398c59643b83c6a5f4b46c772a8e42aeaba7a0f37" exitCode=0 Feb 02 09:13:12 crc kubenswrapper[4764]: I0202 09:13:12.471872 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vp4vx" event={"ID":"7e8b7ce6-dc1a-474e-83e0-659a275c7aaa","Type":"ContainerDied","Data":"bcc1fde8869b8c80d9af527398c59643b83c6a5f4b46c772a8e42aeaba7a0f37"} Feb 02 09:13:12 crc kubenswrapper[4764]: I0202 09:13:12.472296 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vp4vx" event={"ID":"7e8b7ce6-dc1a-474e-83e0-659a275c7aaa","Type":"ContainerStarted","Data":"8a4922eee7c3e8cd96c531aeda44d1c801228a3812e76f2de84f0543e4755ab9"} Feb 02 09:13:12 crc kubenswrapper[4764]: I0202 09:13:12.790474 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-5kwbp"] Feb 02 09:13:12 crc kubenswrapper[4764]: I0202 09:13:12.791573 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5kwbp" Feb 02 09:13:12 crc kubenswrapper[4764]: I0202 09:13:12.793515 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Feb 02 09:13:12 crc kubenswrapper[4764]: I0202 09:13:12.804673 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5kwbp"] Feb 02 09:13:12 crc kubenswrapper[4764]: I0202 09:13:12.854453 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hclwc\" (UniqueName: \"kubernetes.io/projected/4fe67111-6fe1-4886-b192-b75fea98586d-kube-api-access-hclwc\") pod \"certified-operators-5kwbp\" (UID: \"4fe67111-6fe1-4886-b192-b75fea98586d\") " pod="openshift-marketplace/certified-operators-5kwbp" Feb 02 09:13:12 crc kubenswrapper[4764]: I0202 09:13:12.854555 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4fe67111-6fe1-4886-b192-b75fea98586d-utilities\") pod \"certified-operators-5kwbp\" (UID: \"4fe67111-6fe1-4886-b192-b75fea98586d\") " pod="openshift-marketplace/certified-operators-5kwbp" Feb 02 09:13:12 crc kubenswrapper[4764]: I0202 09:13:12.854587 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4fe67111-6fe1-4886-b192-b75fea98586d-catalog-content\") pod \"certified-operators-5kwbp\" (UID: \"4fe67111-6fe1-4886-b192-b75fea98586d\") " pod="openshift-marketplace/certified-operators-5kwbp" Feb 02 09:13:12 crc kubenswrapper[4764]: I0202 09:13:12.955794 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4fe67111-6fe1-4886-b192-b75fea98586d-utilities\") pod \"certified-operators-5kwbp\" (UID: \"4fe67111-6fe1-4886-b192-b75fea98586d\") " pod="openshift-marketplace/certified-operators-5kwbp" Feb 02 09:13:12 crc kubenswrapper[4764]: I0202 09:13:12.955858 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4fe67111-6fe1-4886-b192-b75fea98586d-catalog-content\") pod \"certified-operators-5kwbp\" (UID: \"4fe67111-6fe1-4886-b192-b75fea98586d\") " pod="openshift-marketplace/certified-operators-5kwbp" Feb 02 09:13:12 crc kubenswrapper[4764]: I0202 09:13:12.955905 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hclwc\" (UniqueName: \"kubernetes.io/projected/4fe67111-6fe1-4886-b192-b75fea98586d-kube-api-access-hclwc\") pod \"certified-operators-5kwbp\" (UID: \"4fe67111-6fe1-4886-b192-b75fea98586d\") " pod="openshift-marketplace/certified-operators-5kwbp" Feb 02 09:13:12 crc kubenswrapper[4764]: I0202 09:13:12.956476 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4fe67111-6fe1-4886-b192-b75fea98586d-catalog-content\") pod \"certified-operators-5kwbp\" (UID: \"4fe67111-6fe1-4886-b192-b75fea98586d\") " pod="openshift-marketplace/certified-operators-5kwbp" Feb 02 09:13:12 crc kubenswrapper[4764]: I0202 09:13:12.956603 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4fe67111-6fe1-4886-b192-b75fea98586d-utilities\") pod \"certified-operators-5kwbp\" (UID: \"4fe67111-6fe1-4886-b192-b75fea98586d\") " pod="openshift-marketplace/certified-operators-5kwbp" Feb 02 09:13:12 crc kubenswrapper[4764]: I0202 09:13:12.998721 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hclwc\" (UniqueName: \"kubernetes.io/projected/4fe67111-6fe1-4886-b192-b75fea98586d-kube-api-access-hclwc\") pod \"certified-operators-5kwbp\" (UID: \"4fe67111-6fe1-4886-b192-b75fea98586d\") " pod="openshift-marketplace/certified-operators-5kwbp" Feb 02 09:13:13 crc kubenswrapper[4764]: I0202 09:13:13.166915 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5kwbp" Feb 02 09:13:13 crc kubenswrapper[4764]: I0202 09:13:13.417375 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5kwbp"] Feb 02 09:13:13 crc kubenswrapper[4764]: W0202 09:13:13.426045 4764 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4fe67111_6fe1_4886_b192_b75fea98586d.slice/crio-538d4206ab556fefb8554e39f2c76a21d6c0d0bcf3f7eb4fb48c01c53de39af3 WatchSource:0}: Error finding container 538d4206ab556fefb8554e39f2c76a21d6c0d0bcf3f7eb4fb48c01c53de39af3: Status 404 returned error can't find the container with id 538d4206ab556fefb8554e39f2c76a21d6c0d0bcf3f7eb4fb48c01c53de39af3 Feb 02 09:13:13 crc kubenswrapper[4764]: I0202 09:13:13.477909 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vp4vx" event={"ID":"7e8b7ce6-dc1a-474e-83e0-659a275c7aaa","Type":"ContainerStarted","Data":"5142ae3e8ae1fe78ceac68265da8b41407b74a3824fd2f81fb41ba18790d63de"} Feb 02 09:13:13 crc kubenswrapper[4764]: I0202 09:13:13.479121 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5kwbp" event={"ID":"4fe67111-6fe1-4886-b192-b75fea98586d","Type":"ContainerStarted","Data":"538d4206ab556fefb8554e39f2c76a21d6c0d0bcf3f7eb4fb48c01c53de39af3"} Feb 02 09:13:13 crc kubenswrapper[4764]: I0202 09:13:13.523315 4764 patch_prober.go:28] interesting pod/machine-config-daemon-4ndm4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 09:13:13 crc kubenswrapper[4764]: I0202 09:13:13.523362 4764 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 09:13:13 crc kubenswrapper[4764]: I0202 09:13:13.789589 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-6kgcj"] Feb 02 09:13:13 crc kubenswrapper[4764]: I0202 09:13:13.792531 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6kgcj" Feb 02 09:13:13 crc kubenswrapper[4764]: I0202 09:13:13.794423 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Feb 02 09:13:13 crc kubenswrapper[4764]: I0202 09:13:13.809697 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6kgcj"] Feb 02 09:13:13 crc kubenswrapper[4764]: I0202 09:13:13.865774 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/730067d3-5e02-4aa1-af49-9bb70546886f-utilities\") pod \"community-operators-6kgcj\" (UID: \"730067d3-5e02-4aa1-af49-9bb70546886f\") " pod="openshift-marketplace/community-operators-6kgcj" Feb 02 09:13:13 crc kubenswrapper[4764]: I0202 09:13:13.865834 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/730067d3-5e02-4aa1-af49-9bb70546886f-catalog-content\") pod \"community-operators-6kgcj\" (UID: \"730067d3-5e02-4aa1-af49-9bb70546886f\") " pod="openshift-marketplace/community-operators-6kgcj" Feb 02 09:13:13 crc kubenswrapper[4764]: I0202 09:13:13.866116 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kdx9s\" (UniqueName: \"kubernetes.io/projected/730067d3-5e02-4aa1-af49-9bb70546886f-kube-api-access-kdx9s\") pod \"community-operators-6kgcj\" (UID: \"730067d3-5e02-4aa1-af49-9bb70546886f\") " pod="openshift-marketplace/community-operators-6kgcj" Feb 02 09:13:13 crc kubenswrapper[4764]: I0202 09:13:13.967466 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kdx9s\" (UniqueName: \"kubernetes.io/projected/730067d3-5e02-4aa1-af49-9bb70546886f-kube-api-access-kdx9s\") pod \"community-operators-6kgcj\" (UID: \"730067d3-5e02-4aa1-af49-9bb70546886f\") " pod="openshift-marketplace/community-operators-6kgcj" Feb 02 09:13:13 crc kubenswrapper[4764]: I0202 09:13:13.967523 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/730067d3-5e02-4aa1-af49-9bb70546886f-utilities\") pod \"community-operators-6kgcj\" (UID: \"730067d3-5e02-4aa1-af49-9bb70546886f\") " pod="openshift-marketplace/community-operators-6kgcj" Feb 02 09:13:13 crc kubenswrapper[4764]: I0202 09:13:13.967570 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/730067d3-5e02-4aa1-af49-9bb70546886f-catalog-content\") pod \"community-operators-6kgcj\" (UID: \"730067d3-5e02-4aa1-af49-9bb70546886f\") " pod="openshift-marketplace/community-operators-6kgcj" Feb 02 09:13:13 crc kubenswrapper[4764]: I0202 09:13:13.968037 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/730067d3-5e02-4aa1-af49-9bb70546886f-catalog-content\") pod \"community-operators-6kgcj\" (UID: \"730067d3-5e02-4aa1-af49-9bb70546886f\") " pod="openshift-marketplace/community-operators-6kgcj" Feb 02 09:13:13 crc kubenswrapper[4764]: I0202 09:13:13.968106 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/730067d3-5e02-4aa1-af49-9bb70546886f-utilities\") pod \"community-operators-6kgcj\" (UID: \"730067d3-5e02-4aa1-af49-9bb70546886f\") " pod="openshift-marketplace/community-operators-6kgcj" Feb 02 09:13:13 crc kubenswrapper[4764]: I0202 09:13:13.990279 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kdx9s\" (UniqueName: \"kubernetes.io/projected/730067d3-5e02-4aa1-af49-9bb70546886f-kube-api-access-kdx9s\") pod \"community-operators-6kgcj\" (UID: \"730067d3-5e02-4aa1-af49-9bb70546886f\") " pod="openshift-marketplace/community-operators-6kgcj" Feb 02 09:13:14 crc kubenswrapper[4764]: I0202 09:13:14.107433 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6kgcj" Feb 02 09:13:14 crc kubenswrapper[4764]: I0202 09:13:14.305506 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6kgcj"] Feb 02 09:13:14 crc kubenswrapper[4764]: W0202 09:13:14.315444 4764 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod730067d3_5e02_4aa1_af49_9bb70546886f.slice/crio-1165d054c8494992df1f5e6e8e4c82a1accae68aa47f0155a3e7e513b73fac80 WatchSource:0}: Error finding container 1165d054c8494992df1f5e6e8e4c82a1accae68aa47f0155a3e7e513b73fac80: Status 404 returned error can't find the container with id 1165d054c8494992df1f5e6e8e4c82a1accae68aa47f0155a3e7e513b73fac80 Feb 02 09:13:14 crc kubenswrapper[4764]: I0202 09:13:14.485537 4764 generic.go:334] "Generic (PLEG): container finished" podID="730067d3-5e02-4aa1-af49-9bb70546886f" containerID="52f5ebf795386cbea5e661e0fcef59f2887b3c28030d8b129d309f67777bdec4" exitCode=0 Feb 02 09:13:14 crc kubenswrapper[4764]: I0202 09:13:14.485607 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6kgcj" event={"ID":"730067d3-5e02-4aa1-af49-9bb70546886f","Type":"ContainerDied","Data":"52f5ebf795386cbea5e661e0fcef59f2887b3c28030d8b129d309f67777bdec4"} Feb 02 09:13:14 crc kubenswrapper[4764]: I0202 09:13:14.485636 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6kgcj" event={"ID":"730067d3-5e02-4aa1-af49-9bb70546886f","Type":"ContainerStarted","Data":"1165d054c8494992df1f5e6e8e4c82a1accae68aa47f0155a3e7e513b73fac80"} Feb 02 09:13:14 crc kubenswrapper[4764]: I0202 09:13:14.488961 4764 generic.go:334] "Generic (PLEG): container finished" podID="9443ada1-5f42-40df-b8e2-2c072dd7f264" containerID="dae0308f62f4ef6be76ccaaad8d5d05e4500a0ddfd91a54dba2b6c7d9f61cda8" exitCode=0 Feb 02 09:13:14 crc kubenswrapper[4764]: I0202 09:13:14.489041 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rc4qs" event={"ID":"9443ada1-5f42-40df-b8e2-2c072dd7f264","Type":"ContainerDied","Data":"dae0308f62f4ef6be76ccaaad8d5d05e4500a0ddfd91a54dba2b6c7d9f61cda8"} Feb 02 09:13:14 crc kubenswrapper[4764]: I0202 09:13:14.492122 4764 generic.go:334] "Generic (PLEG): container finished" podID="7e8b7ce6-dc1a-474e-83e0-659a275c7aaa" containerID="5142ae3e8ae1fe78ceac68265da8b41407b74a3824fd2f81fb41ba18790d63de" exitCode=0 Feb 02 09:13:14 crc kubenswrapper[4764]: I0202 09:13:14.492232 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vp4vx" event={"ID":"7e8b7ce6-dc1a-474e-83e0-659a275c7aaa","Type":"ContainerDied","Data":"5142ae3e8ae1fe78ceac68265da8b41407b74a3824fd2f81fb41ba18790d63de"} Feb 02 09:13:14 crc kubenswrapper[4764]: I0202 09:13:14.500172 4764 generic.go:334] "Generic (PLEG): container finished" podID="4fe67111-6fe1-4886-b192-b75fea98586d" containerID="def5f1815342fe5c643a6583b216a34631fdfa46877fda640c9ff17011431841" exitCode=0 Feb 02 09:13:14 crc kubenswrapper[4764]: I0202 09:13:14.500224 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5kwbp" event={"ID":"4fe67111-6fe1-4886-b192-b75fea98586d","Type":"ContainerDied","Data":"def5f1815342fe5c643a6583b216a34631fdfa46877fda640c9ff17011431841"} Feb 02 09:13:15 crc kubenswrapper[4764]: I0202 09:13:15.507127 4764 generic.go:334] "Generic (PLEG): container finished" podID="730067d3-5e02-4aa1-af49-9bb70546886f" containerID="6127b8ae207aa23303a457f1bd085e744eeaca8560c40acb23ca622b99e28dc8" exitCode=0 Feb 02 09:13:15 crc kubenswrapper[4764]: I0202 09:13:15.507342 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6kgcj" event={"ID":"730067d3-5e02-4aa1-af49-9bb70546886f","Type":"ContainerDied","Data":"6127b8ae207aa23303a457f1bd085e744eeaca8560c40acb23ca622b99e28dc8"} Feb 02 09:13:15 crc kubenswrapper[4764]: I0202 09:13:15.512692 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rc4qs" event={"ID":"9443ada1-5f42-40df-b8e2-2c072dd7f264","Type":"ContainerStarted","Data":"1ca7f715f8924dbbd7482b2756d70f72dd31a731d82eebabe7b750824ca38c6f"} Feb 02 09:13:15 crc kubenswrapper[4764]: I0202 09:13:15.516050 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vp4vx" event={"ID":"7e8b7ce6-dc1a-474e-83e0-659a275c7aaa","Type":"ContainerStarted","Data":"411e01b7e2c11b6f507b866b19a984efff1428005cdc42b6321931cdd85513ff"} Feb 02 09:13:15 crc kubenswrapper[4764]: I0202 09:13:15.517828 4764 generic.go:334] "Generic (PLEG): container finished" podID="4fe67111-6fe1-4886-b192-b75fea98586d" containerID="1e80ad8030abbea4a999f0544d15c017753134bdbf8e42c9ab04cf2151559a93" exitCode=0 Feb 02 09:13:15 crc kubenswrapper[4764]: I0202 09:13:15.517858 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5kwbp" event={"ID":"4fe67111-6fe1-4886-b192-b75fea98586d","Type":"ContainerDied","Data":"1e80ad8030abbea4a999f0544d15c017753134bdbf8e42c9ab04cf2151559a93"} Feb 02 09:13:15 crc kubenswrapper[4764]: I0202 09:13:15.549824 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-rc4qs" podStartSLOduration=2.105222669 podStartE2EDuration="4.549805367s" podCreationTimestamp="2026-02-02 09:13:11 +0000 UTC" firstStartedPulling="2026-02-02 09:13:12.467989723 +0000 UTC m=+355.401713811" lastFinishedPulling="2026-02-02 09:13:14.912572411 +0000 UTC m=+357.846296509" observedRunningTime="2026-02-02 09:13:15.549126637 +0000 UTC m=+358.482850735" watchObservedRunningTime="2026-02-02 09:13:15.549805367 +0000 UTC m=+358.483529455" Feb 02 09:13:15 crc kubenswrapper[4764]: I0202 09:13:15.580360 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-vp4vx" podStartSLOduration=3.065579303 podStartE2EDuration="5.580344438s" podCreationTimestamp="2026-02-02 09:13:10 +0000 UTC" firstStartedPulling="2026-02-02 09:13:12.474699965 +0000 UTC m=+355.408424053" lastFinishedPulling="2026-02-02 09:13:14.9894651 +0000 UTC m=+357.923189188" observedRunningTime="2026-02-02 09:13:15.576280386 +0000 UTC m=+358.510004474" watchObservedRunningTime="2026-02-02 09:13:15.580344438 +0000 UTC m=+358.514068526" Feb 02 09:13:16 crc kubenswrapper[4764]: I0202 09:13:16.526200 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6kgcj" event={"ID":"730067d3-5e02-4aa1-af49-9bb70546886f","Type":"ContainerStarted","Data":"d0423aa553ad0cb0a0810aef5dd88e957321030b8e6553f2a5c0269474f3ceb6"} Feb 02 09:13:16 crc kubenswrapper[4764]: I0202 09:13:16.529292 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5kwbp" event={"ID":"4fe67111-6fe1-4886-b192-b75fea98586d","Type":"ContainerStarted","Data":"8ce6047e8be3e410fdabf49f6d80c1296654131c7efa2ab3f951faaf00a4fe06"} Feb 02 09:13:16 crc kubenswrapper[4764]: I0202 09:13:16.570331 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-6kgcj" podStartSLOduration=2.090553669 podStartE2EDuration="3.570316282s" podCreationTimestamp="2026-02-02 09:13:13 +0000 UTC" firstStartedPulling="2026-02-02 09:13:14.486579415 +0000 UTC m=+357.420303503" lastFinishedPulling="2026-02-02 09:13:15.966342028 +0000 UTC m=+358.900066116" observedRunningTime="2026-02-02 09:13:16.554060652 +0000 UTC m=+359.487784750" watchObservedRunningTime="2026-02-02 09:13:16.570316282 +0000 UTC m=+359.504040370" Feb 02 09:13:16 crc kubenswrapper[4764]: I0202 09:13:16.570693 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-5kwbp" podStartSLOduration=3.166054105 podStartE2EDuration="4.570688423s" podCreationTimestamp="2026-02-02 09:13:12 +0000 UTC" firstStartedPulling="2026-02-02 09:13:14.502271428 +0000 UTC m=+357.435995506" lastFinishedPulling="2026-02-02 09:13:15.906905746 +0000 UTC m=+358.840629824" observedRunningTime="2026-02-02 09:13:16.566899479 +0000 UTC m=+359.500623577" watchObservedRunningTime="2026-02-02 09:13:16.570688423 +0000 UTC m=+359.504412511" Feb 02 09:13:20 crc kubenswrapper[4764]: I0202 09:13:20.779888 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-vp4vx" Feb 02 09:13:20 crc kubenswrapper[4764]: I0202 09:13:20.780293 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-vp4vx" Feb 02 09:13:20 crc kubenswrapper[4764]: I0202 09:13:20.827602 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-vp4vx" Feb 02 09:13:21 crc kubenswrapper[4764]: I0202 09:13:21.608660 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-vp4vx" Feb 02 09:13:21 crc kubenswrapper[4764]: I0202 09:13:21.845617 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-rc4qs" Feb 02 09:13:21 crc kubenswrapper[4764]: I0202 09:13:21.846334 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-rc4qs" Feb 02 09:13:21 crc kubenswrapper[4764]: I0202 09:13:21.887896 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-rc4qs" Feb 02 09:13:22 crc kubenswrapper[4764]: I0202 09:13:22.616681 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-rc4qs" Feb 02 09:13:23 crc kubenswrapper[4764]: I0202 09:13:23.167103 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-5kwbp" Feb 02 09:13:23 crc kubenswrapper[4764]: I0202 09:13:23.167984 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-5kwbp" Feb 02 09:13:23 crc kubenswrapper[4764]: I0202 09:13:23.220021 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-5kwbp" Feb 02 09:13:23 crc kubenswrapper[4764]: I0202 09:13:23.609211 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-5kwbp" Feb 02 09:13:24 crc kubenswrapper[4764]: I0202 09:13:24.107984 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-6kgcj" Feb 02 09:13:24 crc kubenswrapper[4764]: I0202 09:13:24.108024 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-6kgcj" Feb 02 09:13:24 crc kubenswrapper[4764]: I0202 09:13:24.164432 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-6kgcj" Feb 02 09:13:24 crc kubenswrapper[4764]: I0202 09:13:24.624337 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-6kgcj" Feb 02 09:13:43 crc kubenswrapper[4764]: I0202 09:13:43.523027 4764 patch_prober.go:28] interesting pod/machine-config-daemon-4ndm4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 09:13:43 crc kubenswrapper[4764]: I0202 09:13:43.523646 4764 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 09:14:13 crc kubenswrapper[4764]: I0202 09:14:13.523682 4764 patch_prober.go:28] interesting pod/machine-config-daemon-4ndm4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 09:14:13 crc kubenswrapper[4764]: I0202 09:14:13.524599 4764 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 09:14:13 crc kubenswrapper[4764]: I0202 09:14:13.524680 4764 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" Feb 02 09:14:13 crc kubenswrapper[4764]: I0202 09:14:13.526032 4764 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c2244e8f664081c8f021b4db06125b0a2f32a0cc2a1d90e1f92743595267e6b2"} pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 02 09:14:13 crc kubenswrapper[4764]: I0202 09:14:13.526166 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" containerName="machine-config-daemon" containerID="cri-o://c2244e8f664081c8f021b4db06125b0a2f32a0cc2a1d90e1f92743595267e6b2" gracePeriod=600 Feb 02 09:14:13 crc kubenswrapper[4764]: I0202 09:14:13.873914 4764 generic.go:334] "Generic (PLEG): container finished" podID="d192f670-9f9d-4539-9641-e4bed73acdd4" containerID="c2244e8f664081c8f021b4db06125b0a2f32a0cc2a1d90e1f92743595267e6b2" exitCode=0 Feb 02 09:14:13 crc kubenswrapper[4764]: I0202 09:14:13.873976 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" event={"ID":"d192f670-9f9d-4539-9641-e4bed73acdd4","Type":"ContainerDied","Data":"c2244e8f664081c8f021b4db06125b0a2f32a0cc2a1d90e1f92743595267e6b2"} Feb 02 09:14:13 crc kubenswrapper[4764]: I0202 09:14:13.874006 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" event={"ID":"d192f670-9f9d-4539-9641-e4bed73acdd4","Type":"ContainerStarted","Data":"d0647a58d2de9a2cdd486044a3db4f79924b0b477fa546211f680159a64217a7"} Feb 02 09:14:13 crc kubenswrapper[4764]: I0202 09:14:13.874022 4764 scope.go:117] "RemoveContainer" containerID="5fdf06410349b7c99aa166e87fbe80bf68e2ae923635a6a978b7e2cd50383f7b" Feb 02 09:15:00 crc kubenswrapper[4764]: I0202 09:15:00.194534 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29500395-v7kj2"] Feb 02 09:15:00 crc kubenswrapper[4764]: I0202 09:15:00.197879 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29500395-v7kj2" Feb 02 09:15:00 crc kubenswrapper[4764]: I0202 09:15:00.202003 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 02 09:15:00 crc kubenswrapper[4764]: I0202 09:15:00.202259 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 02 09:15:00 crc kubenswrapper[4764]: I0202 09:15:00.205448 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29500395-v7kj2"] Feb 02 09:15:00 crc kubenswrapper[4764]: I0202 09:15:00.277053 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jfdp2\" (UniqueName: \"kubernetes.io/projected/60049dcd-be0a-4eda-800a-1ce0c6da3195-kube-api-access-jfdp2\") pod \"collect-profiles-29500395-v7kj2\" (UID: \"60049dcd-be0a-4eda-800a-1ce0c6da3195\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500395-v7kj2" Feb 02 09:15:00 crc kubenswrapper[4764]: I0202 09:15:00.277232 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/60049dcd-be0a-4eda-800a-1ce0c6da3195-config-volume\") pod \"collect-profiles-29500395-v7kj2\" (UID: \"60049dcd-be0a-4eda-800a-1ce0c6da3195\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500395-v7kj2" Feb 02 09:15:00 crc kubenswrapper[4764]: I0202 09:15:00.277307 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/60049dcd-be0a-4eda-800a-1ce0c6da3195-secret-volume\") pod \"collect-profiles-29500395-v7kj2\" (UID: \"60049dcd-be0a-4eda-800a-1ce0c6da3195\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500395-v7kj2" Feb 02 09:15:00 crc kubenswrapper[4764]: I0202 09:15:00.378254 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/60049dcd-be0a-4eda-800a-1ce0c6da3195-secret-volume\") pod \"collect-profiles-29500395-v7kj2\" (UID: \"60049dcd-be0a-4eda-800a-1ce0c6da3195\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500395-v7kj2" Feb 02 09:15:00 crc kubenswrapper[4764]: I0202 09:15:00.378371 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jfdp2\" (UniqueName: \"kubernetes.io/projected/60049dcd-be0a-4eda-800a-1ce0c6da3195-kube-api-access-jfdp2\") pod \"collect-profiles-29500395-v7kj2\" (UID: \"60049dcd-be0a-4eda-800a-1ce0c6da3195\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500395-v7kj2" Feb 02 09:15:00 crc kubenswrapper[4764]: I0202 09:15:00.378412 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/60049dcd-be0a-4eda-800a-1ce0c6da3195-config-volume\") pod \"collect-profiles-29500395-v7kj2\" (UID: \"60049dcd-be0a-4eda-800a-1ce0c6da3195\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500395-v7kj2" Feb 02 09:15:00 crc kubenswrapper[4764]: I0202 09:15:00.379552 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/60049dcd-be0a-4eda-800a-1ce0c6da3195-config-volume\") pod \"collect-profiles-29500395-v7kj2\" (UID: \"60049dcd-be0a-4eda-800a-1ce0c6da3195\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500395-v7kj2" Feb 02 09:15:00 crc kubenswrapper[4764]: I0202 09:15:00.392675 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/60049dcd-be0a-4eda-800a-1ce0c6da3195-secret-volume\") pod \"collect-profiles-29500395-v7kj2\" (UID: \"60049dcd-be0a-4eda-800a-1ce0c6da3195\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500395-v7kj2" Feb 02 09:15:00 crc kubenswrapper[4764]: I0202 09:15:00.403266 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jfdp2\" (UniqueName: \"kubernetes.io/projected/60049dcd-be0a-4eda-800a-1ce0c6da3195-kube-api-access-jfdp2\") pod \"collect-profiles-29500395-v7kj2\" (UID: \"60049dcd-be0a-4eda-800a-1ce0c6da3195\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500395-v7kj2" Feb 02 09:15:00 crc kubenswrapper[4764]: I0202 09:15:00.525985 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29500395-v7kj2" Feb 02 09:15:00 crc kubenswrapper[4764]: I0202 09:15:00.858649 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29500395-v7kj2"] Feb 02 09:15:01 crc kubenswrapper[4764]: I0202 09:15:01.170330 4764 generic.go:334] "Generic (PLEG): container finished" podID="60049dcd-be0a-4eda-800a-1ce0c6da3195" containerID="a86e06b425266081de6db5fd5c39780eeb69d1c8f9406a1f104bd10689e15b82" exitCode=0 Feb 02 09:15:01 crc kubenswrapper[4764]: I0202 09:15:01.170382 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29500395-v7kj2" event={"ID":"60049dcd-be0a-4eda-800a-1ce0c6da3195","Type":"ContainerDied","Data":"a86e06b425266081de6db5fd5c39780eeb69d1c8f9406a1f104bd10689e15b82"} Feb 02 09:15:01 crc kubenswrapper[4764]: I0202 09:15:01.170653 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29500395-v7kj2" event={"ID":"60049dcd-be0a-4eda-800a-1ce0c6da3195","Type":"ContainerStarted","Data":"8bddbf571fba0327d5c95b3b4d28f228a4bfce8c30e94afe48d8e2a18cac8170"} Feb 02 09:15:02 crc kubenswrapper[4764]: I0202 09:15:02.404284 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29500395-v7kj2" Feb 02 09:15:02 crc kubenswrapper[4764]: I0202 09:15:02.508976 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/60049dcd-be0a-4eda-800a-1ce0c6da3195-config-volume\") pod \"60049dcd-be0a-4eda-800a-1ce0c6da3195\" (UID: \"60049dcd-be0a-4eda-800a-1ce0c6da3195\") " Feb 02 09:15:02 crc kubenswrapper[4764]: I0202 09:15:02.509086 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/60049dcd-be0a-4eda-800a-1ce0c6da3195-secret-volume\") pod \"60049dcd-be0a-4eda-800a-1ce0c6da3195\" (UID: \"60049dcd-be0a-4eda-800a-1ce0c6da3195\") " Feb 02 09:15:02 crc kubenswrapper[4764]: I0202 09:15:02.509134 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jfdp2\" (UniqueName: \"kubernetes.io/projected/60049dcd-be0a-4eda-800a-1ce0c6da3195-kube-api-access-jfdp2\") pod \"60049dcd-be0a-4eda-800a-1ce0c6da3195\" (UID: \"60049dcd-be0a-4eda-800a-1ce0c6da3195\") " Feb 02 09:15:02 crc kubenswrapper[4764]: I0202 09:15:02.511737 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/60049dcd-be0a-4eda-800a-1ce0c6da3195-config-volume" (OuterVolumeSpecName: "config-volume") pod "60049dcd-be0a-4eda-800a-1ce0c6da3195" (UID: "60049dcd-be0a-4eda-800a-1ce0c6da3195"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:15:02 crc kubenswrapper[4764]: I0202 09:15:02.515314 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60049dcd-be0a-4eda-800a-1ce0c6da3195-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "60049dcd-be0a-4eda-800a-1ce0c6da3195" (UID: "60049dcd-be0a-4eda-800a-1ce0c6da3195"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:15:02 crc kubenswrapper[4764]: I0202 09:15:02.515841 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/60049dcd-be0a-4eda-800a-1ce0c6da3195-kube-api-access-jfdp2" (OuterVolumeSpecName: "kube-api-access-jfdp2") pod "60049dcd-be0a-4eda-800a-1ce0c6da3195" (UID: "60049dcd-be0a-4eda-800a-1ce0c6da3195"). InnerVolumeSpecName "kube-api-access-jfdp2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:15:02 crc kubenswrapper[4764]: I0202 09:15:02.611219 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jfdp2\" (UniqueName: \"kubernetes.io/projected/60049dcd-be0a-4eda-800a-1ce0c6da3195-kube-api-access-jfdp2\") on node \"crc\" DevicePath \"\"" Feb 02 09:15:02 crc kubenswrapper[4764]: I0202 09:15:02.611337 4764 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/60049dcd-be0a-4eda-800a-1ce0c6da3195-config-volume\") on node \"crc\" DevicePath \"\"" Feb 02 09:15:02 crc kubenswrapper[4764]: I0202 09:15:02.611440 4764 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/60049dcd-be0a-4eda-800a-1ce0c6da3195-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 02 09:15:03 crc kubenswrapper[4764]: I0202 09:15:03.192734 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29500395-v7kj2" event={"ID":"60049dcd-be0a-4eda-800a-1ce0c6da3195","Type":"ContainerDied","Data":"8bddbf571fba0327d5c95b3b4d28f228a4bfce8c30e94afe48d8e2a18cac8170"} Feb 02 09:15:03 crc kubenswrapper[4764]: I0202 09:15:03.192816 4764 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8bddbf571fba0327d5c95b3b4d28f228a4bfce8c30e94afe48d8e2a18cac8170" Feb 02 09:15:03 crc kubenswrapper[4764]: I0202 09:15:03.192768 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29500395-v7kj2" Feb 02 09:15:03 crc kubenswrapper[4764]: E0202 09:15:03.263877 4764 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod60049dcd_be0a_4eda_800a_1ce0c6da3195.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod60049dcd_be0a_4eda_800a_1ce0c6da3195.slice/crio-8bddbf571fba0327d5c95b3b4d28f228a4bfce8c30e94afe48d8e2a18cac8170\": RecentStats: unable to find data in memory cache]" Feb 02 09:16:13 crc kubenswrapper[4764]: I0202 09:16:13.523016 4764 patch_prober.go:28] interesting pod/machine-config-daemon-4ndm4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 09:16:13 crc kubenswrapper[4764]: I0202 09:16:13.523787 4764 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 09:16:43 crc kubenswrapper[4764]: I0202 09:16:43.523731 4764 patch_prober.go:28] interesting pod/machine-config-daemon-4ndm4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 09:16:43 crc kubenswrapper[4764]: I0202 09:16:43.524607 4764 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 09:16:52 crc kubenswrapper[4764]: I0202 09:16:52.603649 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-jksjr"] Feb 02 09:16:52 crc kubenswrapper[4764]: E0202 09:16:52.604247 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60049dcd-be0a-4eda-800a-1ce0c6da3195" containerName="collect-profiles" Feb 02 09:16:52 crc kubenswrapper[4764]: I0202 09:16:52.604258 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="60049dcd-be0a-4eda-800a-1ce0c6da3195" containerName="collect-profiles" Feb 02 09:16:52 crc kubenswrapper[4764]: I0202 09:16:52.604361 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="60049dcd-be0a-4eda-800a-1ce0c6da3195" containerName="collect-profiles" Feb 02 09:16:52 crc kubenswrapper[4764]: I0202 09:16:52.604736 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-jksjr" Feb 02 09:16:52 crc kubenswrapper[4764]: I0202 09:16:52.629254 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-jksjr"] Feb 02 09:16:52 crc kubenswrapper[4764]: I0202 09:16:52.795829 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/0bd6fa78-a55f-4640-a5aa-08d0092de030-registry-tls\") pod \"image-registry-66df7c8f76-jksjr\" (UID: \"0bd6fa78-a55f-4640-a5aa-08d0092de030\") " pod="openshift-image-registry/image-registry-66df7c8f76-jksjr" Feb 02 09:16:52 crc kubenswrapper[4764]: I0202 09:16:52.795882 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-jksjr\" (UID: \"0bd6fa78-a55f-4640-a5aa-08d0092de030\") " pod="openshift-image-registry/image-registry-66df7c8f76-jksjr" Feb 02 09:16:52 crc kubenswrapper[4764]: I0202 09:16:52.795984 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/0bd6fa78-a55f-4640-a5aa-08d0092de030-registry-certificates\") pod \"image-registry-66df7c8f76-jksjr\" (UID: \"0bd6fa78-a55f-4640-a5aa-08d0092de030\") " pod="openshift-image-registry/image-registry-66df7c8f76-jksjr" Feb 02 09:16:52 crc kubenswrapper[4764]: I0202 09:16:52.796033 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/0bd6fa78-a55f-4640-a5aa-08d0092de030-ca-trust-extracted\") pod \"image-registry-66df7c8f76-jksjr\" (UID: \"0bd6fa78-a55f-4640-a5aa-08d0092de030\") " pod="openshift-image-registry/image-registry-66df7c8f76-jksjr" Feb 02 09:16:52 crc kubenswrapper[4764]: I0202 09:16:52.796056 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/0bd6fa78-a55f-4640-a5aa-08d0092de030-bound-sa-token\") pod \"image-registry-66df7c8f76-jksjr\" (UID: \"0bd6fa78-a55f-4640-a5aa-08d0092de030\") " pod="openshift-image-registry/image-registry-66df7c8f76-jksjr" Feb 02 09:16:52 crc kubenswrapper[4764]: I0202 09:16:52.796096 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/0bd6fa78-a55f-4640-a5aa-08d0092de030-installation-pull-secrets\") pod \"image-registry-66df7c8f76-jksjr\" (UID: \"0bd6fa78-a55f-4640-a5aa-08d0092de030\") " pod="openshift-image-registry/image-registry-66df7c8f76-jksjr" Feb 02 09:16:52 crc kubenswrapper[4764]: I0202 09:16:52.796118 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0bd6fa78-a55f-4640-a5aa-08d0092de030-trusted-ca\") pod \"image-registry-66df7c8f76-jksjr\" (UID: \"0bd6fa78-a55f-4640-a5aa-08d0092de030\") " pod="openshift-image-registry/image-registry-66df7c8f76-jksjr" Feb 02 09:16:52 crc kubenswrapper[4764]: I0202 09:16:52.796143 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6d794\" (UniqueName: \"kubernetes.io/projected/0bd6fa78-a55f-4640-a5aa-08d0092de030-kube-api-access-6d794\") pod \"image-registry-66df7c8f76-jksjr\" (UID: \"0bd6fa78-a55f-4640-a5aa-08d0092de030\") " pod="openshift-image-registry/image-registry-66df7c8f76-jksjr" Feb 02 09:16:52 crc kubenswrapper[4764]: I0202 09:16:52.824644 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-jksjr\" (UID: \"0bd6fa78-a55f-4640-a5aa-08d0092de030\") " pod="openshift-image-registry/image-registry-66df7c8f76-jksjr" Feb 02 09:16:52 crc kubenswrapper[4764]: I0202 09:16:52.897077 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/0bd6fa78-a55f-4640-a5aa-08d0092de030-registry-tls\") pod \"image-registry-66df7c8f76-jksjr\" (UID: \"0bd6fa78-a55f-4640-a5aa-08d0092de030\") " pod="openshift-image-registry/image-registry-66df7c8f76-jksjr" Feb 02 09:16:52 crc kubenswrapper[4764]: I0202 09:16:52.897151 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/0bd6fa78-a55f-4640-a5aa-08d0092de030-registry-certificates\") pod \"image-registry-66df7c8f76-jksjr\" (UID: \"0bd6fa78-a55f-4640-a5aa-08d0092de030\") " pod="openshift-image-registry/image-registry-66df7c8f76-jksjr" Feb 02 09:16:52 crc kubenswrapper[4764]: I0202 09:16:52.897206 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/0bd6fa78-a55f-4640-a5aa-08d0092de030-ca-trust-extracted\") pod \"image-registry-66df7c8f76-jksjr\" (UID: \"0bd6fa78-a55f-4640-a5aa-08d0092de030\") " pod="openshift-image-registry/image-registry-66df7c8f76-jksjr" Feb 02 09:16:52 crc kubenswrapper[4764]: I0202 09:16:52.897229 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/0bd6fa78-a55f-4640-a5aa-08d0092de030-bound-sa-token\") pod \"image-registry-66df7c8f76-jksjr\" (UID: \"0bd6fa78-a55f-4640-a5aa-08d0092de030\") " pod="openshift-image-registry/image-registry-66df7c8f76-jksjr" Feb 02 09:16:52 crc kubenswrapper[4764]: I0202 09:16:52.897253 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/0bd6fa78-a55f-4640-a5aa-08d0092de030-installation-pull-secrets\") pod \"image-registry-66df7c8f76-jksjr\" (UID: \"0bd6fa78-a55f-4640-a5aa-08d0092de030\") " pod="openshift-image-registry/image-registry-66df7c8f76-jksjr" Feb 02 09:16:52 crc kubenswrapper[4764]: I0202 09:16:52.897274 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0bd6fa78-a55f-4640-a5aa-08d0092de030-trusted-ca\") pod \"image-registry-66df7c8f76-jksjr\" (UID: \"0bd6fa78-a55f-4640-a5aa-08d0092de030\") " pod="openshift-image-registry/image-registry-66df7c8f76-jksjr" Feb 02 09:16:52 crc kubenswrapper[4764]: I0202 09:16:52.897298 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6d794\" (UniqueName: \"kubernetes.io/projected/0bd6fa78-a55f-4640-a5aa-08d0092de030-kube-api-access-6d794\") pod \"image-registry-66df7c8f76-jksjr\" (UID: \"0bd6fa78-a55f-4640-a5aa-08d0092de030\") " pod="openshift-image-registry/image-registry-66df7c8f76-jksjr" Feb 02 09:16:52 crc kubenswrapper[4764]: I0202 09:16:52.898419 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/0bd6fa78-a55f-4640-a5aa-08d0092de030-ca-trust-extracted\") pod \"image-registry-66df7c8f76-jksjr\" (UID: \"0bd6fa78-a55f-4640-a5aa-08d0092de030\") " pod="openshift-image-registry/image-registry-66df7c8f76-jksjr" Feb 02 09:16:52 crc kubenswrapper[4764]: I0202 09:16:52.898806 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0bd6fa78-a55f-4640-a5aa-08d0092de030-trusted-ca\") pod \"image-registry-66df7c8f76-jksjr\" (UID: \"0bd6fa78-a55f-4640-a5aa-08d0092de030\") " pod="openshift-image-registry/image-registry-66df7c8f76-jksjr" Feb 02 09:16:52 crc kubenswrapper[4764]: I0202 09:16:52.899566 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/0bd6fa78-a55f-4640-a5aa-08d0092de030-registry-certificates\") pod \"image-registry-66df7c8f76-jksjr\" (UID: \"0bd6fa78-a55f-4640-a5aa-08d0092de030\") " pod="openshift-image-registry/image-registry-66df7c8f76-jksjr" Feb 02 09:16:52 crc kubenswrapper[4764]: I0202 09:16:52.909822 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/0bd6fa78-a55f-4640-a5aa-08d0092de030-registry-tls\") pod \"image-registry-66df7c8f76-jksjr\" (UID: \"0bd6fa78-a55f-4640-a5aa-08d0092de030\") " pod="openshift-image-registry/image-registry-66df7c8f76-jksjr" Feb 02 09:16:52 crc kubenswrapper[4764]: I0202 09:16:52.910058 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/0bd6fa78-a55f-4640-a5aa-08d0092de030-installation-pull-secrets\") pod \"image-registry-66df7c8f76-jksjr\" (UID: \"0bd6fa78-a55f-4640-a5aa-08d0092de030\") " pod="openshift-image-registry/image-registry-66df7c8f76-jksjr" Feb 02 09:16:52 crc kubenswrapper[4764]: I0202 09:16:52.915383 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6d794\" (UniqueName: \"kubernetes.io/projected/0bd6fa78-a55f-4640-a5aa-08d0092de030-kube-api-access-6d794\") pod \"image-registry-66df7c8f76-jksjr\" (UID: \"0bd6fa78-a55f-4640-a5aa-08d0092de030\") " pod="openshift-image-registry/image-registry-66df7c8f76-jksjr" Feb 02 09:16:52 crc kubenswrapper[4764]: I0202 09:16:52.919013 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/0bd6fa78-a55f-4640-a5aa-08d0092de030-bound-sa-token\") pod \"image-registry-66df7c8f76-jksjr\" (UID: \"0bd6fa78-a55f-4640-a5aa-08d0092de030\") " pod="openshift-image-registry/image-registry-66df7c8f76-jksjr" Feb 02 09:16:52 crc kubenswrapper[4764]: I0202 09:16:52.921707 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-jksjr" Feb 02 09:16:53 crc kubenswrapper[4764]: I0202 09:16:53.114889 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-jksjr"] Feb 02 09:16:54 crc kubenswrapper[4764]: I0202 09:16:54.106510 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-jksjr" event={"ID":"0bd6fa78-a55f-4640-a5aa-08d0092de030","Type":"ContainerStarted","Data":"956e101fcf71ef5d90ecad5317fbc1ca37c749a44c9ac211117dbb54786c0f27"} Feb 02 09:16:54 crc kubenswrapper[4764]: I0202 09:16:54.106799 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-jksjr" event={"ID":"0bd6fa78-a55f-4640-a5aa-08d0092de030","Type":"ContainerStarted","Data":"3f9cfdd538e94a2b7318f807662d5f90a31762a2c9b966df8d305122c40ceb3b"} Feb 02 09:16:54 crc kubenswrapper[4764]: I0202 09:16:54.106829 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-jksjr" Feb 02 09:17:12 crc kubenswrapper[4764]: I0202 09:17:12.932346 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-jksjr" Feb 02 09:17:12 crc kubenswrapper[4764]: I0202 09:17:12.960301 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-jksjr" podStartSLOduration=20.960279927 podStartE2EDuration="20.960279927s" podCreationTimestamp="2026-02-02 09:16:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:16:54.137516215 +0000 UTC m=+577.071240333" watchObservedRunningTime="2026-02-02 09:17:12.960279927 +0000 UTC m=+595.894004025" Feb 02 09:17:13 crc kubenswrapper[4764]: I0202 09:17:13.020578 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-9nkvv"] Feb 02 09:17:13 crc kubenswrapper[4764]: I0202 09:17:13.523079 4764 patch_prober.go:28] interesting pod/machine-config-daemon-4ndm4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 09:17:13 crc kubenswrapper[4764]: I0202 09:17:13.523497 4764 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 09:17:13 crc kubenswrapper[4764]: I0202 09:17:13.523562 4764 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" Feb 02 09:17:13 crc kubenswrapper[4764]: I0202 09:17:13.524403 4764 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d0647a58d2de9a2cdd486044a3db4f79924b0b477fa546211f680159a64217a7"} pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 02 09:17:13 crc kubenswrapper[4764]: I0202 09:17:13.524506 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" containerName="machine-config-daemon" containerID="cri-o://d0647a58d2de9a2cdd486044a3db4f79924b0b477fa546211f680159a64217a7" gracePeriod=600 Feb 02 09:17:14 crc kubenswrapper[4764]: I0202 09:17:14.250257 4764 generic.go:334] "Generic (PLEG): container finished" podID="d192f670-9f9d-4539-9641-e4bed73acdd4" containerID="d0647a58d2de9a2cdd486044a3db4f79924b0b477fa546211f680159a64217a7" exitCode=0 Feb 02 09:17:14 crc kubenswrapper[4764]: I0202 09:17:14.250343 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" event={"ID":"d192f670-9f9d-4539-9641-e4bed73acdd4","Type":"ContainerDied","Data":"d0647a58d2de9a2cdd486044a3db4f79924b0b477fa546211f680159a64217a7"} Feb 02 09:17:14 crc kubenswrapper[4764]: I0202 09:17:14.250573 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" event={"ID":"d192f670-9f9d-4539-9641-e4bed73acdd4","Type":"ContainerStarted","Data":"942c4d2fe3179033f2460fe16951c41f44faf33abaa4e47531eeac47beeeae75"} Feb 02 09:17:14 crc kubenswrapper[4764]: I0202 09:17:14.250597 4764 scope.go:117] "RemoveContainer" containerID="c2244e8f664081c8f021b4db06125b0a2f32a0cc2a1d90e1f92743595267e6b2" Feb 02 09:17:38 crc kubenswrapper[4764]: I0202 09:17:38.083630 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-9nkvv" podUID="527edab8-f88f-4d82-b42e-b8d52fcaa7d4" containerName="registry" containerID="cri-o://37fa793b00f263b65d2c4b4c33beaeacd1c8c3cc4fac5d41b7d06caa9978028f" gracePeriod=30 Feb 02 09:17:38 crc kubenswrapper[4764]: I0202 09:17:38.423710 4764 generic.go:334] "Generic (PLEG): container finished" podID="527edab8-f88f-4d82-b42e-b8d52fcaa7d4" containerID="37fa793b00f263b65d2c4b4c33beaeacd1c8c3cc4fac5d41b7d06caa9978028f" exitCode=0 Feb 02 09:17:38 crc kubenswrapper[4764]: I0202 09:17:38.423811 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-9nkvv" event={"ID":"527edab8-f88f-4d82-b42e-b8d52fcaa7d4","Type":"ContainerDied","Data":"37fa793b00f263b65d2c4b4c33beaeacd1c8c3cc4fac5d41b7d06caa9978028f"} Feb 02 09:17:38 crc kubenswrapper[4764]: I0202 09:17:38.546526 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-9nkvv" Feb 02 09:17:38 crc kubenswrapper[4764]: I0202 09:17:38.680656 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/527edab8-f88f-4d82-b42e-b8d52fcaa7d4-ca-trust-extracted\") pod \"527edab8-f88f-4d82-b42e-b8d52fcaa7d4\" (UID: \"527edab8-f88f-4d82-b42e-b8d52fcaa7d4\") " Feb 02 09:17:38 crc kubenswrapper[4764]: I0202 09:17:38.680693 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/527edab8-f88f-4d82-b42e-b8d52fcaa7d4-installation-pull-secrets\") pod \"527edab8-f88f-4d82-b42e-b8d52fcaa7d4\" (UID: \"527edab8-f88f-4d82-b42e-b8d52fcaa7d4\") " Feb 02 09:17:38 crc kubenswrapper[4764]: I0202 09:17:38.680714 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-95srz\" (UniqueName: \"kubernetes.io/projected/527edab8-f88f-4d82-b42e-b8d52fcaa7d4-kube-api-access-95srz\") pod \"527edab8-f88f-4d82-b42e-b8d52fcaa7d4\" (UID: \"527edab8-f88f-4d82-b42e-b8d52fcaa7d4\") " Feb 02 09:17:38 crc kubenswrapper[4764]: I0202 09:17:38.680747 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/527edab8-f88f-4d82-b42e-b8d52fcaa7d4-registry-certificates\") pod \"527edab8-f88f-4d82-b42e-b8d52fcaa7d4\" (UID: \"527edab8-f88f-4d82-b42e-b8d52fcaa7d4\") " Feb 02 09:17:38 crc kubenswrapper[4764]: I0202 09:17:38.680764 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/527edab8-f88f-4d82-b42e-b8d52fcaa7d4-bound-sa-token\") pod \"527edab8-f88f-4d82-b42e-b8d52fcaa7d4\" (UID: \"527edab8-f88f-4d82-b42e-b8d52fcaa7d4\") " Feb 02 09:17:38 crc kubenswrapper[4764]: I0202 09:17:38.680982 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"527edab8-f88f-4d82-b42e-b8d52fcaa7d4\" (UID: \"527edab8-f88f-4d82-b42e-b8d52fcaa7d4\") " Feb 02 09:17:38 crc kubenswrapper[4764]: I0202 09:17:38.681002 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/527edab8-f88f-4d82-b42e-b8d52fcaa7d4-registry-tls\") pod \"527edab8-f88f-4d82-b42e-b8d52fcaa7d4\" (UID: \"527edab8-f88f-4d82-b42e-b8d52fcaa7d4\") " Feb 02 09:17:38 crc kubenswrapper[4764]: I0202 09:17:38.681064 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/527edab8-f88f-4d82-b42e-b8d52fcaa7d4-trusted-ca\") pod \"527edab8-f88f-4d82-b42e-b8d52fcaa7d4\" (UID: \"527edab8-f88f-4d82-b42e-b8d52fcaa7d4\") " Feb 02 09:17:38 crc kubenswrapper[4764]: I0202 09:17:38.681733 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/527edab8-f88f-4d82-b42e-b8d52fcaa7d4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "527edab8-f88f-4d82-b42e-b8d52fcaa7d4" (UID: "527edab8-f88f-4d82-b42e-b8d52fcaa7d4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:17:38 crc kubenswrapper[4764]: I0202 09:17:38.681897 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/527edab8-f88f-4d82-b42e-b8d52fcaa7d4-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "527edab8-f88f-4d82-b42e-b8d52fcaa7d4" (UID: "527edab8-f88f-4d82-b42e-b8d52fcaa7d4"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:17:38 crc kubenswrapper[4764]: I0202 09:17:38.694471 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/527edab8-f88f-4d82-b42e-b8d52fcaa7d4-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "527edab8-f88f-4d82-b42e-b8d52fcaa7d4" (UID: "527edab8-f88f-4d82-b42e-b8d52fcaa7d4"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:17:38 crc kubenswrapper[4764]: I0202 09:17:38.694533 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/527edab8-f88f-4d82-b42e-b8d52fcaa7d4-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "527edab8-f88f-4d82-b42e-b8d52fcaa7d4" (UID: "527edab8-f88f-4d82-b42e-b8d52fcaa7d4"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:17:38 crc kubenswrapper[4764]: I0202 09:17:38.699195 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/527edab8-f88f-4d82-b42e-b8d52fcaa7d4-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "527edab8-f88f-4d82-b42e-b8d52fcaa7d4" (UID: "527edab8-f88f-4d82-b42e-b8d52fcaa7d4"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:17:38 crc kubenswrapper[4764]: I0202 09:17:38.699326 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/527edab8-f88f-4d82-b42e-b8d52fcaa7d4-kube-api-access-95srz" (OuterVolumeSpecName: "kube-api-access-95srz") pod "527edab8-f88f-4d82-b42e-b8d52fcaa7d4" (UID: "527edab8-f88f-4d82-b42e-b8d52fcaa7d4"). InnerVolumeSpecName "kube-api-access-95srz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:17:38 crc kubenswrapper[4764]: I0202 09:17:38.700888 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "527edab8-f88f-4d82-b42e-b8d52fcaa7d4" (UID: "527edab8-f88f-4d82-b42e-b8d52fcaa7d4"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Feb 02 09:17:38 crc kubenswrapper[4764]: I0202 09:17:38.711927 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/527edab8-f88f-4d82-b42e-b8d52fcaa7d4-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "527edab8-f88f-4d82-b42e-b8d52fcaa7d4" (UID: "527edab8-f88f-4d82-b42e-b8d52fcaa7d4"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 09:17:38 crc kubenswrapper[4764]: I0202 09:17:38.782335 4764 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/527edab8-f88f-4d82-b42e-b8d52fcaa7d4-bound-sa-token\") on node \"crc\" DevicePath \"\"" Feb 02 09:17:38 crc kubenswrapper[4764]: I0202 09:17:38.782388 4764 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/527edab8-f88f-4d82-b42e-b8d52fcaa7d4-registry-tls\") on node \"crc\" DevicePath \"\"" Feb 02 09:17:38 crc kubenswrapper[4764]: I0202 09:17:38.782407 4764 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/527edab8-f88f-4d82-b42e-b8d52fcaa7d4-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 02 09:17:38 crc kubenswrapper[4764]: I0202 09:17:38.782426 4764 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/527edab8-f88f-4d82-b42e-b8d52fcaa7d4-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Feb 02 09:17:38 crc kubenswrapper[4764]: I0202 09:17:38.782446 4764 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/527edab8-f88f-4d82-b42e-b8d52fcaa7d4-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Feb 02 09:17:38 crc kubenswrapper[4764]: I0202 09:17:38.782468 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-95srz\" (UniqueName: \"kubernetes.io/projected/527edab8-f88f-4d82-b42e-b8d52fcaa7d4-kube-api-access-95srz\") on node \"crc\" DevicePath \"\"" Feb 02 09:17:38 crc kubenswrapper[4764]: I0202 09:17:38.782486 4764 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/527edab8-f88f-4d82-b42e-b8d52fcaa7d4-registry-certificates\") on node \"crc\" DevicePath \"\"" Feb 02 09:17:39 crc kubenswrapper[4764]: I0202 09:17:39.433717 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-9nkvv" event={"ID":"527edab8-f88f-4d82-b42e-b8d52fcaa7d4","Type":"ContainerDied","Data":"0b65b00bc2d8eff15f7d32631379347926baf789e3e9c91110f3de30b1b31c5e"} Feb 02 09:17:39 crc kubenswrapper[4764]: I0202 09:17:39.433768 4764 scope.go:117] "RemoveContainer" containerID="37fa793b00f263b65d2c4b4c33beaeacd1c8c3cc4fac5d41b7d06caa9978028f" Feb 02 09:17:39 crc kubenswrapper[4764]: I0202 09:17:39.433829 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-9nkvv" Feb 02 09:17:39 crc kubenswrapper[4764]: I0202 09:17:39.482038 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-9nkvv"] Feb 02 09:17:39 crc kubenswrapper[4764]: I0202 09:17:39.493383 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-9nkvv"] Feb 02 09:17:39 crc kubenswrapper[4764]: I0202 09:17:39.838652 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="527edab8-f88f-4d82-b42e-b8d52fcaa7d4" path="/var/lib/kubelet/pods/527edab8-f88f-4d82-b42e-b8d52fcaa7d4/volumes" Feb 02 09:17:49 crc kubenswrapper[4764]: I0202 09:17:49.305571 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-cf98fcc89-4wc9w"] Feb 02 09:17:49 crc kubenswrapper[4764]: E0202 09:17:49.307980 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="527edab8-f88f-4d82-b42e-b8d52fcaa7d4" containerName="registry" Feb 02 09:17:49 crc kubenswrapper[4764]: I0202 09:17:49.307998 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="527edab8-f88f-4d82-b42e-b8d52fcaa7d4" containerName="registry" Feb 02 09:17:49 crc kubenswrapper[4764]: I0202 09:17:49.308121 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="527edab8-f88f-4d82-b42e-b8d52fcaa7d4" containerName="registry" Feb 02 09:17:49 crc kubenswrapper[4764]: I0202 09:17:49.308549 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-cf98fcc89-4wc9w" Feb 02 09:17:49 crc kubenswrapper[4764]: I0202 09:17:49.311721 4764 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-gr9tk" Feb 02 09:17:49 crc kubenswrapper[4764]: I0202 09:17:49.315494 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Feb 02 09:17:49 crc kubenswrapper[4764]: I0202 09:17:49.315492 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Feb 02 09:17:49 crc kubenswrapper[4764]: I0202 09:17:49.318074 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-cf98fcc89-4wc9w"] Feb 02 09:17:49 crc kubenswrapper[4764]: I0202 09:17:49.332133 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-858654f9db-mkn7c"] Feb 02 09:17:49 crc kubenswrapper[4764]: I0202 09:17:49.332922 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-858654f9db-mkn7c" Feb 02 09:17:49 crc kubenswrapper[4764]: I0202 09:17:49.336473 4764 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-kkwbp" Feb 02 09:17:49 crc kubenswrapper[4764]: I0202 09:17:49.409994 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-858654f9db-mkn7c"] Feb 02 09:17:49 crc kubenswrapper[4764]: I0202 09:17:49.423022 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-687f57d79b-8gwjv"] Feb 02 09:17:49 crc kubenswrapper[4764]: I0202 09:17:49.423728 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-687f57d79b-8gwjv" Feb 02 09:17:49 crc kubenswrapper[4764]: I0202 09:17:49.424065 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-687f57d79b-8gwjv"] Feb 02 09:17:49 crc kubenswrapper[4764]: I0202 09:17:49.426479 4764 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-2lfg6" Feb 02 09:17:49 crc kubenswrapper[4764]: I0202 09:17:49.477286 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rs5tc\" (UniqueName: \"kubernetes.io/projected/533ed531-c13b-4cbe-9355-e2f1096b905a-kube-api-access-rs5tc\") pod \"cert-manager-858654f9db-mkn7c\" (UID: \"533ed531-c13b-4cbe-9355-e2f1096b905a\") " pod="cert-manager/cert-manager-858654f9db-mkn7c" Feb 02 09:17:49 crc kubenswrapper[4764]: I0202 09:17:49.477541 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-txcxs\" (UniqueName: \"kubernetes.io/projected/1d03ea6c-c159-4e5a-a38c-1bf190d3d454-kube-api-access-txcxs\") pod \"cert-manager-cainjector-cf98fcc89-4wc9w\" (UID: \"1d03ea6c-c159-4e5a-a38c-1bf190d3d454\") " pod="cert-manager/cert-manager-cainjector-cf98fcc89-4wc9w" Feb 02 09:17:49 crc kubenswrapper[4764]: I0202 09:17:49.578927 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j9vrc\" (UniqueName: \"kubernetes.io/projected/6cc98479-e4c6-45e2-912c-c3a479af0321-kube-api-access-j9vrc\") pod \"cert-manager-webhook-687f57d79b-8gwjv\" (UID: \"6cc98479-e4c6-45e2-912c-c3a479af0321\") " pod="cert-manager/cert-manager-webhook-687f57d79b-8gwjv" Feb 02 09:17:49 crc kubenswrapper[4764]: I0202 09:17:49.579015 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rs5tc\" (UniqueName: \"kubernetes.io/projected/533ed531-c13b-4cbe-9355-e2f1096b905a-kube-api-access-rs5tc\") pod \"cert-manager-858654f9db-mkn7c\" (UID: \"533ed531-c13b-4cbe-9355-e2f1096b905a\") " pod="cert-manager/cert-manager-858654f9db-mkn7c" Feb 02 09:17:49 crc kubenswrapper[4764]: I0202 09:17:49.579036 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-txcxs\" (UniqueName: \"kubernetes.io/projected/1d03ea6c-c159-4e5a-a38c-1bf190d3d454-kube-api-access-txcxs\") pod \"cert-manager-cainjector-cf98fcc89-4wc9w\" (UID: \"1d03ea6c-c159-4e5a-a38c-1bf190d3d454\") " pod="cert-manager/cert-manager-cainjector-cf98fcc89-4wc9w" Feb 02 09:17:49 crc kubenswrapper[4764]: I0202 09:17:49.598052 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-txcxs\" (UniqueName: \"kubernetes.io/projected/1d03ea6c-c159-4e5a-a38c-1bf190d3d454-kube-api-access-txcxs\") pod \"cert-manager-cainjector-cf98fcc89-4wc9w\" (UID: \"1d03ea6c-c159-4e5a-a38c-1bf190d3d454\") " pod="cert-manager/cert-manager-cainjector-cf98fcc89-4wc9w" Feb 02 09:17:49 crc kubenswrapper[4764]: I0202 09:17:49.598312 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rs5tc\" (UniqueName: \"kubernetes.io/projected/533ed531-c13b-4cbe-9355-e2f1096b905a-kube-api-access-rs5tc\") pod \"cert-manager-858654f9db-mkn7c\" (UID: \"533ed531-c13b-4cbe-9355-e2f1096b905a\") " pod="cert-manager/cert-manager-858654f9db-mkn7c" Feb 02 09:17:49 crc kubenswrapper[4764]: I0202 09:17:49.678833 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-cf98fcc89-4wc9w" Feb 02 09:17:49 crc kubenswrapper[4764]: I0202 09:17:49.687391 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j9vrc\" (UniqueName: \"kubernetes.io/projected/6cc98479-e4c6-45e2-912c-c3a479af0321-kube-api-access-j9vrc\") pod \"cert-manager-webhook-687f57d79b-8gwjv\" (UID: \"6cc98479-e4c6-45e2-912c-c3a479af0321\") " pod="cert-manager/cert-manager-webhook-687f57d79b-8gwjv" Feb 02 09:17:49 crc kubenswrapper[4764]: I0202 09:17:49.689515 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-858654f9db-mkn7c" Feb 02 09:17:49 crc kubenswrapper[4764]: I0202 09:17:49.715689 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j9vrc\" (UniqueName: \"kubernetes.io/projected/6cc98479-e4c6-45e2-912c-c3a479af0321-kube-api-access-j9vrc\") pod \"cert-manager-webhook-687f57d79b-8gwjv\" (UID: \"6cc98479-e4c6-45e2-912c-c3a479af0321\") " pod="cert-manager/cert-manager-webhook-687f57d79b-8gwjv" Feb 02 09:17:49 crc kubenswrapper[4764]: I0202 09:17:49.737461 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-687f57d79b-8gwjv" Feb 02 09:17:49 crc kubenswrapper[4764]: I0202 09:17:49.911862 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-858654f9db-mkn7c"] Feb 02 09:17:49 crc kubenswrapper[4764]: I0202 09:17:49.918399 4764 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 02 09:17:49 crc kubenswrapper[4764]: I0202 09:17:49.955413 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-cf98fcc89-4wc9w"] Feb 02 09:17:49 crc kubenswrapper[4764]: I0202 09:17:49.998161 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-687f57d79b-8gwjv"] Feb 02 09:17:50 crc kubenswrapper[4764]: I0202 09:17:50.495455 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-858654f9db-mkn7c" event={"ID":"533ed531-c13b-4cbe-9355-e2f1096b905a","Type":"ContainerStarted","Data":"4cc11e24ed501bc8d06dba968682a85ff7f22f7468d809e0b70c219ded480856"} Feb 02 09:17:50 crc kubenswrapper[4764]: I0202 09:17:50.497060 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-cf98fcc89-4wc9w" event={"ID":"1d03ea6c-c159-4e5a-a38c-1bf190d3d454","Type":"ContainerStarted","Data":"b370970e823b090be6021529bf81a00a957d013af40aeac329740bd513758e6c"} Feb 02 09:17:50 crc kubenswrapper[4764]: I0202 09:17:50.498960 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-687f57d79b-8gwjv" event={"ID":"6cc98479-e4c6-45e2-912c-c3a479af0321","Type":"ContainerStarted","Data":"a50305d8f5743807db9e1901ee22a79580eed1ae566147d23c1fa2154083b426"} Feb 02 09:17:54 crc kubenswrapper[4764]: I0202 09:17:54.523692 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-858654f9db-mkn7c" event={"ID":"533ed531-c13b-4cbe-9355-e2f1096b905a","Type":"ContainerStarted","Data":"db8040e668c9d6047159f6bbabf94d1aff1475ddaf70e1674097d5ecd168270b"} Feb 02 09:17:54 crc kubenswrapper[4764]: I0202 09:17:54.525012 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-cf98fcc89-4wc9w" event={"ID":"1d03ea6c-c159-4e5a-a38c-1bf190d3d454","Type":"ContainerStarted","Data":"dbb9648398cce3f32a08bb7bede909d303ee887c5894d980f520af6cc6013f58"} Feb 02 09:17:54 crc kubenswrapper[4764]: I0202 09:17:54.526852 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-687f57d79b-8gwjv" event={"ID":"6cc98479-e4c6-45e2-912c-c3a479af0321","Type":"ContainerStarted","Data":"ebf48462b20128bc4b9deb7996359ac13dbcd0e839932272732a76b66acfe862"} Feb 02 09:17:54 crc kubenswrapper[4764]: I0202 09:17:54.526970 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-687f57d79b-8gwjv" Feb 02 09:17:54 crc kubenswrapper[4764]: I0202 09:17:54.538652 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-858654f9db-mkn7c" podStartSLOduration=1.6889720430000001 podStartE2EDuration="5.538637968s" podCreationTimestamp="2026-02-02 09:17:49 +0000 UTC" firstStartedPulling="2026-02-02 09:17:49.918152338 +0000 UTC m=+632.851876426" lastFinishedPulling="2026-02-02 09:17:53.767818263 +0000 UTC m=+636.701542351" observedRunningTime="2026-02-02 09:17:54.538000901 +0000 UTC m=+637.471724989" watchObservedRunningTime="2026-02-02 09:17:54.538637968 +0000 UTC m=+637.472362056" Feb 02 09:17:54 crc kubenswrapper[4764]: I0202 09:17:54.561184 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-687f57d79b-8gwjv" podStartSLOduration=1.734000429 podStartE2EDuration="5.561163967s" podCreationTimestamp="2026-02-02 09:17:49 +0000 UTC" firstStartedPulling="2026-02-02 09:17:50.006254306 +0000 UTC m=+632.939978394" lastFinishedPulling="2026-02-02 09:17:53.833417844 +0000 UTC m=+636.767141932" observedRunningTime="2026-02-02 09:17:54.559425049 +0000 UTC m=+637.493149137" watchObservedRunningTime="2026-02-02 09:17:54.561163967 +0000 UTC m=+637.494888055" Feb 02 09:17:54 crc kubenswrapper[4764]: I0202 09:17:54.580951 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-cf98fcc89-4wc9w" podStartSLOduration=1.77917609 podStartE2EDuration="5.580911259s" podCreationTimestamp="2026-02-02 09:17:49 +0000 UTC" firstStartedPulling="2026-02-02 09:17:49.964705156 +0000 UTC m=+632.898429254" lastFinishedPulling="2026-02-02 09:17:53.766440335 +0000 UTC m=+636.700164423" observedRunningTime="2026-02-02 09:17:54.577433233 +0000 UTC m=+637.511157321" watchObservedRunningTime="2026-02-02 09:17:54.580911259 +0000 UTC m=+637.514635347" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.284377 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-zhn7j"] Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.285636 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" podUID="24632dda-6100-4ab6-a28e-214ddae4360c" containerName="nbdb" containerID="cri-o://8a989645cf485cd8ff80ffff1a86cd7808e1564eba5667535e161cc68db13b26" gracePeriod=30 Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.285756 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" podUID="24632dda-6100-4ab6-a28e-214ddae4360c" containerName="ovn-acl-logging" containerID="cri-o://c380383b6a014d1b694a13ddc22f54b20b9e4f9d5c9b9000d76187621dce0a21" gracePeriod=30 Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.285790 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" podUID="24632dda-6100-4ab6-a28e-214ddae4360c" containerName="kube-rbac-proxy-node" containerID="cri-o://0a6c67e9b7708cc0d85b0b6ffbbade0d34867e159a9c5487ca2606475c667a50" gracePeriod=30 Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.285848 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" podUID="24632dda-6100-4ab6-a28e-214ddae4360c" containerName="northd" containerID="cri-o://20b71d3ee7a381baa51b0b0d0470283022cfede30f491135737b5e82d631d763" gracePeriod=30 Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.285974 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" podUID="24632dda-6100-4ab6-a28e-214ddae4360c" containerName="sbdb" containerID="cri-o://4b4cfbd4e682bcdcaff785013bdb72580ce36fced733e235538536827036b93d" gracePeriod=30 Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.286134 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" podUID="24632dda-6100-4ab6-a28e-214ddae4360c" containerName="ovn-controller" containerID="cri-o://5529f2bff87753428e9173826fc1880acf96a4c134fa9ecad195b15697cad73a" gracePeriod=30 Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.285694 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" podUID="24632dda-6100-4ab6-a28e-214ddae4360c" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://bffda48d7c881c1470251b501661324859373246cc1a21811bd36ca24887c7aa" gracePeriod=30 Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.344995 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" podUID="24632dda-6100-4ab6-a28e-214ddae4360c" containerName="ovnkube-controller" containerID="cri-o://39395584eacf319103cbe03b647c7c7672646b5619a631dffd7f115435591187" gracePeriod=30 Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.564546 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-nc7qt_6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e/kube-multus/2.log" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.565009 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-nc7qt_6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e/kube-multus/1.log" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.565037 4764 generic.go:334] "Generic (PLEG): container finished" podID="6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e" containerID="b7882e3ed0730ea5035949695b7b94a50060918bee89edc807fd67484b957ee5" exitCode=2 Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.565079 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-nc7qt" event={"ID":"6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e","Type":"ContainerDied","Data":"b7882e3ed0730ea5035949695b7b94a50060918bee89edc807fd67484b957ee5"} Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.565111 4764 scope.go:117] "RemoveContainer" containerID="b52b499d17b4e2e359475bd13730c72ef5cedb052e13def6dfe5bc5fd7487d2d" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.565527 4764 scope.go:117] "RemoveContainer" containerID="b7882e3ed0730ea5035949695b7b94a50060918bee89edc807fd67484b957ee5" Feb 02 09:17:59 crc kubenswrapper[4764]: E0202 09:17:59.565809 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-nc7qt_openshift-multus(6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e)\"" pod="openshift-multus/multus-nc7qt" podUID="6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.571946 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zhn7j_24632dda-6100-4ab6-a28e-214ddae4360c/ovnkube-controller/3.log" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.574794 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zhn7j_24632dda-6100-4ab6-a28e-214ddae4360c/ovn-acl-logging/0.log" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.575480 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zhn7j_24632dda-6100-4ab6-a28e-214ddae4360c/ovn-controller/0.log" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.575828 4764 generic.go:334] "Generic (PLEG): container finished" podID="24632dda-6100-4ab6-a28e-214ddae4360c" containerID="39395584eacf319103cbe03b647c7c7672646b5619a631dffd7f115435591187" exitCode=0 Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.575910 4764 generic.go:334] "Generic (PLEG): container finished" podID="24632dda-6100-4ab6-a28e-214ddae4360c" containerID="4b4cfbd4e682bcdcaff785013bdb72580ce36fced733e235538536827036b93d" exitCode=0 Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.575996 4764 generic.go:334] "Generic (PLEG): container finished" podID="24632dda-6100-4ab6-a28e-214ddae4360c" containerID="8a989645cf485cd8ff80ffff1a86cd7808e1564eba5667535e161cc68db13b26" exitCode=0 Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.576053 4764 generic.go:334] "Generic (PLEG): container finished" podID="24632dda-6100-4ab6-a28e-214ddae4360c" containerID="bffda48d7c881c1470251b501661324859373246cc1a21811bd36ca24887c7aa" exitCode=0 Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.576063 4764 generic.go:334] "Generic (PLEG): container finished" podID="24632dda-6100-4ab6-a28e-214ddae4360c" containerID="0a6c67e9b7708cc0d85b0b6ffbbade0d34867e159a9c5487ca2606475c667a50" exitCode=0 Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.576070 4764 generic.go:334] "Generic (PLEG): container finished" podID="24632dda-6100-4ab6-a28e-214ddae4360c" containerID="c380383b6a014d1b694a13ddc22f54b20b9e4f9d5c9b9000d76187621dce0a21" exitCode=143 Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.576076 4764 generic.go:334] "Generic (PLEG): container finished" podID="24632dda-6100-4ab6-a28e-214ddae4360c" containerID="5529f2bff87753428e9173826fc1880acf96a4c134fa9ecad195b15697cad73a" exitCode=143 Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.575862 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" event={"ID":"24632dda-6100-4ab6-a28e-214ddae4360c","Type":"ContainerDied","Data":"39395584eacf319103cbe03b647c7c7672646b5619a631dffd7f115435591187"} Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.576112 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" event={"ID":"24632dda-6100-4ab6-a28e-214ddae4360c","Type":"ContainerDied","Data":"4b4cfbd4e682bcdcaff785013bdb72580ce36fced733e235538536827036b93d"} Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.576124 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" event={"ID":"24632dda-6100-4ab6-a28e-214ddae4360c","Type":"ContainerDied","Data":"8a989645cf485cd8ff80ffff1a86cd7808e1564eba5667535e161cc68db13b26"} Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.576134 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" event={"ID":"24632dda-6100-4ab6-a28e-214ddae4360c","Type":"ContainerDied","Data":"bffda48d7c881c1470251b501661324859373246cc1a21811bd36ca24887c7aa"} Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.576144 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" event={"ID":"24632dda-6100-4ab6-a28e-214ddae4360c","Type":"ContainerDied","Data":"0a6c67e9b7708cc0d85b0b6ffbbade0d34867e159a9c5487ca2606475c667a50"} Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.576152 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" event={"ID":"24632dda-6100-4ab6-a28e-214ddae4360c","Type":"ContainerDied","Data":"c380383b6a014d1b694a13ddc22f54b20b9e4f9d5c9b9000d76187621dce0a21"} Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.576161 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" event={"ID":"24632dda-6100-4ab6-a28e-214ddae4360c","Type":"ContainerDied","Data":"5529f2bff87753428e9173826fc1880acf96a4c134fa9ecad195b15697cad73a"} Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.612042 4764 scope.go:117] "RemoveContainer" containerID="3cb255e14eb02f7c08364c41fb9906e58b2155eb507f694b77ec8eb1f4251343" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.642132 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zhn7j_24632dda-6100-4ab6-a28e-214ddae4360c/ovn-acl-logging/0.log" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.642536 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zhn7j_24632dda-6100-4ab6-a28e-214ddae4360c/ovn-controller/0.log" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.642899 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.690016 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-gdj8q"] Feb 02 09:17:59 crc kubenswrapper[4764]: E0202 09:17:59.690250 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24632dda-6100-4ab6-a28e-214ddae4360c" containerName="kubecfg-setup" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.690264 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="24632dda-6100-4ab6-a28e-214ddae4360c" containerName="kubecfg-setup" Feb 02 09:17:59 crc kubenswrapper[4764]: E0202 09:17:59.690273 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24632dda-6100-4ab6-a28e-214ddae4360c" containerName="ovnkube-controller" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.690281 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="24632dda-6100-4ab6-a28e-214ddae4360c" containerName="ovnkube-controller" Feb 02 09:17:59 crc kubenswrapper[4764]: E0202 09:17:59.690290 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24632dda-6100-4ab6-a28e-214ddae4360c" containerName="kube-rbac-proxy-node" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.690298 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="24632dda-6100-4ab6-a28e-214ddae4360c" containerName="kube-rbac-proxy-node" Feb 02 09:17:59 crc kubenswrapper[4764]: E0202 09:17:59.690311 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24632dda-6100-4ab6-a28e-214ddae4360c" containerName="ovn-acl-logging" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.690318 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="24632dda-6100-4ab6-a28e-214ddae4360c" containerName="ovn-acl-logging" Feb 02 09:17:59 crc kubenswrapper[4764]: E0202 09:17:59.690330 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24632dda-6100-4ab6-a28e-214ddae4360c" containerName="sbdb" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.690337 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="24632dda-6100-4ab6-a28e-214ddae4360c" containerName="sbdb" Feb 02 09:17:59 crc kubenswrapper[4764]: E0202 09:17:59.690350 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24632dda-6100-4ab6-a28e-214ddae4360c" containerName="ovnkube-controller" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.690359 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="24632dda-6100-4ab6-a28e-214ddae4360c" containerName="ovnkube-controller" Feb 02 09:17:59 crc kubenswrapper[4764]: E0202 09:17:59.690368 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24632dda-6100-4ab6-a28e-214ddae4360c" containerName="ovnkube-controller" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.690376 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="24632dda-6100-4ab6-a28e-214ddae4360c" containerName="ovnkube-controller" Feb 02 09:17:59 crc kubenswrapper[4764]: E0202 09:17:59.690385 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24632dda-6100-4ab6-a28e-214ddae4360c" containerName="northd" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.690394 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="24632dda-6100-4ab6-a28e-214ddae4360c" containerName="northd" Feb 02 09:17:59 crc kubenswrapper[4764]: E0202 09:17:59.690407 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24632dda-6100-4ab6-a28e-214ddae4360c" containerName="ovnkube-controller" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.690414 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="24632dda-6100-4ab6-a28e-214ddae4360c" containerName="ovnkube-controller" Feb 02 09:17:59 crc kubenswrapper[4764]: E0202 09:17:59.690422 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24632dda-6100-4ab6-a28e-214ddae4360c" containerName="nbdb" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.690430 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="24632dda-6100-4ab6-a28e-214ddae4360c" containerName="nbdb" Feb 02 09:17:59 crc kubenswrapper[4764]: E0202 09:17:59.690440 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24632dda-6100-4ab6-a28e-214ddae4360c" containerName="kube-rbac-proxy-ovn-metrics" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.690447 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="24632dda-6100-4ab6-a28e-214ddae4360c" containerName="kube-rbac-proxy-ovn-metrics" Feb 02 09:17:59 crc kubenswrapper[4764]: E0202 09:17:59.690460 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24632dda-6100-4ab6-a28e-214ddae4360c" containerName="ovn-controller" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.690467 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="24632dda-6100-4ab6-a28e-214ddae4360c" containerName="ovn-controller" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.690589 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="24632dda-6100-4ab6-a28e-214ddae4360c" containerName="ovn-controller" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.690603 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="24632dda-6100-4ab6-a28e-214ddae4360c" containerName="nbdb" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.690614 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="24632dda-6100-4ab6-a28e-214ddae4360c" containerName="ovnkube-controller" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.690622 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="24632dda-6100-4ab6-a28e-214ddae4360c" containerName="ovnkube-controller" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.690630 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="24632dda-6100-4ab6-a28e-214ddae4360c" containerName="kube-rbac-proxy-ovn-metrics" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.690640 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="24632dda-6100-4ab6-a28e-214ddae4360c" containerName="sbdb" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.690650 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="24632dda-6100-4ab6-a28e-214ddae4360c" containerName="ovn-acl-logging" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.690663 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="24632dda-6100-4ab6-a28e-214ddae4360c" containerName="kube-rbac-proxy-node" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.690675 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="24632dda-6100-4ab6-a28e-214ddae4360c" containerName="ovnkube-controller" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.690684 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="24632dda-6100-4ab6-a28e-214ddae4360c" containerName="ovnkube-controller" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.690694 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="24632dda-6100-4ab6-a28e-214ddae4360c" containerName="northd" Feb 02 09:17:59 crc kubenswrapper[4764]: E0202 09:17:59.690833 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24632dda-6100-4ab6-a28e-214ddae4360c" containerName="ovnkube-controller" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.690844 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="24632dda-6100-4ab6-a28e-214ddae4360c" containerName="ovnkube-controller" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.690986 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="24632dda-6100-4ab6-a28e-214ddae4360c" containerName="ovnkube-controller" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.692884 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-gdj8q" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.731436 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/24632dda-6100-4ab6-a28e-214ddae4360c-run-systemd\") pod \"24632dda-6100-4ab6-a28e-214ddae4360c\" (UID: \"24632dda-6100-4ab6-a28e-214ddae4360c\") " Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.731481 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/24632dda-6100-4ab6-a28e-214ddae4360c-host-run-netns\") pod \"24632dda-6100-4ab6-a28e-214ddae4360c\" (UID: \"24632dda-6100-4ab6-a28e-214ddae4360c\") " Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.731521 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/24632dda-6100-4ab6-a28e-214ddae4360c-ovn-node-metrics-cert\") pod \"24632dda-6100-4ab6-a28e-214ddae4360c\" (UID: \"24632dda-6100-4ab6-a28e-214ddae4360c\") " Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.731547 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/24632dda-6100-4ab6-a28e-214ddae4360c-etc-openvswitch\") pod \"24632dda-6100-4ab6-a28e-214ddae4360c\" (UID: \"24632dda-6100-4ab6-a28e-214ddae4360c\") " Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.731593 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/24632dda-6100-4ab6-a28e-214ddae4360c-host-cni-bin\") pod \"24632dda-6100-4ab6-a28e-214ddae4360c\" (UID: \"24632dda-6100-4ab6-a28e-214ddae4360c\") " Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.731613 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/24632dda-6100-4ab6-a28e-214ddae4360c-host-var-lib-cni-networks-ovn-kubernetes\") pod \"24632dda-6100-4ab6-a28e-214ddae4360c\" (UID: \"24632dda-6100-4ab6-a28e-214ddae4360c\") " Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.731638 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/24632dda-6100-4ab6-a28e-214ddae4360c-host-kubelet\") pod \"24632dda-6100-4ab6-a28e-214ddae4360c\" (UID: \"24632dda-6100-4ab6-a28e-214ddae4360c\") " Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.731660 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/24632dda-6100-4ab6-a28e-214ddae4360c-var-lib-openvswitch\") pod \"24632dda-6100-4ab6-a28e-214ddae4360c\" (UID: \"24632dda-6100-4ab6-a28e-214ddae4360c\") " Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.731685 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/24632dda-6100-4ab6-a28e-214ddae4360c-run-openvswitch\") pod \"24632dda-6100-4ab6-a28e-214ddae4360c\" (UID: \"24632dda-6100-4ab6-a28e-214ddae4360c\") " Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.731708 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/24632dda-6100-4ab6-a28e-214ddae4360c-host-run-ovn-kubernetes\") pod \"24632dda-6100-4ab6-a28e-214ddae4360c\" (UID: \"24632dda-6100-4ab6-a28e-214ddae4360c\") " Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.731732 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/24632dda-6100-4ab6-a28e-214ddae4360c-run-ovn\") pod \"24632dda-6100-4ab6-a28e-214ddae4360c\" (UID: \"24632dda-6100-4ab6-a28e-214ddae4360c\") " Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.731759 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7r4sk\" (UniqueName: \"kubernetes.io/projected/24632dda-6100-4ab6-a28e-214ddae4360c-kube-api-access-7r4sk\") pod \"24632dda-6100-4ab6-a28e-214ddae4360c\" (UID: \"24632dda-6100-4ab6-a28e-214ddae4360c\") " Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.731778 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/24632dda-6100-4ab6-a28e-214ddae4360c-host-slash\") pod \"24632dda-6100-4ab6-a28e-214ddae4360c\" (UID: \"24632dda-6100-4ab6-a28e-214ddae4360c\") " Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.731806 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/24632dda-6100-4ab6-a28e-214ddae4360c-systemd-units\") pod \"24632dda-6100-4ab6-a28e-214ddae4360c\" (UID: \"24632dda-6100-4ab6-a28e-214ddae4360c\") " Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.731827 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/24632dda-6100-4ab6-a28e-214ddae4360c-log-socket\") pod \"24632dda-6100-4ab6-a28e-214ddae4360c\" (UID: \"24632dda-6100-4ab6-a28e-214ddae4360c\") " Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.731849 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/24632dda-6100-4ab6-a28e-214ddae4360c-node-log\") pod \"24632dda-6100-4ab6-a28e-214ddae4360c\" (UID: \"24632dda-6100-4ab6-a28e-214ddae4360c\") " Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.731869 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/24632dda-6100-4ab6-a28e-214ddae4360c-host-cni-netd\") pod \"24632dda-6100-4ab6-a28e-214ddae4360c\" (UID: \"24632dda-6100-4ab6-a28e-214ddae4360c\") " Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.731896 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/24632dda-6100-4ab6-a28e-214ddae4360c-ovnkube-script-lib\") pod \"24632dda-6100-4ab6-a28e-214ddae4360c\" (UID: \"24632dda-6100-4ab6-a28e-214ddae4360c\") " Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.731940 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/24632dda-6100-4ab6-a28e-214ddae4360c-ovnkube-config\") pod \"24632dda-6100-4ab6-a28e-214ddae4360c\" (UID: \"24632dda-6100-4ab6-a28e-214ddae4360c\") " Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.731998 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/24632dda-6100-4ab6-a28e-214ddae4360c-env-overrides\") pod \"24632dda-6100-4ab6-a28e-214ddae4360c\" (UID: \"24632dda-6100-4ab6-a28e-214ddae4360c\") " Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.732145 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a-ovnkube-config\") pod \"ovnkube-node-gdj8q\" (UID: \"ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-gdj8q" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.732175 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a-var-lib-openvswitch\") pod \"ovnkube-node-gdj8q\" (UID: \"ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-gdj8q" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.732203 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a-run-systemd\") pod \"ovnkube-node-gdj8q\" (UID: \"ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-gdj8q" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.732213 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/24632dda-6100-4ab6-a28e-214ddae4360c-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "24632dda-6100-4ab6-a28e-214ddae4360c" (UID: "24632dda-6100-4ab6-a28e-214ddae4360c"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.732226 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a-node-log\") pod \"ovnkube-node-gdj8q\" (UID: \"ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-gdj8q" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.732314 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a-host-kubelet\") pod \"ovnkube-node-gdj8q\" (UID: \"ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-gdj8q" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.732333 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a-run-openvswitch\") pod \"ovnkube-node-gdj8q\" (UID: \"ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-gdj8q" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.732361 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a-host-slash\") pod \"ovnkube-node-gdj8q\" (UID: \"ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-gdj8q" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.732415 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a-log-socket\") pod \"ovnkube-node-gdj8q\" (UID: \"ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-gdj8q" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.732436 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a-host-run-ovn-kubernetes\") pod \"ovnkube-node-gdj8q\" (UID: \"ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-gdj8q" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.732469 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a-systemd-units\") pod \"ovnkube-node-gdj8q\" (UID: \"ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-gdj8q" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.732505 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s8r58\" (UniqueName: \"kubernetes.io/projected/ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a-kube-api-access-s8r58\") pod \"ovnkube-node-gdj8q\" (UID: \"ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-gdj8q" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.732526 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-gdj8q\" (UID: \"ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-gdj8q" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.732550 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a-host-cni-netd\") pod \"ovnkube-node-gdj8q\" (UID: \"ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-gdj8q" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.732581 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a-env-overrides\") pod \"ovnkube-node-gdj8q\" (UID: \"ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-gdj8q" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.732604 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a-host-run-netns\") pod \"ovnkube-node-gdj8q\" (UID: \"ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-gdj8q" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.732623 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a-etc-openvswitch\") pod \"ovnkube-node-gdj8q\" (UID: \"ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-gdj8q" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.732642 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a-ovnkube-script-lib\") pod \"ovnkube-node-gdj8q\" (UID: \"ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-gdj8q" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.732690 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a-host-cni-bin\") pod \"ovnkube-node-gdj8q\" (UID: \"ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-gdj8q" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.732714 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a-ovn-node-metrics-cert\") pod \"ovnkube-node-gdj8q\" (UID: \"ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-gdj8q" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.732740 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a-run-ovn\") pod \"ovnkube-node-gdj8q\" (UID: \"ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-gdj8q" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.732787 4764 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/24632dda-6100-4ab6-a28e-214ddae4360c-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.732800 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/24632dda-6100-4ab6-a28e-214ddae4360c-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "24632dda-6100-4ab6-a28e-214ddae4360c" (UID: "24632dda-6100-4ab6-a28e-214ddae4360c"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.732828 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/24632dda-6100-4ab6-a28e-214ddae4360c-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "24632dda-6100-4ab6-a28e-214ddae4360c" (UID: "24632dda-6100-4ab6-a28e-214ddae4360c"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.732859 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/24632dda-6100-4ab6-a28e-214ddae4360c-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "24632dda-6100-4ab6-a28e-214ddae4360c" (UID: "24632dda-6100-4ab6-a28e-214ddae4360c"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.732902 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/24632dda-6100-4ab6-a28e-214ddae4360c-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "24632dda-6100-4ab6-a28e-214ddae4360c" (UID: "24632dda-6100-4ab6-a28e-214ddae4360c"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.732925 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/24632dda-6100-4ab6-a28e-214ddae4360c-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "24632dda-6100-4ab6-a28e-214ddae4360c" (UID: "24632dda-6100-4ab6-a28e-214ddae4360c"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.733168 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/24632dda-6100-4ab6-a28e-214ddae4360c-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "24632dda-6100-4ab6-a28e-214ddae4360c" (UID: "24632dda-6100-4ab6-a28e-214ddae4360c"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.733208 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/24632dda-6100-4ab6-a28e-214ddae4360c-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "24632dda-6100-4ab6-a28e-214ddae4360c" (UID: "24632dda-6100-4ab6-a28e-214ddae4360c"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.733230 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/24632dda-6100-4ab6-a28e-214ddae4360c-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "24632dda-6100-4ab6-a28e-214ddae4360c" (UID: "24632dda-6100-4ab6-a28e-214ddae4360c"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.733251 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/24632dda-6100-4ab6-a28e-214ddae4360c-host-slash" (OuterVolumeSpecName: "host-slash") pod "24632dda-6100-4ab6-a28e-214ddae4360c" (UID: "24632dda-6100-4ab6-a28e-214ddae4360c"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.733268 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/24632dda-6100-4ab6-a28e-214ddae4360c-node-log" (OuterVolumeSpecName: "node-log") pod "24632dda-6100-4ab6-a28e-214ddae4360c" (UID: "24632dda-6100-4ab6-a28e-214ddae4360c"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.733277 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/24632dda-6100-4ab6-a28e-214ddae4360c-log-socket" (OuterVolumeSpecName: "log-socket") pod "24632dda-6100-4ab6-a28e-214ddae4360c" (UID: "24632dda-6100-4ab6-a28e-214ddae4360c"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.733316 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/24632dda-6100-4ab6-a28e-214ddae4360c-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "24632dda-6100-4ab6-a28e-214ddae4360c" (UID: "24632dda-6100-4ab6-a28e-214ddae4360c"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.733350 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/24632dda-6100-4ab6-a28e-214ddae4360c-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "24632dda-6100-4ab6-a28e-214ddae4360c" (UID: "24632dda-6100-4ab6-a28e-214ddae4360c"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.733514 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/24632dda-6100-4ab6-a28e-214ddae4360c-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "24632dda-6100-4ab6-a28e-214ddae4360c" (UID: "24632dda-6100-4ab6-a28e-214ddae4360c"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.733672 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/24632dda-6100-4ab6-a28e-214ddae4360c-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "24632dda-6100-4ab6-a28e-214ddae4360c" (UID: "24632dda-6100-4ab6-a28e-214ddae4360c"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.733727 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/24632dda-6100-4ab6-a28e-214ddae4360c-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "24632dda-6100-4ab6-a28e-214ddae4360c" (UID: "24632dda-6100-4ab6-a28e-214ddae4360c"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.737207 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/24632dda-6100-4ab6-a28e-214ddae4360c-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "24632dda-6100-4ab6-a28e-214ddae4360c" (UID: "24632dda-6100-4ab6-a28e-214ddae4360c"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.737572 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/24632dda-6100-4ab6-a28e-214ddae4360c-kube-api-access-7r4sk" (OuterVolumeSpecName: "kube-api-access-7r4sk") pod "24632dda-6100-4ab6-a28e-214ddae4360c" (UID: "24632dda-6100-4ab6-a28e-214ddae4360c"). InnerVolumeSpecName "kube-api-access-7r4sk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.740419 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-687f57d79b-8gwjv" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.744702 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/24632dda-6100-4ab6-a28e-214ddae4360c-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "24632dda-6100-4ab6-a28e-214ddae4360c" (UID: "24632dda-6100-4ab6-a28e-214ddae4360c"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.833463 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s8r58\" (UniqueName: \"kubernetes.io/projected/ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a-kube-api-access-s8r58\") pod \"ovnkube-node-gdj8q\" (UID: \"ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-gdj8q" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.833523 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-gdj8q\" (UID: \"ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-gdj8q" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.833614 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-gdj8q\" (UID: \"ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-gdj8q" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.833649 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a-host-cni-netd\") pod \"ovnkube-node-gdj8q\" (UID: \"ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-gdj8q" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.833671 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a-env-overrides\") pod \"ovnkube-node-gdj8q\" (UID: \"ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-gdj8q" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.833716 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a-host-run-netns\") pod \"ovnkube-node-gdj8q\" (UID: \"ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-gdj8q" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.833732 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a-etc-openvswitch\") pod \"ovnkube-node-gdj8q\" (UID: \"ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-gdj8q" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.833739 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a-host-cni-netd\") pod \"ovnkube-node-gdj8q\" (UID: \"ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-gdj8q" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.833780 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a-host-run-netns\") pod \"ovnkube-node-gdj8q\" (UID: \"ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-gdj8q" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.833808 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a-etc-openvswitch\") pod \"ovnkube-node-gdj8q\" (UID: \"ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-gdj8q" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.834539 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a-env-overrides\") pod \"ovnkube-node-gdj8q\" (UID: \"ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-gdj8q" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.834606 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a-ovnkube-script-lib\") pod \"ovnkube-node-gdj8q\" (UID: \"ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-gdj8q" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.834638 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a-host-cni-bin\") pod \"ovnkube-node-gdj8q\" (UID: \"ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-gdj8q" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.835123 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a-ovnkube-script-lib\") pod \"ovnkube-node-gdj8q\" (UID: \"ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-gdj8q" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.835420 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a-ovn-node-metrics-cert\") pod \"ovnkube-node-gdj8q\" (UID: \"ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-gdj8q" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.835467 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a-host-cni-bin\") pod \"ovnkube-node-gdj8q\" (UID: \"ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-gdj8q" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.835504 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a-run-ovn\") pod \"ovnkube-node-gdj8q\" (UID: \"ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-gdj8q" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.835523 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a-ovnkube-config\") pod \"ovnkube-node-gdj8q\" (UID: \"ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-gdj8q" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.835972 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a-var-lib-openvswitch\") pod \"ovnkube-node-gdj8q\" (UID: \"ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-gdj8q" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.835861 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a-run-ovn\") pod \"ovnkube-node-gdj8q\" (UID: \"ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-gdj8q" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.835998 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a-run-systemd\") pod \"ovnkube-node-gdj8q\" (UID: \"ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-gdj8q" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.836038 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a-var-lib-openvswitch\") pod \"ovnkube-node-gdj8q\" (UID: \"ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-gdj8q" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.836046 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a-node-log\") pod \"ovnkube-node-gdj8q\" (UID: \"ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-gdj8q" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.836068 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a-node-log\") pod \"ovnkube-node-gdj8q\" (UID: \"ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-gdj8q" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.836111 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a-run-systemd\") pod \"ovnkube-node-gdj8q\" (UID: \"ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-gdj8q" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.836151 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a-host-kubelet\") pod \"ovnkube-node-gdj8q\" (UID: \"ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-gdj8q" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.836179 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a-run-openvswitch\") pod \"ovnkube-node-gdj8q\" (UID: \"ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-gdj8q" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.836209 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a-host-slash\") pod \"ovnkube-node-gdj8q\" (UID: \"ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-gdj8q" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.836267 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a-log-socket\") pod \"ovnkube-node-gdj8q\" (UID: \"ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-gdj8q" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.836292 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a-host-run-ovn-kubernetes\") pod \"ovnkube-node-gdj8q\" (UID: \"ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-gdj8q" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.836322 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a-systemd-units\") pod \"ovnkube-node-gdj8q\" (UID: \"ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-gdj8q" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.836394 4764 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/24632dda-6100-4ab6-a28e-214ddae4360c-run-systemd\") on node \"crc\" DevicePath \"\"" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.836409 4764 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/24632dda-6100-4ab6-a28e-214ddae4360c-host-run-netns\") on node \"crc\" DevicePath \"\"" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.836424 4764 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/24632dda-6100-4ab6-a28e-214ddae4360c-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.836437 4764 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/24632dda-6100-4ab6-a28e-214ddae4360c-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.836448 4764 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/24632dda-6100-4ab6-a28e-214ddae4360c-host-cni-bin\") on node \"crc\" DevicePath \"\"" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.836460 4764 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/24632dda-6100-4ab6-a28e-214ddae4360c-host-kubelet\") on node \"crc\" DevicePath \"\"" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.836473 4764 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/24632dda-6100-4ab6-a28e-214ddae4360c-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.836480 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a-host-kubelet\") pod \"ovnkube-node-gdj8q\" (UID: \"ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-gdj8q" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.836486 4764 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/24632dda-6100-4ab6-a28e-214ddae4360c-run-openvswitch\") on node \"crc\" DevicePath \"\"" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.836529 4764 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/24632dda-6100-4ab6-a28e-214ddae4360c-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.836543 4764 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/24632dda-6100-4ab6-a28e-214ddae4360c-run-ovn\") on node \"crc\" DevicePath \"\"" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.836552 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7r4sk\" (UniqueName: \"kubernetes.io/projected/24632dda-6100-4ab6-a28e-214ddae4360c-kube-api-access-7r4sk\") on node \"crc\" DevicePath \"\"" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.836560 4764 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/24632dda-6100-4ab6-a28e-214ddae4360c-host-slash\") on node \"crc\" DevicePath \"\"" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.836627 4764 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/24632dda-6100-4ab6-a28e-214ddae4360c-log-socket\") on node \"crc\" DevicePath \"\"" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.836636 4764 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/24632dda-6100-4ab6-a28e-214ddae4360c-systemd-units\") on node \"crc\" DevicePath \"\"" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.836644 4764 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/24632dda-6100-4ab6-a28e-214ddae4360c-node-log\") on node \"crc\" DevicePath \"\"" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.836652 4764 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/24632dda-6100-4ab6-a28e-214ddae4360c-host-cni-netd\") on node \"crc\" DevicePath \"\"" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.836685 4764 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/24632dda-6100-4ab6-a28e-214ddae4360c-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.836696 4764 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/24632dda-6100-4ab6-a28e-214ddae4360c-ovnkube-config\") on node \"crc\" DevicePath \"\"" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.836706 4764 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/24632dda-6100-4ab6-a28e-214ddae4360c-env-overrides\") on node \"crc\" DevicePath \"\"" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.836461 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a-host-slash\") pod \"ovnkube-node-gdj8q\" (UID: \"ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-gdj8q" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.836718 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a-systemd-units\") pod \"ovnkube-node-gdj8q\" (UID: \"ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-gdj8q" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.836481 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a-run-openvswitch\") pod \"ovnkube-node-gdj8q\" (UID: \"ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-gdj8q" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.836682 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a-host-run-ovn-kubernetes\") pod \"ovnkube-node-gdj8q\" (UID: \"ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-gdj8q" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.836434 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a-ovnkube-config\") pod \"ovnkube-node-gdj8q\" (UID: \"ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-gdj8q" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.836706 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a-log-socket\") pod \"ovnkube-node-gdj8q\" (UID: \"ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-gdj8q" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.839154 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a-ovn-node-metrics-cert\") pod \"ovnkube-node-gdj8q\" (UID: \"ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-gdj8q" Feb 02 09:17:59 crc kubenswrapper[4764]: I0202 09:17:59.850543 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s8r58\" (UniqueName: \"kubernetes.io/projected/ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a-kube-api-access-s8r58\") pod \"ovnkube-node-gdj8q\" (UID: \"ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-gdj8q" Feb 02 09:18:00 crc kubenswrapper[4764]: I0202 09:18:00.006334 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-gdj8q" Feb 02 09:18:00 crc kubenswrapper[4764]: I0202 09:18:00.583621 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-nc7qt_6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e/kube-multus/2.log" Feb 02 09:18:00 crc kubenswrapper[4764]: I0202 09:18:00.587763 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zhn7j_24632dda-6100-4ab6-a28e-214ddae4360c/ovn-acl-logging/0.log" Feb 02 09:18:00 crc kubenswrapper[4764]: I0202 09:18:00.588237 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zhn7j_24632dda-6100-4ab6-a28e-214ddae4360c/ovn-controller/0.log" Feb 02 09:18:00 crc kubenswrapper[4764]: I0202 09:18:00.588547 4764 generic.go:334] "Generic (PLEG): container finished" podID="24632dda-6100-4ab6-a28e-214ddae4360c" containerID="20b71d3ee7a381baa51b0b0d0470283022cfede30f491135737b5e82d631d763" exitCode=0 Feb 02 09:18:00 crc kubenswrapper[4764]: I0202 09:18:00.588649 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" event={"ID":"24632dda-6100-4ab6-a28e-214ddae4360c","Type":"ContainerDied","Data":"20b71d3ee7a381baa51b0b0d0470283022cfede30f491135737b5e82d631d763"} Feb 02 09:18:00 crc kubenswrapper[4764]: I0202 09:18:00.588729 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" event={"ID":"24632dda-6100-4ab6-a28e-214ddae4360c","Type":"ContainerDied","Data":"56f8167826c412be9cb1f1f95757c0e5a8d5cc8a132658744a99a4bbe5ca1c84"} Feb 02 09:18:00 crc kubenswrapper[4764]: I0202 09:18:00.588789 4764 scope.go:117] "RemoveContainer" containerID="39395584eacf319103cbe03b647c7c7672646b5619a631dffd7f115435591187" Feb 02 09:18:00 crc kubenswrapper[4764]: I0202 09:18:00.588974 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-zhn7j" Feb 02 09:18:00 crc kubenswrapper[4764]: I0202 09:18:00.590622 4764 generic.go:334] "Generic (PLEG): container finished" podID="ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a" containerID="0f72ebe19f85dcdfbf3e0f81369691742f6aab459a74e405fd3e49c1b53435c7" exitCode=0 Feb 02 09:18:00 crc kubenswrapper[4764]: I0202 09:18:00.590661 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gdj8q" event={"ID":"ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a","Type":"ContainerDied","Data":"0f72ebe19f85dcdfbf3e0f81369691742f6aab459a74e405fd3e49c1b53435c7"} Feb 02 09:18:00 crc kubenswrapper[4764]: I0202 09:18:00.590687 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gdj8q" event={"ID":"ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a","Type":"ContainerStarted","Data":"92c54e9ec2807e1eaca5563e0441d415399c0bdad0973ffc585631d7ac25bd7e"} Feb 02 09:18:00 crc kubenswrapper[4764]: I0202 09:18:00.610708 4764 scope.go:117] "RemoveContainer" containerID="4b4cfbd4e682bcdcaff785013bdb72580ce36fced733e235538536827036b93d" Feb 02 09:18:00 crc kubenswrapper[4764]: I0202 09:18:00.640716 4764 scope.go:117] "RemoveContainer" containerID="8a989645cf485cd8ff80ffff1a86cd7808e1564eba5667535e161cc68db13b26" Feb 02 09:18:00 crc kubenswrapper[4764]: I0202 09:18:00.673898 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-zhn7j"] Feb 02 09:18:00 crc kubenswrapper[4764]: I0202 09:18:00.677966 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-zhn7j"] Feb 02 09:18:00 crc kubenswrapper[4764]: I0202 09:18:00.699558 4764 scope.go:117] "RemoveContainer" containerID="20b71d3ee7a381baa51b0b0d0470283022cfede30f491135737b5e82d631d763" Feb 02 09:18:00 crc kubenswrapper[4764]: I0202 09:18:00.728978 4764 scope.go:117] "RemoveContainer" containerID="bffda48d7c881c1470251b501661324859373246cc1a21811bd36ca24887c7aa" Feb 02 09:18:00 crc kubenswrapper[4764]: I0202 09:18:00.741119 4764 scope.go:117] "RemoveContainer" containerID="0a6c67e9b7708cc0d85b0b6ffbbade0d34867e159a9c5487ca2606475c667a50" Feb 02 09:18:00 crc kubenswrapper[4764]: I0202 09:18:00.754280 4764 scope.go:117] "RemoveContainer" containerID="c380383b6a014d1b694a13ddc22f54b20b9e4f9d5c9b9000d76187621dce0a21" Feb 02 09:18:00 crc kubenswrapper[4764]: I0202 09:18:00.768360 4764 scope.go:117] "RemoveContainer" containerID="5529f2bff87753428e9173826fc1880acf96a4c134fa9ecad195b15697cad73a" Feb 02 09:18:00 crc kubenswrapper[4764]: I0202 09:18:00.786711 4764 scope.go:117] "RemoveContainer" containerID="579903eadb183466a2564737e6b678e04308239c0cd75b6808892053b97afbc0" Feb 02 09:18:00 crc kubenswrapper[4764]: I0202 09:18:00.808138 4764 scope.go:117] "RemoveContainer" containerID="39395584eacf319103cbe03b647c7c7672646b5619a631dffd7f115435591187" Feb 02 09:18:00 crc kubenswrapper[4764]: E0202 09:18:00.808923 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"39395584eacf319103cbe03b647c7c7672646b5619a631dffd7f115435591187\": container with ID starting with 39395584eacf319103cbe03b647c7c7672646b5619a631dffd7f115435591187 not found: ID does not exist" containerID="39395584eacf319103cbe03b647c7c7672646b5619a631dffd7f115435591187" Feb 02 09:18:00 crc kubenswrapper[4764]: I0202 09:18:00.808963 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"39395584eacf319103cbe03b647c7c7672646b5619a631dffd7f115435591187"} err="failed to get container status \"39395584eacf319103cbe03b647c7c7672646b5619a631dffd7f115435591187\": rpc error: code = NotFound desc = could not find container \"39395584eacf319103cbe03b647c7c7672646b5619a631dffd7f115435591187\": container with ID starting with 39395584eacf319103cbe03b647c7c7672646b5619a631dffd7f115435591187 not found: ID does not exist" Feb 02 09:18:00 crc kubenswrapper[4764]: I0202 09:18:00.808983 4764 scope.go:117] "RemoveContainer" containerID="4b4cfbd4e682bcdcaff785013bdb72580ce36fced733e235538536827036b93d" Feb 02 09:18:00 crc kubenswrapper[4764]: E0202 09:18:00.809325 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4b4cfbd4e682bcdcaff785013bdb72580ce36fced733e235538536827036b93d\": container with ID starting with 4b4cfbd4e682bcdcaff785013bdb72580ce36fced733e235538536827036b93d not found: ID does not exist" containerID="4b4cfbd4e682bcdcaff785013bdb72580ce36fced733e235538536827036b93d" Feb 02 09:18:00 crc kubenswrapper[4764]: I0202 09:18:00.809379 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4b4cfbd4e682bcdcaff785013bdb72580ce36fced733e235538536827036b93d"} err="failed to get container status \"4b4cfbd4e682bcdcaff785013bdb72580ce36fced733e235538536827036b93d\": rpc error: code = NotFound desc = could not find container \"4b4cfbd4e682bcdcaff785013bdb72580ce36fced733e235538536827036b93d\": container with ID starting with 4b4cfbd4e682bcdcaff785013bdb72580ce36fced733e235538536827036b93d not found: ID does not exist" Feb 02 09:18:00 crc kubenswrapper[4764]: I0202 09:18:00.809394 4764 scope.go:117] "RemoveContainer" containerID="8a989645cf485cd8ff80ffff1a86cd7808e1564eba5667535e161cc68db13b26" Feb 02 09:18:00 crc kubenswrapper[4764]: E0202 09:18:00.809812 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8a989645cf485cd8ff80ffff1a86cd7808e1564eba5667535e161cc68db13b26\": container with ID starting with 8a989645cf485cd8ff80ffff1a86cd7808e1564eba5667535e161cc68db13b26 not found: ID does not exist" containerID="8a989645cf485cd8ff80ffff1a86cd7808e1564eba5667535e161cc68db13b26" Feb 02 09:18:00 crc kubenswrapper[4764]: I0202 09:18:00.809879 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8a989645cf485cd8ff80ffff1a86cd7808e1564eba5667535e161cc68db13b26"} err="failed to get container status \"8a989645cf485cd8ff80ffff1a86cd7808e1564eba5667535e161cc68db13b26\": rpc error: code = NotFound desc = could not find container \"8a989645cf485cd8ff80ffff1a86cd7808e1564eba5667535e161cc68db13b26\": container with ID starting with 8a989645cf485cd8ff80ffff1a86cd7808e1564eba5667535e161cc68db13b26 not found: ID does not exist" Feb 02 09:18:00 crc kubenswrapper[4764]: I0202 09:18:00.809910 4764 scope.go:117] "RemoveContainer" containerID="20b71d3ee7a381baa51b0b0d0470283022cfede30f491135737b5e82d631d763" Feb 02 09:18:00 crc kubenswrapper[4764]: E0202 09:18:00.810312 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"20b71d3ee7a381baa51b0b0d0470283022cfede30f491135737b5e82d631d763\": container with ID starting with 20b71d3ee7a381baa51b0b0d0470283022cfede30f491135737b5e82d631d763 not found: ID does not exist" containerID="20b71d3ee7a381baa51b0b0d0470283022cfede30f491135737b5e82d631d763" Feb 02 09:18:00 crc kubenswrapper[4764]: I0202 09:18:00.810331 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"20b71d3ee7a381baa51b0b0d0470283022cfede30f491135737b5e82d631d763"} err="failed to get container status \"20b71d3ee7a381baa51b0b0d0470283022cfede30f491135737b5e82d631d763\": rpc error: code = NotFound desc = could not find container \"20b71d3ee7a381baa51b0b0d0470283022cfede30f491135737b5e82d631d763\": container with ID starting with 20b71d3ee7a381baa51b0b0d0470283022cfede30f491135737b5e82d631d763 not found: ID does not exist" Feb 02 09:18:00 crc kubenswrapper[4764]: I0202 09:18:00.810345 4764 scope.go:117] "RemoveContainer" containerID="bffda48d7c881c1470251b501661324859373246cc1a21811bd36ca24887c7aa" Feb 02 09:18:00 crc kubenswrapper[4764]: E0202 09:18:00.810540 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bffda48d7c881c1470251b501661324859373246cc1a21811bd36ca24887c7aa\": container with ID starting with bffda48d7c881c1470251b501661324859373246cc1a21811bd36ca24887c7aa not found: ID does not exist" containerID="bffda48d7c881c1470251b501661324859373246cc1a21811bd36ca24887c7aa" Feb 02 09:18:00 crc kubenswrapper[4764]: I0202 09:18:00.810556 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bffda48d7c881c1470251b501661324859373246cc1a21811bd36ca24887c7aa"} err="failed to get container status \"bffda48d7c881c1470251b501661324859373246cc1a21811bd36ca24887c7aa\": rpc error: code = NotFound desc = could not find container \"bffda48d7c881c1470251b501661324859373246cc1a21811bd36ca24887c7aa\": container with ID starting with bffda48d7c881c1470251b501661324859373246cc1a21811bd36ca24887c7aa not found: ID does not exist" Feb 02 09:18:00 crc kubenswrapper[4764]: I0202 09:18:00.810595 4764 scope.go:117] "RemoveContainer" containerID="0a6c67e9b7708cc0d85b0b6ffbbade0d34867e159a9c5487ca2606475c667a50" Feb 02 09:18:00 crc kubenswrapper[4764]: E0202 09:18:00.810886 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0a6c67e9b7708cc0d85b0b6ffbbade0d34867e159a9c5487ca2606475c667a50\": container with ID starting with 0a6c67e9b7708cc0d85b0b6ffbbade0d34867e159a9c5487ca2606475c667a50 not found: ID does not exist" containerID="0a6c67e9b7708cc0d85b0b6ffbbade0d34867e159a9c5487ca2606475c667a50" Feb 02 09:18:00 crc kubenswrapper[4764]: I0202 09:18:00.810905 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0a6c67e9b7708cc0d85b0b6ffbbade0d34867e159a9c5487ca2606475c667a50"} err="failed to get container status \"0a6c67e9b7708cc0d85b0b6ffbbade0d34867e159a9c5487ca2606475c667a50\": rpc error: code = NotFound desc = could not find container \"0a6c67e9b7708cc0d85b0b6ffbbade0d34867e159a9c5487ca2606475c667a50\": container with ID starting with 0a6c67e9b7708cc0d85b0b6ffbbade0d34867e159a9c5487ca2606475c667a50 not found: ID does not exist" Feb 02 09:18:00 crc kubenswrapper[4764]: I0202 09:18:00.810916 4764 scope.go:117] "RemoveContainer" containerID="c380383b6a014d1b694a13ddc22f54b20b9e4f9d5c9b9000d76187621dce0a21" Feb 02 09:18:00 crc kubenswrapper[4764]: E0202 09:18:00.811119 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c380383b6a014d1b694a13ddc22f54b20b9e4f9d5c9b9000d76187621dce0a21\": container with ID starting with c380383b6a014d1b694a13ddc22f54b20b9e4f9d5c9b9000d76187621dce0a21 not found: ID does not exist" containerID="c380383b6a014d1b694a13ddc22f54b20b9e4f9d5c9b9000d76187621dce0a21" Feb 02 09:18:00 crc kubenswrapper[4764]: I0202 09:18:00.811135 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c380383b6a014d1b694a13ddc22f54b20b9e4f9d5c9b9000d76187621dce0a21"} err="failed to get container status \"c380383b6a014d1b694a13ddc22f54b20b9e4f9d5c9b9000d76187621dce0a21\": rpc error: code = NotFound desc = could not find container \"c380383b6a014d1b694a13ddc22f54b20b9e4f9d5c9b9000d76187621dce0a21\": container with ID starting with c380383b6a014d1b694a13ddc22f54b20b9e4f9d5c9b9000d76187621dce0a21 not found: ID does not exist" Feb 02 09:18:00 crc kubenswrapper[4764]: I0202 09:18:00.811147 4764 scope.go:117] "RemoveContainer" containerID="5529f2bff87753428e9173826fc1880acf96a4c134fa9ecad195b15697cad73a" Feb 02 09:18:00 crc kubenswrapper[4764]: E0202 09:18:00.811333 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5529f2bff87753428e9173826fc1880acf96a4c134fa9ecad195b15697cad73a\": container with ID starting with 5529f2bff87753428e9173826fc1880acf96a4c134fa9ecad195b15697cad73a not found: ID does not exist" containerID="5529f2bff87753428e9173826fc1880acf96a4c134fa9ecad195b15697cad73a" Feb 02 09:18:00 crc kubenswrapper[4764]: I0202 09:18:00.811352 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5529f2bff87753428e9173826fc1880acf96a4c134fa9ecad195b15697cad73a"} err="failed to get container status \"5529f2bff87753428e9173826fc1880acf96a4c134fa9ecad195b15697cad73a\": rpc error: code = NotFound desc = could not find container \"5529f2bff87753428e9173826fc1880acf96a4c134fa9ecad195b15697cad73a\": container with ID starting with 5529f2bff87753428e9173826fc1880acf96a4c134fa9ecad195b15697cad73a not found: ID does not exist" Feb 02 09:18:00 crc kubenswrapper[4764]: I0202 09:18:00.811363 4764 scope.go:117] "RemoveContainer" containerID="579903eadb183466a2564737e6b678e04308239c0cd75b6808892053b97afbc0" Feb 02 09:18:00 crc kubenswrapper[4764]: E0202 09:18:00.811562 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"579903eadb183466a2564737e6b678e04308239c0cd75b6808892053b97afbc0\": container with ID starting with 579903eadb183466a2564737e6b678e04308239c0cd75b6808892053b97afbc0 not found: ID does not exist" containerID="579903eadb183466a2564737e6b678e04308239c0cd75b6808892053b97afbc0" Feb 02 09:18:00 crc kubenswrapper[4764]: I0202 09:18:00.811578 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"579903eadb183466a2564737e6b678e04308239c0cd75b6808892053b97afbc0"} err="failed to get container status \"579903eadb183466a2564737e6b678e04308239c0cd75b6808892053b97afbc0\": rpc error: code = NotFound desc = could not find container \"579903eadb183466a2564737e6b678e04308239c0cd75b6808892053b97afbc0\": container with ID starting with 579903eadb183466a2564737e6b678e04308239c0cd75b6808892053b97afbc0 not found: ID does not exist" Feb 02 09:18:01 crc kubenswrapper[4764]: I0202 09:18:01.608488 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gdj8q" event={"ID":"ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a","Type":"ContainerStarted","Data":"384cf6e60e8c9096f1b72d170439950ef4b87aa5bcddc18dbe7437a8c03a5157"} Feb 02 09:18:01 crc kubenswrapper[4764]: I0202 09:18:01.608798 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gdj8q" event={"ID":"ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a","Type":"ContainerStarted","Data":"c53b0ca667dd2422a8a87066419fffbcf887da917fc8c78f810480e6d07810b1"} Feb 02 09:18:01 crc kubenswrapper[4764]: I0202 09:18:01.608813 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gdj8q" event={"ID":"ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a","Type":"ContainerStarted","Data":"5c3d26f84e7d6d426abb98f79ac6cd8857ec6dce02d6b2c83d2362a3db00d08b"} Feb 02 09:18:01 crc kubenswrapper[4764]: I0202 09:18:01.608823 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gdj8q" event={"ID":"ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a","Type":"ContainerStarted","Data":"2d5a5d71b183f5292dcbbb21bcdb2fab4414530f38d205fd84ced4302d41bb83"} Feb 02 09:18:01 crc kubenswrapper[4764]: I0202 09:18:01.608831 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gdj8q" event={"ID":"ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a","Type":"ContainerStarted","Data":"7a312e2ad2ff5489536ced5b94f373e92769b90d797ceae0a56b0c6e7608757b"} Feb 02 09:18:01 crc kubenswrapper[4764]: I0202 09:18:01.608840 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gdj8q" event={"ID":"ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a","Type":"ContainerStarted","Data":"93d6398b598ee8e418f6f8518379f44836b5dc2d846d31bb06df0b8ae600e65f"} Feb 02 09:18:01 crc kubenswrapper[4764]: I0202 09:18:01.837209 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="24632dda-6100-4ab6-a28e-214ddae4360c" path="/var/lib/kubelet/pods/24632dda-6100-4ab6-a28e-214ddae4360c/volumes" Feb 02 09:18:04 crc kubenswrapper[4764]: I0202 09:18:04.634135 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gdj8q" event={"ID":"ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a","Type":"ContainerStarted","Data":"084fd1d2400cafa3522229cf6271363f67ec96eda2546e9ab70a803ee160a012"} Feb 02 09:18:06 crc kubenswrapper[4764]: I0202 09:18:06.652456 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gdj8q" event={"ID":"ff4b8d7a-7ba0-43ab-ad3b-67a8adc9ac2a","Type":"ContainerStarted","Data":"23c60481d07a2debc800044786a98fc3f1515f9544205c05175a353980c640ea"} Feb 02 09:18:06 crc kubenswrapper[4764]: I0202 09:18:06.654173 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-gdj8q" Feb 02 09:18:06 crc kubenswrapper[4764]: I0202 09:18:06.654214 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-gdj8q" Feb 02 09:18:06 crc kubenswrapper[4764]: I0202 09:18:06.654226 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-gdj8q" Feb 02 09:18:06 crc kubenswrapper[4764]: I0202 09:18:06.680726 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-gdj8q" Feb 02 09:18:06 crc kubenswrapper[4764]: I0202 09:18:06.685590 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-gdj8q" Feb 02 09:18:06 crc kubenswrapper[4764]: I0202 09:18:06.686274 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-gdj8q" podStartSLOduration=7.686260578 podStartE2EDuration="7.686260578s" podCreationTimestamp="2026-02-02 09:17:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:18:06.680755309 +0000 UTC m=+649.614479397" watchObservedRunningTime="2026-02-02 09:18:06.686260578 +0000 UTC m=+649.619984676" Feb 02 09:18:10 crc kubenswrapper[4764]: I0202 09:18:10.827500 4764 scope.go:117] "RemoveContainer" containerID="b7882e3ed0730ea5035949695b7b94a50060918bee89edc807fd67484b957ee5" Feb 02 09:18:10 crc kubenswrapper[4764]: E0202 09:18:10.828378 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-nc7qt_openshift-multus(6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e)\"" pod="openshift-multus/multus-nc7qt" podUID="6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e" Feb 02 09:18:22 crc kubenswrapper[4764]: I0202 09:18:22.825989 4764 scope.go:117] "RemoveContainer" containerID="b7882e3ed0730ea5035949695b7b94a50060918bee89edc807fd67484b957ee5" Feb 02 09:18:23 crc kubenswrapper[4764]: I0202 09:18:23.798846 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-nc7qt_6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e/kube-multus/2.log" Feb 02 09:18:23 crc kubenswrapper[4764]: I0202 09:18:23.799272 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-nc7qt" event={"ID":"6ae2eac6-1e25-4166-bb6b-7e0e0dd4be6e","Type":"ContainerStarted","Data":"d7c141006f08bda9de384b2880c994d59bce5256fd29b688ec4d60eaa68da770"} Feb 02 09:18:30 crc kubenswrapper[4764]: I0202 09:18:30.030316 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-gdj8q" Feb 02 09:18:39 crc kubenswrapper[4764]: I0202 09:18:39.625586 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec71324cpz"] Feb 02 09:18:39 crc kubenswrapper[4764]: I0202 09:18:39.628199 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec71324cpz" Feb 02 09:18:39 crc kubenswrapper[4764]: I0202 09:18:39.630273 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Feb 02 09:18:39 crc kubenswrapper[4764]: I0202 09:18:39.641504 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec71324cpz"] Feb 02 09:18:39 crc kubenswrapper[4764]: I0202 09:18:39.716025 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/08e7a6ba-02aa-4f59-8e6d-b967cb0f9715-bundle\") pod \"53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec71324cpz\" (UID: \"08e7a6ba-02aa-4f59-8e6d-b967cb0f9715\") " pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec71324cpz" Feb 02 09:18:39 crc kubenswrapper[4764]: I0202 09:18:39.716100 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/08e7a6ba-02aa-4f59-8e6d-b967cb0f9715-util\") pod \"53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec71324cpz\" (UID: \"08e7a6ba-02aa-4f59-8e6d-b967cb0f9715\") " pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec71324cpz" Feb 02 09:18:39 crc kubenswrapper[4764]: I0202 09:18:39.716200 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b6jhk\" (UniqueName: \"kubernetes.io/projected/08e7a6ba-02aa-4f59-8e6d-b967cb0f9715-kube-api-access-b6jhk\") pod \"53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec71324cpz\" (UID: \"08e7a6ba-02aa-4f59-8e6d-b967cb0f9715\") " pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec71324cpz" Feb 02 09:18:39 crc kubenswrapper[4764]: I0202 09:18:39.817395 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/08e7a6ba-02aa-4f59-8e6d-b967cb0f9715-bundle\") pod \"53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec71324cpz\" (UID: \"08e7a6ba-02aa-4f59-8e6d-b967cb0f9715\") " pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec71324cpz" Feb 02 09:18:39 crc kubenswrapper[4764]: I0202 09:18:39.817450 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/08e7a6ba-02aa-4f59-8e6d-b967cb0f9715-util\") pod \"53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec71324cpz\" (UID: \"08e7a6ba-02aa-4f59-8e6d-b967cb0f9715\") " pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec71324cpz" Feb 02 09:18:39 crc kubenswrapper[4764]: I0202 09:18:39.817495 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b6jhk\" (UniqueName: \"kubernetes.io/projected/08e7a6ba-02aa-4f59-8e6d-b967cb0f9715-kube-api-access-b6jhk\") pod \"53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec71324cpz\" (UID: \"08e7a6ba-02aa-4f59-8e6d-b967cb0f9715\") " pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec71324cpz" Feb 02 09:18:39 crc kubenswrapper[4764]: I0202 09:18:39.817880 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/08e7a6ba-02aa-4f59-8e6d-b967cb0f9715-bundle\") pod \"53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec71324cpz\" (UID: \"08e7a6ba-02aa-4f59-8e6d-b967cb0f9715\") " pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec71324cpz" Feb 02 09:18:39 crc kubenswrapper[4764]: I0202 09:18:39.818032 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/08e7a6ba-02aa-4f59-8e6d-b967cb0f9715-util\") pod \"53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec71324cpz\" (UID: \"08e7a6ba-02aa-4f59-8e6d-b967cb0f9715\") " pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec71324cpz" Feb 02 09:18:39 crc kubenswrapper[4764]: I0202 09:18:39.842513 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b6jhk\" (UniqueName: \"kubernetes.io/projected/08e7a6ba-02aa-4f59-8e6d-b967cb0f9715-kube-api-access-b6jhk\") pod \"53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec71324cpz\" (UID: \"08e7a6ba-02aa-4f59-8e6d-b967cb0f9715\") " pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec71324cpz" Feb 02 09:18:39 crc kubenswrapper[4764]: I0202 09:18:39.986781 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec71324cpz" Feb 02 09:18:40 crc kubenswrapper[4764]: I0202 09:18:40.242640 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec71324cpz"] Feb 02 09:18:40 crc kubenswrapper[4764]: W0202 09:18:40.252335 4764 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod08e7a6ba_02aa_4f59_8e6d_b967cb0f9715.slice/crio-b637de479091b3efdec666be6fcdb4f4ec326df16ee0013f0554d800a2d6309e WatchSource:0}: Error finding container b637de479091b3efdec666be6fcdb4f4ec326df16ee0013f0554d800a2d6309e: Status 404 returned error can't find the container with id b637de479091b3efdec666be6fcdb4f4ec326df16ee0013f0554d800a2d6309e Feb 02 09:18:40 crc kubenswrapper[4764]: I0202 09:18:40.931442 4764 generic.go:334] "Generic (PLEG): container finished" podID="08e7a6ba-02aa-4f59-8e6d-b967cb0f9715" containerID="536345e16ca7100afa45a0d82ec845ab865929c4364849b568797d71c4c8aa4f" exitCode=0 Feb 02 09:18:40 crc kubenswrapper[4764]: I0202 09:18:40.931554 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec71324cpz" event={"ID":"08e7a6ba-02aa-4f59-8e6d-b967cb0f9715","Type":"ContainerDied","Data":"536345e16ca7100afa45a0d82ec845ab865929c4364849b568797d71c4c8aa4f"} Feb 02 09:18:40 crc kubenswrapper[4764]: I0202 09:18:40.931753 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec71324cpz" event={"ID":"08e7a6ba-02aa-4f59-8e6d-b967cb0f9715","Type":"ContainerStarted","Data":"b637de479091b3efdec666be6fcdb4f4ec326df16ee0013f0554d800a2d6309e"} Feb 02 09:18:42 crc kubenswrapper[4764]: I0202 09:18:42.945064 4764 generic.go:334] "Generic (PLEG): container finished" podID="08e7a6ba-02aa-4f59-8e6d-b967cb0f9715" containerID="98937b93366f96883741ca8a9e57954818013514b0b510caafd913da89469766" exitCode=0 Feb 02 09:18:42 crc kubenswrapper[4764]: I0202 09:18:42.945164 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec71324cpz" event={"ID":"08e7a6ba-02aa-4f59-8e6d-b967cb0f9715","Type":"ContainerDied","Data":"98937b93366f96883741ca8a9e57954818013514b0b510caafd913da89469766"} Feb 02 09:18:43 crc kubenswrapper[4764]: I0202 09:18:43.957985 4764 generic.go:334] "Generic (PLEG): container finished" podID="08e7a6ba-02aa-4f59-8e6d-b967cb0f9715" containerID="de85c4627abce4218a11ccbad13f73112d78bced9c8e713a35aa318aad54b32b" exitCode=0 Feb 02 09:18:43 crc kubenswrapper[4764]: I0202 09:18:43.958134 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec71324cpz" event={"ID":"08e7a6ba-02aa-4f59-8e6d-b967cb0f9715","Type":"ContainerDied","Data":"de85c4627abce4218a11ccbad13f73112d78bced9c8e713a35aa318aad54b32b"} Feb 02 09:18:45 crc kubenswrapper[4764]: I0202 09:18:45.289131 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec71324cpz" Feb 02 09:18:45 crc kubenswrapper[4764]: I0202 09:18:45.304387 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b6jhk\" (UniqueName: \"kubernetes.io/projected/08e7a6ba-02aa-4f59-8e6d-b967cb0f9715-kube-api-access-b6jhk\") pod \"08e7a6ba-02aa-4f59-8e6d-b967cb0f9715\" (UID: \"08e7a6ba-02aa-4f59-8e6d-b967cb0f9715\") " Feb 02 09:18:45 crc kubenswrapper[4764]: I0202 09:18:45.304462 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/08e7a6ba-02aa-4f59-8e6d-b967cb0f9715-bundle\") pod \"08e7a6ba-02aa-4f59-8e6d-b967cb0f9715\" (UID: \"08e7a6ba-02aa-4f59-8e6d-b967cb0f9715\") " Feb 02 09:18:45 crc kubenswrapper[4764]: I0202 09:18:45.304541 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/08e7a6ba-02aa-4f59-8e6d-b967cb0f9715-util\") pod \"08e7a6ba-02aa-4f59-8e6d-b967cb0f9715\" (UID: \"08e7a6ba-02aa-4f59-8e6d-b967cb0f9715\") " Feb 02 09:18:45 crc kubenswrapper[4764]: I0202 09:18:45.305442 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/08e7a6ba-02aa-4f59-8e6d-b967cb0f9715-bundle" (OuterVolumeSpecName: "bundle") pod "08e7a6ba-02aa-4f59-8e6d-b967cb0f9715" (UID: "08e7a6ba-02aa-4f59-8e6d-b967cb0f9715"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 09:18:45 crc kubenswrapper[4764]: I0202 09:18:45.318272 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/08e7a6ba-02aa-4f59-8e6d-b967cb0f9715-kube-api-access-b6jhk" (OuterVolumeSpecName: "kube-api-access-b6jhk") pod "08e7a6ba-02aa-4f59-8e6d-b967cb0f9715" (UID: "08e7a6ba-02aa-4f59-8e6d-b967cb0f9715"). InnerVolumeSpecName "kube-api-access-b6jhk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:18:45 crc kubenswrapper[4764]: I0202 09:18:45.366821 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/08e7a6ba-02aa-4f59-8e6d-b967cb0f9715-util" (OuterVolumeSpecName: "util") pod "08e7a6ba-02aa-4f59-8e6d-b967cb0f9715" (UID: "08e7a6ba-02aa-4f59-8e6d-b967cb0f9715"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 09:18:45 crc kubenswrapper[4764]: I0202 09:18:45.405372 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b6jhk\" (UniqueName: \"kubernetes.io/projected/08e7a6ba-02aa-4f59-8e6d-b967cb0f9715-kube-api-access-b6jhk\") on node \"crc\" DevicePath \"\"" Feb 02 09:18:45 crc kubenswrapper[4764]: I0202 09:18:45.405408 4764 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/08e7a6ba-02aa-4f59-8e6d-b967cb0f9715-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 09:18:45 crc kubenswrapper[4764]: I0202 09:18:45.405420 4764 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/08e7a6ba-02aa-4f59-8e6d-b967cb0f9715-util\") on node \"crc\" DevicePath \"\"" Feb 02 09:18:45 crc kubenswrapper[4764]: I0202 09:18:45.978092 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec71324cpz" event={"ID":"08e7a6ba-02aa-4f59-8e6d-b967cb0f9715","Type":"ContainerDied","Data":"b637de479091b3efdec666be6fcdb4f4ec326df16ee0013f0554d800a2d6309e"} Feb 02 09:18:45 crc kubenswrapper[4764]: I0202 09:18:45.978151 4764 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b637de479091b3efdec666be6fcdb4f4ec326df16ee0013f0554d800a2d6309e" Feb 02 09:18:45 crc kubenswrapper[4764]: I0202 09:18:45.978249 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec71324cpz" Feb 02 09:18:48 crc kubenswrapper[4764]: I0202 09:18:48.606972 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-646758c888-mv59b"] Feb 02 09:18:48 crc kubenswrapper[4764]: E0202 09:18:48.607599 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08e7a6ba-02aa-4f59-8e6d-b967cb0f9715" containerName="util" Feb 02 09:18:48 crc kubenswrapper[4764]: I0202 09:18:48.607610 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="08e7a6ba-02aa-4f59-8e6d-b967cb0f9715" containerName="util" Feb 02 09:18:48 crc kubenswrapper[4764]: E0202 09:18:48.607624 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08e7a6ba-02aa-4f59-8e6d-b967cb0f9715" containerName="extract" Feb 02 09:18:48 crc kubenswrapper[4764]: I0202 09:18:48.607630 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="08e7a6ba-02aa-4f59-8e6d-b967cb0f9715" containerName="extract" Feb 02 09:18:48 crc kubenswrapper[4764]: E0202 09:18:48.607640 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08e7a6ba-02aa-4f59-8e6d-b967cb0f9715" containerName="pull" Feb 02 09:18:48 crc kubenswrapper[4764]: I0202 09:18:48.607648 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="08e7a6ba-02aa-4f59-8e6d-b967cb0f9715" containerName="pull" Feb 02 09:18:48 crc kubenswrapper[4764]: I0202 09:18:48.607737 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="08e7a6ba-02aa-4f59-8e6d-b967cb0f9715" containerName="extract" Feb 02 09:18:48 crc kubenswrapper[4764]: I0202 09:18:48.608069 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-646758c888-mv59b" Feb 02 09:18:48 crc kubenswrapper[4764]: W0202 09:18:48.611567 4764 reflector.go:561] object-"openshift-nmstate"/"kube-root-ca.crt": failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-nmstate": no relationship found between node 'crc' and this object Feb 02 09:18:48 crc kubenswrapper[4764]: E0202 09:18:48.611619 4764 reflector.go:158] "Unhandled Error" err="object-\"openshift-nmstate\"/\"kube-root-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-root-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-nmstate\": no relationship found between node 'crc' and this object" logger="UnhandledError" Feb 02 09:18:48 crc kubenswrapper[4764]: I0202 09:18:48.612698 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Feb 02 09:18:48 crc kubenswrapper[4764]: I0202 09:18:48.612995 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-hqkkk" Feb 02 09:18:48 crc kubenswrapper[4764]: I0202 09:18:48.628157 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-646758c888-mv59b"] Feb 02 09:18:48 crc kubenswrapper[4764]: I0202 09:18:48.661414 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xkk9c\" (UniqueName: \"kubernetes.io/projected/2c1baf80-192e-47b6-852c-394b7e4ade4e-kube-api-access-xkk9c\") pod \"nmstate-operator-646758c888-mv59b\" (UID: \"2c1baf80-192e-47b6-852c-394b7e4ade4e\") " pod="openshift-nmstate/nmstate-operator-646758c888-mv59b" Feb 02 09:18:48 crc kubenswrapper[4764]: I0202 09:18:48.762716 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xkk9c\" (UniqueName: \"kubernetes.io/projected/2c1baf80-192e-47b6-852c-394b7e4ade4e-kube-api-access-xkk9c\") pod \"nmstate-operator-646758c888-mv59b\" (UID: \"2c1baf80-192e-47b6-852c-394b7e4ade4e\") " pod="openshift-nmstate/nmstate-operator-646758c888-mv59b" Feb 02 09:18:49 crc kubenswrapper[4764]: I0202 09:18:49.437984 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Feb 02 09:18:49 crc kubenswrapper[4764]: I0202 09:18:49.451145 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xkk9c\" (UniqueName: \"kubernetes.io/projected/2c1baf80-192e-47b6-852c-394b7e4ade4e-kube-api-access-xkk9c\") pod \"nmstate-operator-646758c888-mv59b\" (UID: \"2c1baf80-192e-47b6-852c-394b7e4ade4e\") " pod="openshift-nmstate/nmstate-operator-646758c888-mv59b" Feb 02 09:18:49 crc kubenswrapper[4764]: I0202 09:18:49.521979 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-646758c888-mv59b" Feb 02 09:18:49 crc kubenswrapper[4764]: I0202 09:18:49.943349 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-646758c888-mv59b"] Feb 02 09:18:50 crc kubenswrapper[4764]: I0202 09:18:50.000921 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-646758c888-mv59b" event={"ID":"2c1baf80-192e-47b6-852c-394b7e4ade4e","Type":"ContainerStarted","Data":"3eac8d2338d4df433b583db467b9e05cc97da1e569c83a86eb3791edbdb09b43"} Feb 02 09:18:52 crc kubenswrapper[4764]: I0202 09:18:52.015106 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-646758c888-mv59b" event={"ID":"2c1baf80-192e-47b6-852c-394b7e4ade4e","Type":"ContainerStarted","Data":"0e5fe6b3d375df29692d17387efb1dc530d7d5e19515b031ef3bdc9fa7f585f0"} Feb 02 09:18:53 crc kubenswrapper[4764]: I0202 09:18:53.037787 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-646758c888-mv59b" podStartSLOduration=3.112219189 podStartE2EDuration="5.037765914s" podCreationTimestamp="2026-02-02 09:18:48 +0000 UTC" firstStartedPulling="2026-02-02 09:18:49.951629371 +0000 UTC m=+692.885353449" lastFinishedPulling="2026-02-02 09:18:51.877176066 +0000 UTC m=+694.810900174" observedRunningTime="2026-02-02 09:18:53.036672065 +0000 UTC m=+695.970396153" watchObservedRunningTime="2026-02-02 09:18:53.037765914 +0000 UTC m=+695.971490012" Feb 02 09:18:54 crc kubenswrapper[4764]: I0202 09:18:54.058810 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-54757c584b-jz8q9"] Feb 02 09:18:54 crc kubenswrapper[4764]: I0202 09:18:54.059837 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-54757c584b-jz8q9" Feb 02 09:18:54 crc kubenswrapper[4764]: I0202 09:18:54.062077 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-zjqhp" Feb 02 09:18:54 crc kubenswrapper[4764]: I0202 09:18:54.083460 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-8474b5b9d8-gz969"] Feb 02 09:18:54 crc kubenswrapper[4764]: I0202 09:18:54.084206 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-gz969" Feb 02 09:18:54 crc kubenswrapper[4764]: I0202 09:18:54.089467 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Feb 02 09:18:54 crc kubenswrapper[4764]: I0202 09:18:54.104975 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-gswb4"] Feb 02 09:18:54 crc kubenswrapper[4764]: I0202 09:18:54.108859 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-gswb4" Feb 02 09:18:54 crc kubenswrapper[4764]: I0202 09:18:54.139449 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/7e5e0b71-cd10-4ccf-8100-f15a063cceb7-dbus-socket\") pod \"nmstate-handler-gswb4\" (UID: \"7e5e0b71-cd10-4ccf-8100-f15a063cceb7\") " pod="openshift-nmstate/nmstate-handler-gswb4" Feb 02 09:18:54 crc kubenswrapper[4764]: I0202 09:18:54.139493 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/1a77ee37-2514-4a73-b2f1-565442966b16-tls-key-pair\") pod \"nmstate-webhook-8474b5b9d8-gz969\" (UID: \"1a77ee37-2514-4a73-b2f1-565442966b16\") " pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-gz969" Feb 02 09:18:54 crc kubenswrapper[4764]: I0202 09:18:54.139525 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/7e5e0b71-cd10-4ccf-8100-f15a063cceb7-ovs-socket\") pod \"nmstate-handler-gswb4\" (UID: \"7e5e0b71-cd10-4ccf-8100-f15a063cceb7\") " pod="openshift-nmstate/nmstate-handler-gswb4" Feb 02 09:18:54 crc kubenswrapper[4764]: I0202 09:18:54.139600 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9bpcj\" (UniqueName: \"kubernetes.io/projected/1a77ee37-2514-4a73-b2f1-565442966b16-kube-api-access-9bpcj\") pod \"nmstate-webhook-8474b5b9d8-gz969\" (UID: \"1a77ee37-2514-4a73-b2f1-565442966b16\") " pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-gz969" Feb 02 09:18:54 crc kubenswrapper[4764]: I0202 09:18:54.139621 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/7e5e0b71-cd10-4ccf-8100-f15a063cceb7-nmstate-lock\") pod \"nmstate-handler-gswb4\" (UID: \"7e5e0b71-cd10-4ccf-8100-f15a063cceb7\") " pod="openshift-nmstate/nmstate-handler-gswb4" Feb 02 09:18:54 crc kubenswrapper[4764]: I0202 09:18:54.139652 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-swtqr\" (UniqueName: \"kubernetes.io/projected/dd14d932-c6a2-4d7e-a1bb-bf7cf4c30f00-kube-api-access-swtqr\") pod \"nmstate-metrics-54757c584b-jz8q9\" (UID: \"dd14d932-c6a2-4d7e-a1bb-bf7cf4c30f00\") " pod="openshift-nmstate/nmstate-metrics-54757c584b-jz8q9" Feb 02 09:18:54 crc kubenswrapper[4764]: I0202 09:18:54.139670 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-57p89\" (UniqueName: \"kubernetes.io/projected/7e5e0b71-cd10-4ccf-8100-f15a063cceb7-kube-api-access-57p89\") pod \"nmstate-handler-gswb4\" (UID: \"7e5e0b71-cd10-4ccf-8100-f15a063cceb7\") " pod="openshift-nmstate/nmstate-handler-gswb4" Feb 02 09:18:54 crc kubenswrapper[4764]: I0202 09:18:54.192194 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-54757c584b-jz8q9"] Feb 02 09:18:54 crc kubenswrapper[4764]: I0202 09:18:54.197213 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-8474b5b9d8-gz969"] Feb 02 09:18:54 crc kubenswrapper[4764]: I0202 09:18:54.240538 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9bpcj\" (UniqueName: \"kubernetes.io/projected/1a77ee37-2514-4a73-b2f1-565442966b16-kube-api-access-9bpcj\") pod \"nmstate-webhook-8474b5b9d8-gz969\" (UID: \"1a77ee37-2514-4a73-b2f1-565442966b16\") " pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-gz969" Feb 02 09:18:54 crc kubenswrapper[4764]: I0202 09:18:54.240791 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/7e5e0b71-cd10-4ccf-8100-f15a063cceb7-nmstate-lock\") pod \"nmstate-handler-gswb4\" (UID: \"7e5e0b71-cd10-4ccf-8100-f15a063cceb7\") " pod="openshift-nmstate/nmstate-handler-gswb4" Feb 02 09:18:54 crc kubenswrapper[4764]: I0202 09:18:54.240875 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/7e5e0b71-cd10-4ccf-8100-f15a063cceb7-nmstate-lock\") pod \"nmstate-handler-gswb4\" (UID: \"7e5e0b71-cd10-4ccf-8100-f15a063cceb7\") " pod="openshift-nmstate/nmstate-handler-gswb4" Feb 02 09:18:54 crc kubenswrapper[4764]: I0202 09:18:54.240952 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-swtqr\" (UniqueName: \"kubernetes.io/projected/dd14d932-c6a2-4d7e-a1bb-bf7cf4c30f00-kube-api-access-swtqr\") pod \"nmstate-metrics-54757c584b-jz8q9\" (UID: \"dd14d932-c6a2-4d7e-a1bb-bf7cf4c30f00\") " pod="openshift-nmstate/nmstate-metrics-54757c584b-jz8q9" Feb 02 09:18:54 crc kubenswrapper[4764]: I0202 09:18:54.241010 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-57p89\" (UniqueName: \"kubernetes.io/projected/7e5e0b71-cd10-4ccf-8100-f15a063cceb7-kube-api-access-57p89\") pod \"nmstate-handler-gswb4\" (UID: \"7e5e0b71-cd10-4ccf-8100-f15a063cceb7\") " pod="openshift-nmstate/nmstate-handler-gswb4" Feb 02 09:18:54 crc kubenswrapper[4764]: I0202 09:18:54.241071 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/7e5e0b71-cd10-4ccf-8100-f15a063cceb7-dbus-socket\") pod \"nmstate-handler-gswb4\" (UID: \"7e5e0b71-cd10-4ccf-8100-f15a063cceb7\") " pod="openshift-nmstate/nmstate-handler-gswb4" Feb 02 09:18:54 crc kubenswrapper[4764]: I0202 09:18:54.241107 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/1a77ee37-2514-4a73-b2f1-565442966b16-tls-key-pair\") pod \"nmstate-webhook-8474b5b9d8-gz969\" (UID: \"1a77ee37-2514-4a73-b2f1-565442966b16\") " pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-gz969" Feb 02 09:18:54 crc kubenswrapper[4764]: I0202 09:18:54.241152 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/7e5e0b71-cd10-4ccf-8100-f15a063cceb7-ovs-socket\") pod \"nmstate-handler-gswb4\" (UID: \"7e5e0b71-cd10-4ccf-8100-f15a063cceb7\") " pod="openshift-nmstate/nmstate-handler-gswb4" Feb 02 09:18:54 crc kubenswrapper[4764]: I0202 09:18:54.241299 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/7e5e0b71-cd10-4ccf-8100-f15a063cceb7-ovs-socket\") pod \"nmstate-handler-gswb4\" (UID: \"7e5e0b71-cd10-4ccf-8100-f15a063cceb7\") " pod="openshift-nmstate/nmstate-handler-gswb4" Feb 02 09:18:54 crc kubenswrapper[4764]: I0202 09:18:54.241620 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/7e5e0b71-cd10-4ccf-8100-f15a063cceb7-dbus-socket\") pod \"nmstate-handler-gswb4\" (UID: \"7e5e0b71-cd10-4ccf-8100-f15a063cceb7\") " pod="openshift-nmstate/nmstate-handler-gswb4" Feb 02 09:18:54 crc kubenswrapper[4764]: E0202 09:18:54.241783 4764 secret.go:188] Couldn't get secret openshift-nmstate/openshift-nmstate-webhook: secret "openshift-nmstate-webhook" not found Feb 02 09:18:54 crc kubenswrapper[4764]: E0202 09:18:54.241919 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1a77ee37-2514-4a73-b2f1-565442966b16-tls-key-pair podName:1a77ee37-2514-4a73-b2f1-565442966b16 nodeName:}" failed. No retries permitted until 2026-02-02 09:18:54.741898484 +0000 UTC m=+697.675622572 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-key-pair" (UniqueName: "kubernetes.io/secret/1a77ee37-2514-4a73-b2f1-565442966b16-tls-key-pair") pod "nmstate-webhook-8474b5b9d8-gz969" (UID: "1a77ee37-2514-4a73-b2f1-565442966b16") : secret "openshift-nmstate-webhook" not found Feb 02 09:18:54 crc kubenswrapper[4764]: I0202 09:18:54.251217 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7754f76f8b-mxdpd"] Feb 02 09:18:54 crc kubenswrapper[4764]: I0202 09:18:54.251799 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-mxdpd" Feb 02 09:18:54 crc kubenswrapper[4764]: I0202 09:18:54.259201 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Feb 02 09:18:54 crc kubenswrapper[4764]: I0202 09:18:54.259376 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Feb 02 09:18:54 crc kubenswrapper[4764]: I0202 09:18:54.259475 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-xckg7" Feb 02 09:18:54 crc kubenswrapper[4764]: I0202 09:18:54.266692 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7754f76f8b-mxdpd"] Feb 02 09:18:54 crc kubenswrapper[4764]: I0202 09:18:54.268339 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9bpcj\" (UniqueName: \"kubernetes.io/projected/1a77ee37-2514-4a73-b2f1-565442966b16-kube-api-access-9bpcj\") pod \"nmstate-webhook-8474b5b9d8-gz969\" (UID: \"1a77ee37-2514-4a73-b2f1-565442966b16\") " pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-gz969" Feb 02 09:18:54 crc kubenswrapper[4764]: I0202 09:18:54.272827 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-swtqr\" (UniqueName: \"kubernetes.io/projected/dd14d932-c6a2-4d7e-a1bb-bf7cf4c30f00-kube-api-access-swtqr\") pod \"nmstate-metrics-54757c584b-jz8q9\" (UID: \"dd14d932-c6a2-4d7e-a1bb-bf7cf4c30f00\") " pod="openshift-nmstate/nmstate-metrics-54757c584b-jz8q9" Feb 02 09:18:54 crc kubenswrapper[4764]: I0202 09:18:54.286340 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-57p89\" (UniqueName: \"kubernetes.io/projected/7e5e0b71-cd10-4ccf-8100-f15a063cceb7-kube-api-access-57p89\") pod \"nmstate-handler-gswb4\" (UID: \"7e5e0b71-cd10-4ccf-8100-f15a063cceb7\") " pod="openshift-nmstate/nmstate-handler-gswb4" Feb 02 09:18:54 crc kubenswrapper[4764]: I0202 09:18:54.342179 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/67f7c8f7-535f-4a07-a330-adf7c41da701-nginx-conf\") pod \"nmstate-console-plugin-7754f76f8b-mxdpd\" (UID: \"67f7c8f7-535f-4a07-a330-adf7c41da701\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-mxdpd" Feb 02 09:18:54 crc kubenswrapper[4764]: I0202 09:18:54.342277 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j9ts8\" (UniqueName: \"kubernetes.io/projected/67f7c8f7-535f-4a07-a330-adf7c41da701-kube-api-access-j9ts8\") pod \"nmstate-console-plugin-7754f76f8b-mxdpd\" (UID: \"67f7c8f7-535f-4a07-a330-adf7c41da701\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-mxdpd" Feb 02 09:18:54 crc kubenswrapper[4764]: I0202 09:18:54.342336 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/67f7c8f7-535f-4a07-a330-adf7c41da701-plugin-serving-cert\") pod \"nmstate-console-plugin-7754f76f8b-mxdpd\" (UID: \"67f7c8f7-535f-4a07-a330-adf7c41da701\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-mxdpd" Feb 02 09:18:54 crc kubenswrapper[4764]: I0202 09:18:54.380180 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-54757c584b-jz8q9" Feb 02 09:18:54 crc kubenswrapper[4764]: I0202 09:18:54.434085 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-6bc694f886-fp5xj"] Feb 02 09:18:54 crc kubenswrapper[4764]: I0202 09:18:54.434723 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-6bc694f886-fp5xj" Feb 02 09:18:54 crc kubenswrapper[4764]: I0202 09:18:54.437036 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-gswb4" Feb 02 09:18:54 crc kubenswrapper[4764]: I0202 09:18:54.443075 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/67f7c8f7-535f-4a07-a330-adf7c41da701-nginx-conf\") pod \"nmstate-console-plugin-7754f76f8b-mxdpd\" (UID: \"67f7c8f7-535f-4a07-a330-adf7c41da701\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-mxdpd" Feb 02 09:18:54 crc kubenswrapper[4764]: I0202 09:18:54.443136 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j9ts8\" (UniqueName: \"kubernetes.io/projected/67f7c8f7-535f-4a07-a330-adf7c41da701-kube-api-access-j9ts8\") pod \"nmstate-console-plugin-7754f76f8b-mxdpd\" (UID: \"67f7c8f7-535f-4a07-a330-adf7c41da701\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-mxdpd" Feb 02 09:18:54 crc kubenswrapper[4764]: I0202 09:18:54.443439 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/67f7c8f7-535f-4a07-a330-adf7c41da701-plugin-serving-cert\") pod \"nmstate-console-plugin-7754f76f8b-mxdpd\" (UID: \"67f7c8f7-535f-4a07-a330-adf7c41da701\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-mxdpd" Feb 02 09:18:54 crc kubenswrapper[4764]: E0202 09:18:54.443535 4764 secret.go:188] Couldn't get secret openshift-nmstate/plugin-serving-cert: secret "plugin-serving-cert" not found Feb 02 09:18:54 crc kubenswrapper[4764]: E0202 09:18:54.443571 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/67f7c8f7-535f-4a07-a330-adf7c41da701-plugin-serving-cert podName:67f7c8f7-535f-4a07-a330-adf7c41da701 nodeName:}" failed. No retries permitted until 2026-02-02 09:18:54.943558812 +0000 UTC m=+697.877282890 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "plugin-serving-cert" (UniqueName: "kubernetes.io/secret/67f7c8f7-535f-4a07-a330-adf7c41da701-plugin-serving-cert") pod "nmstate-console-plugin-7754f76f8b-mxdpd" (UID: "67f7c8f7-535f-4a07-a330-adf7c41da701") : secret "plugin-serving-cert" not found Feb 02 09:18:54 crc kubenswrapper[4764]: I0202 09:18:54.444540 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/67f7c8f7-535f-4a07-a330-adf7c41da701-nginx-conf\") pod \"nmstate-console-plugin-7754f76f8b-mxdpd\" (UID: \"67f7c8f7-535f-4a07-a330-adf7c41da701\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-mxdpd" Feb 02 09:18:54 crc kubenswrapper[4764]: I0202 09:18:54.454862 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-6bc694f886-fp5xj"] Feb 02 09:18:54 crc kubenswrapper[4764]: I0202 09:18:54.472858 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j9ts8\" (UniqueName: \"kubernetes.io/projected/67f7c8f7-535f-4a07-a330-adf7c41da701-kube-api-access-j9ts8\") pod \"nmstate-console-plugin-7754f76f8b-mxdpd\" (UID: \"67f7c8f7-535f-4a07-a330-adf7c41da701\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-mxdpd" Feb 02 09:18:54 crc kubenswrapper[4764]: I0202 09:18:54.544221 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/e5ff98c7-5fc8-44d1-950e-24dd175739d6-oauth-serving-cert\") pod \"console-6bc694f886-fp5xj\" (UID: \"e5ff98c7-5fc8-44d1-950e-24dd175739d6\") " pod="openshift-console/console-6bc694f886-fp5xj" Feb 02 09:18:54 crc kubenswrapper[4764]: I0202 09:18:54.544267 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/e5ff98c7-5fc8-44d1-950e-24dd175739d6-console-config\") pod \"console-6bc694f886-fp5xj\" (UID: \"e5ff98c7-5fc8-44d1-950e-24dd175739d6\") " pod="openshift-console/console-6bc694f886-fp5xj" Feb 02 09:18:54 crc kubenswrapper[4764]: I0202 09:18:54.544286 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/e5ff98c7-5fc8-44d1-950e-24dd175739d6-console-oauth-config\") pod \"console-6bc694f886-fp5xj\" (UID: \"e5ff98c7-5fc8-44d1-950e-24dd175739d6\") " pod="openshift-console/console-6bc694f886-fp5xj" Feb 02 09:18:54 crc kubenswrapper[4764]: I0202 09:18:54.544309 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/e5ff98c7-5fc8-44d1-950e-24dd175739d6-console-serving-cert\") pod \"console-6bc694f886-fp5xj\" (UID: \"e5ff98c7-5fc8-44d1-950e-24dd175739d6\") " pod="openshift-console/console-6bc694f886-fp5xj" Feb 02 09:18:54 crc kubenswrapper[4764]: I0202 09:18:54.544346 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lcrp9\" (UniqueName: \"kubernetes.io/projected/e5ff98c7-5fc8-44d1-950e-24dd175739d6-kube-api-access-lcrp9\") pod \"console-6bc694f886-fp5xj\" (UID: \"e5ff98c7-5fc8-44d1-950e-24dd175739d6\") " pod="openshift-console/console-6bc694f886-fp5xj" Feb 02 09:18:54 crc kubenswrapper[4764]: I0202 09:18:54.544392 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e5ff98c7-5fc8-44d1-950e-24dd175739d6-trusted-ca-bundle\") pod \"console-6bc694f886-fp5xj\" (UID: \"e5ff98c7-5fc8-44d1-950e-24dd175739d6\") " pod="openshift-console/console-6bc694f886-fp5xj" Feb 02 09:18:54 crc kubenswrapper[4764]: I0202 09:18:54.544412 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e5ff98c7-5fc8-44d1-950e-24dd175739d6-service-ca\") pod \"console-6bc694f886-fp5xj\" (UID: \"e5ff98c7-5fc8-44d1-950e-24dd175739d6\") " pod="openshift-console/console-6bc694f886-fp5xj" Feb 02 09:18:54 crc kubenswrapper[4764]: I0202 09:18:54.646600 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e5ff98c7-5fc8-44d1-950e-24dd175739d6-trusted-ca-bundle\") pod \"console-6bc694f886-fp5xj\" (UID: \"e5ff98c7-5fc8-44d1-950e-24dd175739d6\") " pod="openshift-console/console-6bc694f886-fp5xj" Feb 02 09:18:54 crc kubenswrapper[4764]: I0202 09:18:54.646657 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e5ff98c7-5fc8-44d1-950e-24dd175739d6-service-ca\") pod \"console-6bc694f886-fp5xj\" (UID: \"e5ff98c7-5fc8-44d1-950e-24dd175739d6\") " pod="openshift-console/console-6bc694f886-fp5xj" Feb 02 09:18:54 crc kubenswrapper[4764]: I0202 09:18:54.646684 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/e5ff98c7-5fc8-44d1-950e-24dd175739d6-oauth-serving-cert\") pod \"console-6bc694f886-fp5xj\" (UID: \"e5ff98c7-5fc8-44d1-950e-24dd175739d6\") " pod="openshift-console/console-6bc694f886-fp5xj" Feb 02 09:18:54 crc kubenswrapper[4764]: I0202 09:18:54.646708 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/e5ff98c7-5fc8-44d1-950e-24dd175739d6-console-config\") pod \"console-6bc694f886-fp5xj\" (UID: \"e5ff98c7-5fc8-44d1-950e-24dd175739d6\") " pod="openshift-console/console-6bc694f886-fp5xj" Feb 02 09:18:54 crc kubenswrapper[4764]: I0202 09:18:54.646728 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/e5ff98c7-5fc8-44d1-950e-24dd175739d6-console-oauth-config\") pod \"console-6bc694f886-fp5xj\" (UID: \"e5ff98c7-5fc8-44d1-950e-24dd175739d6\") " pod="openshift-console/console-6bc694f886-fp5xj" Feb 02 09:18:54 crc kubenswrapper[4764]: I0202 09:18:54.646752 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/e5ff98c7-5fc8-44d1-950e-24dd175739d6-console-serving-cert\") pod \"console-6bc694f886-fp5xj\" (UID: \"e5ff98c7-5fc8-44d1-950e-24dd175739d6\") " pod="openshift-console/console-6bc694f886-fp5xj" Feb 02 09:18:54 crc kubenswrapper[4764]: I0202 09:18:54.646792 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lcrp9\" (UniqueName: \"kubernetes.io/projected/e5ff98c7-5fc8-44d1-950e-24dd175739d6-kube-api-access-lcrp9\") pod \"console-6bc694f886-fp5xj\" (UID: \"e5ff98c7-5fc8-44d1-950e-24dd175739d6\") " pod="openshift-console/console-6bc694f886-fp5xj" Feb 02 09:18:54 crc kubenswrapper[4764]: I0202 09:18:54.647817 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/e5ff98c7-5fc8-44d1-950e-24dd175739d6-oauth-serving-cert\") pod \"console-6bc694f886-fp5xj\" (UID: \"e5ff98c7-5fc8-44d1-950e-24dd175739d6\") " pod="openshift-console/console-6bc694f886-fp5xj" Feb 02 09:18:54 crc kubenswrapper[4764]: I0202 09:18:54.647865 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/e5ff98c7-5fc8-44d1-950e-24dd175739d6-console-config\") pod \"console-6bc694f886-fp5xj\" (UID: \"e5ff98c7-5fc8-44d1-950e-24dd175739d6\") " pod="openshift-console/console-6bc694f886-fp5xj" Feb 02 09:18:54 crc kubenswrapper[4764]: I0202 09:18:54.648557 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e5ff98c7-5fc8-44d1-950e-24dd175739d6-trusted-ca-bundle\") pod \"console-6bc694f886-fp5xj\" (UID: \"e5ff98c7-5fc8-44d1-950e-24dd175739d6\") " pod="openshift-console/console-6bc694f886-fp5xj" Feb 02 09:18:54 crc kubenswrapper[4764]: I0202 09:18:54.649073 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e5ff98c7-5fc8-44d1-950e-24dd175739d6-service-ca\") pod \"console-6bc694f886-fp5xj\" (UID: \"e5ff98c7-5fc8-44d1-950e-24dd175739d6\") " pod="openshift-console/console-6bc694f886-fp5xj" Feb 02 09:18:54 crc kubenswrapper[4764]: I0202 09:18:54.651715 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/e5ff98c7-5fc8-44d1-950e-24dd175739d6-console-oauth-config\") pod \"console-6bc694f886-fp5xj\" (UID: \"e5ff98c7-5fc8-44d1-950e-24dd175739d6\") " pod="openshift-console/console-6bc694f886-fp5xj" Feb 02 09:18:54 crc kubenswrapper[4764]: I0202 09:18:54.652388 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/e5ff98c7-5fc8-44d1-950e-24dd175739d6-console-serving-cert\") pod \"console-6bc694f886-fp5xj\" (UID: \"e5ff98c7-5fc8-44d1-950e-24dd175739d6\") " pod="openshift-console/console-6bc694f886-fp5xj" Feb 02 09:18:54 crc kubenswrapper[4764]: I0202 09:18:54.662248 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lcrp9\" (UniqueName: \"kubernetes.io/projected/e5ff98c7-5fc8-44d1-950e-24dd175739d6-kube-api-access-lcrp9\") pod \"console-6bc694f886-fp5xj\" (UID: \"e5ff98c7-5fc8-44d1-950e-24dd175739d6\") " pod="openshift-console/console-6bc694f886-fp5xj" Feb 02 09:18:54 crc kubenswrapper[4764]: I0202 09:18:54.688198 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-54757c584b-jz8q9"] Feb 02 09:18:54 crc kubenswrapper[4764]: W0202 09:18:54.696482 4764 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddd14d932_c6a2_4d7e_a1bb_bf7cf4c30f00.slice/crio-84fa2de41910c42837ac205f72e3a88d9a3eacae1f500eec2343b1c04682f24e WatchSource:0}: Error finding container 84fa2de41910c42837ac205f72e3a88d9a3eacae1f500eec2343b1c04682f24e: Status 404 returned error can't find the container with id 84fa2de41910c42837ac205f72e3a88d9a3eacae1f500eec2343b1c04682f24e Feb 02 09:18:54 crc kubenswrapper[4764]: I0202 09:18:54.747479 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/1a77ee37-2514-4a73-b2f1-565442966b16-tls-key-pair\") pod \"nmstate-webhook-8474b5b9d8-gz969\" (UID: \"1a77ee37-2514-4a73-b2f1-565442966b16\") " pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-gz969" Feb 02 09:18:54 crc kubenswrapper[4764]: I0202 09:18:54.747664 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-6bc694f886-fp5xj" Feb 02 09:18:54 crc kubenswrapper[4764]: I0202 09:18:54.751519 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/1a77ee37-2514-4a73-b2f1-565442966b16-tls-key-pair\") pod \"nmstate-webhook-8474b5b9d8-gz969\" (UID: \"1a77ee37-2514-4a73-b2f1-565442966b16\") " pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-gz969" Feb 02 09:18:54 crc kubenswrapper[4764]: I0202 09:18:54.950799 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/67f7c8f7-535f-4a07-a330-adf7c41da701-plugin-serving-cert\") pod \"nmstate-console-plugin-7754f76f8b-mxdpd\" (UID: \"67f7c8f7-535f-4a07-a330-adf7c41da701\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-mxdpd" Feb 02 09:18:54 crc kubenswrapper[4764]: I0202 09:18:54.957501 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/67f7c8f7-535f-4a07-a330-adf7c41da701-plugin-serving-cert\") pod \"nmstate-console-plugin-7754f76f8b-mxdpd\" (UID: \"67f7c8f7-535f-4a07-a330-adf7c41da701\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-mxdpd" Feb 02 09:18:54 crc kubenswrapper[4764]: I0202 09:18:54.960170 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-6bc694f886-fp5xj"] Feb 02 09:18:54 crc kubenswrapper[4764]: W0202 09:18:54.965881 4764 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode5ff98c7_5fc8_44d1_950e_24dd175739d6.slice/crio-4f8537c2ed2587d615990beee0f988476a42f0392b117b0dca9a66653b121452 WatchSource:0}: Error finding container 4f8537c2ed2587d615990beee0f988476a42f0392b117b0dca9a66653b121452: Status 404 returned error can't find the container with id 4f8537c2ed2587d615990beee0f988476a42f0392b117b0dca9a66653b121452 Feb 02 09:18:55 crc kubenswrapper[4764]: I0202 09:18:55.029866 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-gz969" Feb 02 09:18:55 crc kubenswrapper[4764]: I0202 09:18:55.039343 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-6bc694f886-fp5xj" event={"ID":"e5ff98c7-5fc8-44d1-950e-24dd175739d6","Type":"ContainerStarted","Data":"4f8537c2ed2587d615990beee0f988476a42f0392b117b0dca9a66653b121452"} Feb 02 09:18:55 crc kubenswrapper[4764]: I0202 09:18:55.042201 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-54757c584b-jz8q9" event={"ID":"dd14d932-c6a2-4d7e-a1bb-bf7cf4c30f00","Type":"ContainerStarted","Data":"84fa2de41910c42837ac205f72e3a88d9a3eacae1f500eec2343b1c04682f24e"} Feb 02 09:18:55 crc kubenswrapper[4764]: I0202 09:18:55.043186 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-gswb4" event={"ID":"7e5e0b71-cd10-4ccf-8100-f15a063cceb7","Type":"ContainerStarted","Data":"f2bd0bea4dac75d4651a00f3bc0e12d8f9bb4418b9c75e6bca71aa8943f890ab"} Feb 02 09:18:55 crc kubenswrapper[4764]: I0202 09:18:55.214244 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-mxdpd" Feb 02 09:18:55 crc kubenswrapper[4764]: I0202 09:18:55.238244 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-8474b5b9d8-gz969"] Feb 02 09:18:55 crc kubenswrapper[4764]: I0202 09:18:55.425316 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7754f76f8b-mxdpd"] Feb 02 09:18:56 crc kubenswrapper[4764]: I0202 09:18:56.049981 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-6bc694f886-fp5xj" event={"ID":"e5ff98c7-5fc8-44d1-950e-24dd175739d6","Type":"ContainerStarted","Data":"a5209c4a9f3e0b114e0532f3cdc7539781b29ef5926318498f105ae1cf711c44"} Feb 02 09:18:56 crc kubenswrapper[4764]: I0202 09:18:56.051123 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-mxdpd" event={"ID":"67f7c8f7-535f-4a07-a330-adf7c41da701","Type":"ContainerStarted","Data":"dd9eeb8b1fc136ca8ceb226cdebcca8694e4206c414171877a0f08e6ab12c858"} Feb 02 09:18:56 crc kubenswrapper[4764]: I0202 09:18:56.052388 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-gz969" event={"ID":"1a77ee37-2514-4a73-b2f1-565442966b16","Type":"ContainerStarted","Data":"6cec4c23daa1144f6acf727bc11fa98c949eab48325c74c25a9650edf77cc72b"} Feb 02 09:18:56 crc kubenswrapper[4764]: I0202 09:18:56.068807 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-6bc694f886-fp5xj" podStartSLOduration=2.068788603 podStartE2EDuration="2.068788603s" podCreationTimestamp="2026-02-02 09:18:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:18:56.06832055 +0000 UTC m=+699.002044638" watchObservedRunningTime="2026-02-02 09:18:56.068788603 +0000 UTC m=+699.002512691" Feb 02 09:18:58 crc kubenswrapper[4764]: I0202 09:18:58.067346 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-54757c584b-jz8q9" event={"ID":"dd14d932-c6a2-4d7e-a1bb-bf7cf4c30f00","Type":"ContainerStarted","Data":"bbd3aaee39f89616819a7d65a96a529014a56b26a2004649fe5302f11db2cf32"} Feb 02 09:18:58 crc kubenswrapper[4764]: I0202 09:18:58.069368 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-gswb4" event={"ID":"7e5e0b71-cd10-4ccf-8100-f15a063cceb7","Type":"ContainerStarted","Data":"ae98d25edc16297d2ffdc3f847495e1c5c416c4258c6e7f7f983ac5be1f89a75"} Feb 02 09:18:58 crc kubenswrapper[4764]: I0202 09:18:58.069528 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-gswb4" Feb 02 09:18:58 crc kubenswrapper[4764]: I0202 09:18:58.071903 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-gz969" event={"ID":"1a77ee37-2514-4a73-b2f1-565442966b16","Type":"ContainerStarted","Data":"ed0da4e3abae66f2318c30230bb3706488e029b5ddd8b175f0220f4318761294"} Feb 02 09:18:58 crc kubenswrapper[4764]: I0202 09:18:58.072355 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-gz969" Feb 02 09:18:58 crc kubenswrapper[4764]: I0202 09:18:58.092379 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-gswb4" podStartSLOduration=1.517445898 podStartE2EDuration="4.092358407s" podCreationTimestamp="2026-02-02 09:18:54 +0000 UTC" firstStartedPulling="2026-02-02 09:18:54.54451838 +0000 UTC m=+697.478242458" lastFinishedPulling="2026-02-02 09:18:57.119430849 +0000 UTC m=+700.053154967" observedRunningTime="2026-02-02 09:18:58.089098328 +0000 UTC m=+701.022822436" watchObservedRunningTime="2026-02-02 09:18:58.092358407 +0000 UTC m=+701.026082505" Feb 02 09:18:58 crc kubenswrapper[4764]: I0202 09:18:58.120695 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-gz969" podStartSLOduration=2.177261724 podStartE2EDuration="4.120675355s" podCreationTimestamp="2026-02-02 09:18:54 +0000 UTC" firstStartedPulling="2026-02-02 09:18:55.252623406 +0000 UTC m=+698.186347494" lastFinishedPulling="2026-02-02 09:18:57.196037037 +0000 UTC m=+700.129761125" observedRunningTime="2026-02-02 09:18:58.119503733 +0000 UTC m=+701.053227821" watchObservedRunningTime="2026-02-02 09:18:58.120675355 +0000 UTC m=+701.054399453" Feb 02 09:19:00 crc kubenswrapper[4764]: I0202 09:19:00.083893 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-54757c584b-jz8q9" event={"ID":"dd14d932-c6a2-4d7e-a1bb-bf7cf4c30f00","Type":"ContainerStarted","Data":"f76669b5fb935256755dccb8b53af1dd961e8a0c2c39afe9df57013d3f62d001"} Feb 02 09:19:00 crc kubenswrapper[4764]: I0202 09:19:00.085828 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-mxdpd" event={"ID":"67f7c8f7-535f-4a07-a330-adf7c41da701","Type":"ContainerStarted","Data":"13396c1bdaaa710597dc1c4c9557558f1bf56e991a0927fe69eea840e2e81fb0"} Feb 02 09:19:00 crc kubenswrapper[4764]: I0202 09:19:00.098858 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-54757c584b-jz8q9" podStartSLOduration=0.965369535 podStartE2EDuration="6.098840217s" podCreationTimestamp="2026-02-02 09:18:54 +0000 UTC" firstStartedPulling="2026-02-02 09:18:54.698317272 +0000 UTC m=+697.632041360" lastFinishedPulling="2026-02-02 09:18:59.831787954 +0000 UTC m=+702.765512042" observedRunningTime="2026-02-02 09:19:00.097345227 +0000 UTC m=+703.031069315" watchObservedRunningTime="2026-02-02 09:19:00.098840217 +0000 UTC m=+703.032564305" Feb 02 09:19:00 crc kubenswrapper[4764]: I0202 09:19:00.116847 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-mxdpd" podStartSLOduration=1.7125284299999999 podStartE2EDuration="6.116832245s" podCreationTimestamp="2026-02-02 09:18:54 +0000 UTC" firstStartedPulling="2026-02-02 09:18:55.431468807 +0000 UTC m=+698.365192895" lastFinishedPulling="2026-02-02 09:18:59.835772622 +0000 UTC m=+702.769496710" observedRunningTime="2026-02-02 09:19:00.114201234 +0000 UTC m=+703.047925322" watchObservedRunningTime="2026-02-02 09:19:00.116832245 +0000 UTC m=+703.050556333" Feb 02 09:19:04 crc kubenswrapper[4764]: I0202 09:19:04.471690 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-gswb4" Feb 02 09:19:04 crc kubenswrapper[4764]: I0202 09:19:04.748136 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-6bc694f886-fp5xj" Feb 02 09:19:04 crc kubenswrapper[4764]: I0202 09:19:04.748202 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-6bc694f886-fp5xj" Feb 02 09:19:04 crc kubenswrapper[4764]: I0202 09:19:04.757863 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-6bc694f886-fp5xj" Feb 02 09:19:05 crc kubenswrapper[4764]: I0202 09:19:05.127391 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-6bc694f886-fp5xj" Feb 02 09:19:05 crc kubenswrapper[4764]: I0202 09:19:05.242182 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-cnkss"] Feb 02 09:19:13 crc kubenswrapper[4764]: I0202 09:19:13.523822 4764 patch_prober.go:28] interesting pod/machine-config-daemon-4ndm4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 09:19:13 crc kubenswrapper[4764]: I0202 09:19:13.524504 4764 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 09:19:15 crc kubenswrapper[4764]: I0202 09:19:15.038656 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-gz969" Feb 02 09:19:30 crc kubenswrapper[4764]: I0202 09:19:30.297992 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcv6r4w"] Feb 02 09:19:30 crc kubenswrapper[4764]: I0202 09:19:30.299679 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcv6r4w" Feb 02 09:19:30 crc kubenswrapper[4764]: I0202 09:19:30.301328 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Feb 02 09:19:30 crc kubenswrapper[4764]: I0202 09:19:30.326057 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcv6r4w"] Feb 02 09:19:30 crc kubenswrapper[4764]: I0202 09:19:30.326170 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-cnkss" podUID="d874a84d-48e3-4cb2-ac74-4f0db52e9d75" containerName="console" containerID="cri-o://7f0399b6927158c0f8102427612492ca9041d1130c1c9f64b4196a6117b4f9c2" gracePeriod=15 Feb 02 09:19:30 crc kubenswrapper[4764]: I0202 09:19:30.414014 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sv2q8\" (UniqueName: \"kubernetes.io/projected/6021fc8f-ba41-4247-9b13-991cee9cb2a6-kube-api-access-sv2q8\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcv6r4w\" (UID: \"6021fc8f-ba41-4247-9b13-991cee9cb2a6\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcv6r4w" Feb 02 09:19:30 crc kubenswrapper[4764]: I0202 09:19:30.414089 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6021fc8f-ba41-4247-9b13-991cee9cb2a6-util\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcv6r4w\" (UID: \"6021fc8f-ba41-4247-9b13-991cee9cb2a6\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcv6r4w" Feb 02 09:19:30 crc kubenswrapper[4764]: I0202 09:19:30.414165 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6021fc8f-ba41-4247-9b13-991cee9cb2a6-bundle\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcv6r4w\" (UID: \"6021fc8f-ba41-4247-9b13-991cee9cb2a6\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcv6r4w" Feb 02 09:19:30 crc kubenswrapper[4764]: I0202 09:19:30.462269 4764 patch_prober.go:28] interesting pod/console-f9d7485db-cnkss container/console namespace/openshift-console: Readiness probe status=failure output="Get \"https://10.217.0.10:8443/health\": dial tcp 10.217.0.10:8443: connect: connection refused" start-of-body= Feb 02 09:19:30 crc kubenswrapper[4764]: I0202 09:19:30.462316 4764 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/console-f9d7485db-cnkss" podUID="d874a84d-48e3-4cb2-ac74-4f0db52e9d75" containerName="console" probeResult="failure" output="Get \"https://10.217.0.10:8443/health\": dial tcp 10.217.0.10:8443: connect: connection refused" Feb 02 09:19:30 crc kubenswrapper[4764]: I0202 09:19:30.514865 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sv2q8\" (UniqueName: \"kubernetes.io/projected/6021fc8f-ba41-4247-9b13-991cee9cb2a6-kube-api-access-sv2q8\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcv6r4w\" (UID: \"6021fc8f-ba41-4247-9b13-991cee9cb2a6\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcv6r4w" Feb 02 09:19:30 crc kubenswrapper[4764]: I0202 09:19:30.514911 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6021fc8f-ba41-4247-9b13-991cee9cb2a6-util\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcv6r4w\" (UID: \"6021fc8f-ba41-4247-9b13-991cee9cb2a6\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcv6r4w" Feb 02 09:19:30 crc kubenswrapper[4764]: I0202 09:19:30.514984 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6021fc8f-ba41-4247-9b13-991cee9cb2a6-bundle\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcv6r4w\" (UID: \"6021fc8f-ba41-4247-9b13-991cee9cb2a6\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcv6r4w" Feb 02 09:19:30 crc kubenswrapper[4764]: I0202 09:19:30.515559 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6021fc8f-ba41-4247-9b13-991cee9cb2a6-bundle\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcv6r4w\" (UID: \"6021fc8f-ba41-4247-9b13-991cee9cb2a6\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcv6r4w" Feb 02 09:19:30 crc kubenswrapper[4764]: I0202 09:19:30.515723 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6021fc8f-ba41-4247-9b13-991cee9cb2a6-util\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcv6r4w\" (UID: \"6021fc8f-ba41-4247-9b13-991cee9cb2a6\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcv6r4w" Feb 02 09:19:30 crc kubenswrapper[4764]: I0202 09:19:30.544833 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sv2q8\" (UniqueName: \"kubernetes.io/projected/6021fc8f-ba41-4247-9b13-991cee9cb2a6-kube-api-access-sv2q8\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcv6r4w\" (UID: \"6021fc8f-ba41-4247-9b13-991cee9cb2a6\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcv6r4w" Feb 02 09:19:30 crc kubenswrapper[4764]: I0202 09:19:30.622609 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcv6r4w" Feb 02 09:19:30 crc kubenswrapper[4764]: I0202 09:19:30.665090 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-cnkss_d874a84d-48e3-4cb2-ac74-4f0db52e9d75/console/0.log" Feb 02 09:19:30 crc kubenswrapper[4764]: I0202 09:19:30.665147 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-cnkss" Feb 02 09:19:30 crc kubenswrapper[4764]: I0202 09:19:30.717146 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lpkq9\" (UniqueName: \"kubernetes.io/projected/d874a84d-48e3-4cb2-ac74-4f0db52e9d75-kube-api-access-lpkq9\") pod \"d874a84d-48e3-4cb2-ac74-4f0db52e9d75\" (UID: \"d874a84d-48e3-4cb2-ac74-4f0db52e9d75\") " Feb 02 09:19:30 crc kubenswrapper[4764]: I0202 09:19:30.717191 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/d874a84d-48e3-4cb2-ac74-4f0db52e9d75-service-ca\") pod \"d874a84d-48e3-4cb2-ac74-4f0db52e9d75\" (UID: \"d874a84d-48e3-4cb2-ac74-4f0db52e9d75\") " Feb 02 09:19:30 crc kubenswrapper[4764]: I0202 09:19:30.717226 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/d874a84d-48e3-4cb2-ac74-4f0db52e9d75-oauth-serving-cert\") pod \"d874a84d-48e3-4cb2-ac74-4f0db52e9d75\" (UID: \"d874a84d-48e3-4cb2-ac74-4f0db52e9d75\") " Feb 02 09:19:30 crc kubenswrapper[4764]: I0202 09:19:30.717258 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/d874a84d-48e3-4cb2-ac74-4f0db52e9d75-console-serving-cert\") pod \"d874a84d-48e3-4cb2-ac74-4f0db52e9d75\" (UID: \"d874a84d-48e3-4cb2-ac74-4f0db52e9d75\") " Feb 02 09:19:30 crc kubenswrapper[4764]: I0202 09:19:30.717283 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/d874a84d-48e3-4cb2-ac74-4f0db52e9d75-console-config\") pod \"d874a84d-48e3-4cb2-ac74-4f0db52e9d75\" (UID: \"d874a84d-48e3-4cb2-ac74-4f0db52e9d75\") " Feb 02 09:19:30 crc kubenswrapper[4764]: I0202 09:19:30.717314 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/d874a84d-48e3-4cb2-ac74-4f0db52e9d75-console-oauth-config\") pod \"d874a84d-48e3-4cb2-ac74-4f0db52e9d75\" (UID: \"d874a84d-48e3-4cb2-ac74-4f0db52e9d75\") " Feb 02 09:19:30 crc kubenswrapper[4764]: I0202 09:19:30.717349 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d874a84d-48e3-4cb2-ac74-4f0db52e9d75-trusted-ca-bundle\") pod \"d874a84d-48e3-4cb2-ac74-4f0db52e9d75\" (UID: \"d874a84d-48e3-4cb2-ac74-4f0db52e9d75\") " Feb 02 09:19:30 crc kubenswrapper[4764]: I0202 09:19:30.718157 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d874a84d-48e3-4cb2-ac74-4f0db52e9d75-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "d874a84d-48e3-4cb2-ac74-4f0db52e9d75" (UID: "d874a84d-48e3-4cb2-ac74-4f0db52e9d75"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:19:30 crc kubenswrapper[4764]: I0202 09:19:30.718464 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d874a84d-48e3-4cb2-ac74-4f0db52e9d75-service-ca" (OuterVolumeSpecName: "service-ca") pod "d874a84d-48e3-4cb2-ac74-4f0db52e9d75" (UID: "d874a84d-48e3-4cb2-ac74-4f0db52e9d75"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:19:30 crc kubenswrapper[4764]: I0202 09:19:30.718567 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d874a84d-48e3-4cb2-ac74-4f0db52e9d75-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "d874a84d-48e3-4cb2-ac74-4f0db52e9d75" (UID: "d874a84d-48e3-4cb2-ac74-4f0db52e9d75"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:19:30 crc kubenswrapper[4764]: I0202 09:19:30.719184 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d874a84d-48e3-4cb2-ac74-4f0db52e9d75-console-config" (OuterVolumeSpecName: "console-config") pod "d874a84d-48e3-4cb2-ac74-4f0db52e9d75" (UID: "d874a84d-48e3-4cb2-ac74-4f0db52e9d75"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:19:30 crc kubenswrapper[4764]: I0202 09:19:30.723303 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d874a84d-48e3-4cb2-ac74-4f0db52e9d75-kube-api-access-lpkq9" (OuterVolumeSpecName: "kube-api-access-lpkq9") pod "d874a84d-48e3-4cb2-ac74-4f0db52e9d75" (UID: "d874a84d-48e3-4cb2-ac74-4f0db52e9d75"). InnerVolumeSpecName "kube-api-access-lpkq9". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:19:30 crc kubenswrapper[4764]: I0202 09:19:30.723893 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d874a84d-48e3-4cb2-ac74-4f0db52e9d75-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "d874a84d-48e3-4cb2-ac74-4f0db52e9d75" (UID: "d874a84d-48e3-4cb2-ac74-4f0db52e9d75"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:19:30 crc kubenswrapper[4764]: I0202 09:19:30.724114 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d874a84d-48e3-4cb2-ac74-4f0db52e9d75-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "d874a84d-48e3-4cb2-ac74-4f0db52e9d75" (UID: "d874a84d-48e3-4cb2-ac74-4f0db52e9d75"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:19:30 crc kubenswrapper[4764]: I0202 09:19:30.812421 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcv6r4w"] Feb 02 09:19:30 crc kubenswrapper[4764]: I0202 09:19:30.818786 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lpkq9\" (UniqueName: \"kubernetes.io/projected/d874a84d-48e3-4cb2-ac74-4f0db52e9d75-kube-api-access-lpkq9\") on node \"crc\" DevicePath \"\"" Feb 02 09:19:30 crc kubenswrapper[4764]: I0202 09:19:30.818811 4764 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/d874a84d-48e3-4cb2-ac74-4f0db52e9d75-service-ca\") on node \"crc\" DevicePath \"\"" Feb 02 09:19:30 crc kubenswrapper[4764]: I0202 09:19:30.818822 4764 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/d874a84d-48e3-4cb2-ac74-4f0db52e9d75-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 02 09:19:30 crc kubenswrapper[4764]: I0202 09:19:30.818831 4764 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/d874a84d-48e3-4cb2-ac74-4f0db52e9d75-console-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 02 09:19:30 crc kubenswrapper[4764]: I0202 09:19:30.818839 4764 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/d874a84d-48e3-4cb2-ac74-4f0db52e9d75-console-config\") on node \"crc\" DevicePath \"\"" Feb 02 09:19:30 crc kubenswrapper[4764]: I0202 09:19:30.818847 4764 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/d874a84d-48e3-4cb2-ac74-4f0db52e9d75-console-oauth-config\") on node \"crc\" DevicePath \"\"" Feb 02 09:19:30 crc kubenswrapper[4764]: I0202 09:19:30.818855 4764 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d874a84d-48e3-4cb2-ac74-4f0db52e9d75-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 09:19:31 crc kubenswrapper[4764]: I0202 09:19:31.311121 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-cnkss_d874a84d-48e3-4cb2-ac74-4f0db52e9d75/console/0.log" Feb 02 09:19:31 crc kubenswrapper[4764]: I0202 09:19:31.311489 4764 generic.go:334] "Generic (PLEG): container finished" podID="d874a84d-48e3-4cb2-ac74-4f0db52e9d75" containerID="7f0399b6927158c0f8102427612492ca9041d1130c1c9f64b4196a6117b4f9c2" exitCode=2 Feb 02 09:19:31 crc kubenswrapper[4764]: I0202 09:19:31.311560 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-cnkss" event={"ID":"d874a84d-48e3-4cb2-ac74-4f0db52e9d75","Type":"ContainerDied","Data":"7f0399b6927158c0f8102427612492ca9041d1130c1c9f64b4196a6117b4f9c2"} Feb 02 09:19:31 crc kubenswrapper[4764]: I0202 09:19:31.311592 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-cnkss" event={"ID":"d874a84d-48e3-4cb2-ac74-4f0db52e9d75","Type":"ContainerDied","Data":"b1c6c13f4eacde51e30158c53fd5ee984a9d48d0b801c239460e97316334fa43"} Feb 02 09:19:31 crc kubenswrapper[4764]: I0202 09:19:31.311614 4764 scope.go:117] "RemoveContainer" containerID="7f0399b6927158c0f8102427612492ca9041d1130c1c9f64b4196a6117b4f9c2" Feb 02 09:19:31 crc kubenswrapper[4764]: I0202 09:19:31.311621 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-cnkss" Feb 02 09:19:31 crc kubenswrapper[4764]: I0202 09:19:31.334619 4764 generic.go:334] "Generic (PLEG): container finished" podID="6021fc8f-ba41-4247-9b13-991cee9cb2a6" containerID="5cf3d25f25c25f2f66f692dc94a20d612dae14176b96317a8a1fb3013202efbb" exitCode=0 Feb 02 09:19:31 crc kubenswrapper[4764]: I0202 09:19:31.334669 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcv6r4w" event={"ID":"6021fc8f-ba41-4247-9b13-991cee9cb2a6","Type":"ContainerDied","Data":"5cf3d25f25c25f2f66f692dc94a20d612dae14176b96317a8a1fb3013202efbb"} Feb 02 09:19:31 crc kubenswrapper[4764]: I0202 09:19:31.334703 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcv6r4w" event={"ID":"6021fc8f-ba41-4247-9b13-991cee9cb2a6","Type":"ContainerStarted","Data":"afce74ecfd9cda6239f542c12d23b9b1363a3e99123aa8f51a288c4228e617fe"} Feb 02 09:19:31 crc kubenswrapper[4764]: I0202 09:19:31.364018 4764 scope.go:117] "RemoveContainer" containerID="7f0399b6927158c0f8102427612492ca9041d1130c1c9f64b4196a6117b4f9c2" Feb 02 09:19:31 crc kubenswrapper[4764]: E0202 09:19:31.369176 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7f0399b6927158c0f8102427612492ca9041d1130c1c9f64b4196a6117b4f9c2\": container with ID starting with 7f0399b6927158c0f8102427612492ca9041d1130c1c9f64b4196a6117b4f9c2 not found: ID does not exist" containerID="7f0399b6927158c0f8102427612492ca9041d1130c1c9f64b4196a6117b4f9c2" Feb 02 09:19:31 crc kubenswrapper[4764]: I0202 09:19:31.369219 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7f0399b6927158c0f8102427612492ca9041d1130c1c9f64b4196a6117b4f9c2"} err="failed to get container status \"7f0399b6927158c0f8102427612492ca9041d1130c1c9f64b4196a6117b4f9c2\": rpc error: code = NotFound desc = could not find container \"7f0399b6927158c0f8102427612492ca9041d1130c1c9f64b4196a6117b4f9c2\": container with ID starting with 7f0399b6927158c0f8102427612492ca9041d1130c1c9f64b4196a6117b4f9c2 not found: ID does not exist" Feb 02 09:19:31 crc kubenswrapper[4764]: I0202 09:19:31.403988 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-cnkss"] Feb 02 09:19:31 crc kubenswrapper[4764]: I0202 09:19:31.411854 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-cnkss"] Feb 02 09:19:31 crc kubenswrapper[4764]: I0202 09:19:31.836040 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d874a84d-48e3-4cb2-ac74-4f0db52e9d75" path="/var/lib/kubelet/pods/d874a84d-48e3-4cb2-ac74-4f0db52e9d75/volumes" Feb 02 09:19:33 crc kubenswrapper[4764]: I0202 09:19:33.353662 4764 generic.go:334] "Generic (PLEG): container finished" podID="6021fc8f-ba41-4247-9b13-991cee9cb2a6" containerID="392744ae7e8e3e7abb6ea387a7e430f7a59a515b9b7204285cc3ba5233624127" exitCode=0 Feb 02 09:19:33 crc kubenswrapper[4764]: I0202 09:19:33.353796 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcv6r4w" event={"ID":"6021fc8f-ba41-4247-9b13-991cee9cb2a6","Type":"ContainerDied","Data":"392744ae7e8e3e7abb6ea387a7e430f7a59a515b9b7204285cc3ba5233624127"} Feb 02 09:19:34 crc kubenswrapper[4764]: I0202 09:19:34.363990 4764 generic.go:334] "Generic (PLEG): container finished" podID="6021fc8f-ba41-4247-9b13-991cee9cb2a6" containerID="abb84bbd0aea45afc47f4b4c01aa3cce43e2e6e92f3f655d337613158042d0b8" exitCode=0 Feb 02 09:19:34 crc kubenswrapper[4764]: I0202 09:19:34.364081 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcv6r4w" event={"ID":"6021fc8f-ba41-4247-9b13-991cee9cb2a6","Type":"ContainerDied","Data":"abb84bbd0aea45afc47f4b4c01aa3cce43e2e6e92f3f655d337613158042d0b8"} Feb 02 09:19:35 crc kubenswrapper[4764]: I0202 09:19:35.642761 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcv6r4w" Feb 02 09:19:35 crc kubenswrapper[4764]: I0202 09:19:35.797261 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6021fc8f-ba41-4247-9b13-991cee9cb2a6-util\") pod \"6021fc8f-ba41-4247-9b13-991cee9cb2a6\" (UID: \"6021fc8f-ba41-4247-9b13-991cee9cb2a6\") " Feb 02 09:19:35 crc kubenswrapper[4764]: I0202 09:19:35.797333 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sv2q8\" (UniqueName: \"kubernetes.io/projected/6021fc8f-ba41-4247-9b13-991cee9cb2a6-kube-api-access-sv2q8\") pod \"6021fc8f-ba41-4247-9b13-991cee9cb2a6\" (UID: \"6021fc8f-ba41-4247-9b13-991cee9cb2a6\") " Feb 02 09:19:35 crc kubenswrapper[4764]: I0202 09:19:35.797458 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6021fc8f-ba41-4247-9b13-991cee9cb2a6-bundle\") pod \"6021fc8f-ba41-4247-9b13-991cee9cb2a6\" (UID: \"6021fc8f-ba41-4247-9b13-991cee9cb2a6\") " Feb 02 09:19:35 crc kubenswrapper[4764]: I0202 09:19:35.799511 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6021fc8f-ba41-4247-9b13-991cee9cb2a6-bundle" (OuterVolumeSpecName: "bundle") pod "6021fc8f-ba41-4247-9b13-991cee9cb2a6" (UID: "6021fc8f-ba41-4247-9b13-991cee9cb2a6"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 09:19:35 crc kubenswrapper[4764]: I0202 09:19:35.802893 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6021fc8f-ba41-4247-9b13-991cee9cb2a6-kube-api-access-sv2q8" (OuterVolumeSpecName: "kube-api-access-sv2q8") pod "6021fc8f-ba41-4247-9b13-991cee9cb2a6" (UID: "6021fc8f-ba41-4247-9b13-991cee9cb2a6"). InnerVolumeSpecName "kube-api-access-sv2q8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:19:35 crc kubenswrapper[4764]: I0202 09:19:35.816768 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6021fc8f-ba41-4247-9b13-991cee9cb2a6-util" (OuterVolumeSpecName: "util") pod "6021fc8f-ba41-4247-9b13-991cee9cb2a6" (UID: "6021fc8f-ba41-4247-9b13-991cee9cb2a6"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 09:19:35 crc kubenswrapper[4764]: I0202 09:19:35.899473 4764 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6021fc8f-ba41-4247-9b13-991cee9cb2a6-util\") on node \"crc\" DevicePath \"\"" Feb 02 09:19:35 crc kubenswrapper[4764]: I0202 09:19:35.899515 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sv2q8\" (UniqueName: \"kubernetes.io/projected/6021fc8f-ba41-4247-9b13-991cee9cb2a6-kube-api-access-sv2q8\") on node \"crc\" DevicePath \"\"" Feb 02 09:19:35 crc kubenswrapper[4764]: I0202 09:19:35.899528 4764 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6021fc8f-ba41-4247-9b13-991cee9cb2a6-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 09:19:36 crc kubenswrapper[4764]: I0202 09:19:36.381627 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcv6r4w" event={"ID":"6021fc8f-ba41-4247-9b13-991cee9cb2a6","Type":"ContainerDied","Data":"afce74ecfd9cda6239f542c12d23b9b1363a3e99123aa8f51a288c4228e617fe"} Feb 02 09:19:36 crc kubenswrapper[4764]: I0202 09:19:36.381682 4764 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="afce74ecfd9cda6239f542c12d23b9b1363a3e99123aa8f51a288c4228e617fe" Feb 02 09:19:36 crc kubenswrapper[4764]: I0202 09:19:36.381681 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcv6r4w" Feb 02 09:19:43 crc kubenswrapper[4764]: I0202 09:19:43.523002 4764 patch_prober.go:28] interesting pod/machine-config-daemon-4ndm4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 09:19:43 crc kubenswrapper[4764]: I0202 09:19:43.523617 4764 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 09:19:46 crc kubenswrapper[4764]: I0202 09:19:46.912865 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-7848c48b6-2kvbb"] Feb 02 09:19:46 crc kubenswrapper[4764]: E0202 09:19:46.913313 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d874a84d-48e3-4cb2-ac74-4f0db52e9d75" containerName="console" Feb 02 09:19:46 crc kubenswrapper[4764]: I0202 09:19:46.913326 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="d874a84d-48e3-4cb2-ac74-4f0db52e9d75" containerName="console" Feb 02 09:19:46 crc kubenswrapper[4764]: E0202 09:19:46.913347 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6021fc8f-ba41-4247-9b13-991cee9cb2a6" containerName="pull" Feb 02 09:19:46 crc kubenswrapper[4764]: I0202 09:19:46.913353 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="6021fc8f-ba41-4247-9b13-991cee9cb2a6" containerName="pull" Feb 02 09:19:46 crc kubenswrapper[4764]: E0202 09:19:46.913360 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6021fc8f-ba41-4247-9b13-991cee9cb2a6" containerName="util" Feb 02 09:19:46 crc kubenswrapper[4764]: I0202 09:19:46.913365 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="6021fc8f-ba41-4247-9b13-991cee9cb2a6" containerName="util" Feb 02 09:19:46 crc kubenswrapper[4764]: E0202 09:19:46.913372 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6021fc8f-ba41-4247-9b13-991cee9cb2a6" containerName="extract" Feb 02 09:19:46 crc kubenswrapper[4764]: I0202 09:19:46.913377 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="6021fc8f-ba41-4247-9b13-991cee9cb2a6" containerName="extract" Feb 02 09:19:46 crc kubenswrapper[4764]: I0202 09:19:46.913459 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="d874a84d-48e3-4cb2-ac74-4f0db52e9d75" containerName="console" Feb 02 09:19:46 crc kubenswrapper[4764]: I0202 09:19:46.913478 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="6021fc8f-ba41-4247-9b13-991cee9cb2a6" containerName="extract" Feb 02 09:19:46 crc kubenswrapper[4764]: I0202 09:19:46.913793 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-7848c48b6-2kvbb" Feb 02 09:19:46 crc kubenswrapper[4764]: I0202 09:19:46.922904 4764 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Feb 02 09:19:46 crc kubenswrapper[4764]: I0202 09:19:46.923051 4764 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Feb 02 09:19:46 crc kubenswrapper[4764]: I0202 09:19:46.924638 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Feb 02 09:19:46 crc kubenswrapper[4764]: I0202 09:19:46.926537 4764 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-7h252" Feb 02 09:19:46 crc kubenswrapper[4764]: I0202 09:19:46.927884 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Feb 02 09:19:46 crc kubenswrapper[4764]: I0202 09:19:46.939358 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-7848c48b6-2kvbb"] Feb 02 09:19:47 crc kubenswrapper[4764]: I0202 09:19:47.031515 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f479cadf-cdb2-423f-b9c5-204bef98bf8a-apiservice-cert\") pod \"metallb-operator-controller-manager-7848c48b6-2kvbb\" (UID: \"f479cadf-cdb2-423f-b9c5-204bef98bf8a\") " pod="metallb-system/metallb-operator-controller-manager-7848c48b6-2kvbb" Feb 02 09:19:47 crc kubenswrapper[4764]: I0202 09:19:47.031565 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8tgvs\" (UniqueName: \"kubernetes.io/projected/f479cadf-cdb2-423f-b9c5-204bef98bf8a-kube-api-access-8tgvs\") pod \"metallb-operator-controller-manager-7848c48b6-2kvbb\" (UID: \"f479cadf-cdb2-423f-b9c5-204bef98bf8a\") " pod="metallb-system/metallb-operator-controller-manager-7848c48b6-2kvbb" Feb 02 09:19:47 crc kubenswrapper[4764]: I0202 09:19:47.031592 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f479cadf-cdb2-423f-b9c5-204bef98bf8a-webhook-cert\") pod \"metallb-operator-controller-manager-7848c48b6-2kvbb\" (UID: \"f479cadf-cdb2-423f-b9c5-204bef98bf8a\") " pod="metallb-system/metallb-operator-controller-manager-7848c48b6-2kvbb" Feb 02 09:19:47 crc kubenswrapper[4764]: I0202 09:19:47.132839 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f479cadf-cdb2-423f-b9c5-204bef98bf8a-apiservice-cert\") pod \"metallb-operator-controller-manager-7848c48b6-2kvbb\" (UID: \"f479cadf-cdb2-423f-b9c5-204bef98bf8a\") " pod="metallb-system/metallb-operator-controller-manager-7848c48b6-2kvbb" Feb 02 09:19:47 crc kubenswrapper[4764]: I0202 09:19:47.132900 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8tgvs\" (UniqueName: \"kubernetes.io/projected/f479cadf-cdb2-423f-b9c5-204bef98bf8a-kube-api-access-8tgvs\") pod \"metallb-operator-controller-manager-7848c48b6-2kvbb\" (UID: \"f479cadf-cdb2-423f-b9c5-204bef98bf8a\") " pod="metallb-system/metallb-operator-controller-manager-7848c48b6-2kvbb" Feb 02 09:19:47 crc kubenswrapper[4764]: I0202 09:19:47.132958 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f479cadf-cdb2-423f-b9c5-204bef98bf8a-webhook-cert\") pod \"metallb-operator-controller-manager-7848c48b6-2kvbb\" (UID: \"f479cadf-cdb2-423f-b9c5-204bef98bf8a\") " pod="metallb-system/metallb-operator-controller-manager-7848c48b6-2kvbb" Feb 02 09:19:47 crc kubenswrapper[4764]: I0202 09:19:47.139470 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f479cadf-cdb2-423f-b9c5-204bef98bf8a-webhook-cert\") pod \"metallb-operator-controller-manager-7848c48b6-2kvbb\" (UID: \"f479cadf-cdb2-423f-b9c5-204bef98bf8a\") " pod="metallb-system/metallb-operator-controller-manager-7848c48b6-2kvbb" Feb 02 09:19:47 crc kubenswrapper[4764]: I0202 09:19:47.139713 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f479cadf-cdb2-423f-b9c5-204bef98bf8a-apiservice-cert\") pod \"metallb-operator-controller-manager-7848c48b6-2kvbb\" (UID: \"f479cadf-cdb2-423f-b9c5-204bef98bf8a\") " pod="metallb-system/metallb-operator-controller-manager-7848c48b6-2kvbb" Feb 02 09:19:47 crc kubenswrapper[4764]: I0202 09:19:47.151810 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8tgvs\" (UniqueName: \"kubernetes.io/projected/f479cadf-cdb2-423f-b9c5-204bef98bf8a-kube-api-access-8tgvs\") pod \"metallb-operator-controller-manager-7848c48b6-2kvbb\" (UID: \"f479cadf-cdb2-423f-b9c5-204bef98bf8a\") " pod="metallb-system/metallb-operator-controller-manager-7848c48b6-2kvbb" Feb 02 09:19:47 crc kubenswrapper[4764]: I0202 09:19:47.232927 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-7848c48b6-2kvbb" Feb 02 09:19:47 crc kubenswrapper[4764]: I0202 09:19:47.257200 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-5645b7d98c-hhmxc"] Feb 02 09:19:47 crc kubenswrapper[4764]: I0202 09:19:47.264130 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-5645b7d98c-hhmxc" Feb 02 09:19:47 crc kubenswrapper[4764]: I0202 09:19:47.266775 4764 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Feb 02 09:19:47 crc kubenswrapper[4764]: I0202 09:19:47.266787 4764 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Feb 02 09:19:47 crc kubenswrapper[4764]: I0202 09:19:47.267012 4764 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-c8mvm" Feb 02 09:19:47 crc kubenswrapper[4764]: I0202 09:19:47.291779 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-5645b7d98c-hhmxc"] Feb 02 09:19:47 crc kubenswrapper[4764]: I0202 09:19:47.335655 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gmt64\" (UniqueName: \"kubernetes.io/projected/1436d9d2-a684-4204-91d6-9da1a7df5fa2-kube-api-access-gmt64\") pod \"metallb-operator-webhook-server-5645b7d98c-hhmxc\" (UID: \"1436d9d2-a684-4204-91d6-9da1a7df5fa2\") " pod="metallb-system/metallb-operator-webhook-server-5645b7d98c-hhmxc" Feb 02 09:19:47 crc kubenswrapper[4764]: I0202 09:19:47.335709 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/1436d9d2-a684-4204-91d6-9da1a7df5fa2-apiservice-cert\") pod \"metallb-operator-webhook-server-5645b7d98c-hhmxc\" (UID: \"1436d9d2-a684-4204-91d6-9da1a7df5fa2\") " pod="metallb-system/metallb-operator-webhook-server-5645b7d98c-hhmxc" Feb 02 09:19:47 crc kubenswrapper[4764]: I0202 09:19:47.335748 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/1436d9d2-a684-4204-91d6-9da1a7df5fa2-webhook-cert\") pod \"metallb-operator-webhook-server-5645b7d98c-hhmxc\" (UID: \"1436d9d2-a684-4204-91d6-9da1a7df5fa2\") " pod="metallb-system/metallb-operator-webhook-server-5645b7d98c-hhmxc" Feb 02 09:19:47 crc kubenswrapper[4764]: I0202 09:19:47.439555 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/1436d9d2-a684-4204-91d6-9da1a7df5fa2-apiservice-cert\") pod \"metallb-operator-webhook-server-5645b7d98c-hhmxc\" (UID: \"1436d9d2-a684-4204-91d6-9da1a7df5fa2\") " pod="metallb-system/metallb-operator-webhook-server-5645b7d98c-hhmxc" Feb 02 09:19:47 crc kubenswrapper[4764]: I0202 09:19:47.439920 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/1436d9d2-a684-4204-91d6-9da1a7df5fa2-webhook-cert\") pod \"metallb-operator-webhook-server-5645b7d98c-hhmxc\" (UID: \"1436d9d2-a684-4204-91d6-9da1a7df5fa2\") " pod="metallb-system/metallb-operator-webhook-server-5645b7d98c-hhmxc" Feb 02 09:19:47 crc kubenswrapper[4764]: I0202 09:19:47.439981 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gmt64\" (UniqueName: \"kubernetes.io/projected/1436d9d2-a684-4204-91d6-9da1a7df5fa2-kube-api-access-gmt64\") pod \"metallb-operator-webhook-server-5645b7d98c-hhmxc\" (UID: \"1436d9d2-a684-4204-91d6-9da1a7df5fa2\") " pod="metallb-system/metallb-operator-webhook-server-5645b7d98c-hhmxc" Feb 02 09:19:47 crc kubenswrapper[4764]: I0202 09:19:47.466164 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/1436d9d2-a684-4204-91d6-9da1a7df5fa2-webhook-cert\") pod \"metallb-operator-webhook-server-5645b7d98c-hhmxc\" (UID: \"1436d9d2-a684-4204-91d6-9da1a7df5fa2\") " pod="metallb-system/metallb-operator-webhook-server-5645b7d98c-hhmxc" Feb 02 09:19:47 crc kubenswrapper[4764]: I0202 09:19:47.466248 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/1436d9d2-a684-4204-91d6-9da1a7df5fa2-apiservice-cert\") pod \"metallb-operator-webhook-server-5645b7d98c-hhmxc\" (UID: \"1436d9d2-a684-4204-91d6-9da1a7df5fa2\") " pod="metallb-system/metallb-operator-webhook-server-5645b7d98c-hhmxc" Feb 02 09:19:47 crc kubenswrapper[4764]: I0202 09:19:47.497966 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gmt64\" (UniqueName: \"kubernetes.io/projected/1436d9d2-a684-4204-91d6-9da1a7df5fa2-kube-api-access-gmt64\") pod \"metallb-operator-webhook-server-5645b7d98c-hhmxc\" (UID: \"1436d9d2-a684-4204-91d6-9da1a7df5fa2\") " pod="metallb-system/metallb-operator-webhook-server-5645b7d98c-hhmxc" Feb 02 09:19:47 crc kubenswrapper[4764]: I0202 09:19:47.575722 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-7848c48b6-2kvbb"] Feb 02 09:19:47 crc kubenswrapper[4764]: I0202 09:19:47.612649 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-5645b7d98c-hhmxc" Feb 02 09:19:47 crc kubenswrapper[4764]: I0202 09:19:47.832654 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-5645b7d98c-hhmxc"] Feb 02 09:19:47 crc kubenswrapper[4764]: W0202 09:19:47.837627 4764 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1436d9d2_a684_4204_91d6_9da1a7df5fa2.slice/crio-5c20bf5d1fb1889e0f47b6def4a3c044d624a6a146e25b1b8c283d52f8f22cb0 WatchSource:0}: Error finding container 5c20bf5d1fb1889e0f47b6def4a3c044d624a6a146e25b1b8c283d52f8f22cb0: Status 404 returned error can't find the container with id 5c20bf5d1fb1889e0f47b6def4a3c044d624a6a146e25b1b8c283d52f8f22cb0 Feb 02 09:19:48 crc kubenswrapper[4764]: I0202 09:19:48.463894 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-5645b7d98c-hhmxc" event={"ID":"1436d9d2-a684-4204-91d6-9da1a7df5fa2","Type":"ContainerStarted","Data":"5c20bf5d1fb1889e0f47b6def4a3c044d624a6a146e25b1b8c283d52f8f22cb0"} Feb 02 09:19:48 crc kubenswrapper[4764]: I0202 09:19:48.465215 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-7848c48b6-2kvbb" event={"ID":"f479cadf-cdb2-423f-b9c5-204bef98bf8a","Type":"ContainerStarted","Data":"53f124b20b13f4f264c6b99ded12a044ab541f002d2f29b98109c505767152be"} Feb 02 09:19:53 crc kubenswrapper[4764]: I0202 09:19:53.499769 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-5645b7d98c-hhmxc" event={"ID":"1436d9d2-a684-4204-91d6-9da1a7df5fa2","Type":"ContainerStarted","Data":"99f825bf0e4473c7100879afaaf274d4e6b5c3d84b3bc068e42688d28cfa0fbe"} Feb 02 09:19:53 crc kubenswrapper[4764]: I0202 09:19:53.500320 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-5645b7d98c-hhmxc" Feb 02 09:19:53 crc kubenswrapper[4764]: I0202 09:19:53.501629 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-7848c48b6-2kvbb" event={"ID":"f479cadf-cdb2-423f-b9c5-204bef98bf8a","Type":"ContainerStarted","Data":"00b0336f19b6ef2b99d8552655212fe532eb678e0671d6c9f020dc3ad932a92e"} Feb 02 09:19:53 crc kubenswrapper[4764]: I0202 09:19:53.501717 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-7848c48b6-2kvbb" Feb 02 09:19:53 crc kubenswrapper[4764]: I0202 09:19:53.524796 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-5645b7d98c-hhmxc" podStartSLOduration=1.966234939 podStartE2EDuration="6.524780232s" podCreationTimestamp="2026-02-02 09:19:47 +0000 UTC" firstStartedPulling="2026-02-02 09:19:47.839260543 +0000 UTC m=+750.772984631" lastFinishedPulling="2026-02-02 09:19:52.397805836 +0000 UTC m=+755.331529924" observedRunningTime="2026-02-02 09:19:53.523522388 +0000 UTC m=+756.457246486" watchObservedRunningTime="2026-02-02 09:19:53.524780232 +0000 UTC m=+756.458504310" Feb 02 09:19:53 crc kubenswrapper[4764]: I0202 09:19:53.549806 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-7848c48b6-2kvbb" podStartSLOduration=2.810686881 podStartE2EDuration="7.549785782s" podCreationTimestamp="2026-02-02 09:19:46 +0000 UTC" firstStartedPulling="2026-02-02 09:19:47.587047598 +0000 UTC m=+750.520771686" lastFinishedPulling="2026-02-02 09:19:52.326146499 +0000 UTC m=+755.259870587" observedRunningTime="2026-02-02 09:19:53.542384218 +0000 UTC m=+756.476108306" watchObservedRunningTime="2026-02-02 09:19:53.549785782 +0000 UTC m=+756.483509870" Feb 02 09:19:55 crc kubenswrapper[4764]: I0202 09:19:55.436476 4764 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Feb 02 09:20:07 crc kubenswrapper[4764]: I0202 09:20:07.618966 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-5645b7d98c-hhmxc" Feb 02 09:20:13 crc kubenswrapper[4764]: I0202 09:20:13.523417 4764 patch_prober.go:28] interesting pod/machine-config-daemon-4ndm4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 09:20:13 crc kubenswrapper[4764]: I0202 09:20:13.524426 4764 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 09:20:13 crc kubenswrapper[4764]: I0202 09:20:13.524533 4764 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" Feb 02 09:20:13 crc kubenswrapper[4764]: I0202 09:20:13.525476 4764 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"942c4d2fe3179033f2460fe16951c41f44faf33abaa4e47531eeac47beeeae75"} pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 02 09:20:13 crc kubenswrapper[4764]: I0202 09:20:13.525589 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" containerName="machine-config-daemon" containerID="cri-o://942c4d2fe3179033f2460fe16951c41f44faf33abaa4e47531eeac47beeeae75" gracePeriod=600 Feb 02 09:20:14 crc kubenswrapper[4764]: I0202 09:20:14.637497 4764 generic.go:334] "Generic (PLEG): container finished" podID="d192f670-9f9d-4539-9641-e4bed73acdd4" containerID="942c4d2fe3179033f2460fe16951c41f44faf33abaa4e47531eeac47beeeae75" exitCode=0 Feb 02 09:20:14 crc kubenswrapper[4764]: I0202 09:20:14.637591 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" event={"ID":"d192f670-9f9d-4539-9641-e4bed73acdd4","Type":"ContainerDied","Data":"942c4d2fe3179033f2460fe16951c41f44faf33abaa4e47531eeac47beeeae75"} Feb 02 09:20:14 crc kubenswrapper[4764]: I0202 09:20:14.638165 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" event={"ID":"d192f670-9f9d-4539-9641-e4bed73acdd4","Type":"ContainerStarted","Data":"a4c68bd9cbe36bd80883876d850d402edf3b17e0660281a9f5527447a8167a49"} Feb 02 09:20:14 crc kubenswrapper[4764]: I0202 09:20:14.638189 4764 scope.go:117] "RemoveContainer" containerID="d0647a58d2de9a2cdd486044a3db4f79924b0b477fa546211f680159a64217a7" Feb 02 09:20:27 crc kubenswrapper[4764]: I0202 09:20:27.243053 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-7848c48b6-2kvbb" Feb 02 09:20:28 crc kubenswrapper[4764]: I0202 09:20:28.048219 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-rtp9n"] Feb 02 09:20:28 crc kubenswrapper[4764]: I0202 09:20:28.051540 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-rtp9n" Feb 02 09:20:28 crc kubenswrapper[4764]: I0202 09:20:28.054143 4764 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Feb 02 09:20:28 crc kubenswrapper[4764]: I0202 09:20:28.054655 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Feb 02 09:20:28 crc kubenswrapper[4764]: I0202 09:20:28.058975 4764 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-7wdbf" Feb 02 09:20:28 crc kubenswrapper[4764]: I0202 09:20:28.070534 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-7df86c4f6c-z9ptz"] Feb 02 09:20:28 crc kubenswrapper[4764]: I0202 09:20:28.071393 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-z9ptz" Feb 02 09:20:28 crc kubenswrapper[4764]: I0202 09:20:28.073151 4764 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Feb 02 09:20:28 crc kubenswrapper[4764]: I0202 09:20:28.087300 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7df86c4f6c-z9ptz"] Feb 02 09:20:28 crc kubenswrapper[4764]: I0202 09:20:28.117611 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/9d0a4296-e773-4500-8a43-83796d835ad9-metrics\") pod \"frr-k8s-rtp9n\" (UID: \"9d0a4296-e773-4500-8a43-83796d835ad9\") " pod="metallb-system/frr-k8s-rtp9n" Feb 02 09:20:28 crc kubenswrapper[4764]: I0202 09:20:28.117664 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/9d0a4296-e773-4500-8a43-83796d835ad9-reloader\") pod \"frr-k8s-rtp9n\" (UID: \"9d0a4296-e773-4500-8a43-83796d835ad9\") " pod="metallb-system/frr-k8s-rtp9n" Feb 02 09:20:28 crc kubenswrapper[4764]: I0202 09:20:28.117692 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9d0a4296-e773-4500-8a43-83796d835ad9-metrics-certs\") pod \"frr-k8s-rtp9n\" (UID: \"9d0a4296-e773-4500-8a43-83796d835ad9\") " pod="metallb-system/frr-k8s-rtp9n" Feb 02 09:20:28 crc kubenswrapper[4764]: I0202 09:20:28.117733 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/9d0a4296-e773-4500-8a43-83796d835ad9-frr-conf\") pod \"frr-k8s-rtp9n\" (UID: \"9d0a4296-e773-4500-8a43-83796d835ad9\") " pod="metallb-system/frr-k8s-rtp9n" Feb 02 09:20:28 crc kubenswrapper[4764]: I0202 09:20:28.117761 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/9d0a4296-e773-4500-8a43-83796d835ad9-frr-startup\") pod \"frr-k8s-rtp9n\" (UID: \"9d0a4296-e773-4500-8a43-83796d835ad9\") " pod="metallb-system/frr-k8s-rtp9n" Feb 02 09:20:28 crc kubenswrapper[4764]: I0202 09:20:28.117833 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hwp5w\" (UniqueName: \"kubernetes.io/projected/9d0a4296-e773-4500-8a43-83796d835ad9-kube-api-access-hwp5w\") pod \"frr-k8s-rtp9n\" (UID: \"9d0a4296-e773-4500-8a43-83796d835ad9\") " pod="metallb-system/frr-k8s-rtp9n" Feb 02 09:20:28 crc kubenswrapper[4764]: I0202 09:20:28.117855 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/9d0a4296-e773-4500-8a43-83796d835ad9-frr-sockets\") pod \"frr-k8s-rtp9n\" (UID: \"9d0a4296-e773-4500-8a43-83796d835ad9\") " pod="metallb-system/frr-k8s-rtp9n" Feb 02 09:20:28 crc kubenswrapper[4764]: I0202 09:20:28.156985 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-jw8j4"] Feb 02 09:20:28 crc kubenswrapper[4764]: I0202 09:20:28.157780 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-jw8j4" Feb 02 09:20:28 crc kubenswrapper[4764]: I0202 09:20:28.162073 4764 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Feb 02 09:20:28 crc kubenswrapper[4764]: I0202 09:20:28.162498 4764 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Feb 02 09:20:28 crc kubenswrapper[4764]: I0202 09:20:28.162669 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Feb 02 09:20:28 crc kubenswrapper[4764]: I0202 09:20:28.162736 4764 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-nx667" Feb 02 09:20:28 crc kubenswrapper[4764]: I0202 09:20:28.193584 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-6968d8fdc4-qxqpn"] Feb 02 09:20:28 crc kubenswrapper[4764]: I0202 09:20:28.194415 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-6968d8fdc4-qxqpn" Feb 02 09:20:28 crc kubenswrapper[4764]: I0202 09:20:28.196583 4764 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Feb 02 09:20:28 crc kubenswrapper[4764]: I0202 09:20:28.216811 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-6968d8fdc4-qxqpn"] Feb 02 09:20:28 crc kubenswrapper[4764]: I0202 09:20:28.225641 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4757c3fb-ae2f-498d-a551-f6c2daee3a23-cert\") pod \"frr-k8s-webhook-server-7df86c4f6c-z9ptz\" (UID: \"4757c3fb-ae2f-498d-a551-f6c2daee3a23\") " pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-z9ptz" Feb 02 09:20:28 crc kubenswrapper[4764]: I0202 09:20:28.225679 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/9d0a4296-e773-4500-8a43-83796d835ad9-metrics\") pod \"frr-k8s-rtp9n\" (UID: \"9d0a4296-e773-4500-8a43-83796d835ad9\") " pod="metallb-system/frr-k8s-rtp9n" Feb 02 09:20:28 crc kubenswrapper[4764]: I0202 09:20:28.225696 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/9d0a4296-e773-4500-8a43-83796d835ad9-reloader\") pod \"frr-k8s-rtp9n\" (UID: \"9d0a4296-e773-4500-8a43-83796d835ad9\") " pod="metallb-system/frr-k8s-rtp9n" Feb 02 09:20:28 crc kubenswrapper[4764]: I0202 09:20:28.225715 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9d0a4296-e773-4500-8a43-83796d835ad9-metrics-certs\") pod \"frr-k8s-rtp9n\" (UID: \"9d0a4296-e773-4500-8a43-83796d835ad9\") " pod="metallb-system/frr-k8s-rtp9n" Feb 02 09:20:28 crc kubenswrapper[4764]: I0202 09:20:28.225734 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/c8444bdb-052f-4c56-841a-b3c6718759d8-memberlist\") pod \"speaker-jw8j4\" (UID: \"c8444bdb-052f-4c56-841a-b3c6718759d8\") " pod="metallb-system/speaker-jw8j4" Feb 02 09:20:28 crc kubenswrapper[4764]: I0202 09:20:28.225760 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/9d0a4296-e773-4500-8a43-83796d835ad9-frr-conf\") pod \"frr-k8s-rtp9n\" (UID: \"9d0a4296-e773-4500-8a43-83796d835ad9\") " pod="metallb-system/frr-k8s-rtp9n" Feb 02 09:20:28 crc kubenswrapper[4764]: I0202 09:20:28.225775 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/9d0a4296-e773-4500-8a43-83796d835ad9-frr-startup\") pod \"frr-k8s-rtp9n\" (UID: \"9d0a4296-e773-4500-8a43-83796d835ad9\") " pod="metallb-system/frr-k8s-rtp9n" Feb 02 09:20:28 crc kubenswrapper[4764]: I0202 09:20:28.225793 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/c8444bdb-052f-4c56-841a-b3c6718759d8-metallb-excludel2\") pod \"speaker-jw8j4\" (UID: \"c8444bdb-052f-4c56-841a-b3c6718759d8\") " pod="metallb-system/speaker-jw8j4" Feb 02 09:20:28 crc kubenswrapper[4764]: I0202 09:20:28.225820 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lddth\" (UniqueName: \"kubernetes.io/projected/4757c3fb-ae2f-498d-a551-f6c2daee3a23-kube-api-access-lddth\") pod \"frr-k8s-webhook-server-7df86c4f6c-z9ptz\" (UID: \"4757c3fb-ae2f-498d-a551-f6c2daee3a23\") " pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-z9ptz" Feb 02 09:20:28 crc kubenswrapper[4764]: I0202 09:20:28.225863 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hwp5w\" (UniqueName: \"kubernetes.io/projected/9d0a4296-e773-4500-8a43-83796d835ad9-kube-api-access-hwp5w\") pod \"frr-k8s-rtp9n\" (UID: \"9d0a4296-e773-4500-8a43-83796d835ad9\") " pod="metallb-system/frr-k8s-rtp9n" Feb 02 09:20:28 crc kubenswrapper[4764]: I0202 09:20:28.225877 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c8444bdb-052f-4c56-841a-b3c6718759d8-metrics-certs\") pod \"speaker-jw8j4\" (UID: \"c8444bdb-052f-4c56-841a-b3c6718759d8\") " pod="metallb-system/speaker-jw8j4" Feb 02 09:20:28 crc kubenswrapper[4764]: I0202 09:20:28.225891 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/9d0a4296-e773-4500-8a43-83796d835ad9-frr-sockets\") pod \"frr-k8s-rtp9n\" (UID: \"9d0a4296-e773-4500-8a43-83796d835ad9\") " pod="metallb-system/frr-k8s-rtp9n" Feb 02 09:20:28 crc kubenswrapper[4764]: I0202 09:20:28.225907 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-55ml5\" (UniqueName: \"kubernetes.io/projected/c8444bdb-052f-4c56-841a-b3c6718759d8-kube-api-access-55ml5\") pod \"speaker-jw8j4\" (UID: \"c8444bdb-052f-4c56-841a-b3c6718759d8\") " pod="metallb-system/speaker-jw8j4" Feb 02 09:20:28 crc kubenswrapper[4764]: I0202 09:20:28.229021 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/9d0a4296-e773-4500-8a43-83796d835ad9-metrics\") pod \"frr-k8s-rtp9n\" (UID: \"9d0a4296-e773-4500-8a43-83796d835ad9\") " pod="metallb-system/frr-k8s-rtp9n" Feb 02 09:20:28 crc kubenswrapper[4764]: I0202 09:20:28.237234 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/9d0a4296-e773-4500-8a43-83796d835ad9-frr-sockets\") pod \"frr-k8s-rtp9n\" (UID: \"9d0a4296-e773-4500-8a43-83796d835ad9\") " pod="metallb-system/frr-k8s-rtp9n" Feb 02 09:20:28 crc kubenswrapper[4764]: I0202 09:20:28.237717 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/9d0a4296-e773-4500-8a43-83796d835ad9-frr-startup\") pod \"frr-k8s-rtp9n\" (UID: \"9d0a4296-e773-4500-8a43-83796d835ad9\") " pod="metallb-system/frr-k8s-rtp9n" Feb 02 09:20:28 crc kubenswrapper[4764]: I0202 09:20:28.239521 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/9d0a4296-e773-4500-8a43-83796d835ad9-reloader\") pod \"frr-k8s-rtp9n\" (UID: \"9d0a4296-e773-4500-8a43-83796d835ad9\") " pod="metallb-system/frr-k8s-rtp9n" Feb 02 09:20:28 crc kubenswrapper[4764]: I0202 09:20:28.239623 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/9d0a4296-e773-4500-8a43-83796d835ad9-frr-conf\") pod \"frr-k8s-rtp9n\" (UID: \"9d0a4296-e773-4500-8a43-83796d835ad9\") " pod="metallb-system/frr-k8s-rtp9n" Feb 02 09:20:28 crc kubenswrapper[4764]: I0202 09:20:28.245466 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9d0a4296-e773-4500-8a43-83796d835ad9-metrics-certs\") pod \"frr-k8s-rtp9n\" (UID: \"9d0a4296-e773-4500-8a43-83796d835ad9\") " pod="metallb-system/frr-k8s-rtp9n" Feb 02 09:20:28 crc kubenswrapper[4764]: I0202 09:20:28.250828 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hwp5w\" (UniqueName: \"kubernetes.io/projected/9d0a4296-e773-4500-8a43-83796d835ad9-kube-api-access-hwp5w\") pod \"frr-k8s-rtp9n\" (UID: \"9d0a4296-e773-4500-8a43-83796d835ad9\") " pod="metallb-system/frr-k8s-rtp9n" Feb 02 09:20:28 crc kubenswrapper[4764]: I0202 09:20:28.326670 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lddth\" (UniqueName: \"kubernetes.io/projected/4757c3fb-ae2f-498d-a551-f6c2daee3a23-kube-api-access-lddth\") pod \"frr-k8s-webhook-server-7df86c4f6c-z9ptz\" (UID: \"4757c3fb-ae2f-498d-a551-f6c2daee3a23\") " pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-z9ptz" Feb 02 09:20:28 crc kubenswrapper[4764]: I0202 09:20:28.326722 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7vfkj\" (UniqueName: \"kubernetes.io/projected/aa337cba-937a-44e6-af3b-50f8f3438489-kube-api-access-7vfkj\") pod \"controller-6968d8fdc4-qxqpn\" (UID: \"aa337cba-937a-44e6-af3b-50f8f3438489\") " pod="metallb-system/controller-6968d8fdc4-qxqpn" Feb 02 09:20:28 crc kubenswrapper[4764]: I0202 09:20:28.326772 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/aa337cba-937a-44e6-af3b-50f8f3438489-cert\") pod \"controller-6968d8fdc4-qxqpn\" (UID: \"aa337cba-937a-44e6-af3b-50f8f3438489\") " pod="metallb-system/controller-6968d8fdc4-qxqpn" Feb 02 09:20:28 crc kubenswrapper[4764]: I0202 09:20:28.326798 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c8444bdb-052f-4c56-841a-b3c6718759d8-metrics-certs\") pod \"speaker-jw8j4\" (UID: \"c8444bdb-052f-4c56-841a-b3c6718759d8\") " pod="metallb-system/speaker-jw8j4" Feb 02 09:20:28 crc kubenswrapper[4764]: I0202 09:20:28.326816 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-55ml5\" (UniqueName: \"kubernetes.io/projected/c8444bdb-052f-4c56-841a-b3c6718759d8-kube-api-access-55ml5\") pod \"speaker-jw8j4\" (UID: \"c8444bdb-052f-4c56-841a-b3c6718759d8\") " pod="metallb-system/speaker-jw8j4" Feb 02 09:20:28 crc kubenswrapper[4764]: I0202 09:20:28.326836 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4757c3fb-ae2f-498d-a551-f6c2daee3a23-cert\") pod \"frr-k8s-webhook-server-7df86c4f6c-z9ptz\" (UID: \"4757c3fb-ae2f-498d-a551-f6c2daee3a23\") " pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-z9ptz" Feb 02 09:20:28 crc kubenswrapper[4764]: I0202 09:20:28.326858 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/aa337cba-937a-44e6-af3b-50f8f3438489-metrics-certs\") pod \"controller-6968d8fdc4-qxqpn\" (UID: \"aa337cba-937a-44e6-af3b-50f8f3438489\") " pod="metallb-system/controller-6968d8fdc4-qxqpn" Feb 02 09:20:28 crc kubenswrapper[4764]: I0202 09:20:28.326874 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/c8444bdb-052f-4c56-841a-b3c6718759d8-memberlist\") pod \"speaker-jw8j4\" (UID: \"c8444bdb-052f-4c56-841a-b3c6718759d8\") " pod="metallb-system/speaker-jw8j4" Feb 02 09:20:28 crc kubenswrapper[4764]: I0202 09:20:28.326901 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/c8444bdb-052f-4c56-841a-b3c6718759d8-metallb-excludel2\") pod \"speaker-jw8j4\" (UID: \"c8444bdb-052f-4c56-841a-b3c6718759d8\") " pod="metallb-system/speaker-jw8j4" Feb 02 09:20:28 crc kubenswrapper[4764]: I0202 09:20:28.327572 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/c8444bdb-052f-4c56-841a-b3c6718759d8-metallb-excludel2\") pod \"speaker-jw8j4\" (UID: \"c8444bdb-052f-4c56-841a-b3c6718759d8\") " pod="metallb-system/speaker-jw8j4" Feb 02 09:20:28 crc kubenswrapper[4764]: E0202 09:20:28.327962 4764 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Feb 02 09:20:28 crc kubenswrapper[4764]: E0202 09:20:28.328006 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c8444bdb-052f-4c56-841a-b3c6718759d8-memberlist podName:c8444bdb-052f-4c56-841a-b3c6718759d8 nodeName:}" failed. No retries permitted until 2026-02-02 09:20:28.827994504 +0000 UTC m=+791.761718592 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/c8444bdb-052f-4c56-841a-b3c6718759d8-memberlist") pod "speaker-jw8j4" (UID: "c8444bdb-052f-4c56-841a-b3c6718759d8") : secret "metallb-memberlist" not found Feb 02 09:20:28 crc kubenswrapper[4764]: I0202 09:20:28.330724 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4757c3fb-ae2f-498d-a551-f6c2daee3a23-cert\") pod \"frr-k8s-webhook-server-7df86c4f6c-z9ptz\" (UID: \"4757c3fb-ae2f-498d-a551-f6c2daee3a23\") " pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-z9ptz" Feb 02 09:20:28 crc kubenswrapper[4764]: I0202 09:20:28.332303 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c8444bdb-052f-4c56-841a-b3c6718759d8-metrics-certs\") pod \"speaker-jw8j4\" (UID: \"c8444bdb-052f-4c56-841a-b3c6718759d8\") " pod="metallb-system/speaker-jw8j4" Feb 02 09:20:28 crc kubenswrapper[4764]: I0202 09:20:28.345321 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-55ml5\" (UniqueName: \"kubernetes.io/projected/c8444bdb-052f-4c56-841a-b3c6718759d8-kube-api-access-55ml5\") pod \"speaker-jw8j4\" (UID: \"c8444bdb-052f-4c56-841a-b3c6718759d8\") " pod="metallb-system/speaker-jw8j4" Feb 02 09:20:28 crc kubenswrapper[4764]: I0202 09:20:28.345782 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lddth\" (UniqueName: \"kubernetes.io/projected/4757c3fb-ae2f-498d-a551-f6c2daee3a23-kube-api-access-lddth\") pod \"frr-k8s-webhook-server-7df86c4f6c-z9ptz\" (UID: \"4757c3fb-ae2f-498d-a551-f6c2daee3a23\") " pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-z9ptz" Feb 02 09:20:28 crc kubenswrapper[4764]: I0202 09:20:28.369770 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-rtp9n" Feb 02 09:20:28 crc kubenswrapper[4764]: I0202 09:20:28.401395 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-z9ptz" Feb 02 09:20:28 crc kubenswrapper[4764]: I0202 09:20:28.428304 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/aa337cba-937a-44e6-af3b-50f8f3438489-metrics-certs\") pod \"controller-6968d8fdc4-qxqpn\" (UID: \"aa337cba-937a-44e6-af3b-50f8f3438489\") " pod="metallb-system/controller-6968d8fdc4-qxqpn" Feb 02 09:20:28 crc kubenswrapper[4764]: I0202 09:20:28.428392 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7vfkj\" (UniqueName: \"kubernetes.io/projected/aa337cba-937a-44e6-af3b-50f8f3438489-kube-api-access-7vfkj\") pod \"controller-6968d8fdc4-qxqpn\" (UID: \"aa337cba-937a-44e6-af3b-50f8f3438489\") " pod="metallb-system/controller-6968d8fdc4-qxqpn" Feb 02 09:20:28 crc kubenswrapper[4764]: I0202 09:20:28.428429 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/aa337cba-937a-44e6-af3b-50f8f3438489-cert\") pod \"controller-6968d8fdc4-qxqpn\" (UID: \"aa337cba-937a-44e6-af3b-50f8f3438489\") " pod="metallb-system/controller-6968d8fdc4-qxqpn" Feb 02 09:20:28 crc kubenswrapper[4764]: I0202 09:20:28.431973 4764 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Feb 02 09:20:28 crc kubenswrapper[4764]: I0202 09:20:28.434221 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/aa337cba-937a-44e6-af3b-50f8f3438489-metrics-certs\") pod \"controller-6968d8fdc4-qxqpn\" (UID: \"aa337cba-937a-44e6-af3b-50f8f3438489\") " pod="metallb-system/controller-6968d8fdc4-qxqpn" Feb 02 09:20:28 crc kubenswrapper[4764]: I0202 09:20:28.441880 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/aa337cba-937a-44e6-af3b-50f8f3438489-cert\") pod \"controller-6968d8fdc4-qxqpn\" (UID: \"aa337cba-937a-44e6-af3b-50f8f3438489\") " pod="metallb-system/controller-6968d8fdc4-qxqpn" Feb 02 09:20:28 crc kubenswrapper[4764]: I0202 09:20:28.447336 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7vfkj\" (UniqueName: \"kubernetes.io/projected/aa337cba-937a-44e6-af3b-50f8f3438489-kube-api-access-7vfkj\") pod \"controller-6968d8fdc4-qxqpn\" (UID: \"aa337cba-937a-44e6-af3b-50f8f3438489\") " pod="metallb-system/controller-6968d8fdc4-qxqpn" Feb 02 09:20:28 crc kubenswrapper[4764]: I0202 09:20:28.512624 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-6968d8fdc4-qxqpn" Feb 02 09:20:28 crc kubenswrapper[4764]: I0202 09:20:28.702492 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7df86c4f6c-z9ptz"] Feb 02 09:20:28 crc kubenswrapper[4764]: I0202 09:20:28.760235 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-z9ptz" event={"ID":"4757c3fb-ae2f-498d-a551-f6c2daee3a23","Type":"ContainerStarted","Data":"c5a8928d0ce43ba27c5e9986fbc375f2978d60da39108fdbb5bb3d72861e5b4d"} Feb 02 09:20:28 crc kubenswrapper[4764]: I0202 09:20:28.761000 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-rtp9n" event={"ID":"9d0a4296-e773-4500-8a43-83796d835ad9","Type":"ContainerStarted","Data":"ca7347edb5233024749da1e3a23479005e6cb226bebcb712da97b63bb66bdf70"} Feb 02 09:20:28 crc kubenswrapper[4764]: I0202 09:20:28.818811 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-6968d8fdc4-qxqpn"] Feb 02 09:20:28 crc kubenswrapper[4764]: W0202 09:20:28.821557 4764 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaa337cba_937a_44e6_af3b_50f8f3438489.slice/crio-d7bf59b4a6ac17c2b3d460b38c91911747ea4d13441c5ae713e466dff0a1b432 WatchSource:0}: Error finding container d7bf59b4a6ac17c2b3d460b38c91911747ea4d13441c5ae713e466dff0a1b432: Status 404 returned error can't find the container with id d7bf59b4a6ac17c2b3d460b38c91911747ea4d13441c5ae713e466dff0a1b432 Feb 02 09:20:28 crc kubenswrapper[4764]: I0202 09:20:28.833630 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/c8444bdb-052f-4c56-841a-b3c6718759d8-memberlist\") pod \"speaker-jw8j4\" (UID: \"c8444bdb-052f-4c56-841a-b3c6718759d8\") " pod="metallb-system/speaker-jw8j4" Feb 02 09:20:28 crc kubenswrapper[4764]: E0202 09:20:28.833809 4764 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Feb 02 09:20:28 crc kubenswrapper[4764]: E0202 09:20:28.833859 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c8444bdb-052f-4c56-841a-b3c6718759d8-memberlist podName:c8444bdb-052f-4c56-841a-b3c6718759d8 nodeName:}" failed. No retries permitted until 2026-02-02 09:20:29.833842514 +0000 UTC m=+792.767566602 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/c8444bdb-052f-4c56-841a-b3c6718759d8-memberlist") pod "speaker-jw8j4" (UID: "c8444bdb-052f-4c56-841a-b3c6718759d8") : secret "metallb-memberlist" not found Feb 02 09:20:29 crc kubenswrapper[4764]: I0202 09:20:29.769344 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6968d8fdc4-qxqpn" event={"ID":"aa337cba-937a-44e6-af3b-50f8f3438489","Type":"ContainerStarted","Data":"2122da9d56800f4a7b7c524f8aba9ebf82da3fb37ace897dc69e27988357bc89"} Feb 02 09:20:29 crc kubenswrapper[4764]: I0202 09:20:29.769761 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6968d8fdc4-qxqpn" event={"ID":"aa337cba-937a-44e6-af3b-50f8f3438489","Type":"ContainerStarted","Data":"6c5be0bd0ccab3ee6b582f84bfeed88389157e2b6f636724ea1195738375f8b6"} Feb 02 09:20:29 crc kubenswrapper[4764]: I0202 09:20:29.769782 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6968d8fdc4-qxqpn" event={"ID":"aa337cba-937a-44e6-af3b-50f8f3438489","Type":"ContainerStarted","Data":"d7bf59b4a6ac17c2b3d460b38c91911747ea4d13441c5ae713e466dff0a1b432"} Feb 02 09:20:29 crc kubenswrapper[4764]: I0202 09:20:29.773234 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-6968d8fdc4-qxqpn" Feb 02 09:20:29 crc kubenswrapper[4764]: I0202 09:20:29.798869 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-6968d8fdc4-qxqpn" podStartSLOduration=1.798849513 podStartE2EDuration="1.798849513s" podCreationTimestamp="2026-02-02 09:20:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:20:29.795724867 +0000 UTC m=+792.729448975" watchObservedRunningTime="2026-02-02 09:20:29.798849513 +0000 UTC m=+792.732573611" Feb 02 09:20:29 crc kubenswrapper[4764]: I0202 09:20:29.851175 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/c8444bdb-052f-4c56-841a-b3c6718759d8-memberlist\") pod \"speaker-jw8j4\" (UID: \"c8444bdb-052f-4c56-841a-b3c6718759d8\") " pod="metallb-system/speaker-jw8j4" Feb 02 09:20:29 crc kubenswrapper[4764]: I0202 09:20:29.865663 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/c8444bdb-052f-4c56-841a-b3c6718759d8-memberlist\") pod \"speaker-jw8j4\" (UID: \"c8444bdb-052f-4c56-841a-b3c6718759d8\") " pod="metallb-system/speaker-jw8j4" Feb 02 09:20:29 crc kubenswrapper[4764]: I0202 09:20:29.972014 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-jw8j4" Feb 02 09:20:30 crc kubenswrapper[4764]: I0202 09:20:30.783009 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-jw8j4" event={"ID":"c8444bdb-052f-4c56-841a-b3c6718759d8","Type":"ContainerStarted","Data":"392332d5e0a943669b581ec3ef195bf32c7c9f9606c8625da50df42d3abe1f24"} Feb 02 09:20:30 crc kubenswrapper[4764]: I0202 09:20:30.783061 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-jw8j4" event={"ID":"c8444bdb-052f-4c56-841a-b3c6718759d8","Type":"ContainerStarted","Data":"1bcbd129696416d019736bebd3b90484270df8f5d1661b99bd5680f75e40073f"} Feb 02 09:20:30 crc kubenswrapper[4764]: I0202 09:20:30.783073 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-jw8j4" event={"ID":"c8444bdb-052f-4c56-841a-b3c6718759d8","Type":"ContainerStarted","Data":"cbc78ef4c8686ec81d7fcab15d1dea62bceeb85dd9ba798ddd4949b3b6cc93a7"} Feb 02 09:20:30 crc kubenswrapper[4764]: I0202 09:20:30.783296 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-jw8j4" Feb 02 09:20:30 crc kubenswrapper[4764]: I0202 09:20:30.834111 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-jw8j4" podStartSLOduration=2.834095495 podStartE2EDuration="2.834095495s" podCreationTimestamp="2026-02-02 09:20:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:20:30.830107614 +0000 UTC m=+793.763831702" watchObservedRunningTime="2026-02-02 09:20:30.834095495 +0000 UTC m=+793.767819583" Feb 02 09:20:36 crc kubenswrapper[4764]: I0202 09:20:36.815489 4764 generic.go:334] "Generic (PLEG): container finished" podID="9d0a4296-e773-4500-8a43-83796d835ad9" containerID="eb2816cad1ca509b4f937aabca7fe191f95b6e44e5112607a1039133e7c04d12" exitCode=0 Feb 02 09:20:36 crc kubenswrapper[4764]: I0202 09:20:36.815573 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-rtp9n" event={"ID":"9d0a4296-e773-4500-8a43-83796d835ad9","Type":"ContainerDied","Data":"eb2816cad1ca509b4f937aabca7fe191f95b6e44e5112607a1039133e7c04d12"} Feb 02 09:20:36 crc kubenswrapper[4764]: I0202 09:20:36.819540 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-z9ptz" event={"ID":"4757c3fb-ae2f-498d-a551-f6c2daee3a23","Type":"ContainerStarted","Data":"7f9a9223f72d0b2b73ebb1f92e291f8d9fd057a099169dcb918bcfa33e020403"} Feb 02 09:20:36 crc kubenswrapper[4764]: I0202 09:20:36.819794 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-z9ptz" Feb 02 09:20:36 crc kubenswrapper[4764]: I0202 09:20:36.858600 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-z9ptz" podStartSLOduration=0.95047715 podStartE2EDuration="8.85858361s" podCreationTimestamp="2026-02-02 09:20:28 +0000 UTC" firstStartedPulling="2026-02-02 09:20:28.713421574 +0000 UTC m=+791.647145662" lastFinishedPulling="2026-02-02 09:20:36.621528034 +0000 UTC m=+799.555252122" observedRunningTime="2026-02-02 09:20:36.856681847 +0000 UTC m=+799.790405935" watchObservedRunningTime="2026-02-02 09:20:36.85858361 +0000 UTC m=+799.792307698" Feb 02 09:20:37 crc kubenswrapper[4764]: I0202 09:20:37.438763 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-p7xkn"] Feb 02 09:20:37 crc kubenswrapper[4764]: I0202 09:20:37.440767 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-p7xkn" Feb 02 09:20:37 crc kubenswrapper[4764]: I0202 09:20:37.460356 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-p7xkn"] Feb 02 09:20:37 crc kubenswrapper[4764]: I0202 09:20:37.555188 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/554335f8-e3ed-405c-82f3-6243c32e1f86-catalog-content\") pod \"community-operators-p7xkn\" (UID: \"554335f8-e3ed-405c-82f3-6243c32e1f86\") " pod="openshift-marketplace/community-operators-p7xkn" Feb 02 09:20:37 crc kubenswrapper[4764]: I0202 09:20:37.555251 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9qmc8\" (UniqueName: \"kubernetes.io/projected/554335f8-e3ed-405c-82f3-6243c32e1f86-kube-api-access-9qmc8\") pod \"community-operators-p7xkn\" (UID: \"554335f8-e3ed-405c-82f3-6243c32e1f86\") " pod="openshift-marketplace/community-operators-p7xkn" Feb 02 09:20:37 crc kubenswrapper[4764]: I0202 09:20:37.555277 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/554335f8-e3ed-405c-82f3-6243c32e1f86-utilities\") pod \"community-operators-p7xkn\" (UID: \"554335f8-e3ed-405c-82f3-6243c32e1f86\") " pod="openshift-marketplace/community-operators-p7xkn" Feb 02 09:20:37 crc kubenswrapper[4764]: I0202 09:20:37.656685 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/554335f8-e3ed-405c-82f3-6243c32e1f86-catalog-content\") pod \"community-operators-p7xkn\" (UID: \"554335f8-e3ed-405c-82f3-6243c32e1f86\") " pod="openshift-marketplace/community-operators-p7xkn" Feb 02 09:20:37 crc kubenswrapper[4764]: I0202 09:20:37.656732 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9qmc8\" (UniqueName: \"kubernetes.io/projected/554335f8-e3ed-405c-82f3-6243c32e1f86-kube-api-access-9qmc8\") pod \"community-operators-p7xkn\" (UID: \"554335f8-e3ed-405c-82f3-6243c32e1f86\") " pod="openshift-marketplace/community-operators-p7xkn" Feb 02 09:20:37 crc kubenswrapper[4764]: I0202 09:20:37.656750 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/554335f8-e3ed-405c-82f3-6243c32e1f86-utilities\") pod \"community-operators-p7xkn\" (UID: \"554335f8-e3ed-405c-82f3-6243c32e1f86\") " pod="openshift-marketplace/community-operators-p7xkn" Feb 02 09:20:37 crc kubenswrapper[4764]: I0202 09:20:37.657173 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/554335f8-e3ed-405c-82f3-6243c32e1f86-catalog-content\") pod \"community-operators-p7xkn\" (UID: \"554335f8-e3ed-405c-82f3-6243c32e1f86\") " pod="openshift-marketplace/community-operators-p7xkn" Feb 02 09:20:37 crc kubenswrapper[4764]: I0202 09:20:37.657217 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/554335f8-e3ed-405c-82f3-6243c32e1f86-utilities\") pod \"community-operators-p7xkn\" (UID: \"554335f8-e3ed-405c-82f3-6243c32e1f86\") " pod="openshift-marketplace/community-operators-p7xkn" Feb 02 09:20:37 crc kubenswrapper[4764]: I0202 09:20:37.680627 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9qmc8\" (UniqueName: \"kubernetes.io/projected/554335f8-e3ed-405c-82f3-6243c32e1f86-kube-api-access-9qmc8\") pod \"community-operators-p7xkn\" (UID: \"554335f8-e3ed-405c-82f3-6243c32e1f86\") " pod="openshift-marketplace/community-operators-p7xkn" Feb 02 09:20:37 crc kubenswrapper[4764]: I0202 09:20:37.755190 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-p7xkn" Feb 02 09:20:37 crc kubenswrapper[4764]: I0202 09:20:37.834989 4764 generic.go:334] "Generic (PLEG): container finished" podID="9d0a4296-e773-4500-8a43-83796d835ad9" containerID="16c721e456c547033b3523e903fbfe0360c2ec58f3b1b23ccefcafa71d4f259a" exitCode=0 Feb 02 09:20:37 crc kubenswrapper[4764]: I0202 09:20:37.835091 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-rtp9n" event={"ID":"9d0a4296-e773-4500-8a43-83796d835ad9","Type":"ContainerDied","Data":"16c721e456c547033b3523e903fbfe0360c2ec58f3b1b23ccefcafa71d4f259a"} Feb 02 09:20:38 crc kubenswrapper[4764]: I0202 09:20:38.060544 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-p7xkn"] Feb 02 09:20:38 crc kubenswrapper[4764]: W0202 09:20:38.072487 4764 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod554335f8_e3ed_405c_82f3_6243c32e1f86.slice/crio-bee8733755fb5a91bf0b60243a5cd896c005b2295a38c660e4468975af515070 WatchSource:0}: Error finding container bee8733755fb5a91bf0b60243a5cd896c005b2295a38c660e4468975af515070: Status 404 returned error can't find the container with id bee8733755fb5a91bf0b60243a5cd896c005b2295a38c660e4468975af515070 Feb 02 09:20:38 crc kubenswrapper[4764]: I0202 09:20:38.519570 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-6968d8fdc4-qxqpn" Feb 02 09:20:38 crc kubenswrapper[4764]: I0202 09:20:38.841814 4764 generic.go:334] "Generic (PLEG): container finished" podID="9d0a4296-e773-4500-8a43-83796d835ad9" containerID="ef86dee9652f02d7c2efb8782ff5fdf6a9f56d3e0783a9b9be33365434fdcd29" exitCode=0 Feb 02 09:20:38 crc kubenswrapper[4764]: I0202 09:20:38.841875 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-rtp9n" event={"ID":"9d0a4296-e773-4500-8a43-83796d835ad9","Type":"ContainerDied","Data":"ef86dee9652f02d7c2efb8782ff5fdf6a9f56d3e0783a9b9be33365434fdcd29"} Feb 02 09:20:38 crc kubenswrapper[4764]: I0202 09:20:38.844170 4764 generic.go:334] "Generic (PLEG): container finished" podID="554335f8-e3ed-405c-82f3-6243c32e1f86" containerID="f9d4c4b675b39c41c558e20628aacb6e277f0dae42f5e3a45e511de4e5f03a50" exitCode=0 Feb 02 09:20:38 crc kubenswrapper[4764]: I0202 09:20:38.844217 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p7xkn" event={"ID":"554335f8-e3ed-405c-82f3-6243c32e1f86","Type":"ContainerDied","Data":"f9d4c4b675b39c41c558e20628aacb6e277f0dae42f5e3a45e511de4e5f03a50"} Feb 02 09:20:38 crc kubenswrapper[4764]: I0202 09:20:38.844247 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p7xkn" event={"ID":"554335f8-e3ed-405c-82f3-6243c32e1f86","Type":"ContainerStarted","Data":"bee8733755fb5a91bf0b60243a5cd896c005b2295a38c660e4468975af515070"} Feb 02 09:20:39 crc kubenswrapper[4764]: I0202 09:20:39.855531 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-rtp9n" event={"ID":"9d0a4296-e773-4500-8a43-83796d835ad9","Type":"ContainerStarted","Data":"dd9377faa42bee2818b64a8dc979cf030045eb2558ddafe26683b26cb29ddf59"} Feb 02 09:20:39 crc kubenswrapper[4764]: I0202 09:20:39.856144 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-rtp9n" event={"ID":"9d0a4296-e773-4500-8a43-83796d835ad9","Type":"ContainerStarted","Data":"e56a63ce3b329f312a18c5202be4f48d593f73b18e2e4095bbdc95da9010504d"} Feb 02 09:20:39 crc kubenswrapper[4764]: I0202 09:20:39.856168 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-rtp9n" event={"ID":"9d0a4296-e773-4500-8a43-83796d835ad9","Type":"ContainerStarted","Data":"07507ff6735274abab8183e6f6234beb43edf7f6c74dff0f86918ec3d9c70032"} Feb 02 09:20:39 crc kubenswrapper[4764]: I0202 09:20:39.856186 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-rtp9n" event={"ID":"9d0a4296-e773-4500-8a43-83796d835ad9","Type":"ContainerStarted","Data":"9bb58ed948a9d7cc746abca7e6c3a4f58570fdbc0184ee08ccfa68eabec5f79d"} Feb 02 09:20:39 crc kubenswrapper[4764]: I0202 09:20:39.856204 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-rtp9n" event={"ID":"9d0a4296-e773-4500-8a43-83796d835ad9","Type":"ContainerStarted","Data":"136f5f6702aa4ce7ba1174c4e5e4ceb6440ffb528b1f2fdb30be976c91b08323"} Feb 02 09:20:40 crc kubenswrapper[4764]: I0202 09:20:40.867024 4764 generic.go:334] "Generic (PLEG): container finished" podID="554335f8-e3ed-405c-82f3-6243c32e1f86" containerID="7d6f825bae02a241570471aa23c891bc251246b8825d7601cb6430ae100bdb28" exitCode=0 Feb 02 09:20:40 crc kubenswrapper[4764]: I0202 09:20:40.867169 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p7xkn" event={"ID":"554335f8-e3ed-405c-82f3-6243c32e1f86","Type":"ContainerDied","Data":"7d6f825bae02a241570471aa23c891bc251246b8825d7601cb6430ae100bdb28"} Feb 02 09:20:40 crc kubenswrapper[4764]: I0202 09:20:40.873849 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-rtp9n" event={"ID":"9d0a4296-e773-4500-8a43-83796d835ad9","Type":"ContainerStarted","Data":"6ac3008e9d348ff7a464eace968a0a85a669209c108998fd6bf6588f03d4bbaf"} Feb 02 09:20:40 crc kubenswrapper[4764]: I0202 09:20:40.874734 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-rtp9n" Feb 02 09:20:41 crc kubenswrapper[4764]: I0202 09:20:41.903426 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p7xkn" event={"ID":"554335f8-e3ed-405c-82f3-6243c32e1f86","Type":"ContainerStarted","Data":"28214c15bc63f0ed8efce6c0191169bbab58fa6f1cc0489c2f312086a7a032f6"} Feb 02 09:20:41 crc kubenswrapper[4764]: I0202 09:20:41.927826 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-rtp9n" podStartSLOduration=5.779829641 podStartE2EDuration="13.927810886s" podCreationTimestamp="2026-02-02 09:20:28 +0000 UTC" firstStartedPulling="2026-02-02 09:20:28.507420106 +0000 UTC m=+791.441144184" lastFinishedPulling="2026-02-02 09:20:36.655401341 +0000 UTC m=+799.589125429" observedRunningTime="2026-02-02 09:20:40.934439614 +0000 UTC m=+803.868163742" watchObservedRunningTime="2026-02-02 09:20:41.927810886 +0000 UTC m=+804.861534974" Feb 02 09:20:41 crc kubenswrapper[4764]: I0202 09:20:41.929558 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-p7xkn" podStartSLOduration=2.387257203 podStartE2EDuration="4.929553404s" podCreationTimestamp="2026-02-02 09:20:37 +0000 UTC" firstStartedPulling="2026-02-02 09:20:38.846335064 +0000 UTC m=+801.780059152" lastFinishedPulling="2026-02-02 09:20:41.388631255 +0000 UTC m=+804.322355353" observedRunningTime="2026-02-02 09:20:41.927275801 +0000 UTC m=+804.860999909" watchObservedRunningTime="2026-02-02 09:20:41.929553404 +0000 UTC m=+804.863277502" Feb 02 09:20:43 crc kubenswrapper[4764]: I0202 09:20:43.370795 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-rtp9n" Feb 02 09:20:43 crc kubenswrapper[4764]: I0202 09:20:43.445518 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-rtp9n" Feb 02 09:20:43 crc kubenswrapper[4764]: I0202 09:20:43.611052 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-t2pkp"] Feb 02 09:20:43 crc kubenswrapper[4764]: I0202 09:20:43.612402 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t2pkp" Feb 02 09:20:43 crc kubenswrapper[4764]: I0202 09:20:43.631897 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-t2pkp"] Feb 02 09:20:43 crc kubenswrapper[4764]: I0202 09:20:43.738435 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cbb6142d-f15d-45d8-9b8d-cafbb96ebfbd-catalog-content\") pod \"redhat-operators-t2pkp\" (UID: \"cbb6142d-f15d-45d8-9b8d-cafbb96ebfbd\") " pod="openshift-marketplace/redhat-operators-t2pkp" Feb 02 09:20:43 crc kubenswrapper[4764]: I0202 09:20:43.738479 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cbb6142d-f15d-45d8-9b8d-cafbb96ebfbd-utilities\") pod \"redhat-operators-t2pkp\" (UID: \"cbb6142d-f15d-45d8-9b8d-cafbb96ebfbd\") " pod="openshift-marketplace/redhat-operators-t2pkp" Feb 02 09:20:43 crc kubenswrapper[4764]: I0202 09:20:43.738505 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g596k\" (UniqueName: \"kubernetes.io/projected/cbb6142d-f15d-45d8-9b8d-cafbb96ebfbd-kube-api-access-g596k\") pod \"redhat-operators-t2pkp\" (UID: \"cbb6142d-f15d-45d8-9b8d-cafbb96ebfbd\") " pod="openshift-marketplace/redhat-operators-t2pkp" Feb 02 09:20:43 crc kubenswrapper[4764]: I0202 09:20:43.839619 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cbb6142d-f15d-45d8-9b8d-cafbb96ebfbd-catalog-content\") pod \"redhat-operators-t2pkp\" (UID: \"cbb6142d-f15d-45d8-9b8d-cafbb96ebfbd\") " pod="openshift-marketplace/redhat-operators-t2pkp" Feb 02 09:20:43 crc kubenswrapper[4764]: I0202 09:20:43.839898 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cbb6142d-f15d-45d8-9b8d-cafbb96ebfbd-utilities\") pod \"redhat-operators-t2pkp\" (UID: \"cbb6142d-f15d-45d8-9b8d-cafbb96ebfbd\") " pod="openshift-marketplace/redhat-operators-t2pkp" Feb 02 09:20:43 crc kubenswrapper[4764]: I0202 09:20:43.840008 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g596k\" (UniqueName: \"kubernetes.io/projected/cbb6142d-f15d-45d8-9b8d-cafbb96ebfbd-kube-api-access-g596k\") pod \"redhat-operators-t2pkp\" (UID: \"cbb6142d-f15d-45d8-9b8d-cafbb96ebfbd\") " pod="openshift-marketplace/redhat-operators-t2pkp" Feb 02 09:20:43 crc kubenswrapper[4764]: I0202 09:20:43.840173 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cbb6142d-f15d-45d8-9b8d-cafbb96ebfbd-catalog-content\") pod \"redhat-operators-t2pkp\" (UID: \"cbb6142d-f15d-45d8-9b8d-cafbb96ebfbd\") " pod="openshift-marketplace/redhat-operators-t2pkp" Feb 02 09:20:43 crc kubenswrapper[4764]: I0202 09:20:43.840375 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cbb6142d-f15d-45d8-9b8d-cafbb96ebfbd-utilities\") pod \"redhat-operators-t2pkp\" (UID: \"cbb6142d-f15d-45d8-9b8d-cafbb96ebfbd\") " pod="openshift-marketplace/redhat-operators-t2pkp" Feb 02 09:20:43 crc kubenswrapper[4764]: I0202 09:20:43.864753 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g596k\" (UniqueName: \"kubernetes.io/projected/cbb6142d-f15d-45d8-9b8d-cafbb96ebfbd-kube-api-access-g596k\") pod \"redhat-operators-t2pkp\" (UID: \"cbb6142d-f15d-45d8-9b8d-cafbb96ebfbd\") " pod="openshift-marketplace/redhat-operators-t2pkp" Feb 02 09:20:43 crc kubenswrapper[4764]: I0202 09:20:43.937618 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t2pkp" Feb 02 09:20:44 crc kubenswrapper[4764]: I0202 09:20:44.192873 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-t2pkp"] Feb 02 09:20:44 crc kubenswrapper[4764]: W0202 09:20:44.200030 4764 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcbb6142d_f15d_45d8_9b8d_cafbb96ebfbd.slice/crio-8f24c3118bc471d0982c70f762e8e21791b98cb968ec6c32bdc1f23f4a7193a6 WatchSource:0}: Error finding container 8f24c3118bc471d0982c70f762e8e21791b98cb968ec6c32bdc1f23f4a7193a6: Status 404 returned error can't find the container with id 8f24c3118bc471d0982c70f762e8e21791b98cb968ec6c32bdc1f23f4a7193a6 Feb 02 09:20:44 crc kubenswrapper[4764]: I0202 09:20:44.917417 4764 generic.go:334] "Generic (PLEG): container finished" podID="cbb6142d-f15d-45d8-9b8d-cafbb96ebfbd" containerID="aea4b8619a07fa6e9e25f19ed61b90b8b46beef9f2a2ff88504362caa72aa431" exitCode=0 Feb 02 09:20:44 crc kubenswrapper[4764]: I0202 09:20:44.917504 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t2pkp" event={"ID":"cbb6142d-f15d-45d8-9b8d-cafbb96ebfbd","Type":"ContainerDied","Data":"aea4b8619a07fa6e9e25f19ed61b90b8b46beef9f2a2ff88504362caa72aa431"} Feb 02 09:20:44 crc kubenswrapper[4764]: I0202 09:20:44.917708 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t2pkp" event={"ID":"cbb6142d-f15d-45d8-9b8d-cafbb96ebfbd","Type":"ContainerStarted","Data":"8f24c3118bc471d0982c70f762e8e21791b98cb968ec6c32bdc1f23f4a7193a6"} Feb 02 09:20:45 crc kubenswrapper[4764]: I0202 09:20:45.930491 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t2pkp" event={"ID":"cbb6142d-f15d-45d8-9b8d-cafbb96ebfbd","Type":"ContainerStarted","Data":"a0243c747887515934803d4f9cab422d5d60d3dcf0db4e51780addbccb98e133"} Feb 02 09:20:46 crc kubenswrapper[4764]: I0202 09:20:46.942192 4764 generic.go:334] "Generic (PLEG): container finished" podID="cbb6142d-f15d-45d8-9b8d-cafbb96ebfbd" containerID="a0243c747887515934803d4f9cab422d5d60d3dcf0db4e51780addbccb98e133" exitCode=0 Feb 02 09:20:46 crc kubenswrapper[4764]: I0202 09:20:46.942248 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t2pkp" event={"ID":"cbb6142d-f15d-45d8-9b8d-cafbb96ebfbd","Type":"ContainerDied","Data":"a0243c747887515934803d4f9cab422d5d60d3dcf0db4e51780addbccb98e133"} Feb 02 09:20:47 crc kubenswrapper[4764]: I0202 09:20:47.755878 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-p7xkn" Feb 02 09:20:47 crc kubenswrapper[4764]: I0202 09:20:47.756316 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-p7xkn" Feb 02 09:20:47 crc kubenswrapper[4764]: I0202 09:20:47.805702 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-p7xkn" Feb 02 09:20:47 crc kubenswrapper[4764]: I0202 09:20:47.953514 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t2pkp" event={"ID":"cbb6142d-f15d-45d8-9b8d-cafbb96ebfbd","Type":"ContainerStarted","Data":"3edeff5eafe3470f3ee7441c35f3ed36b967745258f9a1ebaed0a3b37037412b"} Feb 02 09:20:47 crc kubenswrapper[4764]: I0202 09:20:47.983412 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-t2pkp" podStartSLOduration=2.455969033 podStartE2EDuration="4.983396522s" podCreationTimestamp="2026-02-02 09:20:43 +0000 UTC" firstStartedPulling="2026-02-02 09:20:44.919163897 +0000 UTC m=+807.852887985" lastFinishedPulling="2026-02-02 09:20:47.446591346 +0000 UTC m=+810.380315474" observedRunningTime="2026-02-02 09:20:47.979057202 +0000 UTC m=+810.912781290" watchObservedRunningTime="2026-02-02 09:20:47.983396522 +0000 UTC m=+810.917120610" Feb 02 09:20:48 crc kubenswrapper[4764]: I0202 09:20:48.001051 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-p7xkn" Feb 02 09:20:48 crc kubenswrapper[4764]: I0202 09:20:48.372692 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-rtp9n" Feb 02 09:20:48 crc kubenswrapper[4764]: I0202 09:20:48.424495 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-z9ptz" Feb 02 09:20:49 crc kubenswrapper[4764]: I0202 09:20:49.975616 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-jw8j4" Feb 02 09:20:50 crc kubenswrapper[4764]: I0202 09:20:50.202358 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-p7xkn"] Feb 02 09:20:50 crc kubenswrapper[4764]: I0202 09:20:50.202777 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-p7xkn" podUID="554335f8-e3ed-405c-82f3-6243c32e1f86" containerName="registry-server" containerID="cri-o://28214c15bc63f0ed8efce6c0191169bbab58fa6f1cc0489c2f312086a7a032f6" gracePeriod=2 Feb 02 09:20:50 crc kubenswrapper[4764]: I0202 09:20:50.622927 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-p7xkn" Feb 02 09:20:50 crc kubenswrapper[4764]: I0202 09:20:50.735640 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/554335f8-e3ed-405c-82f3-6243c32e1f86-utilities\") pod \"554335f8-e3ed-405c-82f3-6243c32e1f86\" (UID: \"554335f8-e3ed-405c-82f3-6243c32e1f86\") " Feb 02 09:20:50 crc kubenswrapper[4764]: I0202 09:20:50.735688 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9qmc8\" (UniqueName: \"kubernetes.io/projected/554335f8-e3ed-405c-82f3-6243c32e1f86-kube-api-access-9qmc8\") pod \"554335f8-e3ed-405c-82f3-6243c32e1f86\" (UID: \"554335f8-e3ed-405c-82f3-6243c32e1f86\") " Feb 02 09:20:50 crc kubenswrapper[4764]: I0202 09:20:50.735772 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/554335f8-e3ed-405c-82f3-6243c32e1f86-catalog-content\") pod \"554335f8-e3ed-405c-82f3-6243c32e1f86\" (UID: \"554335f8-e3ed-405c-82f3-6243c32e1f86\") " Feb 02 09:20:50 crc kubenswrapper[4764]: I0202 09:20:50.740396 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/554335f8-e3ed-405c-82f3-6243c32e1f86-kube-api-access-9qmc8" (OuterVolumeSpecName: "kube-api-access-9qmc8") pod "554335f8-e3ed-405c-82f3-6243c32e1f86" (UID: "554335f8-e3ed-405c-82f3-6243c32e1f86"). InnerVolumeSpecName "kube-api-access-9qmc8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:20:50 crc kubenswrapper[4764]: I0202 09:20:50.749516 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/554335f8-e3ed-405c-82f3-6243c32e1f86-utilities" (OuterVolumeSpecName: "utilities") pod "554335f8-e3ed-405c-82f3-6243c32e1f86" (UID: "554335f8-e3ed-405c-82f3-6243c32e1f86"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 09:20:50 crc kubenswrapper[4764]: I0202 09:20:50.837415 4764 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/554335f8-e3ed-405c-82f3-6243c32e1f86-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 09:20:50 crc kubenswrapper[4764]: I0202 09:20:50.837442 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9qmc8\" (UniqueName: \"kubernetes.io/projected/554335f8-e3ed-405c-82f3-6243c32e1f86-kube-api-access-9qmc8\") on node \"crc\" DevicePath \"\"" Feb 02 09:20:50 crc kubenswrapper[4764]: I0202 09:20:50.975304 4764 generic.go:334] "Generic (PLEG): container finished" podID="554335f8-e3ed-405c-82f3-6243c32e1f86" containerID="28214c15bc63f0ed8efce6c0191169bbab58fa6f1cc0489c2f312086a7a032f6" exitCode=0 Feb 02 09:20:50 crc kubenswrapper[4764]: I0202 09:20:50.975342 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p7xkn" event={"ID":"554335f8-e3ed-405c-82f3-6243c32e1f86","Type":"ContainerDied","Data":"28214c15bc63f0ed8efce6c0191169bbab58fa6f1cc0489c2f312086a7a032f6"} Feb 02 09:20:50 crc kubenswrapper[4764]: I0202 09:20:50.975371 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p7xkn" event={"ID":"554335f8-e3ed-405c-82f3-6243c32e1f86","Type":"ContainerDied","Data":"bee8733755fb5a91bf0b60243a5cd896c005b2295a38c660e4468975af515070"} Feb 02 09:20:50 crc kubenswrapper[4764]: I0202 09:20:50.975391 4764 scope.go:117] "RemoveContainer" containerID="28214c15bc63f0ed8efce6c0191169bbab58fa6f1cc0489c2f312086a7a032f6" Feb 02 09:20:50 crc kubenswrapper[4764]: I0202 09:20:50.975402 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-p7xkn" Feb 02 09:20:51 crc kubenswrapper[4764]: I0202 09:20:51.004372 4764 scope.go:117] "RemoveContainer" containerID="7d6f825bae02a241570471aa23c891bc251246b8825d7601cb6430ae100bdb28" Feb 02 09:20:51 crc kubenswrapper[4764]: I0202 09:20:51.020010 4764 scope.go:117] "RemoveContainer" containerID="f9d4c4b675b39c41c558e20628aacb6e277f0dae42f5e3a45e511de4e5f03a50" Feb 02 09:20:51 crc kubenswrapper[4764]: I0202 09:20:51.041882 4764 scope.go:117] "RemoveContainer" containerID="28214c15bc63f0ed8efce6c0191169bbab58fa6f1cc0489c2f312086a7a032f6" Feb 02 09:20:51 crc kubenswrapper[4764]: E0202 09:20:51.042301 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"28214c15bc63f0ed8efce6c0191169bbab58fa6f1cc0489c2f312086a7a032f6\": container with ID starting with 28214c15bc63f0ed8efce6c0191169bbab58fa6f1cc0489c2f312086a7a032f6 not found: ID does not exist" containerID="28214c15bc63f0ed8efce6c0191169bbab58fa6f1cc0489c2f312086a7a032f6" Feb 02 09:20:51 crc kubenswrapper[4764]: I0202 09:20:51.042382 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"28214c15bc63f0ed8efce6c0191169bbab58fa6f1cc0489c2f312086a7a032f6"} err="failed to get container status \"28214c15bc63f0ed8efce6c0191169bbab58fa6f1cc0489c2f312086a7a032f6\": rpc error: code = NotFound desc = could not find container \"28214c15bc63f0ed8efce6c0191169bbab58fa6f1cc0489c2f312086a7a032f6\": container with ID starting with 28214c15bc63f0ed8efce6c0191169bbab58fa6f1cc0489c2f312086a7a032f6 not found: ID does not exist" Feb 02 09:20:51 crc kubenswrapper[4764]: I0202 09:20:51.042458 4764 scope.go:117] "RemoveContainer" containerID="7d6f825bae02a241570471aa23c891bc251246b8825d7601cb6430ae100bdb28" Feb 02 09:20:51 crc kubenswrapper[4764]: E0202 09:20:51.043365 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7d6f825bae02a241570471aa23c891bc251246b8825d7601cb6430ae100bdb28\": container with ID starting with 7d6f825bae02a241570471aa23c891bc251246b8825d7601cb6430ae100bdb28 not found: ID does not exist" containerID="7d6f825bae02a241570471aa23c891bc251246b8825d7601cb6430ae100bdb28" Feb 02 09:20:51 crc kubenswrapper[4764]: I0202 09:20:51.043385 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d6f825bae02a241570471aa23c891bc251246b8825d7601cb6430ae100bdb28"} err="failed to get container status \"7d6f825bae02a241570471aa23c891bc251246b8825d7601cb6430ae100bdb28\": rpc error: code = NotFound desc = could not find container \"7d6f825bae02a241570471aa23c891bc251246b8825d7601cb6430ae100bdb28\": container with ID starting with 7d6f825bae02a241570471aa23c891bc251246b8825d7601cb6430ae100bdb28 not found: ID does not exist" Feb 02 09:20:51 crc kubenswrapper[4764]: I0202 09:20:51.043398 4764 scope.go:117] "RemoveContainer" containerID="f9d4c4b675b39c41c558e20628aacb6e277f0dae42f5e3a45e511de4e5f03a50" Feb 02 09:20:51 crc kubenswrapper[4764]: E0202 09:20:51.043650 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f9d4c4b675b39c41c558e20628aacb6e277f0dae42f5e3a45e511de4e5f03a50\": container with ID starting with f9d4c4b675b39c41c558e20628aacb6e277f0dae42f5e3a45e511de4e5f03a50 not found: ID does not exist" containerID="f9d4c4b675b39c41c558e20628aacb6e277f0dae42f5e3a45e511de4e5f03a50" Feb 02 09:20:51 crc kubenswrapper[4764]: I0202 09:20:51.043716 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f9d4c4b675b39c41c558e20628aacb6e277f0dae42f5e3a45e511de4e5f03a50"} err="failed to get container status \"f9d4c4b675b39c41c558e20628aacb6e277f0dae42f5e3a45e511de4e5f03a50\": rpc error: code = NotFound desc = could not find container \"f9d4c4b675b39c41c558e20628aacb6e277f0dae42f5e3a45e511de4e5f03a50\": container with ID starting with f9d4c4b675b39c41c558e20628aacb6e277f0dae42f5e3a45e511de4e5f03a50 not found: ID does not exist" Feb 02 09:20:53 crc kubenswrapper[4764]: I0202 09:20:53.938514 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-t2pkp" Feb 02 09:20:53 crc kubenswrapper[4764]: I0202 09:20:53.939334 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-t2pkp" Feb 02 09:20:53 crc kubenswrapper[4764]: I0202 09:20:53.981521 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-t2pkp" Feb 02 09:20:54 crc kubenswrapper[4764]: I0202 09:20:54.034538 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-t2pkp" Feb 02 09:20:54 crc kubenswrapper[4764]: I0202 09:20:54.151189 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/554335f8-e3ed-405c-82f3-6243c32e1f86-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "554335f8-e3ed-405c-82f3-6243c32e1f86" (UID: "554335f8-e3ed-405c-82f3-6243c32e1f86"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 09:20:54 crc kubenswrapper[4764]: I0202 09:20:54.185227 4764 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/554335f8-e3ed-405c-82f3-6243c32e1f86-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 09:20:54 crc kubenswrapper[4764]: I0202 09:20:54.306602 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-p7xkn"] Feb 02 09:20:54 crc kubenswrapper[4764]: I0202 09:20:54.311959 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-p7xkn"] Feb 02 09:20:55 crc kubenswrapper[4764]: I0202 09:20:55.837774 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="554335f8-e3ed-405c-82f3-6243c32e1f86" path="/var/lib/kubelet/pods/554335f8-e3ed-405c-82f3-6243c32e1f86/volumes" Feb 02 09:20:56 crc kubenswrapper[4764]: I0202 09:20:56.016652 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-8ctp4"] Feb 02 09:20:56 crc kubenswrapper[4764]: E0202 09:20:56.017018 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="554335f8-e3ed-405c-82f3-6243c32e1f86" containerName="registry-server" Feb 02 09:20:56 crc kubenswrapper[4764]: I0202 09:20:56.017038 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="554335f8-e3ed-405c-82f3-6243c32e1f86" containerName="registry-server" Feb 02 09:20:56 crc kubenswrapper[4764]: E0202 09:20:56.017064 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="554335f8-e3ed-405c-82f3-6243c32e1f86" containerName="extract-content" Feb 02 09:20:56 crc kubenswrapper[4764]: I0202 09:20:56.017076 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="554335f8-e3ed-405c-82f3-6243c32e1f86" containerName="extract-content" Feb 02 09:20:56 crc kubenswrapper[4764]: E0202 09:20:56.017103 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="554335f8-e3ed-405c-82f3-6243c32e1f86" containerName="extract-utilities" Feb 02 09:20:56 crc kubenswrapper[4764]: I0202 09:20:56.017116 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="554335f8-e3ed-405c-82f3-6243c32e1f86" containerName="extract-utilities" Feb 02 09:20:56 crc kubenswrapper[4764]: I0202 09:20:56.017324 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="554335f8-e3ed-405c-82f3-6243c32e1f86" containerName="registry-server" Feb 02 09:20:56 crc kubenswrapper[4764]: I0202 09:20:56.018023 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-8ctp4" Feb 02 09:20:56 crc kubenswrapper[4764]: I0202 09:20:56.023919 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Feb 02 09:20:56 crc kubenswrapper[4764]: I0202 09:20:56.024283 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Feb 02 09:20:56 crc kubenswrapper[4764]: I0202 09:20:56.024388 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-kgjhk" Feb 02 09:20:56 crc kubenswrapper[4764]: I0202 09:20:56.033260 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-8ctp4"] Feb 02 09:20:56 crc kubenswrapper[4764]: I0202 09:20:56.113854 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cbjhk\" (UniqueName: \"kubernetes.io/projected/50551dd4-71f5-4eb4-8736-ed04f0845c7e-kube-api-access-cbjhk\") pod \"openstack-operator-index-8ctp4\" (UID: \"50551dd4-71f5-4eb4-8736-ed04f0845c7e\") " pod="openstack-operators/openstack-operator-index-8ctp4" Feb 02 09:20:56 crc kubenswrapper[4764]: I0202 09:20:56.215008 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cbjhk\" (UniqueName: \"kubernetes.io/projected/50551dd4-71f5-4eb4-8736-ed04f0845c7e-kube-api-access-cbjhk\") pod \"openstack-operator-index-8ctp4\" (UID: \"50551dd4-71f5-4eb4-8736-ed04f0845c7e\") " pod="openstack-operators/openstack-operator-index-8ctp4" Feb 02 09:20:56 crc kubenswrapper[4764]: I0202 09:20:56.246961 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cbjhk\" (UniqueName: \"kubernetes.io/projected/50551dd4-71f5-4eb4-8736-ed04f0845c7e-kube-api-access-cbjhk\") pod \"openstack-operator-index-8ctp4\" (UID: \"50551dd4-71f5-4eb4-8736-ed04f0845c7e\") " pod="openstack-operators/openstack-operator-index-8ctp4" Feb 02 09:20:56 crc kubenswrapper[4764]: I0202 09:20:56.355158 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-8ctp4" Feb 02 09:20:56 crc kubenswrapper[4764]: I0202 09:20:56.741120 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-8ctp4"] Feb 02 09:20:57 crc kubenswrapper[4764]: I0202 09:20:57.016764 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-8ctp4" event={"ID":"50551dd4-71f5-4eb4-8736-ed04f0845c7e","Type":"ContainerStarted","Data":"4d2b08b02dc9b00d5c96bf36bfa00264cf3be9e69ef5ee8ee417ae24e46feaa3"} Feb 02 09:20:58 crc kubenswrapper[4764]: I0202 09:20:58.806693 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-t2pkp"] Feb 02 09:20:58 crc kubenswrapper[4764]: I0202 09:20:58.807412 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-t2pkp" podUID="cbb6142d-f15d-45d8-9b8d-cafbb96ebfbd" containerName="registry-server" containerID="cri-o://3edeff5eafe3470f3ee7441c35f3ed36b967745258f9a1ebaed0a3b37037412b" gracePeriod=2 Feb 02 09:20:59 crc kubenswrapper[4764]: I0202 09:20:59.054341 4764 generic.go:334] "Generic (PLEG): container finished" podID="cbb6142d-f15d-45d8-9b8d-cafbb96ebfbd" containerID="3edeff5eafe3470f3ee7441c35f3ed36b967745258f9a1ebaed0a3b37037412b" exitCode=0 Feb 02 09:20:59 crc kubenswrapper[4764]: I0202 09:20:59.054415 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t2pkp" event={"ID":"cbb6142d-f15d-45d8-9b8d-cafbb96ebfbd","Type":"ContainerDied","Data":"3edeff5eafe3470f3ee7441c35f3ed36b967745258f9a1ebaed0a3b37037412b"} Feb 02 09:20:59 crc kubenswrapper[4764]: I0202 09:20:59.060382 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-8ctp4" event={"ID":"50551dd4-71f5-4eb4-8736-ed04f0845c7e","Type":"ContainerStarted","Data":"eebd1d0bb316935f51279285f4d029118c988f2793b4cba21af3b647016a95df"} Feb 02 09:20:59 crc kubenswrapper[4764]: I0202 09:20:59.086061 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-8ctp4" podStartSLOduration=2.637535581 podStartE2EDuration="4.086046062s" podCreationTimestamp="2026-02-02 09:20:55 +0000 UTC" firstStartedPulling="2026-02-02 09:20:56.753599904 +0000 UTC m=+819.687323992" lastFinishedPulling="2026-02-02 09:20:58.202110375 +0000 UTC m=+821.135834473" observedRunningTime="2026-02-02 09:20:59.082490494 +0000 UTC m=+822.016214572" watchObservedRunningTime="2026-02-02 09:20:59.086046062 +0000 UTC m=+822.019770150" Feb 02 09:20:59 crc kubenswrapper[4764]: I0202 09:20:59.216304 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t2pkp" Feb 02 09:20:59 crc kubenswrapper[4764]: I0202 09:20:59.255851 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g596k\" (UniqueName: \"kubernetes.io/projected/cbb6142d-f15d-45d8-9b8d-cafbb96ebfbd-kube-api-access-g596k\") pod \"cbb6142d-f15d-45d8-9b8d-cafbb96ebfbd\" (UID: \"cbb6142d-f15d-45d8-9b8d-cafbb96ebfbd\") " Feb 02 09:20:59 crc kubenswrapper[4764]: I0202 09:20:59.255965 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cbb6142d-f15d-45d8-9b8d-cafbb96ebfbd-utilities\") pod \"cbb6142d-f15d-45d8-9b8d-cafbb96ebfbd\" (UID: \"cbb6142d-f15d-45d8-9b8d-cafbb96ebfbd\") " Feb 02 09:20:59 crc kubenswrapper[4764]: I0202 09:20:59.256003 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cbb6142d-f15d-45d8-9b8d-cafbb96ebfbd-catalog-content\") pod \"cbb6142d-f15d-45d8-9b8d-cafbb96ebfbd\" (UID: \"cbb6142d-f15d-45d8-9b8d-cafbb96ebfbd\") " Feb 02 09:20:59 crc kubenswrapper[4764]: I0202 09:20:59.257289 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cbb6142d-f15d-45d8-9b8d-cafbb96ebfbd-utilities" (OuterVolumeSpecName: "utilities") pod "cbb6142d-f15d-45d8-9b8d-cafbb96ebfbd" (UID: "cbb6142d-f15d-45d8-9b8d-cafbb96ebfbd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 09:20:59 crc kubenswrapper[4764]: I0202 09:20:59.262194 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cbb6142d-f15d-45d8-9b8d-cafbb96ebfbd-kube-api-access-g596k" (OuterVolumeSpecName: "kube-api-access-g596k") pod "cbb6142d-f15d-45d8-9b8d-cafbb96ebfbd" (UID: "cbb6142d-f15d-45d8-9b8d-cafbb96ebfbd"). InnerVolumeSpecName "kube-api-access-g596k". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:20:59 crc kubenswrapper[4764]: I0202 09:20:59.357821 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g596k\" (UniqueName: \"kubernetes.io/projected/cbb6142d-f15d-45d8-9b8d-cafbb96ebfbd-kube-api-access-g596k\") on node \"crc\" DevicePath \"\"" Feb 02 09:20:59 crc kubenswrapper[4764]: I0202 09:20:59.357867 4764 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cbb6142d-f15d-45d8-9b8d-cafbb96ebfbd-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 09:20:59 crc kubenswrapper[4764]: I0202 09:20:59.374126 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cbb6142d-f15d-45d8-9b8d-cafbb96ebfbd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cbb6142d-f15d-45d8-9b8d-cafbb96ebfbd" (UID: "cbb6142d-f15d-45d8-9b8d-cafbb96ebfbd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 09:20:59 crc kubenswrapper[4764]: I0202 09:20:59.459461 4764 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cbb6142d-f15d-45d8-9b8d-cafbb96ebfbd-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 09:21:00 crc kubenswrapper[4764]: I0202 09:21:00.075101 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t2pkp" Feb 02 09:21:00 crc kubenswrapper[4764]: I0202 09:21:00.075609 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t2pkp" event={"ID":"cbb6142d-f15d-45d8-9b8d-cafbb96ebfbd","Type":"ContainerDied","Data":"8f24c3118bc471d0982c70f762e8e21791b98cb968ec6c32bdc1f23f4a7193a6"} Feb 02 09:21:00 crc kubenswrapper[4764]: I0202 09:21:00.075670 4764 scope.go:117] "RemoveContainer" containerID="3edeff5eafe3470f3ee7441c35f3ed36b967745258f9a1ebaed0a3b37037412b" Feb 02 09:21:00 crc kubenswrapper[4764]: I0202 09:21:00.094164 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-t2pkp"] Feb 02 09:21:00 crc kubenswrapper[4764]: I0202 09:21:00.103410 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-t2pkp"] Feb 02 09:21:00 crc kubenswrapper[4764]: I0202 09:21:00.109400 4764 scope.go:117] "RemoveContainer" containerID="a0243c747887515934803d4f9cab422d5d60d3dcf0db4e51780addbccb98e133" Feb 02 09:21:00 crc kubenswrapper[4764]: I0202 09:21:00.127900 4764 scope.go:117] "RemoveContainer" containerID="aea4b8619a07fa6e9e25f19ed61b90b8b46beef9f2a2ff88504362caa72aa431" Feb 02 09:21:01 crc kubenswrapper[4764]: I0202 09:21:01.839182 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cbb6142d-f15d-45d8-9b8d-cafbb96ebfbd" path="/var/lib/kubelet/pods/cbb6142d-f15d-45d8-9b8d-cafbb96ebfbd/volumes" Feb 02 09:21:06 crc kubenswrapper[4764]: I0202 09:21:06.355996 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-8ctp4" Feb 02 09:21:06 crc kubenswrapper[4764]: I0202 09:21:06.356293 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-8ctp4" Feb 02 09:21:06 crc kubenswrapper[4764]: I0202 09:21:06.388349 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-8ctp4" Feb 02 09:21:07 crc kubenswrapper[4764]: I0202 09:21:07.159115 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-8ctp4" Feb 02 09:21:11 crc kubenswrapper[4764]: I0202 09:21:11.075793 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/805de4b7130efd4fbe2290d0c51d78e6295b83d77efc3ff2d1b014ffe7z8r74"] Feb 02 09:21:11 crc kubenswrapper[4764]: E0202 09:21:11.076671 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cbb6142d-f15d-45d8-9b8d-cafbb96ebfbd" containerName="registry-server" Feb 02 09:21:11 crc kubenswrapper[4764]: I0202 09:21:11.076690 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="cbb6142d-f15d-45d8-9b8d-cafbb96ebfbd" containerName="registry-server" Feb 02 09:21:11 crc kubenswrapper[4764]: E0202 09:21:11.076701 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cbb6142d-f15d-45d8-9b8d-cafbb96ebfbd" containerName="extract-content" Feb 02 09:21:11 crc kubenswrapper[4764]: I0202 09:21:11.076708 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="cbb6142d-f15d-45d8-9b8d-cafbb96ebfbd" containerName="extract-content" Feb 02 09:21:11 crc kubenswrapper[4764]: E0202 09:21:11.076720 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cbb6142d-f15d-45d8-9b8d-cafbb96ebfbd" containerName="extract-utilities" Feb 02 09:21:11 crc kubenswrapper[4764]: I0202 09:21:11.076731 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="cbb6142d-f15d-45d8-9b8d-cafbb96ebfbd" containerName="extract-utilities" Feb 02 09:21:11 crc kubenswrapper[4764]: I0202 09:21:11.076898 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="cbb6142d-f15d-45d8-9b8d-cafbb96ebfbd" containerName="registry-server" Feb 02 09:21:11 crc kubenswrapper[4764]: I0202 09:21:11.077916 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/805de4b7130efd4fbe2290d0c51d78e6295b83d77efc3ff2d1b014ffe7z8r74" Feb 02 09:21:11 crc kubenswrapper[4764]: I0202 09:21:11.081590 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-j5ndf" Feb 02 09:21:11 crc kubenswrapper[4764]: I0202 09:21:11.088749 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/805de4b7130efd4fbe2290d0c51d78e6295b83d77efc3ff2d1b014ffe7z8r74"] Feb 02 09:21:11 crc kubenswrapper[4764]: I0202 09:21:11.125337 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m9nhw\" (UniqueName: \"kubernetes.io/projected/894b5d6a-cddd-43e6-a59e-6076f985e5a3-kube-api-access-m9nhw\") pod \"805de4b7130efd4fbe2290d0c51d78e6295b83d77efc3ff2d1b014ffe7z8r74\" (UID: \"894b5d6a-cddd-43e6-a59e-6076f985e5a3\") " pod="openstack-operators/805de4b7130efd4fbe2290d0c51d78e6295b83d77efc3ff2d1b014ffe7z8r74" Feb 02 09:21:11 crc kubenswrapper[4764]: I0202 09:21:11.125940 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/894b5d6a-cddd-43e6-a59e-6076f985e5a3-bundle\") pod \"805de4b7130efd4fbe2290d0c51d78e6295b83d77efc3ff2d1b014ffe7z8r74\" (UID: \"894b5d6a-cddd-43e6-a59e-6076f985e5a3\") " pod="openstack-operators/805de4b7130efd4fbe2290d0c51d78e6295b83d77efc3ff2d1b014ffe7z8r74" Feb 02 09:21:11 crc kubenswrapper[4764]: I0202 09:21:11.126007 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/894b5d6a-cddd-43e6-a59e-6076f985e5a3-util\") pod \"805de4b7130efd4fbe2290d0c51d78e6295b83d77efc3ff2d1b014ffe7z8r74\" (UID: \"894b5d6a-cddd-43e6-a59e-6076f985e5a3\") " pod="openstack-operators/805de4b7130efd4fbe2290d0c51d78e6295b83d77efc3ff2d1b014ffe7z8r74" Feb 02 09:21:11 crc kubenswrapper[4764]: I0202 09:21:11.228223 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m9nhw\" (UniqueName: \"kubernetes.io/projected/894b5d6a-cddd-43e6-a59e-6076f985e5a3-kube-api-access-m9nhw\") pod \"805de4b7130efd4fbe2290d0c51d78e6295b83d77efc3ff2d1b014ffe7z8r74\" (UID: \"894b5d6a-cddd-43e6-a59e-6076f985e5a3\") " pod="openstack-operators/805de4b7130efd4fbe2290d0c51d78e6295b83d77efc3ff2d1b014ffe7z8r74" Feb 02 09:21:11 crc kubenswrapper[4764]: I0202 09:21:11.228355 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/894b5d6a-cddd-43e6-a59e-6076f985e5a3-bundle\") pod \"805de4b7130efd4fbe2290d0c51d78e6295b83d77efc3ff2d1b014ffe7z8r74\" (UID: \"894b5d6a-cddd-43e6-a59e-6076f985e5a3\") " pod="openstack-operators/805de4b7130efd4fbe2290d0c51d78e6295b83d77efc3ff2d1b014ffe7z8r74" Feb 02 09:21:11 crc kubenswrapper[4764]: I0202 09:21:11.228388 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/894b5d6a-cddd-43e6-a59e-6076f985e5a3-util\") pod \"805de4b7130efd4fbe2290d0c51d78e6295b83d77efc3ff2d1b014ffe7z8r74\" (UID: \"894b5d6a-cddd-43e6-a59e-6076f985e5a3\") " pod="openstack-operators/805de4b7130efd4fbe2290d0c51d78e6295b83d77efc3ff2d1b014ffe7z8r74" Feb 02 09:21:11 crc kubenswrapper[4764]: I0202 09:21:11.228881 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/894b5d6a-cddd-43e6-a59e-6076f985e5a3-bundle\") pod \"805de4b7130efd4fbe2290d0c51d78e6295b83d77efc3ff2d1b014ffe7z8r74\" (UID: \"894b5d6a-cddd-43e6-a59e-6076f985e5a3\") " pod="openstack-operators/805de4b7130efd4fbe2290d0c51d78e6295b83d77efc3ff2d1b014ffe7z8r74" Feb 02 09:21:11 crc kubenswrapper[4764]: I0202 09:21:11.229098 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/894b5d6a-cddd-43e6-a59e-6076f985e5a3-util\") pod \"805de4b7130efd4fbe2290d0c51d78e6295b83d77efc3ff2d1b014ffe7z8r74\" (UID: \"894b5d6a-cddd-43e6-a59e-6076f985e5a3\") " pod="openstack-operators/805de4b7130efd4fbe2290d0c51d78e6295b83d77efc3ff2d1b014ffe7z8r74" Feb 02 09:21:11 crc kubenswrapper[4764]: I0202 09:21:11.254400 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m9nhw\" (UniqueName: \"kubernetes.io/projected/894b5d6a-cddd-43e6-a59e-6076f985e5a3-kube-api-access-m9nhw\") pod \"805de4b7130efd4fbe2290d0c51d78e6295b83d77efc3ff2d1b014ffe7z8r74\" (UID: \"894b5d6a-cddd-43e6-a59e-6076f985e5a3\") " pod="openstack-operators/805de4b7130efd4fbe2290d0c51d78e6295b83d77efc3ff2d1b014ffe7z8r74" Feb 02 09:21:11 crc kubenswrapper[4764]: I0202 09:21:11.394173 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/805de4b7130efd4fbe2290d0c51d78e6295b83d77efc3ff2d1b014ffe7z8r74" Feb 02 09:21:11 crc kubenswrapper[4764]: I0202 09:21:11.688387 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/805de4b7130efd4fbe2290d0c51d78e6295b83d77efc3ff2d1b014ffe7z8r74"] Feb 02 09:21:12 crc kubenswrapper[4764]: I0202 09:21:12.167478 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/805de4b7130efd4fbe2290d0c51d78e6295b83d77efc3ff2d1b014ffe7z8r74" event={"ID":"894b5d6a-cddd-43e6-a59e-6076f985e5a3","Type":"ContainerStarted","Data":"e45df77eb45320d908e3fa959df0a43ed00611f12d6772996b62615f524e4f47"} Feb 02 09:21:12 crc kubenswrapper[4764]: I0202 09:21:12.167568 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/805de4b7130efd4fbe2290d0c51d78e6295b83d77efc3ff2d1b014ffe7z8r74" event={"ID":"894b5d6a-cddd-43e6-a59e-6076f985e5a3","Type":"ContainerStarted","Data":"24db46963e934a19c60451e2e43a43f4740bc76daa438082911b77b7311f54f9"} Feb 02 09:21:13 crc kubenswrapper[4764]: I0202 09:21:13.176705 4764 generic.go:334] "Generic (PLEG): container finished" podID="894b5d6a-cddd-43e6-a59e-6076f985e5a3" containerID="e45df77eb45320d908e3fa959df0a43ed00611f12d6772996b62615f524e4f47" exitCode=0 Feb 02 09:21:13 crc kubenswrapper[4764]: I0202 09:21:13.177293 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/805de4b7130efd4fbe2290d0c51d78e6295b83d77efc3ff2d1b014ffe7z8r74" event={"ID":"894b5d6a-cddd-43e6-a59e-6076f985e5a3","Type":"ContainerDied","Data":"e45df77eb45320d908e3fa959df0a43ed00611f12d6772996b62615f524e4f47"} Feb 02 09:21:15 crc kubenswrapper[4764]: I0202 09:21:15.200998 4764 generic.go:334] "Generic (PLEG): container finished" podID="894b5d6a-cddd-43e6-a59e-6076f985e5a3" containerID="172153c0381111610ad629e335618d2fbb0fdf822e5bacf0c8571eafbf3dec8d" exitCode=0 Feb 02 09:21:15 crc kubenswrapper[4764]: I0202 09:21:15.201082 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/805de4b7130efd4fbe2290d0c51d78e6295b83d77efc3ff2d1b014ffe7z8r74" event={"ID":"894b5d6a-cddd-43e6-a59e-6076f985e5a3","Type":"ContainerDied","Data":"172153c0381111610ad629e335618d2fbb0fdf822e5bacf0c8571eafbf3dec8d"} Feb 02 09:21:16 crc kubenswrapper[4764]: I0202 09:21:16.209099 4764 generic.go:334] "Generic (PLEG): container finished" podID="894b5d6a-cddd-43e6-a59e-6076f985e5a3" containerID="7328a6d14b2e990dc130150ad36e7b8f8d459aced4c01231f30957568988c707" exitCode=0 Feb 02 09:21:16 crc kubenswrapper[4764]: I0202 09:21:16.209147 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/805de4b7130efd4fbe2290d0c51d78e6295b83d77efc3ff2d1b014ffe7z8r74" event={"ID":"894b5d6a-cddd-43e6-a59e-6076f985e5a3","Type":"ContainerDied","Data":"7328a6d14b2e990dc130150ad36e7b8f8d459aced4c01231f30957568988c707"} Feb 02 09:21:17 crc kubenswrapper[4764]: I0202 09:21:17.478614 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/805de4b7130efd4fbe2290d0c51d78e6295b83d77efc3ff2d1b014ffe7z8r74" Feb 02 09:21:17 crc kubenswrapper[4764]: I0202 09:21:17.631412 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m9nhw\" (UniqueName: \"kubernetes.io/projected/894b5d6a-cddd-43e6-a59e-6076f985e5a3-kube-api-access-m9nhw\") pod \"894b5d6a-cddd-43e6-a59e-6076f985e5a3\" (UID: \"894b5d6a-cddd-43e6-a59e-6076f985e5a3\") " Feb 02 09:21:17 crc kubenswrapper[4764]: I0202 09:21:17.631522 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/894b5d6a-cddd-43e6-a59e-6076f985e5a3-bundle\") pod \"894b5d6a-cddd-43e6-a59e-6076f985e5a3\" (UID: \"894b5d6a-cddd-43e6-a59e-6076f985e5a3\") " Feb 02 09:21:17 crc kubenswrapper[4764]: I0202 09:21:17.631558 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/894b5d6a-cddd-43e6-a59e-6076f985e5a3-util\") pod \"894b5d6a-cddd-43e6-a59e-6076f985e5a3\" (UID: \"894b5d6a-cddd-43e6-a59e-6076f985e5a3\") " Feb 02 09:21:17 crc kubenswrapper[4764]: I0202 09:21:17.632365 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/894b5d6a-cddd-43e6-a59e-6076f985e5a3-bundle" (OuterVolumeSpecName: "bundle") pod "894b5d6a-cddd-43e6-a59e-6076f985e5a3" (UID: "894b5d6a-cddd-43e6-a59e-6076f985e5a3"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 09:21:17 crc kubenswrapper[4764]: I0202 09:21:17.637570 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/894b5d6a-cddd-43e6-a59e-6076f985e5a3-kube-api-access-m9nhw" (OuterVolumeSpecName: "kube-api-access-m9nhw") pod "894b5d6a-cddd-43e6-a59e-6076f985e5a3" (UID: "894b5d6a-cddd-43e6-a59e-6076f985e5a3"). InnerVolumeSpecName "kube-api-access-m9nhw". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:21:17 crc kubenswrapper[4764]: I0202 09:21:17.644477 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/894b5d6a-cddd-43e6-a59e-6076f985e5a3-util" (OuterVolumeSpecName: "util") pod "894b5d6a-cddd-43e6-a59e-6076f985e5a3" (UID: "894b5d6a-cddd-43e6-a59e-6076f985e5a3"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 09:21:17 crc kubenswrapper[4764]: I0202 09:21:17.732673 4764 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/894b5d6a-cddd-43e6-a59e-6076f985e5a3-util\") on node \"crc\" DevicePath \"\"" Feb 02 09:21:17 crc kubenswrapper[4764]: I0202 09:21:17.732711 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m9nhw\" (UniqueName: \"kubernetes.io/projected/894b5d6a-cddd-43e6-a59e-6076f985e5a3-kube-api-access-m9nhw\") on node \"crc\" DevicePath \"\"" Feb 02 09:21:17 crc kubenswrapper[4764]: I0202 09:21:17.732724 4764 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/894b5d6a-cddd-43e6-a59e-6076f985e5a3-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 09:21:18 crc kubenswrapper[4764]: I0202 09:21:18.228212 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/805de4b7130efd4fbe2290d0c51d78e6295b83d77efc3ff2d1b014ffe7z8r74" event={"ID":"894b5d6a-cddd-43e6-a59e-6076f985e5a3","Type":"ContainerDied","Data":"24db46963e934a19c60451e2e43a43f4740bc76daa438082911b77b7311f54f9"} Feb 02 09:21:18 crc kubenswrapper[4764]: I0202 09:21:18.228269 4764 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="24db46963e934a19c60451e2e43a43f4740bc76daa438082911b77b7311f54f9" Feb 02 09:21:18 crc kubenswrapper[4764]: I0202 09:21:18.228415 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/805de4b7130efd4fbe2290d0c51d78e6295b83d77efc3ff2d1b014ffe7z8r74" Feb 02 09:21:22 crc kubenswrapper[4764]: I0202 09:21:22.056880 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-init-6bf6665fd-l67kz"] Feb 02 09:21:22 crc kubenswrapper[4764]: E0202 09:21:22.057513 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="894b5d6a-cddd-43e6-a59e-6076f985e5a3" containerName="pull" Feb 02 09:21:22 crc kubenswrapper[4764]: I0202 09:21:22.057527 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="894b5d6a-cddd-43e6-a59e-6076f985e5a3" containerName="pull" Feb 02 09:21:22 crc kubenswrapper[4764]: E0202 09:21:22.057537 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="894b5d6a-cddd-43e6-a59e-6076f985e5a3" containerName="extract" Feb 02 09:21:22 crc kubenswrapper[4764]: I0202 09:21:22.057543 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="894b5d6a-cddd-43e6-a59e-6076f985e5a3" containerName="extract" Feb 02 09:21:22 crc kubenswrapper[4764]: E0202 09:21:22.057552 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="894b5d6a-cddd-43e6-a59e-6076f985e5a3" containerName="util" Feb 02 09:21:22 crc kubenswrapper[4764]: I0202 09:21:22.057558 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="894b5d6a-cddd-43e6-a59e-6076f985e5a3" containerName="util" Feb 02 09:21:22 crc kubenswrapper[4764]: I0202 09:21:22.057651 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="894b5d6a-cddd-43e6-a59e-6076f985e5a3" containerName="extract" Feb 02 09:21:22 crc kubenswrapper[4764]: I0202 09:21:22.058080 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-init-6bf6665fd-l67kz" Feb 02 09:21:22 crc kubenswrapper[4764]: I0202 09:21:22.061666 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-init-dockercfg-kfg9f" Feb 02 09:21:22 crc kubenswrapper[4764]: I0202 09:21:22.077767 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-init-6bf6665fd-l67kz"] Feb 02 09:21:22 crc kubenswrapper[4764]: I0202 09:21:22.196213 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z87pz\" (UniqueName: \"kubernetes.io/projected/ee6621b4-2d57-402e-bfda-df7f7f922dc8-kube-api-access-z87pz\") pod \"openstack-operator-controller-init-6bf6665fd-l67kz\" (UID: \"ee6621b4-2d57-402e-bfda-df7f7f922dc8\") " pod="openstack-operators/openstack-operator-controller-init-6bf6665fd-l67kz" Feb 02 09:21:22 crc kubenswrapper[4764]: I0202 09:21:22.297463 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z87pz\" (UniqueName: \"kubernetes.io/projected/ee6621b4-2d57-402e-bfda-df7f7f922dc8-kube-api-access-z87pz\") pod \"openstack-operator-controller-init-6bf6665fd-l67kz\" (UID: \"ee6621b4-2d57-402e-bfda-df7f7f922dc8\") " pod="openstack-operators/openstack-operator-controller-init-6bf6665fd-l67kz" Feb 02 09:21:22 crc kubenswrapper[4764]: I0202 09:21:22.327291 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z87pz\" (UniqueName: \"kubernetes.io/projected/ee6621b4-2d57-402e-bfda-df7f7f922dc8-kube-api-access-z87pz\") pod \"openstack-operator-controller-init-6bf6665fd-l67kz\" (UID: \"ee6621b4-2d57-402e-bfda-df7f7f922dc8\") " pod="openstack-operators/openstack-operator-controller-init-6bf6665fd-l67kz" Feb 02 09:21:22 crc kubenswrapper[4764]: I0202 09:21:22.374533 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-init-6bf6665fd-l67kz" Feb 02 09:21:22 crc kubenswrapper[4764]: I0202 09:21:22.770614 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-init-6bf6665fd-l67kz"] Feb 02 09:21:23 crc kubenswrapper[4764]: I0202 09:21:23.261762 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-init-6bf6665fd-l67kz" event={"ID":"ee6621b4-2d57-402e-bfda-df7f7f922dc8","Type":"ContainerStarted","Data":"cfed13f48f141cadbd63bd74738c2bff536cf220f3799e7dd348b0ed6f604811"} Feb 02 09:21:29 crc kubenswrapper[4764]: I0202 09:21:29.308815 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-init-6bf6665fd-l67kz" event={"ID":"ee6621b4-2d57-402e-bfda-df7f7f922dc8","Type":"ContainerStarted","Data":"251efb4e657aa641c0e4cdfb50a93135ac12986ed1040d386cdabedcad9cffdf"} Feb 02 09:21:29 crc kubenswrapper[4764]: I0202 09:21:29.310373 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-init-6bf6665fd-l67kz" Feb 02 09:21:29 crc kubenswrapper[4764]: I0202 09:21:29.370090 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-init-6bf6665fd-l67kz" podStartSLOduration=1.901746752 podStartE2EDuration="7.370065429s" podCreationTimestamp="2026-02-02 09:21:22 +0000 UTC" firstStartedPulling="2026-02-02 09:21:22.773237981 +0000 UTC m=+845.706962069" lastFinishedPulling="2026-02-02 09:21:28.241556658 +0000 UTC m=+851.175280746" observedRunningTime="2026-02-02 09:21:29.341871624 +0000 UTC m=+852.275595752" watchObservedRunningTime="2026-02-02 09:21:29.370065429 +0000 UTC m=+852.303789527" Feb 02 09:21:42 crc kubenswrapper[4764]: I0202 09:21:42.381566 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-init-6bf6665fd-l67kz" Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.039998 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7b6c4d8c5f-wtkdr"] Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.041211 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7b6c4d8c5f-wtkdr" Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.046033 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-gjxsn" Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.053055 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-8d874c8fc-czpmn"] Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.054006 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-8d874c8fc-czpmn" Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.055555 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-79fw6" Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.066501 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7b6c4d8c5f-wtkdr"] Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.071430 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-6d9697b7f4-w4qf5"] Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.072309 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-6d9697b7f4-w4qf5" Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.074252 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-wttnj" Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.086330 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-8886f4c47-hvtx6"] Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.087286 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-8886f4c47-hvtx6" Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.106129 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-phgjn" Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.112042 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pqvfw\" (UniqueName: \"kubernetes.io/projected/9cbd726b-9dbb-46b3-9bcd-f3e9813b4f07-kube-api-access-pqvfw\") pod \"barbican-operator-controller-manager-7b6c4d8c5f-wtkdr\" (UID: \"9cbd726b-9dbb-46b3-9bcd-f3e9813b4f07\") " pod="openstack-operators/barbican-operator-controller-manager-7b6c4d8c5f-wtkdr" Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.112129 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jxd7n\" (UniqueName: \"kubernetes.io/projected/7c8a2462-306c-43da-a538-9e34a11b5c21-kube-api-access-jxd7n\") pod \"cinder-operator-controller-manager-8d874c8fc-czpmn\" (UID: \"7c8a2462-306c-43da-a538-9e34a11b5c21\") " pod="openstack-operators/cinder-operator-controller-manager-8d874c8fc-czpmn" Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.122785 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-8d874c8fc-czpmn"] Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.145078 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-6d9697b7f4-w4qf5"] Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.199980 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-69d6db494d-5f766"] Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.200777 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-69d6db494d-5f766" Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.205780 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-j9bpq" Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.208617 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-69d6db494d-5f766"] Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.213992 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jxd7n\" (UniqueName: \"kubernetes.io/projected/7c8a2462-306c-43da-a538-9e34a11b5c21-kube-api-access-jxd7n\") pod \"cinder-operator-controller-manager-8d874c8fc-czpmn\" (UID: \"7c8a2462-306c-43da-a538-9e34a11b5c21\") " pod="openstack-operators/cinder-operator-controller-manager-8d874c8fc-czpmn" Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.214044 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qlmjs\" (UniqueName: \"kubernetes.io/projected/8b0e7cb4-257b-403b-8b92-4578ce77c388-kube-api-access-qlmjs\") pod \"designate-operator-controller-manager-6d9697b7f4-w4qf5\" (UID: \"8b0e7cb4-257b-403b-8b92-4578ce77c388\") " pod="openstack-operators/designate-operator-controller-manager-6d9697b7f4-w4qf5" Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.214085 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pqvfw\" (UniqueName: \"kubernetes.io/projected/9cbd726b-9dbb-46b3-9bcd-f3e9813b4f07-kube-api-access-pqvfw\") pod \"barbican-operator-controller-manager-7b6c4d8c5f-wtkdr\" (UID: \"9cbd726b-9dbb-46b3-9bcd-f3e9813b4f07\") " pod="openstack-operators/barbican-operator-controller-manager-7b6c4d8c5f-wtkdr" Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.214132 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2nmj8\" (UniqueName: \"kubernetes.io/projected/ee5cb41c-4d3d-4a9d-b1f7-1ede61e7b21c-kube-api-access-2nmj8\") pod \"glance-operator-controller-manager-8886f4c47-hvtx6\" (UID: \"ee5cb41c-4d3d-4a9d-b1f7-1ede61e7b21c\") " pod="openstack-operators/glance-operator-controller-manager-8886f4c47-hvtx6" Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.262334 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-5fb775575f-ntj47"] Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.263155 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-5fb775575f-ntj47" Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.265548 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-k86bs" Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.265988 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pqvfw\" (UniqueName: \"kubernetes.io/projected/9cbd726b-9dbb-46b3-9bcd-f3e9813b4f07-kube-api-access-pqvfw\") pod \"barbican-operator-controller-manager-7b6c4d8c5f-wtkdr\" (UID: \"9cbd726b-9dbb-46b3-9bcd-f3e9813b4f07\") " pod="openstack-operators/barbican-operator-controller-manager-7b6c4d8c5f-wtkdr" Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.273789 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jxd7n\" (UniqueName: \"kubernetes.io/projected/7c8a2462-306c-43da-a538-9e34a11b5c21-kube-api-access-jxd7n\") pod \"cinder-operator-controller-manager-8d874c8fc-czpmn\" (UID: \"7c8a2462-306c-43da-a538-9e34a11b5c21\") " pod="openstack-operators/cinder-operator-controller-manager-8d874c8fc-czpmn" Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.279975 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-5fb775575f-ntj47"] Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.296484 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-79955696d6-mxvzm"] Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.297281 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-79955696d6-mxvzm" Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.301615 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.301715 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-f96wx" Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.303997 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-79955696d6-mxvzm"] Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.315966 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2nmj8\" (UniqueName: \"kubernetes.io/projected/ee5cb41c-4d3d-4a9d-b1f7-1ede61e7b21c-kube-api-access-2nmj8\") pod \"glance-operator-controller-manager-8886f4c47-hvtx6\" (UID: \"ee5cb41c-4d3d-4a9d-b1f7-1ede61e7b21c\") " pod="openstack-operators/glance-operator-controller-manager-8886f4c47-hvtx6" Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.316019 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cq8gf\" (UniqueName: \"kubernetes.io/projected/becd2ba3-ccad-4ea9-b3c7-6b39f6f8cb3d-kube-api-access-cq8gf\") pod \"heat-operator-controller-manager-69d6db494d-5f766\" (UID: \"becd2ba3-ccad-4ea9-b3c7-6b39f6f8cb3d\") " pod="openstack-operators/heat-operator-controller-manager-69d6db494d-5f766" Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.316060 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qlmjs\" (UniqueName: \"kubernetes.io/projected/8b0e7cb4-257b-403b-8b92-4578ce77c388-kube-api-access-qlmjs\") pod \"designate-operator-controller-manager-6d9697b7f4-w4qf5\" (UID: \"8b0e7cb4-257b-403b-8b92-4578ce77c388\") " pod="openstack-operators/designate-operator-controller-manager-6d9697b7f4-w4qf5" Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.316120 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hdtzz\" (UniqueName: \"kubernetes.io/projected/82e16b73-9dfe-413b-b3dd-65da79b3d9dd-kube-api-access-hdtzz\") pod \"horizon-operator-controller-manager-5fb775575f-ntj47\" (UID: \"82e16b73-9dfe-413b-b3dd-65da79b3d9dd\") " pod="openstack-operators/horizon-operator-controller-manager-5fb775575f-ntj47" Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.333250 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-5f4b8bd54d-h4n52"] Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.334227 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-5f4b8bd54d-h4n52" Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.340626 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-p2vcf" Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.348764 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-8886f4c47-hvtx6"] Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.349279 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2nmj8\" (UniqueName: \"kubernetes.io/projected/ee5cb41c-4d3d-4a9d-b1f7-1ede61e7b21c-kube-api-access-2nmj8\") pod \"glance-operator-controller-manager-8886f4c47-hvtx6\" (UID: \"ee5cb41c-4d3d-4a9d-b1f7-1ede61e7b21c\") " pod="openstack-operators/glance-operator-controller-manager-8886f4c47-hvtx6" Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.366589 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qlmjs\" (UniqueName: \"kubernetes.io/projected/8b0e7cb4-257b-403b-8b92-4578ce77c388-kube-api-access-qlmjs\") pod \"designate-operator-controller-manager-6d9697b7f4-w4qf5\" (UID: \"8b0e7cb4-257b-403b-8b92-4578ce77c388\") " pod="openstack-operators/designate-operator-controller-manager-6d9697b7f4-w4qf5" Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.366971 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7b6c4d8c5f-wtkdr" Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.381803 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-8d874c8fc-czpmn" Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.389545 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-84f48565d4-qqnp6"] Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.390340 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-84f48565d4-qqnp6" Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.404981 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-lfpmd" Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.407200 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-6d9697b7f4-w4qf5" Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.416205 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-5f4b8bd54d-h4n52"] Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.416798 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cq8gf\" (UniqueName: \"kubernetes.io/projected/becd2ba3-ccad-4ea9-b3c7-6b39f6f8cb3d-kube-api-access-cq8gf\") pod \"heat-operator-controller-manager-69d6db494d-5f766\" (UID: \"becd2ba3-ccad-4ea9-b3c7-6b39f6f8cb3d\") " pod="openstack-operators/heat-operator-controller-manager-69d6db494d-5f766" Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.416827 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6sjlh\" (UniqueName: \"kubernetes.io/projected/b593c3f3-a225-4cb6-b967-4c056e3326a0-kube-api-access-6sjlh\") pod \"infra-operator-controller-manager-79955696d6-mxvzm\" (UID: \"b593c3f3-a225-4cb6-b967-4c056e3326a0\") " pod="openstack-operators/infra-operator-controller-manager-79955696d6-mxvzm" Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.416883 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b593c3f3-a225-4cb6-b967-4c056e3326a0-cert\") pod \"infra-operator-controller-manager-79955696d6-mxvzm\" (UID: \"b593c3f3-a225-4cb6-b967-4c056e3326a0\") " pod="openstack-operators/infra-operator-controller-manager-79955696d6-mxvzm" Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.416914 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hpfm5\" (UniqueName: \"kubernetes.io/projected/8564e79b-9abd-41e1-a0a7-0d4b71011300-kube-api-access-hpfm5\") pod \"keystone-operator-controller-manager-84f48565d4-qqnp6\" (UID: \"8564e79b-9abd-41e1-a0a7-0d4b71011300\") " pod="openstack-operators/keystone-operator-controller-manager-84f48565d4-qqnp6" Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.416948 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nxd9k\" (UniqueName: \"kubernetes.io/projected/ad8e7100-d5b8-4a56-bbda-056483f79113-kube-api-access-nxd9k\") pod \"ironic-operator-controller-manager-5f4b8bd54d-h4n52\" (UID: \"ad8e7100-d5b8-4a56-bbda-056483f79113\") " pod="openstack-operators/ironic-operator-controller-manager-5f4b8bd54d-h4n52" Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.416973 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hdtzz\" (UniqueName: \"kubernetes.io/projected/82e16b73-9dfe-413b-b3dd-65da79b3d9dd-kube-api-access-hdtzz\") pod \"horizon-operator-controller-manager-5fb775575f-ntj47\" (UID: \"82e16b73-9dfe-413b-b3dd-65da79b3d9dd\") " pod="openstack-operators/horizon-operator-controller-manager-5fb775575f-ntj47" Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.429019 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-7dd968899f-g2p6c"] Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.430009 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-7dd968899f-g2p6c" Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.460485 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-7jw7s" Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.463355 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-8886f4c47-hvtx6" Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.472162 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hdtzz\" (UniqueName: \"kubernetes.io/projected/82e16b73-9dfe-413b-b3dd-65da79b3d9dd-kube-api-access-hdtzz\") pod \"horizon-operator-controller-manager-5fb775575f-ntj47\" (UID: \"82e16b73-9dfe-413b-b3dd-65da79b3d9dd\") " pod="openstack-operators/horizon-operator-controller-manager-5fb775575f-ntj47" Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.479890 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cq8gf\" (UniqueName: \"kubernetes.io/projected/becd2ba3-ccad-4ea9-b3c7-6b39f6f8cb3d-kube-api-access-cq8gf\") pod \"heat-operator-controller-manager-69d6db494d-5f766\" (UID: \"becd2ba3-ccad-4ea9-b3c7-6b39f6f8cb3d\") " pod="openstack-operators/heat-operator-controller-manager-69d6db494d-5f766" Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.479973 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-84f48565d4-qqnp6"] Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.509067 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-7dd968899f-g2p6c"] Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.520688 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-67bf948998-jhbzc"] Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.521492 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-67bf948998-jhbzc" Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.522595 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nxd9k\" (UniqueName: \"kubernetes.io/projected/ad8e7100-d5b8-4a56-bbda-056483f79113-kube-api-access-nxd9k\") pod \"ironic-operator-controller-manager-5f4b8bd54d-h4n52\" (UID: \"ad8e7100-d5b8-4a56-bbda-056483f79113\") " pod="openstack-operators/ironic-operator-controller-manager-5f4b8bd54d-h4n52" Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.522644 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6sjlh\" (UniqueName: \"kubernetes.io/projected/b593c3f3-a225-4cb6-b967-4c056e3326a0-kube-api-access-6sjlh\") pod \"infra-operator-controller-manager-79955696d6-mxvzm\" (UID: \"b593c3f3-a225-4cb6-b967-4c056e3326a0\") " pod="openstack-operators/infra-operator-controller-manager-79955696d6-mxvzm" Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.522677 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wwxnc\" (UniqueName: \"kubernetes.io/projected/6d4fffad-a306-4f0b-ac40-7fa839f75aa3-kube-api-access-wwxnc\") pod \"manila-operator-controller-manager-7dd968899f-g2p6c\" (UID: \"6d4fffad-a306-4f0b-ac40-7fa839f75aa3\") " pod="openstack-operators/manila-operator-controller-manager-7dd968899f-g2p6c" Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.522713 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b593c3f3-a225-4cb6-b967-4c056e3326a0-cert\") pod \"infra-operator-controller-manager-79955696d6-mxvzm\" (UID: \"b593c3f3-a225-4cb6-b967-4c056e3326a0\") " pod="openstack-operators/infra-operator-controller-manager-79955696d6-mxvzm" Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.522737 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hpfm5\" (UniqueName: \"kubernetes.io/projected/8564e79b-9abd-41e1-a0a7-0d4b71011300-kube-api-access-hpfm5\") pod \"keystone-operator-controller-manager-84f48565d4-qqnp6\" (UID: \"8564e79b-9abd-41e1-a0a7-0d4b71011300\") " pod="openstack-operators/keystone-operator-controller-manager-84f48565d4-qqnp6" Feb 02 09:22:02 crc kubenswrapper[4764]: E0202 09:22:02.527440 4764 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Feb 02 09:22:02 crc kubenswrapper[4764]: E0202 09:22:02.527494 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b593c3f3-a225-4cb6-b967-4c056e3326a0-cert podName:b593c3f3-a225-4cb6-b967-4c056e3326a0 nodeName:}" failed. No retries permitted until 2026-02-02 09:22:03.02747892 +0000 UTC m=+885.961203008 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/b593c3f3-a225-4cb6-b967-4c056e3326a0-cert") pod "infra-operator-controller-manager-79955696d6-mxvzm" (UID: "b593c3f3-a225-4cb6-b967-4c056e3326a0") : secret "infra-operator-webhook-server-cert" not found Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.527848 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-qdrxq" Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.528212 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-69d6db494d-5f766" Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.561624 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-585dbc889-7nrsg"] Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.562676 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-585dbc889-7nrsg" Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.576036 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-67bf948998-jhbzc"] Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.580782 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-jjs7g" Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.592583 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6sjlh\" (UniqueName: \"kubernetes.io/projected/b593c3f3-a225-4cb6-b967-4c056e3326a0-kube-api-access-6sjlh\") pod \"infra-operator-controller-manager-79955696d6-mxvzm\" (UID: \"b593c3f3-a225-4cb6-b967-4c056e3326a0\") " pod="openstack-operators/infra-operator-controller-manager-79955696d6-mxvzm" Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.599282 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hpfm5\" (UniqueName: \"kubernetes.io/projected/8564e79b-9abd-41e1-a0a7-0d4b71011300-kube-api-access-hpfm5\") pod \"keystone-operator-controller-manager-84f48565d4-qqnp6\" (UID: \"8564e79b-9abd-41e1-a0a7-0d4b71011300\") " pod="openstack-operators/keystone-operator-controller-manager-84f48565d4-qqnp6" Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.599540 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nxd9k\" (UniqueName: \"kubernetes.io/projected/ad8e7100-d5b8-4a56-bbda-056483f79113-kube-api-access-nxd9k\") pod \"ironic-operator-controller-manager-5f4b8bd54d-h4n52\" (UID: \"ad8e7100-d5b8-4a56-bbda-056483f79113\") " pod="openstack-operators/ironic-operator-controller-manager-5f4b8bd54d-h4n52" Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.624082 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wwxnc\" (UniqueName: \"kubernetes.io/projected/6d4fffad-a306-4f0b-ac40-7fa839f75aa3-kube-api-access-wwxnc\") pod \"manila-operator-controller-manager-7dd968899f-g2p6c\" (UID: \"6d4fffad-a306-4f0b-ac40-7fa839f75aa3\") " pod="openstack-operators/manila-operator-controller-manager-7dd968899f-g2p6c" Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.624465 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dmclv\" (UniqueName: \"kubernetes.io/projected/f5cc9236-c70f-4449-a902-ef416aaf8781-kube-api-access-dmclv\") pod \"neutron-operator-controller-manager-585dbc889-7nrsg\" (UID: \"f5cc9236-c70f-4449-a902-ef416aaf8781\") " pod="openstack-operators/neutron-operator-controller-manager-585dbc889-7nrsg" Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.624515 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t2srq\" (UniqueName: \"kubernetes.io/projected/9e55cb1e-ed3b-4b7b-8b52-42436eb48223-kube-api-access-t2srq\") pod \"mariadb-operator-controller-manager-67bf948998-jhbzc\" (UID: \"9e55cb1e-ed3b-4b7b-8b52-42436eb48223\") " pod="openstack-operators/mariadb-operator-controller-manager-67bf948998-jhbzc" Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.625026 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-5fb775575f-ntj47" Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.627841 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-585dbc889-7nrsg"] Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.635849 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-55bff696bd-p7nmb"] Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.636608 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-55bff696bd-p7nmb" Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.649689 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-g9nr6" Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.650088 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-55bff696bd-p7nmb"] Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.654110 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-6687f8d877-5ncb8"] Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.654882 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-6687f8d877-5ncb8" Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.663995 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-6687f8d877-5ncb8"] Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.677464 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-8xd4b" Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.689721 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-788c46999f-pk2hh"] Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.690410 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-788c46999f-pk2hh" Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.707080 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-5f4b8bd54d-h4n52" Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.707524 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-f5xsb" Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.722604 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wwxnc\" (UniqueName: \"kubernetes.io/projected/6d4fffad-a306-4f0b-ac40-7fa839f75aa3-kube-api-access-wwxnc\") pod \"manila-operator-controller-manager-7dd968899f-g2p6c\" (UID: \"6d4fffad-a306-4f0b-ac40-7fa839f75aa3\") " pod="openstack-operators/manila-operator-controller-manager-7dd968899f-g2p6c" Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.725723 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-788c46999f-pk2hh"] Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.726880 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dmclv\" (UniqueName: \"kubernetes.io/projected/f5cc9236-c70f-4449-a902-ef416aaf8781-kube-api-access-dmclv\") pod \"neutron-operator-controller-manager-585dbc889-7nrsg\" (UID: \"f5cc9236-c70f-4449-a902-ef416aaf8781\") " pod="openstack-operators/neutron-operator-controller-manager-585dbc889-7nrsg" Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.728886 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t2srq\" (UniqueName: \"kubernetes.io/projected/9e55cb1e-ed3b-4b7b-8b52-42436eb48223-kube-api-access-t2srq\") pod \"mariadb-operator-controller-manager-67bf948998-jhbzc\" (UID: \"9e55cb1e-ed3b-4b7b-8b52-42436eb48223\") " pod="openstack-operators/mariadb-operator-controller-manager-67bf948998-jhbzc" Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.729077 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nphdp\" (UniqueName: \"kubernetes.io/projected/e98e75d4-963f-4a67-894e-302b7429eb01-kube-api-access-nphdp\") pod \"nova-operator-controller-manager-55bff696bd-p7nmb\" (UID: \"e98e75d4-963f-4a67-894e-302b7429eb01\") " pod="openstack-operators/nova-operator-controller-manager-55bff696bd-p7nmb" Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.729193 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ppzmn\" (UniqueName: \"kubernetes.io/projected/1844ef02-28fb-45ce-a697-54f1560c9087-kube-api-access-ppzmn\") pod \"octavia-operator-controller-manager-6687f8d877-5ncb8\" (UID: \"1844ef02-28fb-45ce-a697-54f1560c9087\") " pod="openstack-operators/octavia-operator-controller-manager-6687f8d877-5ncb8" Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.729299 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mps2g\" (UniqueName: \"kubernetes.io/projected/674c8820-f063-487b-9122-1ed46d1bc39f-kube-api-access-mps2g\") pod \"ovn-operator-controller-manager-788c46999f-pk2hh\" (UID: \"674c8820-f063-487b-9122-1ed46d1bc39f\") " pod="openstack-operators/ovn-operator-controller-manager-788c46999f-pk2hh" Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.757918 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-5ff45dfdbfb5fxl"] Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.760133 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5ff45dfdbfb5fxl" Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.766368 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-xjjq7" Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.766528 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.768590 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t2srq\" (UniqueName: \"kubernetes.io/projected/9e55cb1e-ed3b-4b7b-8b52-42436eb48223-kube-api-access-t2srq\") pod \"mariadb-operator-controller-manager-67bf948998-jhbzc\" (UID: \"9e55cb1e-ed3b-4b7b-8b52-42436eb48223\") " pod="openstack-operators/mariadb-operator-controller-manager-67bf948998-jhbzc" Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.801858 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-84f48565d4-qqnp6" Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.829817 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dmclv\" (UniqueName: \"kubernetes.io/projected/f5cc9236-c70f-4449-a902-ef416aaf8781-kube-api-access-dmclv\") pod \"neutron-operator-controller-manager-585dbc889-7nrsg\" (UID: \"f5cc9236-c70f-4449-a902-ef416aaf8781\") " pod="openstack-operators/neutron-operator-controller-manager-585dbc889-7nrsg" Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.835507 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nphdp\" (UniqueName: \"kubernetes.io/projected/e98e75d4-963f-4a67-894e-302b7429eb01-kube-api-access-nphdp\") pod \"nova-operator-controller-manager-55bff696bd-p7nmb\" (UID: \"e98e75d4-963f-4a67-894e-302b7429eb01\") " pod="openstack-operators/nova-operator-controller-manager-55bff696bd-p7nmb" Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.835743 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ppzmn\" (UniqueName: \"kubernetes.io/projected/1844ef02-28fb-45ce-a697-54f1560c9087-kube-api-access-ppzmn\") pod \"octavia-operator-controller-manager-6687f8d877-5ncb8\" (UID: \"1844ef02-28fb-45ce-a697-54f1560c9087\") " pod="openstack-operators/octavia-operator-controller-manager-6687f8d877-5ncb8" Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.835942 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mps2g\" (UniqueName: \"kubernetes.io/projected/674c8820-f063-487b-9122-1ed46d1bc39f-kube-api-access-mps2g\") pod \"ovn-operator-controller-manager-788c46999f-pk2hh\" (UID: \"674c8820-f063-487b-9122-1ed46d1bc39f\") " pod="openstack-operators/ovn-operator-controller-manager-788c46999f-pk2hh" Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.855639 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-7dd968899f-g2p6c" Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.871608 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-5b964cf4cd-kskcp"] Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.873183 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-5b964cf4cd-kskcp" Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.892524 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-54z5q" Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.933756 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-67bf948998-jhbzc" Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.968858 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-585dbc889-7nrsg" Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.972527 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zvvzx\" (UniqueName: \"kubernetes.io/projected/1464055e-4745-45e3-be46-5b7700dce733-kube-api-access-zvvzx\") pod \"placement-operator-controller-manager-5b964cf4cd-kskcp\" (UID: \"1464055e-4745-45e3-be46-5b7700dce733\") " pod="openstack-operators/placement-operator-controller-manager-5b964cf4cd-kskcp" Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.972558 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1b05c36c-dbb6-4ad4-8985-8d64551812ff-cert\") pod \"openstack-baremetal-operator-controller-manager-5ff45dfdbfb5fxl\" (UID: \"1b05c36c-dbb6-4ad4-8985-8d64551812ff\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5ff45dfdbfb5fxl" Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.972620 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rlddz\" (UniqueName: \"kubernetes.io/projected/1b05c36c-dbb6-4ad4-8985-8d64551812ff-kube-api-access-rlddz\") pod \"openstack-baremetal-operator-controller-manager-5ff45dfdbfb5fxl\" (UID: \"1b05c36c-dbb6-4ad4-8985-8d64551812ff\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5ff45dfdbfb5fxl" Feb 02 09:22:02 crc kubenswrapper[4764]: I0202 09:22:02.988658 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ppzmn\" (UniqueName: \"kubernetes.io/projected/1844ef02-28fb-45ce-a697-54f1560c9087-kube-api-access-ppzmn\") pod \"octavia-operator-controller-manager-6687f8d877-5ncb8\" (UID: \"1844ef02-28fb-45ce-a697-54f1560c9087\") " pod="openstack-operators/octavia-operator-controller-manager-6687f8d877-5ncb8" Feb 02 09:22:03 crc kubenswrapper[4764]: I0202 09:22:03.013678 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nphdp\" (UniqueName: \"kubernetes.io/projected/e98e75d4-963f-4a67-894e-302b7429eb01-kube-api-access-nphdp\") pod \"nova-operator-controller-manager-55bff696bd-p7nmb\" (UID: \"e98e75d4-963f-4a67-894e-302b7429eb01\") " pod="openstack-operators/nova-operator-controller-manager-55bff696bd-p7nmb" Feb 02 09:22:03 crc kubenswrapper[4764]: I0202 09:22:03.014144 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mps2g\" (UniqueName: \"kubernetes.io/projected/674c8820-f063-487b-9122-1ed46d1bc39f-kube-api-access-mps2g\") pod \"ovn-operator-controller-manager-788c46999f-pk2hh\" (UID: \"674c8820-f063-487b-9122-1ed46d1bc39f\") " pod="openstack-operators/ovn-operator-controller-manager-788c46999f-pk2hh" Feb 02 09:22:03 crc kubenswrapper[4764]: I0202 09:22:02.999061 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-64b5b76f97-98xsk"] Feb 02 09:22:03 crc kubenswrapper[4764]: I0202 09:22:03.021721 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-55bff696bd-p7nmb" Feb 02 09:22:03 crc kubenswrapper[4764]: I0202 09:22:03.022038 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-68fc8c869-fb6zn"] Feb 02 09:22:03 crc kubenswrapper[4764]: I0202 09:22:03.022416 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-64b5b76f97-98xsk" Feb 02 09:22:03 crc kubenswrapper[4764]: I0202 09:22:03.092872 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-556ww" Feb 02 09:22:03 crc kubenswrapper[4764]: I0202 09:22:03.095330 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-68fc8c869-fb6zn" Feb 02 09:22:03 crc kubenswrapper[4764]: I0202 09:22:03.096276 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zvvzx\" (UniqueName: \"kubernetes.io/projected/1464055e-4745-45e3-be46-5b7700dce733-kube-api-access-zvvzx\") pod \"placement-operator-controller-manager-5b964cf4cd-kskcp\" (UID: \"1464055e-4745-45e3-be46-5b7700dce733\") " pod="openstack-operators/placement-operator-controller-manager-5b964cf4cd-kskcp" Feb 02 09:22:03 crc kubenswrapper[4764]: I0202 09:22:03.096301 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1b05c36c-dbb6-4ad4-8985-8d64551812ff-cert\") pod \"openstack-baremetal-operator-controller-manager-5ff45dfdbfb5fxl\" (UID: \"1b05c36c-dbb6-4ad4-8985-8d64551812ff\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5ff45dfdbfb5fxl" Feb 02 09:22:03 crc kubenswrapper[4764]: I0202 09:22:03.096351 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rlddz\" (UniqueName: \"kubernetes.io/projected/1b05c36c-dbb6-4ad4-8985-8d64551812ff-kube-api-access-rlddz\") pod \"openstack-baremetal-operator-controller-manager-5ff45dfdbfb5fxl\" (UID: \"1b05c36c-dbb6-4ad4-8985-8d64551812ff\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5ff45dfdbfb5fxl" Feb 02 09:22:03 crc kubenswrapper[4764]: I0202 09:22:03.096401 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b593c3f3-a225-4cb6-b967-4c056e3326a0-cert\") pod \"infra-operator-controller-manager-79955696d6-mxvzm\" (UID: \"b593c3f3-a225-4cb6-b967-4c056e3326a0\") " pod="openstack-operators/infra-operator-controller-manager-79955696d6-mxvzm" Feb 02 09:22:03 crc kubenswrapper[4764]: E0202 09:22:03.096506 4764 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Feb 02 09:22:03 crc kubenswrapper[4764]: E0202 09:22:03.096547 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b593c3f3-a225-4cb6-b967-4c056e3326a0-cert podName:b593c3f3-a225-4cb6-b967-4c056e3326a0 nodeName:}" failed. No retries permitted until 2026-02-02 09:22:04.096533858 +0000 UTC m=+887.030257946 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/b593c3f3-a225-4cb6-b967-4c056e3326a0-cert") pod "infra-operator-controller-manager-79955696d6-mxvzm" (UID: "b593c3f3-a225-4cb6-b967-4c056e3326a0") : secret "infra-operator-webhook-server-cert" not found Feb 02 09:22:03 crc kubenswrapper[4764]: I0202 09:22:03.098847 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-6687f8d877-5ncb8" Feb 02 09:22:03 crc kubenswrapper[4764]: E0202 09:22:03.102849 4764 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 02 09:22:03 crc kubenswrapper[4764]: E0202 09:22:03.109116 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1b05c36c-dbb6-4ad4-8985-8d64551812ff-cert podName:1b05c36c-dbb6-4ad4-8985-8d64551812ff nodeName:}" failed. No retries permitted until 2026-02-02 09:22:03.609077883 +0000 UTC m=+886.542801971 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/1b05c36c-dbb6-4ad4-8985-8d64551812ff-cert") pod "openstack-baremetal-operator-controller-manager-5ff45dfdbfb5fxl" (UID: "1b05c36c-dbb6-4ad4-8985-8d64551812ff") : secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 02 09:22:03 crc kubenswrapper[4764]: I0202 09:22:03.103349 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-788c46999f-pk2hh" Feb 02 09:22:03 crc kubenswrapper[4764]: I0202 09:22:03.110194 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-fz7qq" Feb 02 09:22:03 crc kubenswrapper[4764]: I0202 09:22:03.117745 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-5ff45dfdbfb5fxl"] Feb 02 09:22:03 crc kubenswrapper[4764]: I0202 09:22:03.177626 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rlddz\" (UniqueName: \"kubernetes.io/projected/1b05c36c-dbb6-4ad4-8985-8d64551812ff-kube-api-access-rlddz\") pod \"openstack-baremetal-operator-controller-manager-5ff45dfdbfb5fxl\" (UID: \"1b05c36c-dbb6-4ad4-8985-8d64551812ff\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5ff45dfdbfb5fxl" Feb 02 09:22:03 crc kubenswrapper[4764]: I0202 09:22:03.201456 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zvvzx\" (UniqueName: \"kubernetes.io/projected/1464055e-4745-45e3-be46-5b7700dce733-kube-api-access-zvvzx\") pod \"placement-operator-controller-manager-5b964cf4cd-kskcp\" (UID: \"1464055e-4745-45e3-be46-5b7700dce733\") " pod="openstack-operators/placement-operator-controller-manager-5b964cf4cd-kskcp" Feb 02 09:22:03 crc kubenswrapper[4764]: I0202 09:22:03.204038 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8nfld\" (UniqueName: \"kubernetes.io/projected/1a100291-702c-4908-8a63-acb05c4ef5e9-kube-api-access-8nfld\") pod \"swift-operator-controller-manager-68fc8c869-fb6zn\" (UID: \"1a100291-702c-4908-8a63-acb05c4ef5e9\") " pod="openstack-operators/swift-operator-controller-manager-68fc8c869-fb6zn" Feb 02 09:22:03 crc kubenswrapper[4764]: I0202 09:22:03.204108 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vtxrr\" (UniqueName: \"kubernetes.io/projected/7c983dd6-f499-41d5-a049-765a15b4f912-kube-api-access-vtxrr\") pod \"telemetry-operator-controller-manager-64b5b76f97-98xsk\" (UID: \"7c983dd6-f499-41d5-a049-765a15b4f912\") " pod="openstack-operators/telemetry-operator-controller-manager-64b5b76f97-98xsk" Feb 02 09:22:03 crc kubenswrapper[4764]: I0202 09:22:03.214727 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-68fc8c869-fb6zn"] Feb 02 09:22:03 crc kubenswrapper[4764]: I0202 09:22:03.234203 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-64b5b76f97-98xsk"] Feb 02 09:22:03 crc kubenswrapper[4764]: I0202 09:22:03.241745 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-5b964cf4cd-kskcp"] Feb 02 09:22:03 crc kubenswrapper[4764]: I0202 09:22:03.278700 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-5b964cf4cd-kskcp" Feb 02 09:22:03 crc kubenswrapper[4764]: I0202 09:22:03.306493 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8nfld\" (UniqueName: \"kubernetes.io/projected/1a100291-702c-4908-8a63-acb05c4ef5e9-kube-api-access-8nfld\") pod \"swift-operator-controller-manager-68fc8c869-fb6zn\" (UID: \"1a100291-702c-4908-8a63-acb05c4ef5e9\") " pod="openstack-operators/swift-operator-controller-manager-68fc8c869-fb6zn" Feb 02 09:22:03 crc kubenswrapper[4764]: I0202 09:22:03.306578 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vtxrr\" (UniqueName: \"kubernetes.io/projected/7c983dd6-f499-41d5-a049-765a15b4f912-kube-api-access-vtxrr\") pod \"telemetry-operator-controller-manager-64b5b76f97-98xsk\" (UID: \"7c983dd6-f499-41d5-a049-765a15b4f912\") " pod="openstack-operators/telemetry-operator-controller-manager-64b5b76f97-98xsk" Feb 02 09:22:03 crc kubenswrapper[4764]: I0202 09:22:03.362257 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vtxrr\" (UniqueName: \"kubernetes.io/projected/7c983dd6-f499-41d5-a049-765a15b4f912-kube-api-access-vtxrr\") pod \"telemetry-operator-controller-manager-64b5b76f97-98xsk\" (UID: \"7c983dd6-f499-41d5-a049-765a15b4f912\") " pod="openstack-operators/telemetry-operator-controller-manager-64b5b76f97-98xsk" Feb 02 09:22:03 crc kubenswrapper[4764]: I0202 09:22:03.372295 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-56f8bfcd9f-bd9cj"] Feb 02 09:22:03 crc kubenswrapper[4764]: I0202 09:22:03.373442 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-56f8bfcd9f-bd9cj" Feb 02 09:22:03 crc kubenswrapper[4764]: I0202 09:22:03.387405 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-kpcfw" Feb 02 09:22:03 crc kubenswrapper[4764]: I0202 09:22:03.397493 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8nfld\" (UniqueName: \"kubernetes.io/projected/1a100291-702c-4908-8a63-acb05c4ef5e9-kube-api-access-8nfld\") pod \"swift-operator-controller-manager-68fc8c869-fb6zn\" (UID: \"1a100291-702c-4908-8a63-acb05c4ef5e9\") " pod="openstack-operators/swift-operator-controller-manager-68fc8c869-fb6zn" Feb 02 09:22:03 crc kubenswrapper[4764]: I0202 09:22:03.399985 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-56f8bfcd9f-bd9cj"] Feb 02 09:22:03 crc kubenswrapper[4764]: I0202 09:22:03.405722 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-564965969-6kcc6"] Feb 02 09:22:03 crc kubenswrapper[4764]: I0202 09:22:03.406666 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-564965969-6kcc6" Feb 02 09:22:03 crc kubenswrapper[4764]: I0202 09:22:03.425549 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-54klv" Feb 02 09:22:03 crc kubenswrapper[4764]: I0202 09:22:03.470110 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-64b5b76f97-98xsk" Feb 02 09:22:03 crc kubenswrapper[4764]: I0202 09:22:03.485728 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-68fc8c869-fb6zn" Feb 02 09:22:03 crc kubenswrapper[4764]: I0202 09:22:03.488092 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-564965969-6kcc6"] Feb 02 09:22:03 crc kubenswrapper[4764]: I0202 09:22:03.513116 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b7bqv\" (UniqueName: \"kubernetes.io/projected/65c22016-b2f2-4a07-87be-480d18870dea-kube-api-access-b7bqv\") pod \"watcher-operator-controller-manager-564965969-6kcc6\" (UID: \"65c22016-b2f2-4a07-87be-480d18870dea\") " pod="openstack-operators/watcher-operator-controller-manager-564965969-6kcc6" Feb 02 09:22:03 crc kubenswrapper[4764]: I0202 09:22:03.521179 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2ztpm\" (UniqueName: \"kubernetes.io/projected/b8748b41-5e08-47bf-a853-ad8c4f2b0908-kube-api-access-2ztpm\") pod \"test-operator-controller-manager-56f8bfcd9f-bd9cj\" (UID: \"b8748b41-5e08-47bf-a853-ad8c4f2b0908\") " pod="openstack-operators/test-operator-controller-manager-56f8bfcd9f-bd9cj" Feb 02 09:22:03 crc kubenswrapper[4764]: I0202 09:22:03.566198 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-646f757d77-g8bhc"] Feb 02 09:22:03 crc kubenswrapper[4764]: I0202 09:22:03.593664 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-646f757d77-g8bhc" Feb 02 09:22:03 crc kubenswrapper[4764]: I0202 09:22:03.606410 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Feb 02 09:22:03 crc kubenswrapper[4764]: I0202 09:22:03.606600 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-ffpxl" Feb 02 09:22:03 crc kubenswrapper[4764]: I0202 09:22:03.606725 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Feb 02 09:22:03 crc kubenswrapper[4764]: I0202 09:22:03.623583 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1b05c36c-dbb6-4ad4-8985-8d64551812ff-cert\") pod \"openstack-baremetal-operator-controller-manager-5ff45dfdbfb5fxl\" (UID: \"1b05c36c-dbb6-4ad4-8985-8d64551812ff\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5ff45dfdbfb5fxl" Feb 02 09:22:03 crc kubenswrapper[4764]: I0202 09:22:03.623639 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b7bqv\" (UniqueName: \"kubernetes.io/projected/65c22016-b2f2-4a07-87be-480d18870dea-kube-api-access-b7bqv\") pod \"watcher-operator-controller-manager-564965969-6kcc6\" (UID: \"65c22016-b2f2-4a07-87be-480d18870dea\") " pod="openstack-operators/watcher-operator-controller-manager-564965969-6kcc6" Feb 02 09:22:03 crc kubenswrapper[4764]: I0202 09:22:03.623663 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2ztpm\" (UniqueName: \"kubernetes.io/projected/b8748b41-5e08-47bf-a853-ad8c4f2b0908-kube-api-access-2ztpm\") pod \"test-operator-controller-manager-56f8bfcd9f-bd9cj\" (UID: \"b8748b41-5e08-47bf-a853-ad8c4f2b0908\") " pod="openstack-operators/test-operator-controller-manager-56f8bfcd9f-bd9cj" Feb 02 09:22:03 crc kubenswrapper[4764]: E0202 09:22:03.627622 4764 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 02 09:22:03 crc kubenswrapper[4764]: E0202 09:22:03.627674 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1b05c36c-dbb6-4ad4-8985-8d64551812ff-cert podName:1b05c36c-dbb6-4ad4-8985-8d64551812ff nodeName:}" failed. No retries permitted until 2026-02-02 09:22:04.627660303 +0000 UTC m=+887.561384391 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/1b05c36c-dbb6-4ad4-8985-8d64551812ff-cert") pod "openstack-baremetal-operator-controller-manager-5ff45dfdbfb5fxl" (UID: "1b05c36c-dbb6-4ad4-8985-8d64551812ff") : secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 02 09:22:03 crc kubenswrapper[4764]: I0202 09:22:03.629580 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-646f757d77-g8bhc"] Feb 02 09:22:03 crc kubenswrapper[4764]: I0202 09:22:03.675563 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b7bqv\" (UniqueName: \"kubernetes.io/projected/65c22016-b2f2-4a07-87be-480d18870dea-kube-api-access-b7bqv\") pod \"watcher-operator-controller-manager-564965969-6kcc6\" (UID: \"65c22016-b2f2-4a07-87be-480d18870dea\") " pod="openstack-operators/watcher-operator-controller-manager-564965969-6kcc6" Feb 02 09:22:03 crc kubenswrapper[4764]: I0202 09:22:03.680508 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2ztpm\" (UniqueName: \"kubernetes.io/projected/b8748b41-5e08-47bf-a853-ad8c4f2b0908-kube-api-access-2ztpm\") pod \"test-operator-controller-manager-56f8bfcd9f-bd9cj\" (UID: \"b8748b41-5e08-47bf-a853-ad8c4f2b0908\") " pod="openstack-operators/test-operator-controller-manager-56f8bfcd9f-bd9cj" Feb 02 09:22:03 crc kubenswrapper[4764]: I0202 09:22:03.708243 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-qmjfr"] Feb 02 09:22:03 crc kubenswrapper[4764]: I0202 09:22:03.715669 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-qmjfr" Feb 02 09:22:03 crc kubenswrapper[4764]: I0202 09:22:03.717215 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-sxzpb" Feb 02 09:22:03 crc kubenswrapper[4764]: I0202 09:22:03.719975 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-qmjfr"] Feb 02 09:22:03 crc kubenswrapper[4764]: I0202 09:22:03.725248 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/a46546b9-e589-4eca-bca7-9150038f9ab8-webhook-certs\") pod \"openstack-operator-controller-manager-646f757d77-g8bhc\" (UID: \"a46546b9-e589-4eca-bca7-9150038f9ab8\") " pod="openstack-operators/openstack-operator-controller-manager-646f757d77-g8bhc" Feb 02 09:22:03 crc kubenswrapper[4764]: I0202 09:22:03.725309 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v45k6\" (UniqueName: \"kubernetes.io/projected/a46546b9-e589-4eca-bca7-9150038f9ab8-kube-api-access-v45k6\") pod \"openstack-operator-controller-manager-646f757d77-g8bhc\" (UID: \"a46546b9-e589-4eca-bca7-9150038f9ab8\") " pod="openstack-operators/openstack-operator-controller-manager-646f757d77-g8bhc" Feb 02 09:22:03 crc kubenswrapper[4764]: I0202 09:22:03.725381 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a46546b9-e589-4eca-bca7-9150038f9ab8-metrics-certs\") pod \"openstack-operator-controller-manager-646f757d77-g8bhc\" (UID: \"a46546b9-e589-4eca-bca7-9150038f9ab8\") " pod="openstack-operators/openstack-operator-controller-manager-646f757d77-g8bhc" Feb 02 09:22:03 crc kubenswrapper[4764]: I0202 09:22:03.732471 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-8886f4c47-hvtx6"] Feb 02 09:22:03 crc kubenswrapper[4764]: I0202 09:22:03.791961 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-56f8bfcd9f-bd9cj" Feb 02 09:22:03 crc kubenswrapper[4764]: I0202 09:22:03.818974 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-564965969-6kcc6" Feb 02 09:22:03 crc kubenswrapper[4764]: I0202 09:22:03.831836 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a46546b9-e589-4eca-bca7-9150038f9ab8-metrics-certs\") pod \"openstack-operator-controller-manager-646f757d77-g8bhc\" (UID: \"a46546b9-e589-4eca-bca7-9150038f9ab8\") " pod="openstack-operators/openstack-operator-controller-manager-646f757d77-g8bhc" Feb 02 09:22:03 crc kubenswrapper[4764]: I0202 09:22:03.831890 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/a46546b9-e589-4eca-bca7-9150038f9ab8-webhook-certs\") pod \"openstack-operator-controller-manager-646f757d77-g8bhc\" (UID: \"a46546b9-e589-4eca-bca7-9150038f9ab8\") " pod="openstack-operators/openstack-operator-controller-manager-646f757d77-g8bhc" Feb 02 09:22:03 crc kubenswrapper[4764]: I0202 09:22:03.831925 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v45k6\" (UniqueName: \"kubernetes.io/projected/a46546b9-e589-4eca-bca7-9150038f9ab8-kube-api-access-v45k6\") pod \"openstack-operator-controller-manager-646f757d77-g8bhc\" (UID: \"a46546b9-e589-4eca-bca7-9150038f9ab8\") " pod="openstack-operators/openstack-operator-controller-manager-646f757d77-g8bhc" Feb 02 09:22:03 crc kubenswrapper[4764]: I0202 09:22:03.831979 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hjd22\" (UniqueName: \"kubernetes.io/projected/36bffbf8-2209-4410-8005-d0164bb1eb34-kube-api-access-hjd22\") pod \"rabbitmq-cluster-operator-manager-668c99d594-qmjfr\" (UID: \"36bffbf8-2209-4410-8005-d0164bb1eb34\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-qmjfr" Feb 02 09:22:03 crc kubenswrapper[4764]: E0202 09:22:03.832103 4764 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Feb 02 09:22:03 crc kubenswrapper[4764]: E0202 09:22:03.832165 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a46546b9-e589-4eca-bca7-9150038f9ab8-metrics-certs podName:a46546b9-e589-4eca-bca7-9150038f9ab8 nodeName:}" failed. No retries permitted until 2026-02-02 09:22:04.332147429 +0000 UTC m=+887.265871517 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/a46546b9-e589-4eca-bca7-9150038f9ab8-metrics-certs") pod "openstack-operator-controller-manager-646f757d77-g8bhc" (UID: "a46546b9-e589-4eca-bca7-9150038f9ab8") : secret "metrics-server-cert" not found Feb 02 09:22:03 crc kubenswrapper[4764]: E0202 09:22:03.832176 4764 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Feb 02 09:22:03 crc kubenswrapper[4764]: E0202 09:22:03.832224 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a46546b9-e589-4eca-bca7-9150038f9ab8-webhook-certs podName:a46546b9-e589-4eca-bca7-9150038f9ab8 nodeName:}" failed. No retries permitted until 2026-02-02 09:22:04.332211301 +0000 UTC m=+887.265935389 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/a46546b9-e589-4eca-bca7-9150038f9ab8-webhook-certs") pod "openstack-operator-controller-manager-646f757d77-g8bhc" (UID: "a46546b9-e589-4eca-bca7-9150038f9ab8") : secret "webhook-server-cert" not found Feb 02 09:22:03 crc kubenswrapper[4764]: I0202 09:22:03.895815 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v45k6\" (UniqueName: \"kubernetes.io/projected/a46546b9-e589-4eca-bca7-9150038f9ab8-kube-api-access-v45k6\") pod \"openstack-operator-controller-manager-646f757d77-g8bhc\" (UID: \"a46546b9-e589-4eca-bca7-9150038f9ab8\") " pod="openstack-operators/openstack-operator-controller-manager-646f757d77-g8bhc" Feb 02 09:22:03 crc kubenswrapper[4764]: I0202 09:22:03.939272 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hjd22\" (UniqueName: \"kubernetes.io/projected/36bffbf8-2209-4410-8005-d0164bb1eb34-kube-api-access-hjd22\") pod \"rabbitmq-cluster-operator-manager-668c99d594-qmjfr\" (UID: \"36bffbf8-2209-4410-8005-d0164bb1eb34\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-qmjfr" Feb 02 09:22:03 crc kubenswrapper[4764]: I0202 09:22:03.961706 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-5fb775575f-ntj47"] Feb 02 09:22:03 crc kubenswrapper[4764]: I0202 09:22:03.985765 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hjd22\" (UniqueName: \"kubernetes.io/projected/36bffbf8-2209-4410-8005-d0164bb1eb34-kube-api-access-hjd22\") pod \"rabbitmq-cluster-operator-manager-668c99d594-qmjfr\" (UID: \"36bffbf8-2209-4410-8005-d0164bb1eb34\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-qmjfr" Feb 02 09:22:04 crc kubenswrapper[4764]: I0202 09:22:04.070231 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-qmjfr" Feb 02 09:22:04 crc kubenswrapper[4764]: I0202 09:22:04.144127 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b593c3f3-a225-4cb6-b967-4c056e3326a0-cert\") pod \"infra-operator-controller-manager-79955696d6-mxvzm\" (UID: \"b593c3f3-a225-4cb6-b967-4c056e3326a0\") " pod="openstack-operators/infra-operator-controller-manager-79955696d6-mxvzm" Feb 02 09:22:04 crc kubenswrapper[4764]: E0202 09:22:04.144257 4764 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Feb 02 09:22:04 crc kubenswrapper[4764]: E0202 09:22:04.144306 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b593c3f3-a225-4cb6-b967-4c056e3326a0-cert podName:b593c3f3-a225-4cb6-b967-4c056e3326a0 nodeName:}" failed. No retries permitted until 2026-02-02 09:22:06.144292687 +0000 UTC m=+889.078016775 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/b593c3f3-a225-4cb6-b967-4c056e3326a0-cert") pod "infra-operator-controller-manager-79955696d6-mxvzm" (UID: "b593c3f3-a225-4cb6-b967-4c056e3326a0") : secret "infra-operator-webhook-server-cert" not found Feb 02 09:22:04 crc kubenswrapper[4764]: I0202 09:22:04.158047 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-69d6db494d-5f766"] Feb 02 09:22:04 crc kubenswrapper[4764]: W0202 09:22:04.177185 4764 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbecd2ba3_ccad_4ea9_b3c7_6b39f6f8cb3d.slice/crio-d66867c06b1805c11eff8242bb87ecec6e4d9f0de68ebe845d077f08d3cc4a51 WatchSource:0}: Error finding container d66867c06b1805c11eff8242bb87ecec6e4d9f0de68ebe845d077f08d3cc4a51: Status 404 returned error can't find the container with id d66867c06b1805c11eff8242bb87ecec6e4d9f0de68ebe845d077f08d3cc4a51 Feb 02 09:22:04 crc kubenswrapper[4764]: I0202 09:22:04.195066 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-84f48565d4-qqnp6"] Feb 02 09:22:04 crc kubenswrapper[4764]: I0202 09:22:04.357919 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a46546b9-e589-4eca-bca7-9150038f9ab8-metrics-certs\") pod \"openstack-operator-controller-manager-646f757d77-g8bhc\" (UID: \"a46546b9-e589-4eca-bca7-9150038f9ab8\") " pod="openstack-operators/openstack-operator-controller-manager-646f757d77-g8bhc" Feb 02 09:22:04 crc kubenswrapper[4764]: E0202 09:22:04.358203 4764 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Feb 02 09:22:04 crc kubenswrapper[4764]: E0202 09:22:04.360066 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a46546b9-e589-4eca-bca7-9150038f9ab8-metrics-certs podName:a46546b9-e589-4eca-bca7-9150038f9ab8 nodeName:}" failed. No retries permitted until 2026-02-02 09:22:05.360037434 +0000 UTC m=+888.293761522 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/a46546b9-e589-4eca-bca7-9150038f9ab8-metrics-certs") pod "openstack-operator-controller-manager-646f757d77-g8bhc" (UID: "a46546b9-e589-4eca-bca7-9150038f9ab8") : secret "metrics-server-cert" not found Feb 02 09:22:04 crc kubenswrapper[4764]: I0202 09:22:04.382154 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/a46546b9-e589-4eca-bca7-9150038f9ab8-webhook-certs\") pod \"openstack-operator-controller-manager-646f757d77-g8bhc\" (UID: \"a46546b9-e589-4eca-bca7-9150038f9ab8\") " pod="openstack-operators/openstack-operator-controller-manager-646f757d77-g8bhc" Feb 02 09:22:04 crc kubenswrapper[4764]: E0202 09:22:04.382354 4764 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Feb 02 09:22:04 crc kubenswrapper[4764]: E0202 09:22:04.384409 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a46546b9-e589-4eca-bca7-9150038f9ab8-webhook-certs podName:a46546b9-e589-4eca-bca7-9150038f9ab8 nodeName:}" failed. No retries permitted until 2026-02-02 09:22:05.384389024 +0000 UTC m=+888.318113112 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/a46546b9-e589-4eca-bca7-9150038f9ab8-webhook-certs") pod "openstack-operator-controller-manager-646f757d77-g8bhc" (UID: "a46546b9-e589-4eca-bca7-9150038f9ab8") : secret "webhook-server-cert" not found Feb 02 09:22:04 crc kubenswrapper[4764]: I0202 09:22:04.577868 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-7dd968899f-g2p6c"] Feb 02 09:22:04 crc kubenswrapper[4764]: I0202 09:22:04.585100 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-8d874c8fc-czpmn"] Feb 02 09:22:04 crc kubenswrapper[4764]: I0202 09:22:04.589870 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-6d9697b7f4-w4qf5"] Feb 02 09:22:04 crc kubenswrapper[4764]: I0202 09:22:04.594161 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-67bf948998-jhbzc"] Feb 02 09:22:04 crc kubenswrapper[4764]: I0202 09:22:04.609275 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7b6c4d8c5f-wtkdr"] Feb 02 09:22:04 crc kubenswrapper[4764]: W0202 09:22:04.612263 4764 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7c8a2462_306c_43da_a538_9e34a11b5c21.slice/crio-9d088c8dd82671ed0bd7472df26011005162cb497d1bc8f9ce2c9961eab1585d WatchSource:0}: Error finding container 9d088c8dd82671ed0bd7472df26011005162cb497d1bc8f9ce2c9961eab1585d: Status 404 returned error can't find the container with id 9d088c8dd82671ed0bd7472df26011005162cb497d1bc8f9ce2c9961eab1585d Feb 02 09:22:04 crc kubenswrapper[4764]: I0202 09:22:04.648111 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-585dbc889-7nrsg"] Feb 02 09:22:04 crc kubenswrapper[4764]: W0202 09:22:04.661489 4764 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1844ef02_28fb_45ce_a697_54f1560c9087.slice/crio-5856813a1e84f4d0e82f6d751114f3cb58400a4f88af2d9c3bccfcd38bd98435 WatchSource:0}: Error finding container 5856813a1e84f4d0e82f6d751114f3cb58400a4f88af2d9c3bccfcd38bd98435: Status 404 returned error can't find the container with id 5856813a1e84f4d0e82f6d751114f3cb58400a4f88af2d9c3bccfcd38bd98435 Feb 02 09:22:04 crc kubenswrapper[4764]: I0202 09:22:04.661589 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-5fb775575f-ntj47" event={"ID":"82e16b73-9dfe-413b-b3dd-65da79b3d9dd","Type":"ContainerStarted","Data":"8b70c6cf8a624f7e439e3484b9a5555d6a2deb63f1f1ebf167e1c3f79cc06e50"} Feb 02 09:22:04 crc kubenswrapper[4764]: I0202 09:22:04.663760 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-5f4b8bd54d-h4n52"] Feb 02 09:22:04 crc kubenswrapper[4764]: I0202 09:22:04.665082 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7dd968899f-g2p6c" event={"ID":"6d4fffad-a306-4f0b-ac40-7fa839f75aa3","Type":"ContainerStarted","Data":"15678de37bce7a61ddc3aa4d8d2528f7a7172f8e2a318976f409af240f572016"} Feb 02 09:22:04 crc kubenswrapper[4764]: I0202 09:22:04.668713 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-8d874c8fc-czpmn" event={"ID":"7c8a2462-306c-43da-a538-9e34a11b5c21","Type":"ContainerStarted","Data":"9d088c8dd82671ed0bd7472df26011005162cb497d1bc8f9ce2c9961eab1585d"} Feb 02 09:22:04 crc kubenswrapper[4764]: I0202 09:22:04.672418 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-788c46999f-pk2hh"] Feb 02 09:22:04 crc kubenswrapper[4764]: I0202 09:22:04.678241 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-67bf948998-jhbzc" event={"ID":"9e55cb1e-ed3b-4b7b-8b52-42436eb48223","Type":"ContainerStarted","Data":"0fbd9253693c1f56b7fb12ae318942f19fc091f79a400ca46daf3403fa3335ca"} Feb 02 09:22:04 crc kubenswrapper[4764]: I0202 09:22:04.679202 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-585dbc889-7nrsg" event={"ID":"f5cc9236-c70f-4449-a902-ef416aaf8781","Type":"ContainerStarted","Data":"d64164e960d821b4eeb914bb52ef7c78217707bb182d13ebbebbdb5e5b383c56"} Feb 02 09:22:04 crc kubenswrapper[4764]: I0202 09:22:04.681543 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-69d6db494d-5f766" event={"ID":"becd2ba3-ccad-4ea9-b3c7-6b39f6f8cb3d","Type":"ContainerStarted","Data":"d66867c06b1805c11eff8242bb87ecec6e4d9f0de68ebe845d077f08d3cc4a51"} Feb 02 09:22:04 crc kubenswrapper[4764]: I0202 09:22:04.682559 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-8886f4c47-hvtx6" event={"ID":"ee5cb41c-4d3d-4a9d-b1f7-1ede61e7b21c","Type":"ContainerStarted","Data":"9377800931dc0601882391901e2c4dff1154601811063a7689d0d81fac826955"} Feb 02 09:22:04 crc kubenswrapper[4764]: I0202 09:22:04.683797 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-84f48565d4-qqnp6" event={"ID":"8564e79b-9abd-41e1-a0a7-0d4b71011300","Type":"ContainerStarted","Data":"443c931fcc07bb0349ad8ea64ed45b956e966645b0488f66a104b500c6c6090f"} Feb 02 09:22:04 crc kubenswrapper[4764]: I0202 09:22:04.684406 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-6687f8d877-5ncb8"] Feb 02 09:22:04 crc kubenswrapper[4764]: I0202 09:22:04.685381 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-6d9697b7f4-w4qf5" event={"ID":"8b0e7cb4-257b-403b-8b92-4578ce77c388","Type":"ContainerStarted","Data":"5a9f920711ca547e4a63e16d6f5c6ce2c30b9b21909e6dd021ff351f8bc22622"} Feb 02 09:22:04 crc kubenswrapper[4764]: I0202 09:22:04.687245 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7b6c4d8c5f-wtkdr" event={"ID":"9cbd726b-9dbb-46b3-9bcd-f3e9813b4f07","Type":"ContainerStarted","Data":"a6a3f917e09c2f47cda7f0f7ec81fcab6c1d5b2de088399c4466349451298b66"} Feb 02 09:22:04 crc kubenswrapper[4764]: I0202 09:22:04.687879 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1b05c36c-dbb6-4ad4-8985-8d64551812ff-cert\") pod \"openstack-baremetal-operator-controller-manager-5ff45dfdbfb5fxl\" (UID: \"1b05c36c-dbb6-4ad4-8985-8d64551812ff\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5ff45dfdbfb5fxl" Feb 02 09:22:04 crc kubenswrapper[4764]: E0202 09:22:04.688052 4764 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 02 09:22:04 crc kubenswrapper[4764]: E0202 09:22:04.688137 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1b05c36c-dbb6-4ad4-8985-8d64551812ff-cert podName:1b05c36c-dbb6-4ad4-8985-8d64551812ff nodeName:}" failed. No retries permitted until 2026-02-02 09:22:06.688121902 +0000 UTC m=+889.621845990 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/1b05c36c-dbb6-4ad4-8985-8d64551812ff-cert") pod "openstack-baremetal-operator-controller-manager-5ff45dfdbfb5fxl" (UID: "1b05c36c-dbb6-4ad4-8985-8d64551812ff") : secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 02 09:22:04 crc kubenswrapper[4764]: I0202 09:22:04.692269 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-5b964cf4cd-kskcp"] Feb 02 09:22:04 crc kubenswrapper[4764]: I0202 09:22:04.804916 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-564965969-6kcc6"] Feb 02 09:22:04 crc kubenswrapper[4764]: I0202 09:22:04.811988 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-68fc8c869-fb6zn"] Feb 02 09:22:04 crc kubenswrapper[4764]: I0202 09:22:04.818155 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-56f8bfcd9f-bd9cj"] Feb 02 09:22:04 crc kubenswrapper[4764]: E0202 09:22:04.819532 4764 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:3e01e99d3ca1b6c20b1bb015b00cfcbffc584f22a93dc6fe4019d63b813c0241,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-2ztpm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-56f8bfcd9f-bd9cj_openstack-operators(b8748b41-5e08-47bf-a853-ad8c4f2b0908): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Feb 02 09:22:04 crc kubenswrapper[4764]: E0202 09:22:04.820810 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/test-operator-controller-manager-56f8bfcd9f-bd9cj" podUID="b8748b41-5e08-47bf-a853-ad8c4f2b0908" Feb 02 09:22:04 crc kubenswrapper[4764]: W0202 09:22:04.819605 4764 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod65c22016_b2f2_4a07_87be_480d18870dea.slice/crio-71974600c592182c067eabd332456913cda50961db3721bc767112990cd6a2da WatchSource:0}: Error finding container 71974600c592182c067eabd332456913cda50961db3721bc767112990cd6a2da: Status 404 returned error can't find the container with id 71974600c592182c067eabd332456913cda50961db3721bc767112990cd6a2da Feb 02 09:22:04 crc kubenswrapper[4764]: E0202 09:22:04.822884 4764 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:7869203f6f97de780368d507636031090fed3b658d2f7771acbd4481bdfc870b,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-b7bqv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-564965969-6kcc6_openstack-operators(65c22016-b2f2-4a07-87be-480d18870dea): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Feb 02 09:22:04 crc kubenswrapper[4764]: E0202 09:22:04.824307 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/watcher-operator-controller-manager-564965969-6kcc6" podUID="65c22016-b2f2-4a07-87be-480d18870dea" Feb 02 09:22:04 crc kubenswrapper[4764]: I0202 09:22:04.838985 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-64b5b76f97-98xsk"] Feb 02 09:22:04 crc kubenswrapper[4764]: I0202 09:22:04.852700 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-qmjfr"] Feb 02 09:22:04 crc kubenswrapper[4764]: E0202 09:22:04.857594 4764 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/telemetry-operator@sha256:f9bf288cd0c13912404027a58ea3b90d4092b641e8265adc5c88644ea7fe901a,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-vtxrr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-64b5b76f97-98xsk_openstack-operators(7c983dd6-f499-41d5-a049-765a15b4f912): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Feb 02 09:22:04 crc kubenswrapper[4764]: E0202 09:22:04.859143 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/telemetry-operator-controller-manager-64b5b76f97-98xsk" podUID="7c983dd6-f499-41d5-a049-765a15b4f912" Feb 02 09:22:04 crc kubenswrapper[4764]: E0202 09:22:04.868776 4764 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-hjd22,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-qmjfr_openstack-operators(36bffbf8-2209-4410-8005-d0164bb1eb34): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Feb 02 09:22:04 crc kubenswrapper[4764]: E0202 09:22:04.870177 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-qmjfr" podUID="36bffbf8-2209-4410-8005-d0164bb1eb34" Feb 02 09:22:04 crc kubenswrapper[4764]: I0202 09:22:04.874399 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-55bff696bd-p7nmb"] Feb 02 09:22:04 crc kubenswrapper[4764]: W0202 09:22:04.879791 4764 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode98e75d4_963f_4a67_894e_302b7429eb01.slice/crio-32921a9eb82d00539fa888876175e5dda8263c868ae38306ef6ca411483fd250 WatchSource:0}: Error finding container 32921a9eb82d00539fa888876175e5dda8263c868ae38306ef6ca411483fd250: Status 404 returned error can't find the container with id 32921a9eb82d00539fa888876175e5dda8263c868ae38306ef6ca411483fd250 Feb 02 09:22:04 crc kubenswrapper[4764]: E0202 09:22:04.883371 4764 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:5340b88039fac393da49ef4e181b2720c809c27a6bb30531a07a49342a1da45e,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-nphdp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-55bff696bd-p7nmb_openstack-operators(e98e75d4-963f-4a67-894e-302b7429eb01): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Feb 02 09:22:04 crc kubenswrapper[4764]: E0202 09:22:04.884544 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/nova-operator-controller-manager-55bff696bd-p7nmb" podUID="e98e75d4-963f-4a67-894e-302b7429eb01" Feb 02 09:22:05 crc kubenswrapper[4764]: I0202 09:22:05.399780 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a46546b9-e589-4eca-bca7-9150038f9ab8-metrics-certs\") pod \"openstack-operator-controller-manager-646f757d77-g8bhc\" (UID: \"a46546b9-e589-4eca-bca7-9150038f9ab8\") " pod="openstack-operators/openstack-operator-controller-manager-646f757d77-g8bhc" Feb 02 09:22:05 crc kubenswrapper[4764]: I0202 09:22:05.400078 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/a46546b9-e589-4eca-bca7-9150038f9ab8-webhook-certs\") pod \"openstack-operator-controller-manager-646f757d77-g8bhc\" (UID: \"a46546b9-e589-4eca-bca7-9150038f9ab8\") " pod="openstack-operators/openstack-operator-controller-manager-646f757d77-g8bhc" Feb 02 09:22:05 crc kubenswrapper[4764]: E0202 09:22:05.400116 4764 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Feb 02 09:22:05 crc kubenswrapper[4764]: E0202 09:22:05.400196 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a46546b9-e589-4eca-bca7-9150038f9ab8-metrics-certs podName:a46546b9-e589-4eca-bca7-9150038f9ab8 nodeName:}" failed. No retries permitted until 2026-02-02 09:22:07.400179135 +0000 UTC m=+890.333903223 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/a46546b9-e589-4eca-bca7-9150038f9ab8-metrics-certs") pod "openstack-operator-controller-manager-646f757d77-g8bhc" (UID: "a46546b9-e589-4eca-bca7-9150038f9ab8") : secret "metrics-server-cert" not found Feb 02 09:22:05 crc kubenswrapper[4764]: E0202 09:22:05.400245 4764 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Feb 02 09:22:05 crc kubenswrapper[4764]: E0202 09:22:05.400292 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a46546b9-e589-4eca-bca7-9150038f9ab8-webhook-certs podName:a46546b9-e589-4eca-bca7-9150038f9ab8 nodeName:}" failed. No retries permitted until 2026-02-02 09:22:07.400278268 +0000 UTC m=+890.334002356 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/a46546b9-e589-4eca-bca7-9150038f9ab8-webhook-certs") pod "openstack-operator-controller-manager-646f757d77-g8bhc" (UID: "a46546b9-e589-4eca-bca7-9150038f9ab8") : secret "webhook-server-cert" not found Feb 02 09:22:05 crc kubenswrapper[4764]: I0202 09:22:05.715091 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-55bff696bd-p7nmb" event={"ID":"e98e75d4-963f-4a67-894e-302b7429eb01","Type":"ContainerStarted","Data":"32921a9eb82d00539fa888876175e5dda8263c868ae38306ef6ca411483fd250"} Feb 02 09:22:05 crc kubenswrapper[4764]: E0202 09:22:05.716657 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:5340b88039fac393da49ef4e181b2720c809c27a6bb30531a07a49342a1da45e\\\"\"" pod="openstack-operators/nova-operator-controller-manager-55bff696bd-p7nmb" podUID="e98e75d4-963f-4a67-894e-302b7429eb01" Feb 02 09:22:05 crc kubenswrapper[4764]: I0202 09:22:05.721240 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-qmjfr" event={"ID":"36bffbf8-2209-4410-8005-d0164bb1eb34","Type":"ContainerStarted","Data":"0cf028e65fe2766312779f17ab9829955d8bfb3f7c164f1cec527be41157f2ee"} Feb 02 09:22:05 crc kubenswrapper[4764]: I0202 09:22:05.722698 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-68fc8c869-fb6zn" event={"ID":"1a100291-702c-4908-8a63-acb05c4ef5e9","Type":"ContainerStarted","Data":"fb35e291a06071c86e2c4121f6fb63bc30d99a5d75800c7df33f80f4af673991"} Feb 02 09:22:05 crc kubenswrapper[4764]: E0202 09:22:05.723707 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-qmjfr" podUID="36bffbf8-2209-4410-8005-d0164bb1eb34" Feb 02 09:22:05 crc kubenswrapper[4764]: I0202 09:22:05.724564 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-56f8bfcd9f-bd9cj" event={"ID":"b8748b41-5e08-47bf-a853-ad8c4f2b0908","Type":"ContainerStarted","Data":"20b274e24052f1cfde04a172694c7d6731d5f0068e28981bc5990124583ffb50"} Feb 02 09:22:05 crc kubenswrapper[4764]: E0202 09:22:05.734076 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:3e01e99d3ca1b6c20b1bb015b00cfcbffc584f22a93dc6fe4019d63b813c0241\\\"\"" pod="openstack-operators/test-operator-controller-manager-56f8bfcd9f-bd9cj" podUID="b8748b41-5e08-47bf-a853-ad8c4f2b0908" Feb 02 09:22:05 crc kubenswrapper[4764]: I0202 09:22:05.735636 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-5f4b8bd54d-h4n52" event={"ID":"ad8e7100-d5b8-4a56-bbda-056483f79113","Type":"ContainerStarted","Data":"b29ddd202fe656200ead9b23f6e6a3220d72ed50737b95b5999cf518acc2a348"} Feb 02 09:22:05 crc kubenswrapper[4764]: I0202 09:22:05.742564 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-564965969-6kcc6" event={"ID":"65c22016-b2f2-4a07-87be-480d18870dea","Type":"ContainerStarted","Data":"71974600c592182c067eabd332456913cda50961db3721bc767112990cd6a2da"} Feb 02 09:22:05 crc kubenswrapper[4764]: E0202 09:22:05.747704 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:7869203f6f97de780368d507636031090fed3b658d2f7771acbd4481bdfc870b\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-564965969-6kcc6" podUID="65c22016-b2f2-4a07-87be-480d18870dea" Feb 02 09:22:05 crc kubenswrapper[4764]: I0202 09:22:05.756061 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-5b964cf4cd-kskcp" event={"ID":"1464055e-4745-45e3-be46-5b7700dce733","Type":"ContainerStarted","Data":"d9fc30a517b82ca58a8a00d8905c9e0f2d04790bc5a805db1cd2bff4034314be"} Feb 02 09:22:05 crc kubenswrapper[4764]: I0202 09:22:05.759834 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-788c46999f-pk2hh" event={"ID":"674c8820-f063-487b-9122-1ed46d1bc39f","Type":"ContainerStarted","Data":"2cf3d094a9fc971170b3d47dd917e31991cc58aaeb307d9a7d5f5943c140a7b7"} Feb 02 09:22:05 crc kubenswrapper[4764]: I0202 09:22:05.761086 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-64b5b76f97-98xsk" event={"ID":"7c983dd6-f499-41d5-a049-765a15b4f912","Type":"ContainerStarted","Data":"93b3d5b04e16ff218d241f3d56ffd6900ec024d9a39d06ce0f0864a22a3ea94f"} Feb 02 09:22:05 crc kubenswrapper[4764]: E0202 09:22:05.764742 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:f9bf288cd0c13912404027a58ea3b90d4092b641e8265adc5c88644ea7fe901a\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-64b5b76f97-98xsk" podUID="7c983dd6-f499-41d5-a049-765a15b4f912" Feb 02 09:22:05 crc kubenswrapper[4764]: I0202 09:22:05.768475 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-6687f8d877-5ncb8" event={"ID":"1844ef02-28fb-45ce-a697-54f1560c9087","Type":"ContainerStarted","Data":"5856813a1e84f4d0e82f6d751114f3cb58400a4f88af2d9c3bccfcd38bd98435"} Feb 02 09:22:06 crc kubenswrapper[4764]: I0202 09:22:06.211047 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b593c3f3-a225-4cb6-b967-4c056e3326a0-cert\") pod \"infra-operator-controller-manager-79955696d6-mxvzm\" (UID: \"b593c3f3-a225-4cb6-b967-4c056e3326a0\") " pod="openstack-operators/infra-operator-controller-manager-79955696d6-mxvzm" Feb 02 09:22:06 crc kubenswrapper[4764]: E0202 09:22:06.211252 4764 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Feb 02 09:22:06 crc kubenswrapper[4764]: E0202 09:22:06.211300 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b593c3f3-a225-4cb6-b967-4c056e3326a0-cert podName:b593c3f3-a225-4cb6-b967-4c056e3326a0 nodeName:}" failed. No retries permitted until 2026-02-02 09:22:10.211285084 +0000 UTC m=+893.145009172 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/b593c3f3-a225-4cb6-b967-4c056e3326a0-cert") pod "infra-operator-controller-manager-79955696d6-mxvzm" (UID: "b593c3f3-a225-4cb6-b967-4c056e3326a0") : secret "infra-operator-webhook-server-cert" not found Feb 02 09:22:06 crc kubenswrapper[4764]: I0202 09:22:06.726670 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1b05c36c-dbb6-4ad4-8985-8d64551812ff-cert\") pod \"openstack-baremetal-operator-controller-manager-5ff45dfdbfb5fxl\" (UID: \"1b05c36c-dbb6-4ad4-8985-8d64551812ff\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5ff45dfdbfb5fxl" Feb 02 09:22:06 crc kubenswrapper[4764]: E0202 09:22:06.726884 4764 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 02 09:22:06 crc kubenswrapper[4764]: E0202 09:22:06.727011 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1b05c36c-dbb6-4ad4-8985-8d64551812ff-cert podName:1b05c36c-dbb6-4ad4-8985-8d64551812ff nodeName:}" failed. No retries permitted until 2026-02-02 09:22:10.726991144 +0000 UTC m=+893.660715232 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/1b05c36c-dbb6-4ad4-8985-8d64551812ff-cert") pod "openstack-baremetal-operator-controller-manager-5ff45dfdbfb5fxl" (UID: "1b05c36c-dbb6-4ad4-8985-8d64551812ff") : secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 02 09:22:06 crc kubenswrapper[4764]: E0202 09:22:06.788068 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:f9bf288cd0c13912404027a58ea3b90d4092b641e8265adc5c88644ea7fe901a\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-64b5b76f97-98xsk" podUID="7c983dd6-f499-41d5-a049-765a15b4f912" Feb 02 09:22:06 crc kubenswrapper[4764]: E0202 09:22:06.788084 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:3e01e99d3ca1b6c20b1bb015b00cfcbffc584f22a93dc6fe4019d63b813c0241\\\"\"" pod="openstack-operators/test-operator-controller-manager-56f8bfcd9f-bd9cj" podUID="b8748b41-5e08-47bf-a853-ad8c4f2b0908" Feb 02 09:22:06 crc kubenswrapper[4764]: E0202 09:22:06.788148 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:7869203f6f97de780368d507636031090fed3b658d2f7771acbd4481bdfc870b\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-564965969-6kcc6" podUID="65c22016-b2f2-4a07-87be-480d18870dea" Feb 02 09:22:06 crc kubenswrapper[4764]: E0202 09:22:06.788175 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:5340b88039fac393da49ef4e181b2720c809c27a6bb30531a07a49342a1da45e\\\"\"" pod="openstack-operators/nova-operator-controller-manager-55bff696bd-p7nmb" podUID="e98e75d4-963f-4a67-894e-302b7429eb01" Feb 02 09:22:06 crc kubenswrapper[4764]: E0202 09:22:06.788203 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-qmjfr" podUID="36bffbf8-2209-4410-8005-d0164bb1eb34" Feb 02 09:22:07 crc kubenswrapper[4764]: I0202 09:22:07.435557 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a46546b9-e589-4eca-bca7-9150038f9ab8-metrics-certs\") pod \"openstack-operator-controller-manager-646f757d77-g8bhc\" (UID: \"a46546b9-e589-4eca-bca7-9150038f9ab8\") " pod="openstack-operators/openstack-operator-controller-manager-646f757d77-g8bhc" Feb 02 09:22:07 crc kubenswrapper[4764]: E0202 09:22:07.435694 4764 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Feb 02 09:22:07 crc kubenswrapper[4764]: E0202 09:22:07.435755 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a46546b9-e589-4eca-bca7-9150038f9ab8-metrics-certs podName:a46546b9-e589-4eca-bca7-9150038f9ab8 nodeName:}" failed. No retries permitted until 2026-02-02 09:22:11.435735646 +0000 UTC m=+894.369459734 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/a46546b9-e589-4eca-bca7-9150038f9ab8-metrics-certs") pod "openstack-operator-controller-manager-646f757d77-g8bhc" (UID: "a46546b9-e589-4eca-bca7-9150038f9ab8") : secret "metrics-server-cert" not found Feb 02 09:22:07 crc kubenswrapper[4764]: I0202 09:22:07.436156 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/a46546b9-e589-4eca-bca7-9150038f9ab8-webhook-certs\") pod \"openstack-operator-controller-manager-646f757d77-g8bhc\" (UID: \"a46546b9-e589-4eca-bca7-9150038f9ab8\") " pod="openstack-operators/openstack-operator-controller-manager-646f757d77-g8bhc" Feb 02 09:22:07 crc kubenswrapper[4764]: E0202 09:22:07.436242 4764 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Feb 02 09:22:07 crc kubenswrapper[4764]: E0202 09:22:07.436272 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a46546b9-e589-4eca-bca7-9150038f9ab8-webhook-certs podName:a46546b9-e589-4eca-bca7-9150038f9ab8 nodeName:}" failed. No retries permitted until 2026-02-02 09:22:11.436264741 +0000 UTC m=+894.369988819 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/a46546b9-e589-4eca-bca7-9150038f9ab8-webhook-certs") pod "openstack-operator-controller-manager-646f757d77-g8bhc" (UID: "a46546b9-e589-4eca-bca7-9150038f9ab8") : secret "webhook-server-cert" not found Feb 02 09:22:10 crc kubenswrapper[4764]: I0202 09:22:10.288799 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b593c3f3-a225-4cb6-b967-4c056e3326a0-cert\") pod \"infra-operator-controller-manager-79955696d6-mxvzm\" (UID: \"b593c3f3-a225-4cb6-b967-4c056e3326a0\") " pod="openstack-operators/infra-operator-controller-manager-79955696d6-mxvzm" Feb 02 09:22:10 crc kubenswrapper[4764]: E0202 09:22:10.289028 4764 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Feb 02 09:22:10 crc kubenswrapper[4764]: E0202 09:22:10.289215 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b593c3f3-a225-4cb6-b967-4c056e3326a0-cert podName:b593c3f3-a225-4cb6-b967-4c056e3326a0 nodeName:}" failed. No retries permitted until 2026-02-02 09:22:18.289193522 +0000 UTC m=+901.222917620 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/b593c3f3-a225-4cb6-b967-4c056e3326a0-cert") pod "infra-operator-controller-manager-79955696d6-mxvzm" (UID: "b593c3f3-a225-4cb6-b967-4c056e3326a0") : secret "infra-operator-webhook-server-cert" not found Feb 02 09:22:10 crc kubenswrapper[4764]: I0202 09:22:10.795103 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1b05c36c-dbb6-4ad4-8985-8d64551812ff-cert\") pod \"openstack-baremetal-operator-controller-manager-5ff45dfdbfb5fxl\" (UID: \"1b05c36c-dbb6-4ad4-8985-8d64551812ff\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5ff45dfdbfb5fxl" Feb 02 09:22:10 crc kubenswrapper[4764]: E0202 09:22:10.795578 4764 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 02 09:22:10 crc kubenswrapper[4764]: E0202 09:22:10.795635 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1b05c36c-dbb6-4ad4-8985-8d64551812ff-cert podName:1b05c36c-dbb6-4ad4-8985-8d64551812ff nodeName:}" failed. No retries permitted until 2026-02-02 09:22:18.795617207 +0000 UTC m=+901.729341295 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/1b05c36c-dbb6-4ad4-8985-8d64551812ff-cert") pod "openstack-baremetal-operator-controller-manager-5ff45dfdbfb5fxl" (UID: "1b05c36c-dbb6-4ad4-8985-8d64551812ff") : secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 02 09:22:11 crc kubenswrapper[4764]: I0202 09:22:11.506231 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a46546b9-e589-4eca-bca7-9150038f9ab8-metrics-certs\") pod \"openstack-operator-controller-manager-646f757d77-g8bhc\" (UID: \"a46546b9-e589-4eca-bca7-9150038f9ab8\") " pod="openstack-operators/openstack-operator-controller-manager-646f757d77-g8bhc" Feb 02 09:22:11 crc kubenswrapper[4764]: I0202 09:22:11.506334 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/a46546b9-e589-4eca-bca7-9150038f9ab8-webhook-certs\") pod \"openstack-operator-controller-manager-646f757d77-g8bhc\" (UID: \"a46546b9-e589-4eca-bca7-9150038f9ab8\") " pod="openstack-operators/openstack-operator-controller-manager-646f757d77-g8bhc" Feb 02 09:22:11 crc kubenswrapper[4764]: E0202 09:22:11.506517 4764 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Feb 02 09:22:11 crc kubenswrapper[4764]: E0202 09:22:11.506573 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a46546b9-e589-4eca-bca7-9150038f9ab8-webhook-certs podName:a46546b9-e589-4eca-bca7-9150038f9ab8 nodeName:}" failed. No retries permitted until 2026-02-02 09:22:19.506554538 +0000 UTC m=+902.440278646 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/a46546b9-e589-4eca-bca7-9150038f9ab8-webhook-certs") pod "openstack-operator-controller-manager-646f757d77-g8bhc" (UID: "a46546b9-e589-4eca-bca7-9150038f9ab8") : secret "webhook-server-cert" not found Feb 02 09:22:11 crc kubenswrapper[4764]: E0202 09:22:11.506878 4764 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Feb 02 09:22:11 crc kubenswrapper[4764]: E0202 09:22:11.507042 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a46546b9-e589-4eca-bca7-9150038f9ab8-metrics-certs podName:a46546b9-e589-4eca-bca7-9150038f9ab8 nodeName:}" failed. No retries permitted until 2026-02-02 09:22:19.50700498 +0000 UTC m=+902.440729108 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/a46546b9-e589-4eca-bca7-9150038f9ab8-metrics-certs") pod "openstack-operator-controller-manager-646f757d77-g8bhc" (UID: "a46546b9-e589-4eca-bca7-9150038f9ab8") : secret "metrics-server-cert" not found Feb 02 09:22:13 crc kubenswrapper[4764]: I0202 09:22:13.522994 4764 patch_prober.go:28] interesting pod/machine-config-daemon-4ndm4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 09:22:13 crc kubenswrapper[4764]: I0202 09:22:13.523050 4764 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 09:22:18 crc kubenswrapper[4764]: E0202 09:22:18.070336 4764 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/designate-operator@sha256:d9f6f8dc6a6dd9b0d7c96e4c89b3056291fd61f11126a1304256a4d6cacd0382" Feb 02 09:22:18 crc kubenswrapper[4764]: E0202 09:22:18.071090 4764 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/designate-operator@sha256:d9f6f8dc6a6dd9b0d7c96e4c89b3056291fd61f11126a1304256a4d6cacd0382,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-qlmjs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod designate-operator-controller-manager-6d9697b7f4-w4qf5_openstack-operators(8b0e7cb4-257b-403b-8b92-4578ce77c388): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 02 09:22:18 crc kubenswrapper[4764]: E0202 09:22:18.072287 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/designate-operator-controller-manager-6d9697b7f4-w4qf5" podUID="8b0e7cb4-257b-403b-8b92-4578ce77c388" Feb 02 09:22:18 crc kubenswrapper[4764]: I0202 09:22:18.323937 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b593c3f3-a225-4cb6-b967-4c056e3326a0-cert\") pod \"infra-operator-controller-manager-79955696d6-mxvzm\" (UID: \"b593c3f3-a225-4cb6-b967-4c056e3326a0\") " pod="openstack-operators/infra-operator-controller-manager-79955696d6-mxvzm" Feb 02 09:22:18 crc kubenswrapper[4764]: E0202 09:22:18.324074 4764 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Feb 02 09:22:18 crc kubenswrapper[4764]: E0202 09:22:18.324142 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b593c3f3-a225-4cb6-b967-4c056e3326a0-cert podName:b593c3f3-a225-4cb6-b967-4c056e3326a0 nodeName:}" failed. No retries permitted until 2026-02-02 09:22:34.324123621 +0000 UTC m=+917.257847709 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/b593c3f3-a225-4cb6-b967-4c056e3326a0-cert") pod "infra-operator-controller-manager-79955696d6-mxvzm" (UID: "b593c3f3-a225-4cb6-b967-4c056e3326a0") : secret "infra-operator-webhook-server-cert" not found Feb 02 09:22:18 crc kubenswrapper[4764]: I0202 09:22:18.831405 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1b05c36c-dbb6-4ad4-8985-8d64551812ff-cert\") pod \"openstack-baremetal-operator-controller-manager-5ff45dfdbfb5fxl\" (UID: \"1b05c36c-dbb6-4ad4-8985-8d64551812ff\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5ff45dfdbfb5fxl" Feb 02 09:22:18 crc kubenswrapper[4764]: E0202 09:22:18.831566 4764 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 02 09:22:18 crc kubenswrapper[4764]: E0202 09:22:18.831610 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1b05c36c-dbb6-4ad4-8985-8d64551812ff-cert podName:1b05c36c-dbb6-4ad4-8985-8d64551812ff nodeName:}" failed. No retries permitted until 2026-02-02 09:22:34.831596534 +0000 UTC m=+917.765320612 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/1b05c36c-dbb6-4ad4-8985-8d64551812ff-cert") pod "openstack-baremetal-operator-controller-manager-5ff45dfdbfb5fxl" (UID: "1b05c36c-dbb6-4ad4-8985-8d64551812ff") : secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 02 09:22:18 crc kubenswrapper[4764]: E0202 09:22:18.890991 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/designate-operator@sha256:d9f6f8dc6a6dd9b0d7c96e4c89b3056291fd61f11126a1304256a4d6cacd0382\\\"\"" pod="openstack-operators/designate-operator-controller-manager-6d9697b7f4-w4qf5" podUID="8b0e7cb4-257b-403b-8b92-4578ce77c388" Feb 02 09:22:18 crc kubenswrapper[4764]: E0202 09:22:18.931336 4764 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/neutron-operator@sha256:bbb46b8b3b69fdfad7bafc10a7e88f6ea58bcdc3c91e30beb79e24417d52e0f6" Feb 02 09:22:18 crc kubenswrapper[4764]: E0202 09:22:18.931655 4764 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/neutron-operator@sha256:bbb46b8b3b69fdfad7bafc10a7e88f6ea58bcdc3c91e30beb79e24417d52e0f6,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-dmclv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-585dbc889-7nrsg_openstack-operators(f5cc9236-c70f-4449-a902-ef416aaf8781): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 02 09:22:18 crc kubenswrapper[4764]: E0202 09:22:18.933476 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/neutron-operator-controller-manager-585dbc889-7nrsg" podUID="f5cc9236-c70f-4449-a902-ef416aaf8781" Feb 02 09:22:19 crc kubenswrapper[4764]: I0202 09:22:19.551982 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a46546b9-e589-4eca-bca7-9150038f9ab8-metrics-certs\") pod \"openstack-operator-controller-manager-646f757d77-g8bhc\" (UID: \"a46546b9-e589-4eca-bca7-9150038f9ab8\") " pod="openstack-operators/openstack-operator-controller-manager-646f757d77-g8bhc" Feb 02 09:22:19 crc kubenswrapper[4764]: I0202 09:22:19.552058 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/a46546b9-e589-4eca-bca7-9150038f9ab8-webhook-certs\") pod \"openstack-operator-controller-manager-646f757d77-g8bhc\" (UID: \"a46546b9-e589-4eca-bca7-9150038f9ab8\") " pod="openstack-operators/openstack-operator-controller-manager-646f757d77-g8bhc" Feb 02 09:22:19 crc kubenswrapper[4764]: E0202 09:22:19.552262 4764 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Feb 02 09:22:19 crc kubenswrapper[4764]: E0202 09:22:19.552319 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a46546b9-e589-4eca-bca7-9150038f9ab8-webhook-certs podName:a46546b9-e589-4eca-bca7-9150038f9ab8 nodeName:}" failed. No retries permitted until 2026-02-02 09:22:35.552301105 +0000 UTC m=+918.486025193 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/a46546b9-e589-4eca-bca7-9150038f9ab8-webhook-certs") pod "openstack-operator-controller-manager-646f757d77-g8bhc" (UID: "a46546b9-e589-4eca-bca7-9150038f9ab8") : secret "webhook-server-cert" not found Feb 02 09:22:19 crc kubenswrapper[4764]: E0202 09:22:19.552368 4764 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Feb 02 09:22:19 crc kubenswrapper[4764]: E0202 09:22:19.552388 4764 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a46546b9-e589-4eca-bca7-9150038f9ab8-metrics-certs podName:a46546b9-e589-4eca-bca7-9150038f9ab8 nodeName:}" failed. No retries permitted until 2026-02-02 09:22:35.552381477 +0000 UTC m=+918.486105565 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/a46546b9-e589-4eca-bca7-9150038f9ab8-metrics-certs") pod "openstack-operator-controller-manager-646f757d77-g8bhc" (UID: "a46546b9-e589-4eca-bca7-9150038f9ab8") : secret "metrics-server-cert" not found Feb 02 09:22:19 crc kubenswrapper[4764]: E0202 09:22:19.588976 4764 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/octavia-operator@sha256:e6f2f361f1dcbb321407a5884951e16ff96e7b88942b10b548f27ad4de14a0be" Feb 02 09:22:19 crc kubenswrapper[4764]: E0202 09:22:19.589147 4764 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/octavia-operator@sha256:e6f2f361f1dcbb321407a5884951e16ff96e7b88942b10b548f27ad4de14a0be,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-ppzmn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-6687f8d877-5ncb8_openstack-operators(1844ef02-28fb-45ce-a697-54f1560c9087): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 02 09:22:19 crc kubenswrapper[4764]: E0202 09:22:19.592272 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/octavia-operator-controller-manager-6687f8d877-5ncb8" podUID="1844ef02-28fb-45ce-a697-54f1560c9087" Feb 02 09:22:19 crc kubenswrapper[4764]: E0202 09:22:19.895521 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:e6f2f361f1dcbb321407a5884951e16ff96e7b88942b10b548f27ad4de14a0be\\\"\"" pod="openstack-operators/octavia-operator-controller-manager-6687f8d877-5ncb8" podUID="1844ef02-28fb-45ce-a697-54f1560c9087" Feb 02 09:22:19 crc kubenswrapper[4764]: E0202 09:22:19.895908 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:bbb46b8b3b69fdfad7bafc10a7e88f6ea58bcdc3c91e30beb79e24417d52e0f6\\\"\"" pod="openstack-operators/neutron-operator-controller-manager-585dbc889-7nrsg" podUID="f5cc9236-c70f-4449-a902-ef416aaf8781" Feb 02 09:22:20 crc kubenswrapper[4764]: E0202 09:22:20.213835 4764 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/cinder-operator@sha256:6e21a1dda86ba365817102d23a5d4d2d5dcd1c4d8e5f8d74bd24548aa8c63898" Feb 02 09:22:20 crc kubenswrapper[4764]: E0202 09:22:20.214054 4764 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/cinder-operator@sha256:6e21a1dda86ba365817102d23a5d4d2d5dcd1c4d8e5f8d74bd24548aa8c63898,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-jxd7n,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-operator-controller-manager-8d874c8fc-czpmn_openstack-operators(7c8a2462-306c-43da-a538-9e34a11b5c21): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 02 09:22:20 crc kubenswrapper[4764]: E0202 09:22:20.215426 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/cinder-operator-controller-manager-8d874c8fc-czpmn" podUID="7c8a2462-306c-43da-a538-9e34a11b5c21" Feb 02 09:22:20 crc kubenswrapper[4764]: E0202 09:22:20.900689 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/cinder-operator@sha256:6e21a1dda86ba365817102d23a5d4d2d5dcd1c4d8e5f8d74bd24548aa8c63898\\\"\"" pod="openstack-operators/cinder-operator-controller-manager-8d874c8fc-czpmn" podUID="7c8a2462-306c-43da-a538-9e34a11b5c21" Feb 02 09:22:22 crc kubenswrapper[4764]: E0202 09:22:22.264836 4764 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/manila-operator@sha256:cd911e8d7a7a1104d77691dbaaf54370015cbb82859337746db5a9186d5dc566" Feb 02 09:22:22 crc kubenswrapper[4764]: E0202 09:22:22.265479 4764 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/manila-operator@sha256:cd911e8d7a7a1104d77691dbaaf54370015cbb82859337746db5a9186d5dc566,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-wwxnc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod manila-operator-controller-manager-7dd968899f-g2p6c_openstack-operators(6d4fffad-a306-4f0b-ac40-7fa839f75aa3): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 02 09:22:22 crc kubenswrapper[4764]: E0202 09:22:22.266616 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/manila-operator-controller-manager-7dd968899f-g2p6c" podUID="6d4fffad-a306-4f0b-ac40-7fa839f75aa3" Feb 02 09:22:22 crc kubenswrapper[4764]: E0202 09:22:22.916269 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/manila-operator@sha256:cd911e8d7a7a1104d77691dbaaf54370015cbb82859337746db5a9186d5dc566\\\"\"" pod="openstack-operators/manila-operator-controller-manager-7dd968899f-g2p6c" podUID="6d4fffad-a306-4f0b-ac40-7fa839f75aa3" Feb 02 09:22:23 crc kubenswrapper[4764]: E0202 09:22:23.698670 4764 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/keystone-operator@sha256:319c969e88f109b26487a9f5a67203682803d7386424703ab7ca0340be99ae17" Feb 02 09:22:23 crc kubenswrapper[4764]: E0202 09:22:23.698984 4764 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/keystone-operator@sha256:319c969e88f109b26487a9f5a67203682803d7386424703ab7ca0340be99ae17,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-hpfm5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-84f48565d4-qqnp6_openstack-operators(8564e79b-9abd-41e1-a0a7-0d4b71011300): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 02 09:22:23 crc kubenswrapper[4764]: E0202 09:22:23.700126 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/keystone-operator-controller-manager-84f48565d4-qqnp6" podUID="8564e79b-9abd-41e1-a0a7-0d4b71011300" Feb 02 09:22:23 crc kubenswrapper[4764]: E0202 09:22:23.920974 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/keystone-operator@sha256:319c969e88f109b26487a9f5a67203682803d7386424703ab7ca0340be99ae17\\\"\"" pod="openstack-operators/keystone-operator-controller-manager-84f48565d4-qqnp6" podUID="8564e79b-9abd-41e1-a0a7-0d4b71011300" Feb 02 09:22:32 crc kubenswrapper[4764]: I0202 09:22:32.983518 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7b6c4d8c5f-wtkdr" event={"ID":"9cbd726b-9dbb-46b3-9bcd-f3e9813b4f07","Type":"ContainerStarted","Data":"aa6a1a58cdfe088d0b17f143eb86237e07628ba0c90e75f9615c9cd5db968cab"} Feb 02 09:22:32 crc kubenswrapper[4764]: I0202 09:22:32.984186 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-7b6c4d8c5f-wtkdr" Feb 02 09:22:32 crc kubenswrapper[4764]: I0202 09:22:32.988171 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-8886f4c47-hvtx6" event={"ID":"ee5cb41c-4d3d-4a9d-b1f7-1ede61e7b21c","Type":"ContainerStarted","Data":"29afd7cdea87fd86f0cb3ec6c7a4c16199e1d6d2da1c2c1db7cdeb396f18eb80"} Feb 02 09:22:32 crc kubenswrapper[4764]: I0202 09:22:32.988490 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-8886f4c47-hvtx6" Feb 02 09:22:32 crc kubenswrapper[4764]: I0202 09:22:32.989665 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-5b964cf4cd-kskcp" event={"ID":"1464055e-4745-45e3-be46-5b7700dce733","Type":"ContainerStarted","Data":"ae440bc47724ab28551663acb078053b240836b4981a3a15a069ced5f802dd29"} Feb 02 09:22:32 crc kubenswrapper[4764]: I0202 09:22:32.990269 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-5b964cf4cd-kskcp" Feb 02 09:22:32 crc kubenswrapper[4764]: I0202 09:22:32.993246 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-67bf948998-jhbzc" event={"ID":"9e55cb1e-ed3b-4b7b-8b52-42436eb48223","Type":"ContainerStarted","Data":"19d34fd5be050f06c9587e6c1facb289c2deaaa9a641cc94033393c831cc6ef2"} Feb 02 09:22:32 crc kubenswrapper[4764]: I0202 09:22:32.993498 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-67bf948998-jhbzc" Feb 02 09:22:32 crc kubenswrapper[4764]: I0202 09:22:32.994562 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-69d6db494d-5f766" event={"ID":"becd2ba3-ccad-4ea9-b3c7-6b39f6f8cb3d","Type":"ContainerStarted","Data":"0b3036726919ae222ebc23143207a6a1c7ef7e933248dbddf760a7d45770d9b0"} Feb 02 09:22:32 crc kubenswrapper[4764]: I0202 09:22:32.994849 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-69d6db494d-5f766" Feb 02 09:22:32 crc kubenswrapper[4764]: I0202 09:22:32.995772 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-5fb775575f-ntj47" event={"ID":"82e16b73-9dfe-413b-b3dd-65da79b3d9dd","Type":"ContainerStarted","Data":"20366c8204a0ba402b5e58e171d2908e68ccb9798a267f66d21420fda064c2d5"} Feb 02 09:22:32 crc kubenswrapper[4764]: I0202 09:22:32.996106 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-5fb775575f-ntj47" Feb 02 09:22:32 crc kubenswrapper[4764]: I0202 09:22:32.997275 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-788c46999f-pk2hh" event={"ID":"674c8820-f063-487b-9122-1ed46d1bc39f","Type":"ContainerStarted","Data":"221d507a0696fdf9277f548e3747990a61890442e2874de59fc94097501c404b"} Feb 02 09:22:32 crc kubenswrapper[4764]: I0202 09:22:32.997640 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-788c46999f-pk2hh" Feb 02 09:22:32 crc kubenswrapper[4764]: I0202 09:22:32.999484 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-6d9697b7f4-w4qf5" event={"ID":"8b0e7cb4-257b-403b-8b92-4578ce77c388","Type":"ContainerStarted","Data":"ef19e3e881c99d04cfb8be09d9c4a2ed208e116758d4bf04345fdc78addbab6f"} Feb 02 09:22:32 crc kubenswrapper[4764]: I0202 09:22:32.999850 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-6d9697b7f4-w4qf5" Feb 02 09:22:33 crc kubenswrapper[4764]: I0202 09:22:33.004330 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-64b5b76f97-98xsk" event={"ID":"7c983dd6-f499-41d5-a049-765a15b4f912","Type":"ContainerStarted","Data":"d2d95b04af140c85fbfd239b58e1796da166b32a4ff11d47bfecb13745db58c0"} Feb 02 09:22:33 crc kubenswrapper[4764]: I0202 09:22:33.004927 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-64b5b76f97-98xsk" Feb 02 09:22:33 crc kubenswrapper[4764]: I0202 09:22:33.006449 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-5f4b8bd54d-h4n52" event={"ID":"ad8e7100-d5b8-4a56-bbda-056483f79113","Type":"ContainerStarted","Data":"b867df2f074403cff7e16e79876e57aba72f296456bb1bb3b95b8cb1dd189fcd"} Feb 02 09:22:33 crc kubenswrapper[4764]: I0202 09:22:33.006781 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-5f4b8bd54d-h4n52" Feb 02 09:22:33 crc kubenswrapper[4764]: I0202 09:22:33.007924 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-8d874c8fc-czpmn" event={"ID":"7c8a2462-306c-43da-a538-9e34a11b5c21","Type":"ContainerStarted","Data":"966303fa4fb56e2045e6f2b6d1b744ad221df3844e603d6f56f934910f06b5d1"} Feb 02 09:22:33 crc kubenswrapper[4764]: I0202 09:22:33.008264 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-8d874c8fc-czpmn" Feb 02 09:22:33 crc kubenswrapper[4764]: I0202 09:22:33.012993 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-7b6c4d8c5f-wtkdr" podStartSLOduration=7.217915963 podStartE2EDuration="31.012979059s" podCreationTimestamp="2026-02-02 09:22:02 +0000 UTC" firstStartedPulling="2026-02-02 09:22:04.642400974 +0000 UTC m=+887.576125062" lastFinishedPulling="2026-02-02 09:22:28.43746405 +0000 UTC m=+911.371188158" observedRunningTime="2026-02-02 09:22:33.011967482 +0000 UTC m=+915.945691560" watchObservedRunningTime="2026-02-02 09:22:33.012979059 +0000 UTC m=+915.946703147" Feb 02 09:22:33 crc kubenswrapper[4764]: I0202 09:22:33.035388 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-8d874c8fc-czpmn" podStartSLOduration=3.258533097 podStartE2EDuration="31.035371906s" podCreationTimestamp="2026-02-02 09:22:02 +0000 UTC" firstStartedPulling="2026-02-02 09:22:04.637610962 +0000 UTC m=+887.571335040" lastFinishedPulling="2026-02-02 09:22:32.414449761 +0000 UTC m=+915.348173849" observedRunningTime="2026-02-02 09:22:33.033963507 +0000 UTC m=+915.967687595" watchObservedRunningTime="2026-02-02 09:22:33.035371906 +0000 UTC m=+915.969095994" Feb 02 09:22:33 crc kubenswrapper[4764]: I0202 09:22:33.062038 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-67bf948998-jhbzc" podStartSLOduration=11.369523228 podStartE2EDuration="31.062020529s" podCreationTimestamp="2026-02-02 09:22:02 +0000 UTC" firstStartedPulling="2026-02-02 09:22:04.614521416 +0000 UTC m=+887.548245504" lastFinishedPulling="2026-02-02 09:22:24.307018717 +0000 UTC m=+907.240742805" observedRunningTime="2026-02-02 09:22:33.056457066 +0000 UTC m=+915.990181154" watchObservedRunningTime="2026-02-02 09:22:33.062020529 +0000 UTC m=+915.995744617" Feb 02 09:22:33 crc kubenswrapper[4764]: I0202 09:22:33.123438 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-64b5b76f97-98xsk" podStartSLOduration=3.634771039 podStartE2EDuration="31.123421119s" podCreationTimestamp="2026-02-02 09:22:02 +0000 UTC" firstStartedPulling="2026-02-02 09:22:04.857140552 +0000 UTC m=+887.790864640" lastFinishedPulling="2026-02-02 09:22:32.345790622 +0000 UTC m=+915.279514720" observedRunningTime="2026-02-02 09:22:33.096280882 +0000 UTC m=+916.030004970" watchObservedRunningTime="2026-02-02 09:22:33.123421119 +0000 UTC m=+916.057145207" Feb 02 09:22:33 crc kubenswrapper[4764]: I0202 09:22:33.162951 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-5f4b8bd54d-h4n52" podStartSLOduration=4.237465072 podStartE2EDuration="31.162917385s" podCreationTimestamp="2026-02-02 09:22:02 +0000 UTC" firstStartedPulling="2026-02-02 09:22:04.664665766 +0000 UTC m=+887.598389844" lastFinishedPulling="2026-02-02 09:22:31.590118049 +0000 UTC m=+914.523842157" observedRunningTime="2026-02-02 09:22:33.162456163 +0000 UTC m=+916.096180251" watchObservedRunningTime="2026-02-02 09:22:33.162917385 +0000 UTC m=+916.096641473" Feb 02 09:22:33 crc kubenswrapper[4764]: I0202 09:22:33.166440 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-6d9697b7f4-w4qf5" podStartSLOduration=3.394110887 podStartE2EDuration="31.166433382s" podCreationTimestamp="2026-02-02 09:22:02 +0000 UTC" firstStartedPulling="2026-02-02 09:22:04.642188648 +0000 UTC m=+887.575912726" lastFinishedPulling="2026-02-02 09:22:32.414511133 +0000 UTC m=+915.348235221" observedRunningTime="2026-02-02 09:22:33.124742035 +0000 UTC m=+916.058466123" watchObservedRunningTime="2026-02-02 09:22:33.166433382 +0000 UTC m=+916.100157460" Feb 02 09:22:33 crc kubenswrapper[4764]: I0202 09:22:33.199025 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-8886f4c47-hvtx6" podStartSLOduration=4.109720998 podStartE2EDuration="31.199005318s" podCreationTimestamp="2026-02-02 09:22:02 +0000 UTC" firstStartedPulling="2026-02-02 09:22:03.825339972 +0000 UTC m=+886.759064050" lastFinishedPulling="2026-02-02 09:22:30.914624272 +0000 UTC m=+913.848348370" observedRunningTime="2026-02-02 09:22:33.19505737 +0000 UTC m=+916.128781458" watchObservedRunningTime="2026-02-02 09:22:33.199005318 +0000 UTC m=+916.132729406" Feb 02 09:22:33 crc kubenswrapper[4764]: I0202 09:22:33.229624 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-5fb775575f-ntj47" podStartSLOduration=11.554138138999999 podStartE2EDuration="31.22960569s" podCreationTimestamp="2026-02-02 09:22:02 +0000 UTC" firstStartedPulling="2026-02-02 09:22:04.00879818 +0000 UTC m=+886.942522268" lastFinishedPulling="2026-02-02 09:22:23.684265701 +0000 UTC m=+906.617989819" observedRunningTime="2026-02-02 09:22:33.227619596 +0000 UTC m=+916.161343684" watchObservedRunningTime="2026-02-02 09:22:33.22960569 +0000 UTC m=+916.163329778" Feb 02 09:22:33 crc kubenswrapper[4764]: I0202 09:22:33.264050 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-788c46999f-pk2hh" podStartSLOduration=12.229630597 podStartE2EDuration="31.264029978s" podCreationTimestamp="2026-02-02 09:22:02 +0000 UTC" firstStartedPulling="2026-02-02 09:22:04.65097904 +0000 UTC m=+887.584703128" lastFinishedPulling="2026-02-02 09:22:23.685378421 +0000 UTC m=+906.619102509" observedRunningTime="2026-02-02 09:22:33.252611013 +0000 UTC m=+916.186335101" watchObservedRunningTime="2026-02-02 09:22:33.264029978 +0000 UTC m=+916.197754066" Feb 02 09:22:33 crc kubenswrapper[4764]: I0202 09:22:33.320617 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-5b964cf4cd-kskcp" podStartSLOduration=7.54487743 podStartE2EDuration="31.320599004s" podCreationTimestamp="2026-02-02 09:22:02 +0000 UTC" firstStartedPulling="2026-02-02 09:22:04.663275268 +0000 UTC m=+887.596999356" lastFinishedPulling="2026-02-02 09:22:28.438996802 +0000 UTC m=+911.372720930" observedRunningTime="2026-02-02 09:22:33.316049079 +0000 UTC m=+916.249773167" watchObservedRunningTime="2026-02-02 09:22:33.320599004 +0000 UTC m=+916.254323082" Feb 02 09:22:33 crc kubenswrapper[4764]: I0202 09:22:33.343990 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-69d6db494d-5f766" podStartSLOduration=5.381517972 podStartE2EDuration="31.343974047s" podCreationTimestamp="2026-02-02 09:22:02 +0000 UTC" firstStartedPulling="2026-02-02 09:22:04.180882504 +0000 UTC m=+887.114606582" lastFinishedPulling="2026-02-02 09:22:30.143338569 +0000 UTC m=+913.077062657" observedRunningTime="2026-02-02 09:22:33.342071235 +0000 UTC m=+916.275795323" watchObservedRunningTime="2026-02-02 09:22:33.343974047 +0000 UTC m=+916.277698135" Feb 02 09:22:34 crc kubenswrapper[4764]: I0202 09:22:34.017165 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-qmjfr" event={"ID":"36bffbf8-2209-4410-8005-d0164bb1eb34","Type":"ContainerStarted","Data":"cae4a71da08c82c2d7f6132aa26ef6888d6e5281a87640c5ed695c43141b94f3"} Feb 02 09:22:34 crc kubenswrapper[4764]: I0202 09:22:34.019504 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-68fc8c869-fb6zn" event={"ID":"1a100291-702c-4908-8a63-acb05c4ef5e9","Type":"ContainerStarted","Data":"97ed123b3cb650b6000563ac30e9254ff529ffe9cfe1595dd8d28376a6115675"} Feb 02 09:22:34 crc kubenswrapper[4764]: I0202 09:22:34.019565 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-68fc8c869-fb6zn" Feb 02 09:22:34 crc kubenswrapper[4764]: I0202 09:22:34.020965 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-56f8bfcd9f-bd9cj" event={"ID":"b8748b41-5e08-47bf-a853-ad8c4f2b0908","Type":"ContainerStarted","Data":"3a0f507347aa57d3f02028fe8449793cbfbfcf06fa8bcf793dac870ba8521c54"} Feb 02 09:22:34 crc kubenswrapper[4764]: I0202 09:22:34.021130 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-56f8bfcd9f-bd9cj" Feb 02 09:22:34 crc kubenswrapper[4764]: I0202 09:22:34.022517 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-564965969-6kcc6" event={"ID":"65c22016-b2f2-4a07-87be-480d18870dea","Type":"ContainerStarted","Data":"558daf6836a7cea542021008e0c7cad5534184b730346e3f15255a84e9d36b00"} Feb 02 09:22:34 crc kubenswrapper[4764]: I0202 09:22:34.022685 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-564965969-6kcc6" Feb 02 09:22:34 crc kubenswrapper[4764]: I0202 09:22:34.023918 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-55bff696bd-p7nmb" event={"ID":"e98e75d4-963f-4a67-894e-302b7429eb01","Type":"ContainerStarted","Data":"d813a255d3e7288c306b8e46c1c6deb525bf2cb2940fba8bceadb6929bb06abc"} Feb 02 09:22:34 crc kubenswrapper[4764]: I0202 09:22:34.139776 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-qmjfr" podStartSLOduration=3.748203276 podStartE2EDuration="31.139758894s" podCreationTimestamp="2026-02-02 09:22:03 +0000 UTC" firstStartedPulling="2026-02-02 09:22:04.86866873 +0000 UTC m=+887.802392808" lastFinishedPulling="2026-02-02 09:22:32.260224338 +0000 UTC m=+915.193948426" observedRunningTime="2026-02-02 09:22:34.099804805 +0000 UTC m=+917.033528893" watchObservedRunningTime="2026-02-02 09:22:34.139758894 +0000 UTC m=+917.073482982" Feb 02 09:22:34 crc kubenswrapper[4764]: I0202 09:22:34.141116 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-56f8bfcd9f-bd9cj" podStartSLOduration=4.54591342 podStartE2EDuration="32.141111942s" podCreationTimestamp="2026-02-02 09:22:02 +0000 UTC" firstStartedPulling="2026-02-02 09:22:04.819374913 +0000 UTC m=+887.753099001" lastFinishedPulling="2026-02-02 09:22:32.414573435 +0000 UTC m=+915.348297523" observedRunningTime="2026-02-02 09:22:34.131044295 +0000 UTC m=+917.064768383" watchObservedRunningTime="2026-02-02 09:22:34.141111942 +0000 UTC m=+917.074836030" Feb 02 09:22:34 crc kubenswrapper[4764]: I0202 09:22:34.164587 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-68fc8c869-fb6zn" podStartSLOduration=13.32402888 podStartE2EDuration="32.164567367s" podCreationTimestamp="2026-02-02 09:22:02 +0000 UTC" firstStartedPulling="2026-02-02 09:22:04.844007561 +0000 UTC m=+887.777731649" lastFinishedPulling="2026-02-02 09:22:23.684546048 +0000 UTC m=+906.618270136" observedRunningTime="2026-02-02 09:22:34.156894746 +0000 UTC m=+917.090618844" watchObservedRunningTime="2026-02-02 09:22:34.164567367 +0000 UTC m=+917.098291455" Feb 02 09:22:34 crc kubenswrapper[4764]: I0202 09:22:34.236566 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-55bff696bd-p7nmb" podStartSLOduration=4.774036988 podStartE2EDuration="32.236549728s" podCreationTimestamp="2026-02-02 09:22:02 +0000 UTC" firstStartedPulling="2026-02-02 09:22:04.883242551 +0000 UTC m=+887.816966629" lastFinishedPulling="2026-02-02 09:22:32.345755241 +0000 UTC m=+915.279479369" observedRunningTime="2026-02-02 09:22:34.234214293 +0000 UTC m=+917.167938381" watchObservedRunningTime="2026-02-02 09:22:34.236549728 +0000 UTC m=+917.170273816" Feb 02 09:22:34 crc kubenswrapper[4764]: I0202 09:22:34.238447 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-564965969-6kcc6" podStartSLOduration=4.85334206 podStartE2EDuration="32.23844169s" podCreationTimestamp="2026-02-02 09:22:02 +0000 UTC" firstStartedPulling="2026-02-02 09:22:04.822768087 +0000 UTC m=+887.756492175" lastFinishedPulling="2026-02-02 09:22:32.207867717 +0000 UTC m=+915.141591805" observedRunningTime="2026-02-02 09:22:34.193286807 +0000 UTC m=+917.127010895" watchObservedRunningTime="2026-02-02 09:22:34.23844169 +0000 UTC m=+917.172165778" Feb 02 09:22:34 crc kubenswrapper[4764]: I0202 09:22:34.422518 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b593c3f3-a225-4cb6-b967-4c056e3326a0-cert\") pod \"infra-operator-controller-manager-79955696d6-mxvzm\" (UID: \"b593c3f3-a225-4cb6-b967-4c056e3326a0\") " pod="openstack-operators/infra-operator-controller-manager-79955696d6-mxvzm" Feb 02 09:22:34 crc kubenswrapper[4764]: I0202 09:22:34.428692 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b593c3f3-a225-4cb6-b967-4c056e3326a0-cert\") pod \"infra-operator-controller-manager-79955696d6-mxvzm\" (UID: \"b593c3f3-a225-4cb6-b967-4c056e3326a0\") " pod="openstack-operators/infra-operator-controller-manager-79955696d6-mxvzm" Feb 02 09:22:34 crc kubenswrapper[4764]: I0202 09:22:34.443167 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-f96wx" Feb 02 09:22:34 crc kubenswrapper[4764]: I0202 09:22:34.451705 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-79955696d6-mxvzm" Feb 02 09:22:34 crc kubenswrapper[4764]: I0202 09:22:34.930894 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1b05c36c-dbb6-4ad4-8985-8d64551812ff-cert\") pod \"openstack-baremetal-operator-controller-manager-5ff45dfdbfb5fxl\" (UID: \"1b05c36c-dbb6-4ad4-8985-8d64551812ff\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5ff45dfdbfb5fxl" Feb 02 09:22:34 crc kubenswrapper[4764]: I0202 09:22:34.940531 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1b05c36c-dbb6-4ad4-8985-8d64551812ff-cert\") pod \"openstack-baremetal-operator-controller-manager-5ff45dfdbfb5fxl\" (UID: \"1b05c36c-dbb6-4ad4-8985-8d64551812ff\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5ff45dfdbfb5fxl" Feb 02 09:22:35 crc kubenswrapper[4764]: I0202 09:22:35.016131 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-xjjq7" Feb 02 09:22:35 crc kubenswrapper[4764]: I0202 09:22:35.024367 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5ff45dfdbfb5fxl" Feb 02 09:22:35 crc kubenswrapper[4764]: I0202 09:22:35.030339 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-585dbc889-7nrsg" event={"ID":"f5cc9236-c70f-4449-a902-ef416aaf8781","Type":"ContainerStarted","Data":"ecf4b53d16ca56e9b40a22288e4e33e41260c6833ad2574100378ac85d18fa34"} Feb 02 09:22:35 crc kubenswrapper[4764]: I0202 09:22:35.030584 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-585dbc889-7nrsg" Feb 02 09:22:35 crc kubenswrapper[4764]: I0202 09:22:35.032640 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-6687f8d877-5ncb8" event={"ID":"1844ef02-28fb-45ce-a697-54f1560c9087","Type":"ContainerStarted","Data":"924b69dcdfcdac01ec9b3098e3a7162e240941a43c8b581c4476ac3b58993e09"} Feb 02 09:22:35 crc kubenswrapper[4764]: I0202 09:22:35.081693 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-6687f8d877-5ncb8" podStartSLOduration=2.991482418 podStartE2EDuration="33.081674722s" podCreationTimestamp="2026-02-02 09:22:02 +0000 UTC" firstStartedPulling="2026-02-02 09:22:04.670125346 +0000 UTC m=+887.603849434" lastFinishedPulling="2026-02-02 09:22:34.76031765 +0000 UTC m=+917.694041738" observedRunningTime="2026-02-02 09:22:35.077792776 +0000 UTC m=+918.011516864" watchObservedRunningTime="2026-02-02 09:22:35.081674722 +0000 UTC m=+918.015398810" Feb 02 09:22:35 crc kubenswrapper[4764]: I0202 09:22:35.082534 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-585dbc889-7nrsg" podStartSLOduration=3.572572179 podStartE2EDuration="33.082528036s" podCreationTimestamp="2026-02-02 09:22:02 +0000 UTC" firstStartedPulling="2026-02-02 09:22:04.64190448 +0000 UTC m=+887.575628568" lastFinishedPulling="2026-02-02 09:22:34.151860337 +0000 UTC m=+917.085584425" observedRunningTime="2026-02-02 09:22:35.051144442 +0000 UTC m=+917.984868530" watchObservedRunningTime="2026-02-02 09:22:35.082528036 +0000 UTC m=+918.016252124" Feb 02 09:22:35 crc kubenswrapper[4764]: I0202 09:22:35.096765 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-79955696d6-mxvzm"] Feb 02 09:22:35 crc kubenswrapper[4764]: I0202 09:22:35.607286 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-5ff45dfdbfb5fxl"] Feb 02 09:22:35 crc kubenswrapper[4764]: I0202 09:22:35.638970 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/a46546b9-e589-4eca-bca7-9150038f9ab8-webhook-certs\") pod \"openstack-operator-controller-manager-646f757d77-g8bhc\" (UID: \"a46546b9-e589-4eca-bca7-9150038f9ab8\") " pod="openstack-operators/openstack-operator-controller-manager-646f757d77-g8bhc" Feb 02 09:22:35 crc kubenswrapper[4764]: I0202 09:22:35.639150 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a46546b9-e589-4eca-bca7-9150038f9ab8-metrics-certs\") pod \"openstack-operator-controller-manager-646f757d77-g8bhc\" (UID: \"a46546b9-e589-4eca-bca7-9150038f9ab8\") " pod="openstack-operators/openstack-operator-controller-manager-646f757d77-g8bhc" Feb 02 09:22:35 crc kubenswrapper[4764]: I0202 09:22:35.644495 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/a46546b9-e589-4eca-bca7-9150038f9ab8-webhook-certs\") pod \"openstack-operator-controller-manager-646f757d77-g8bhc\" (UID: \"a46546b9-e589-4eca-bca7-9150038f9ab8\") " pod="openstack-operators/openstack-operator-controller-manager-646f757d77-g8bhc" Feb 02 09:22:35 crc kubenswrapper[4764]: I0202 09:22:35.644631 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a46546b9-e589-4eca-bca7-9150038f9ab8-metrics-certs\") pod \"openstack-operator-controller-manager-646f757d77-g8bhc\" (UID: \"a46546b9-e589-4eca-bca7-9150038f9ab8\") " pod="openstack-operators/openstack-operator-controller-manager-646f757d77-g8bhc" Feb 02 09:22:35 crc kubenswrapper[4764]: I0202 09:22:35.855461 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-ffpxl" Feb 02 09:22:35 crc kubenswrapper[4764]: I0202 09:22:35.865089 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-646f757d77-g8bhc" Feb 02 09:22:36 crc kubenswrapper[4764]: I0202 09:22:36.048710 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-79955696d6-mxvzm" event={"ID":"b593c3f3-a225-4cb6-b967-4c056e3326a0","Type":"ContainerStarted","Data":"2e073e1fefa0160c612e8c98c4610bbb4fb4121bc4b86dd84939fbf29dde15a0"} Feb 02 09:22:36 crc kubenswrapper[4764]: I0202 09:22:36.056236 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5ff45dfdbfb5fxl" event={"ID":"1b05c36c-dbb6-4ad4-8985-8d64551812ff","Type":"ContainerStarted","Data":"f8dd867b95e6883cc035f0fa0248662265fde72175598522f91088060b8bfdcf"} Feb 02 09:22:36 crc kubenswrapper[4764]: I0202 09:22:36.356078 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-646f757d77-g8bhc"] Feb 02 09:22:37 crc kubenswrapper[4764]: I0202 09:22:37.078951 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-646f757d77-g8bhc" event={"ID":"a46546b9-e589-4eca-bca7-9150038f9ab8","Type":"ContainerStarted","Data":"63e3a97da91d6daf76e4244fb94f24bad514b1f97817ae2a0dd08e689ef4260f"} Feb 02 09:22:37 crc kubenswrapper[4764]: I0202 09:22:37.079323 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-646f757d77-g8bhc" event={"ID":"a46546b9-e589-4eca-bca7-9150038f9ab8","Type":"ContainerStarted","Data":"9f40bcc65ed3b1d03400238309b663bfb271d70f41d500977385c51cb4a12ac0"} Feb 02 09:22:37 crc kubenswrapper[4764]: I0202 09:22:37.079472 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-646f757d77-g8bhc" Feb 02 09:22:37 crc kubenswrapper[4764]: I0202 09:22:37.101488 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-646f757d77-g8bhc" podStartSLOduration=34.101464329 podStartE2EDuration="34.101464329s" podCreationTimestamp="2026-02-02 09:22:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:22:37.098848397 +0000 UTC m=+920.032572495" watchObservedRunningTime="2026-02-02 09:22:37.101464329 +0000 UTC m=+920.035188417" Feb 02 09:22:40 crc kubenswrapper[4764]: I0202 09:22:40.131000 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7dd968899f-g2p6c" event={"ID":"6d4fffad-a306-4f0b-ac40-7fa839f75aa3","Type":"ContainerStarted","Data":"ee9069112c641a92488e49d8d265bf9c40ff64a053a7f3529c5356e5ecc6b1a7"} Feb 02 09:22:40 crc kubenswrapper[4764]: I0202 09:22:40.132042 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-7dd968899f-g2p6c" Feb 02 09:22:40 crc kubenswrapper[4764]: I0202 09:22:40.133287 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5ff45dfdbfb5fxl" event={"ID":"1b05c36c-dbb6-4ad4-8985-8d64551812ff","Type":"ContainerStarted","Data":"10d147ed7abab4c2e420c0848451e875500e192dae1ac31b6968406af86d7a3a"} Feb 02 09:22:40 crc kubenswrapper[4764]: I0202 09:22:40.133690 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5ff45dfdbfb5fxl" Feb 02 09:22:40 crc kubenswrapper[4764]: I0202 09:22:40.139277 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-84f48565d4-qqnp6" event={"ID":"8564e79b-9abd-41e1-a0a7-0d4b71011300","Type":"ContainerStarted","Data":"f9dc25cd03a46bf0af61fe5fadbf9c5c643591610b93d95ccf0ce10763a4ff29"} Feb 02 09:22:40 crc kubenswrapper[4764]: I0202 09:22:40.139847 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-84f48565d4-qqnp6" Feb 02 09:22:40 crc kubenswrapper[4764]: I0202 09:22:40.140866 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-79955696d6-mxvzm" event={"ID":"b593c3f3-a225-4cb6-b967-4c056e3326a0","Type":"ContainerStarted","Data":"faa629291a2ef99a9bca5a0140658704dea2ba12e8a4674d5e4c628229ed2531"} Feb 02 09:22:40 crc kubenswrapper[4764]: I0202 09:22:40.141278 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-79955696d6-mxvzm" Feb 02 09:22:40 crc kubenswrapper[4764]: I0202 09:22:40.169673 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-7dd968899f-g2p6c" podStartSLOduration=3.244932622 podStartE2EDuration="38.169656343s" podCreationTimestamp="2026-02-02 09:22:02 +0000 UTC" firstStartedPulling="2026-02-02 09:22:04.610355092 +0000 UTC m=+887.544079180" lastFinishedPulling="2026-02-02 09:22:39.535078813 +0000 UTC m=+922.468802901" observedRunningTime="2026-02-02 09:22:40.164746718 +0000 UTC m=+923.098470806" watchObservedRunningTime="2026-02-02 09:22:40.169656343 +0000 UTC m=+923.103380431" Feb 02 09:22:40 crc kubenswrapper[4764]: I0202 09:22:40.189191 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-79955696d6-mxvzm" podStartSLOduration=33.781714004 podStartE2EDuration="38.18917708s" podCreationTimestamp="2026-02-02 09:22:02 +0000 UTC" firstStartedPulling="2026-02-02 09:22:35.115517624 +0000 UTC m=+918.049241712" lastFinishedPulling="2026-02-02 09:22:39.52298066 +0000 UTC m=+922.456704788" observedRunningTime="2026-02-02 09:22:40.187287098 +0000 UTC m=+923.121011186" watchObservedRunningTime="2026-02-02 09:22:40.18917708 +0000 UTC m=+923.122901168" Feb 02 09:22:40 crc kubenswrapper[4764]: I0202 09:22:40.228289 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5ff45dfdbfb5fxl" podStartSLOduration=34.317074275 podStartE2EDuration="38.228273886s" podCreationTimestamp="2026-02-02 09:22:02 +0000 UTC" firstStartedPulling="2026-02-02 09:22:35.622435492 +0000 UTC m=+918.556159580" lastFinishedPulling="2026-02-02 09:22:39.533635093 +0000 UTC m=+922.467359191" observedRunningTime="2026-02-02 09:22:40.226343033 +0000 UTC m=+923.160067121" watchObservedRunningTime="2026-02-02 09:22:40.228273886 +0000 UTC m=+923.161997974" Feb 02 09:22:40 crc kubenswrapper[4764]: I0202 09:22:40.254010 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-84f48565d4-qqnp6" podStartSLOduration=2.9330201000000002 podStartE2EDuration="38.253993464s" podCreationTimestamp="2026-02-02 09:22:02 +0000 UTC" firstStartedPulling="2026-02-02 09:22:04.205280096 +0000 UTC m=+887.139004184" lastFinishedPulling="2026-02-02 09:22:39.52625343 +0000 UTC m=+922.459977548" observedRunningTime="2026-02-02 09:22:40.251219987 +0000 UTC m=+923.184944075" watchObservedRunningTime="2026-02-02 09:22:40.253993464 +0000 UTC m=+923.187717552" Feb 02 09:22:42 crc kubenswrapper[4764]: I0202 09:22:42.371531 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-7b6c4d8c5f-wtkdr" Feb 02 09:22:42 crc kubenswrapper[4764]: I0202 09:22:42.385545 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-8d874c8fc-czpmn" Feb 02 09:22:42 crc kubenswrapper[4764]: I0202 09:22:42.412207 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-6d9697b7f4-w4qf5" Feb 02 09:22:42 crc kubenswrapper[4764]: I0202 09:22:42.469051 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-8886f4c47-hvtx6" Feb 02 09:22:42 crc kubenswrapper[4764]: I0202 09:22:42.533568 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-69d6db494d-5f766" Feb 02 09:22:42 crc kubenswrapper[4764]: I0202 09:22:42.627995 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-5fb775575f-ntj47" Feb 02 09:22:42 crc kubenswrapper[4764]: I0202 09:22:42.716544 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-5f4b8bd54d-h4n52" Feb 02 09:22:42 crc kubenswrapper[4764]: I0202 09:22:42.937068 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-67bf948998-jhbzc" Feb 02 09:22:42 crc kubenswrapper[4764]: I0202 09:22:42.976572 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-585dbc889-7nrsg" Feb 02 09:22:43 crc kubenswrapper[4764]: I0202 09:22:43.022763 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-55bff696bd-p7nmb" Feb 02 09:22:43 crc kubenswrapper[4764]: I0202 09:22:43.026485 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-55bff696bd-p7nmb" Feb 02 09:22:43 crc kubenswrapper[4764]: I0202 09:22:43.103385 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-6687f8d877-5ncb8" Feb 02 09:22:43 crc kubenswrapper[4764]: I0202 09:22:43.106136 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-6687f8d877-5ncb8" Feb 02 09:22:43 crc kubenswrapper[4764]: I0202 09:22:43.106350 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-788c46999f-pk2hh" Feb 02 09:22:43 crc kubenswrapper[4764]: I0202 09:22:43.281722 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-5b964cf4cd-kskcp" Feb 02 09:22:43 crc kubenswrapper[4764]: I0202 09:22:43.473492 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-64b5b76f97-98xsk" Feb 02 09:22:43 crc kubenswrapper[4764]: I0202 09:22:43.489957 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-68fc8c869-fb6zn" Feb 02 09:22:43 crc kubenswrapper[4764]: I0202 09:22:43.522687 4764 patch_prober.go:28] interesting pod/machine-config-daemon-4ndm4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 09:22:43 crc kubenswrapper[4764]: I0202 09:22:43.522751 4764 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 09:22:43 crc kubenswrapper[4764]: I0202 09:22:43.796634 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-56f8bfcd9f-bd9cj" Feb 02 09:22:43 crc kubenswrapper[4764]: I0202 09:22:43.838865 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-564965969-6kcc6" Feb 02 09:22:44 crc kubenswrapper[4764]: I0202 09:22:44.458242 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-79955696d6-mxvzm" Feb 02 09:22:45 crc kubenswrapper[4764]: I0202 09:22:45.035542 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5ff45dfdbfb5fxl" Feb 02 09:22:45 crc kubenswrapper[4764]: I0202 09:22:45.871797 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-646f757d77-g8bhc" Feb 02 09:22:52 crc kubenswrapper[4764]: I0202 09:22:52.804443 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-84f48565d4-qqnp6" Feb 02 09:22:52 crc kubenswrapper[4764]: I0202 09:22:52.876481 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-7dd968899f-g2p6c" Feb 02 09:22:56 crc kubenswrapper[4764]: I0202 09:22:56.990943 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-sk7ts"] Feb 02 09:22:56 crc kubenswrapper[4764]: I0202 09:22:56.994068 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-sk7ts" Feb 02 09:22:57 crc kubenswrapper[4764]: I0202 09:22:57.032638 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-sk7ts"] Feb 02 09:22:57 crc kubenswrapper[4764]: I0202 09:22:57.087469 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f495491c-57cd-4bc6-a813-8c3d28bc9bae-utilities\") pod \"redhat-marketplace-sk7ts\" (UID: \"f495491c-57cd-4bc6-a813-8c3d28bc9bae\") " pod="openshift-marketplace/redhat-marketplace-sk7ts" Feb 02 09:22:57 crc kubenswrapper[4764]: I0202 09:22:57.087625 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hlw4p\" (UniqueName: \"kubernetes.io/projected/f495491c-57cd-4bc6-a813-8c3d28bc9bae-kube-api-access-hlw4p\") pod \"redhat-marketplace-sk7ts\" (UID: \"f495491c-57cd-4bc6-a813-8c3d28bc9bae\") " pod="openshift-marketplace/redhat-marketplace-sk7ts" Feb 02 09:22:57 crc kubenswrapper[4764]: I0202 09:22:57.087654 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f495491c-57cd-4bc6-a813-8c3d28bc9bae-catalog-content\") pod \"redhat-marketplace-sk7ts\" (UID: \"f495491c-57cd-4bc6-a813-8c3d28bc9bae\") " pod="openshift-marketplace/redhat-marketplace-sk7ts" Feb 02 09:22:57 crc kubenswrapper[4764]: I0202 09:22:57.188617 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hlw4p\" (UniqueName: \"kubernetes.io/projected/f495491c-57cd-4bc6-a813-8c3d28bc9bae-kube-api-access-hlw4p\") pod \"redhat-marketplace-sk7ts\" (UID: \"f495491c-57cd-4bc6-a813-8c3d28bc9bae\") " pod="openshift-marketplace/redhat-marketplace-sk7ts" Feb 02 09:22:57 crc kubenswrapper[4764]: I0202 09:22:57.188660 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f495491c-57cd-4bc6-a813-8c3d28bc9bae-catalog-content\") pod \"redhat-marketplace-sk7ts\" (UID: \"f495491c-57cd-4bc6-a813-8c3d28bc9bae\") " pod="openshift-marketplace/redhat-marketplace-sk7ts" Feb 02 09:22:57 crc kubenswrapper[4764]: I0202 09:22:57.188724 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f495491c-57cd-4bc6-a813-8c3d28bc9bae-utilities\") pod \"redhat-marketplace-sk7ts\" (UID: \"f495491c-57cd-4bc6-a813-8c3d28bc9bae\") " pod="openshift-marketplace/redhat-marketplace-sk7ts" Feb 02 09:22:57 crc kubenswrapper[4764]: I0202 09:22:57.189176 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f495491c-57cd-4bc6-a813-8c3d28bc9bae-utilities\") pod \"redhat-marketplace-sk7ts\" (UID: \"f495491c-57cd-4bc6-a813-8c3d28bc9bae\") " pod="openshift-marketplace/redhat-marketplace-sk7ts" Feb 02 09:22:57 crc kubenswrapper[4764]: I0202 09:22:57.189266 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f495491c-57cd-4bc6-a813-8c3d28bc9bae-catalog-content\") pod \"redhat-marketplace-sk7ts\" (UID: \"f495491c-57cd-4bc6-a813-8c3d28bc9bae\") " pod="openshift-marketplace/redhat-marketplace-sk7ts" Feb 02 09:22:57 crc kubenswrapper[4764]: I0202 09:22:57.217942 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hlw4p\" (UniqueName: \"kubernetes.io/projected/f495491c-57cd-4bc6-a813-8c3d28bc9bae-kube-api-access-hlw4p\") pod \"redhat-marketplace-sk7ts\" (UID: \"f495491c-57cd-4bc6-a813-8c3d28bc9bae\") " pod="openshift-marketplace/redhat-marketplace-sk7ts" Feb 02 09:22:57 crc kubenswrapper[4764]: I0202 09:22:57.324548 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-sk7ts" Feb 02 09:22:57 crc kubenswrapper[4764]: I0202 09:22:57.540889 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-sk7ts"] Feb 02 09:22:58 crc kubenswrapper[4764]: I0202 09:22:58.274831 4764 generic.go:334] "Generic (PLEG): container finished" podID="f495491c-57cd-4bc6-a813-8c3d28bc9bae" containerID="04ad4fdacca5c93b11686db2a2b628692ca4f22cb6c57f19eb7a041be448ca3d" exitCode=0 Feb 02 09:22:58 crc kubenswrapper[4764]: I0202 09:22:58.275177 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sk7ts" event={"ID":"f495491c-57cd-4bc6-a813-8c3d28bc9bae","Type":"ContainerDied","Data":"04ad4fdacca5c93b11686db2a2b628692ca4f22cb6c57f19eb7a041be448ca3d"} Feb 02 09:22:58 crc kubenswrapper[4764]: I0202 09:22:58.275208 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sk7ts" event={"ID":"f495491c-57cd-4bc6-a813-8c3d28bc9bae","Type":"ContainerStarted","Data":"9883185d8ec11af9737029159018489dc2f677552450583cff703cf1d1d1a541"} Feb 02 09:22:58 crc kubenswrapper[4764]: I0202 09:22:58.279876 4764 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 02 09:22:59 crc kubenswrapper[4764]: I0202 09:22:59.285244 4764 generic.go:334] "Generic (PLEG): container finished" podID="f495491c-57cd-4bc6-a813-8c3d28bc9bae" containerID="bf2f4bdb426e99a02c9dfcd2d8d929760648cef1ad0182dfb14ef376f6f4fbdf" exitCode=0 Feb 02 09:22:59 crc kubenswrapper[4764]: I0202 09:22:59.285298 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sk7ts" event={"ID":"f495491c-57cd-4bc6-a813-8c3d28bc9bae","Type":"ContainerDied","Data":"bf2f4bdb426e99a02c9dfcd2d8d929760648cef1ad0182dfb14ef376f6f4fbdf"} Feb 02 09:23:00 crc kubenswrapper[4764]: I0202 09:23:00.296781 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sk7ts" event={"ID":"f495491c-57cd-4bc6-a813-8c3d28bc9bae","Type":"ContainerStarted","Data":"c30f41cc3905ffc0267bea1b6f3167e74db354c8b727c13410d71d190b973863"} Feb 02 09:23:07 crc kubenswrapper[4764]: I0202 09:23:07.325302 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-sk7ts" Feb 02 09:23:07 crc kubenswrapper[4764]: I0202 09:23:07.325846 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-sk7ts" Feb 02 09:23:07 crc kubenswrapper[4764]: I0202 09:23:07.376409 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-sk7ts" Feb 02 09:23:07 crc kubenswrapper[4764]: I0202 09:23:07.404892 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-sk7ts" podStartSLOduration=10.006266364 podStartE2EDuration="11.404876678s" podCreationTimestamp="2026-02-02 09:22:56 +0000 UTC" firstStartedPulling="2026-02-02 09:22:58.279634328 +0000 UTC m=+941.213358416" lastFinishedPulling="2026-02-02 09:22:59.678244622 +0000 UTC m=+942.611968730" observedRunningTime="2026-02-02 09:23:00.321347398 +0000 UTC m=+943.255071496" watchObservedRunningTime="2026-02-02 09:23:07.404876678 +0000 UTC m=+950.338600766" Feb 02 09:23:07 crc kubenswrapper[4764]: I0202 09:23:07.447045 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-sk7ts" Feb 02 09:23:07 crc kubenswrapper[4764]: I0202 09:23:07.617875 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-sk7ts"] Feb 02 09:23:09 crc kubenswrapper[4764]: I0202 09:23:09.380191 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-sk7ts" podUID="f495491c-57cd-4bc6-a813-8c3d28bc9bae" containerName="registry-server" containerID="cri-o://c30f41cc3905ffc0267bea1b6f3167e74db354c8b727c13410d71d190b973863" gracePeriod=2 Feb 02 09:23:10 crc kubenswrapper[4764]: I0202 09:23:10.388995 4764 generic.go:334] "Generic (PLEG): container finished" podID="f495491c-57cd-4bc6-a813-8c3d28bc9bae" containerID="c30f41cc3905ffc0267bea1b6f3167e74db354c8b727c13410d71d190b973863" exitCode=0 Feb 02 09:23:10 crc kubenswrapper[4764]: I0202 09:23:10.389243 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sk7ts" event={"ID":"f495491c-57cd-4bc6-a813-8c3d28bc9bae","Type":"ContainerDied","Data":"c30f41cc3905ffc0267bea1b6f3167e74db354c8b727c13410d71d190b973863"} Feb 02 09:23:10 crc kubenswrapper[4764]: I0202 09:23:10.659980 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-855cbc58c5-zcszj"] Feb 02 09:23:10 crc kubenswrapper[4764]: I0202 09:23:10.664315 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-855cbc58c5-zcszj" Feb 02 09:23:10 crc kubenswrapper[4764]: I0202 09:23:10.685441 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-855cbc58c5-zcszj"] Feb 02 09:23:10 crc kubenswrapper[4764]: I0202 09:23:10.685594 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Feb 02 09:23:10 crc kubenswrapper[4764]: I0202 09:23:10.685793 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-24g5h" Feb 02 09:23:10 crc kubenswrapper[4764]: I0202 09:23:10.694542 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Feb 02 09:23:10 crc kubenswrapper[4764]: I0202 09:23:10.694707 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Feb 02 09:23:10 crc kubenswrapper[4764]: I0202 09:23:10.746897 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/985e7a8f-2a38-4ee8-a541-f2c8b7ff9fde-config\") pod \"dnsmasq-dns-855cbc58c5-zcszj\" (UID: \"985e7a8f-2a38-4ee8-a541-f2c8b7ff9fde\") " pod="openstack/dnsmasq-dns-855cbc58c5-zcszj" Feb 02 09:23:10 crc kubenswrapper[4764]: I0202 09:23:10.747009 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m62f5\" (UniqueName: \"kubernetes.io/projected/985e7a8f-2a38-4ee8-a541-f2c8b7ff9fde-kube-api-access-m62f5\") pod \"dnsmasq-dns-855cbc58c5-zcszj\" (UID: \"985e7a8f-2a38-4ee8-a541-f2c8b7ff9fde\") " pod="openstack/dnsmasq-dns-855cbc58c5-zcszj" Feb 02 09:23:10 crc kubenswrapper[4764]: I0202 09:23:10.849533 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/985e7a8f-2a38-4ee8-a541-f2c8b7ff9fde-config\") pod \"dnsmasq-dns-855cbc58c5-zcszj\" (UID: \"985e7a8f-2a38-4ee8-a541-f2c8b7ff9fde\") " pod="openstack/dnsmasq-dns-855cbc58c5-zcszj" Feb 02 09:23:10 crc kubenswrapper[4764]: I0202 09:23:10.849639 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m62f5\" (UniqueName: \"kubernetes.io/projected/985e7a8f-2a38-4ee8-a541-f2c8b7ff9fde-kube-api-access-m62f5\") pod \"dnsmasq-dns-855cbc58c5-zcszj\" (UID: \"985e7a8f-2a38-4ee8-a541-f2c8b7ff9fde\") " pod="openstack/dnsmasq-dns-855cbc58c5-zcszj" Feb 02 09:23:10 crc kubenswrapper[4764]: I0202 09:23:10.851233 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/985e7a8f-2a38-4ee8-a541-f2c8b7ff9fde-config\") pod \"dnsmasq-dns-855cbc58c5-zcszj\" (UID: \"985e7a8f-2a38-4ee8-a541-f2c8b7ff9fde\") " pod="openstack/dnsmasq-dns-855cbc58c5-zcszj" Feb 02 09:23:10 crc kubenswrapper[4764]: I0202 09:23:10.889271 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m62f5\" (UniqueName: \"kubernetes.io/projected/985e7a8f-2a38-4ee8-a541-f2c8b7ff9fde-kube-api-access-m62f5\") pod \"dnsmasq-dns-855cbc58c5-zcszj\" (UID: \"985e7a8f-2a38-4ee8-a541-f2c8b7ff9fde\") " pod="openstack/dnsmasq-dns-855cbc58c5-zcszj" Feb 02 09:23:11 crc kubenswrapper[4764]: I0202 09:23:10.998945 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-lx9rg"] Feb 02 09:23:11 crc kubenswrapper[4764]: I0202 09:23:11.003432 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lx9rg" Feb 02 09:23:11 crc kubenswrapper[4764]: I0202 09:23:11.015579 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6fcf94d689-4qtb8"] Feb 02 09:23:11 crc kubenswrapper[4764]: I0202 09:23:11.016704 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6fcf94d689-4qtb8" Feb 02 09:23:11 crc kubenswrapper[4764]: I0202 09:23:11.018363 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Feb 02 09:23:11 crc kubenswrapper[4764]: I0202 09:23:11.018878 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-855cbc58c5-zcszj" Feb 02 09:23:11 crc kubenswrapper[4764]: I0202 09:23:11.042418 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-lx9rg"] Feb 02 09:23:11 crc kubenswrapper[4764]: I0202 09:23:11.056394 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/10c729f0-d54f-4b06-995b-ab35e4d583bb-utilities\") pod \"certified-operators-lx9rg\" (UID: \"10c729f0-d54f-4b06-995b-ab35e4d583bb\") " pod="openshift-marketplace/certified-operators-lx9rg" Feb 02 09:23:11 crc kubenswrapper[4764]: I0202 09:23:11.056446 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/10c729f0-d54f-4b06-995b-ab35e4d583bb-catalog-content\") pod \"certified-operators-lx9rg\" (UID: \"10c729f0-d54f-4b06-995b-ab35e4d583bb\") " pod="openshift-marketplace/certified-operators-lx9rg" Feb 02 09:23:11 crc kubenswrapper[4764]: I0202 09:23:11.056467 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xztz7\" (UniqueName: \"kubernetes.io/projected/10c729f0-d54f-4b06-995b-ab35e4d583bb-kube-api-access-xztz7\") pod \"certified-operators-lx9rg\" (UID: \"10c729f0-d54f-4b06-995b-ab35e4d583bb\") " pod="openshift-marketplace/certified-operators-lx9rg" Feb 02 09:23:11 crc kubenswrapper[4764]: I0202 09:23:11.072262 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6fcf94d689-4qtb8"] Feb 02 09:23:11 crc kubenswrapper[4764]: I0202 09:23:11.158579 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rkqmf\" (UniqueName: \"kubernetes.io/projected/58c314a7-0090-49dd-afbd-88826c8733b0-kube-api-access-rkqmf\") pod \"dnsmasq-dns-6fcf94d689-4qtb8\" (UID: \"58c314a7-0090-49dd-afbd-88826c8733b0\") " pod="openstack/dnsmasq-dns-6fcf94d689-4qtb8" Feb 02 09:23:11 crc kubenswrapper[4764]: I0202 09:23:11.158636 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/58c314a7-0090-49dd-afbd-88826c8733b0-dns-svc\") pod \"dnsmasq-dns-6fcf94d689-4qtb8\" (UID: \"58c314a7-0090-49dd-afbd-88826c8733b0\") " pod="openstack/dnsmasq-dns-6fcf94d689-4qtb8" Feb 02 09:23:11 crc kubenswrapper[4764]: I0202 09:23:11.158682 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/10c729f0-d54f-4b06-995b-ab35e4d583bb-utilities\") pod \"certified-operators-lx9rg\" (UID: \"10c729f0-d54f-4b06-995b-ab35e4d583bb\") " pod="openshift-marketplace/certified-operators-lx9rg" Feb 02 09:23:11 crc kubenswrapper[4764]: I0202 09:23:11.158712 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/58c314a7-0090-49dd-afbd-88826c8733b0-config\") pod \"dnsmasq-dns-6fcf94d689-4qtb8\" (UID: \"58c314a7-0090-49dd-afbd-88826c8733b0\") " pod="openstack/dnsmasq-dns-6fcf94d689-4qtb8" Feb 02 09:23:11 crc kubenswrapper[4764]: I0202 09:23:11.158733 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/10c729f0-d54f-4b06-995b-ab35e4d583bb-catalog-content\") pod \"certified-operators-lx9rg\" (UID: \"10c729f0-d54f-4b06-995b-ab35e4d583bb\") " pod="openshift-marketplace/certified-operators-lx9rg" Feb 02 09:23:11 crc kubenswrapper[4764]: I0202 09:23:11.158752 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xztz7\" (UniqueName: \"kubernetes.io/projected/10c729f0-d54f-4b06-995b-ab35e4d583bb-kube-api-access-xztz7\") pod \"certified-operators-lx9rg\" (UID: \"10c729f0-d54f-4b06-995b-ab35e4d583bb\") " pod="openshift-marketplace/certified-operators-lx9rg" Feb 02 09:23:11 crc kubenswrapper[4764]: I0202 09:23:11.159769 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/10c729f0-d54f-4b06-995b-ab35e4d583bb-utilities\") pod \"certified-operators-lx9rg\" (UID: \"10c729f0-d54f-4b06-995b-ab35e4d583bb\") " pod="openshift-marketplace/certified-operators-lx9rg" Feb 02 09:23:11 crc kubenswrapper[4764]: I0202 09:23:11.160009 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/10c729f0-d54f-4b06-995b-ab35e4d583bb-catalog-content\") pod \"certified-operators-lx9rg\" (UID: \"10c729f0-d54f-4b06-995b-ab35e4d583bb\") " pod="openshift-marketplace/certified-operators-lx9rg" Feb 02 09:23:11 crc kubenswrapper[4764]: I0202 09:23:11.187700 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xztz7\" (UniqueName: \"kubernetes.io/projected/10c729f0-d54f-4b06-995b-ab35e4d583bb-kube-api-access-xztz7\") pod \"certified-operators-lx9rg\" (UID: \"10c729f0-d54f-4b06-995b-ab35e4d583bb\") " pod="openshift-marketplace/certified-operators-lx9rg" Feb 02 09:23:11 crc kubenswrapper[4764]: I0202 09:23:11.189239 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-sk7ts" Feb 02 09:23:11 crc kubenswrapper[4764]: I0202 09:23:11.259681 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rkqmf\" (UniqueName: \"kubernetes.io/projected/58c314a7-0090-49dd-afbd-88826c8733b0-kube-api-access-rkqmf\") pod \"dnsmasq-dns-6fcf94d689-4qtb8\" (UID: \"58c314a7-0090-49dd-afbd-88826c8733b0\") " pod="openstack/dnsmasq-dns-6fcf94d689-4qtb8" Feb 02 09:23:11 crc kubenswrapper[4764]: I0202 09:23:11.259731 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/58c314a7-0090-49dd-afbd-88826c8733b0-dns-svc\") pod \"dnsmasq-dns-6fcf94d689-4qtb8\" (UID: \"58c314a7-0090-49dd-afbd-88826c8733b0\") " pod="openstack/dnsmasq-dns-6fcf94d689-4qtb8" Feb 02 09:23:11 crc kubenswrapper[4764]: I0202 09:23:11.259773 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/58c314a7-0090-49dd-afbd-88826c8733b0-config\") pod \"dnsmasq-dns-6fcf94d689-4qtb8\" (UID: \"58c314a7-0090-49dd-afbd-88826c8733b0\") " pod="openstack/dnsmasq-dns-6fcf94d689-4qtb8" Feb 02 09:23:11 crc kubenswrapper[4764]: I0202 09:23:11.260684 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/58c314a7-0090-49dd-afbd-88826c8733b0-config\") pod \"dnsmasq-dns-6fcf94d689-4qtb8\" (UID: \"58c314a7-0090-49dd-afbd-88826c8733b0\") " pod="openstack/dnsmasq-dns-6fcf94d689-4qtb8" Feb 02 09:23:11 crc kubenswrapper[4764]: I0202 09:23:11.262547 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/58c314a7-0090-49dd-afbd-88826c8733b0-dns-svc\") pod \"dnsmasq-dns-6fcf94d689-4qtb8\" (UID: \"58c314a7-0090-49dd-afbd-88826c8733b0\") " pod="openstack/dnsmasq-dns-6fcf94d689-4qtb8" Feb 02 09:23:11 crc kubenswrapper[4764]: I0202 09:23:11.277701 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rkqmf\" (UniqueName: \"kubernetes.io/projected/58c314a7-0090-49dd-afbd-88826c8733b0-kube-api-access-rkqmf\") pod \"dnsmasq-dns-6fcf94d689-4qtb8\" (UID: \"58c314a7-0090-49dd-afbd-88826c8733b0\") " pod="openstack/dnsmasq-dns-6fcf94d689-4qtb8" Feb 02 09:23:11 crc kubenswrapper[4764]: I0202 09:23:11.346625 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lx9rg" Feb 02 09:23:11 crc kubenswrapper[4764]: I0202 09:23:11.360572 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hlw4p\" (UniqueName: \"kubernetes.io/projected/f495491c-57cd-4bc6-a813-8c3d28bc9bae-kube-api-access-hlw4p\") pod \"f495491c-57cd-4bc6-a813-8c3d28bc9bae\" (UID: \"f495491c-57cd-4bc6-a813-8c3d28bc9bae\") " Feb 02 09:23:11 crc kubenswrapper[4764]: I0202 09:23:11.360697 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f495491c-57cd-4bc6-a813-8c3d28bc9bae-catalog-content\") pod \"f495491c-57cd-4bc6-a813-8c3d28bc9bae\" (UID: \"f495491c-57cd-4bc6-a813-8c3d28bc9bae\") " Feb 02 09:23:11 crc kubenswrapper[4764]: I0202 09:23:11.360784 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f495491c-57cd-4bc6-a813-8c3d28bc9bae-utilities\") pod \"f495491c-57cd-4bc6-a813-8c3d28bc9bae\" (UID: \"f495491c-57cd-4bc6-a813-8c3d28bc9bae\") " Feb 02 09:23:11 crc kubenswrapper[4764]: I0202 09:23:11.361811 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f495491c-57cd-4bc6-a813-8c3d28bc9bae-utilities" (OuterVolumeSpecName: "utilities") pod "f495491c-57cd-4bc6-a813-8c3d28bc9bae" (UID: "f495491c-57cd-4bc6-a813-8c3d28bc9bae"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 09:23:11 crc kubenswrapper[4764]: I0202 09:23:11.366132 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f495491c-57cd-4bc6-a813-8c3d28bc9bae-kube-api-access-hlw4p" (OuterVolumeSpecName: "kube-api-access-hlw4p") pod "f495491c-57cd-4bc6-a813-8c3d28bc9bae" (UID: "f495491c-57cd-4bc6-a813-8c3d28bc9bae"). InnerVolumeSpecName "kube-api-access-hlw4p". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:23:11 crc kubenswrapper[4764]: I0202 09:23:11.376757 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f495491c-57cd-4bc6-a813-8c3d28bc9bae-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f495491c-57cd-4bc6-a813-8c3d28bc9bae" (UID: "f495491c-57cd-4bc6-a813-8c3d28bc9bae"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 09:23:11 crc kubenswrapper[4764]: I0202 09:23:11.399214 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sk7ts" event={"ID":"f495491c-57cd-4bc6-a813-8c3d28bc9bae","Type":"ContainerDied","Data":"9883185d8ec11af9737029159018489dc2f677552450583cff703cf1d1d1a541"} Feb 02 09:23:11 crc kubenswrapper[4764]: I0202 09:23:11.399280 4764 scope.go:117] "RemoveContainer" containerID="c30f41cc3905ffc0267bea1b6f3167e74db354c8b727c13410d71d190b973863" Feb 02 09:23:11 crc kubenswrapper[4764]: I0202 09:23:11.399303 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-sk7ts" Feb 02 09:23:11 crc kubenswrapper[4764]: I0202 09:23:11.417585 4764 scope.go:117] "RemoveContainer" containerID="bf2f4bdb426e99a02c9dfcd2d8d929760648cef1ad0182dfb14ef376f6f4fbdf" Feb 02 09:23:11 crc kubenswrapper[4764]: I0202 09:23:11.454852 4764 scope.go:117] "RemoveContainer" containerID="04ad4fdacca5c93b11686db2a2b628692ca4f22cb6c57f19eb7a041be448ca3d" Feb 02 09:23:11 crc kubenswrapper[4764]: I0202 09:23:11.456194 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-sk7ts"] Feb 02 09:23:11 crc kubenswrapper[4764]: I0202 09:23:11.461893 4764 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f495491c-57cd-4bc6-a813-8c3d28bc9bae-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 09:23:11 crc kubenswrapper[4764]: I0202 09:23:11.461915 4764 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f495491c-57cd-4bc6-a813-8c3d28bc9bae-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 09:23:11 crc kubenswrapper[4764]: I0202 09:23:11.461924 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hlw4p\" (UniqueName: \"kubernetes.io/projected/f495491c-57cd-4bc6-a813-8c3d28bc9bae-kube-api-access-hlw4p\") on node \"crc\" DevicePath \"\"" Feb 02 09:23:11 crc kubenswrapper[4764]: I0202 09:23:11.469871 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6fcf94d689-4qtb8" Feb 02 09:23:11 crc kubenswrapper[4764]: I0202 09:23:11.471557 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-sk7ts"] Feb 02 09:23:11 crc kubenswrapper[4764]: I0202 09:23:11.539387 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-855cbc58c5-zcszj"] Feb 02 09:23:11 crc kubenswrapper[4764]: I0202 09:23:11.595547 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-lx9rg"] Feb 02 09:23:11 crc kubenswrapper[4764]: I0202 09:23:11.832992 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f495491c-57cd-4bc6-a813-8c3d28bc9bae" path="/var/lib/kubelet/pods/f495491c-57cd-4bc6-a813-8c3d28bc9bae/volumes" Feb 02 09:23:12 crc kubenswrapper[4764]: I0202 09:23:12.008638 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6fcf94d689-4qtb8"] Feb 02 09:23:12 crc kubenswrapper[4764]: I0202 09:23:12.422089 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-855cbc58c5-zcszj" event={"ID":"985e7a8f-2a38-4ee8-a541-f2c8b7ff9fde","Type":"ContainerStarted","Data":"e707d1b26af371fcf8f95e95d6de95b9112fbf47773ac827fb844ae58376f42f"} Feb 02 09:23:12 crc kubenswrapper[4764]: I0202 09:23:12.440642 4764 generic.go:334] "Generic (PLEG): container finished" podID="10c729f0-d54f-4b06-995b-ab35e4d583bb" containerID="16b97d45425301e0814fee1ea9af4f1211a46ca037f1c773b9c55603eb0a3de5" exitCode=0 Feb 02 09:23:12 crc kubenswrapper[4764]: I0202 09:23:12.440765 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lx9rg" event={"ID":"10c729f0-d54f-4b06-995b-ab35e4d583bb","Type":"ContainerDied","Data":"16b97d45425301e0814fee1ea9af4f1211a46ca037f1c773b9c55603eb0a3de5"} Feb 02 09:23:12 crc kubenswrapper[4764]: I0202 09:23:12.440799 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lx9rg" event={"ID":"10c729f0-d54f-4b06-995b-ab35e4d583bb","Type":"ContainerStarted","Data":"44ce85ed6b7ce24f7e9e665a4856d6eb03422a9ddeb88c2f6d4222de281cf391"} Feb 02 09:23:12 crc kubenswrapper[4764]: I0202 09:23:12.453477 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6fcf94d689-4qtb8"] Feb 02 09:23:12 crc kubenswrapper[4764]: I0202 09:23:12.462996 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6fcf94d689-4qtb8" event={"ID":"58c314a7-0090-49dd-afbd-88826c8733b0","Type":"ContainerStarted","Data":"3685d9d71b7ceae49d109efc666e0e6c4a7e2c69318c9c0e08c1dc333747764e"} Feb 02 09:23:12 crc kubenswrapper[4764]: I0202 09:23:12.526705 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-f54874ffc-wmsqt"] Feb 02 09:23:12 crc kubenswrapper[4764]: E0202 09:23:12.527026 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f495491c-57cd-4bc6-a813-8c3d28bc9bae" containerName="extract-utilities" Feb 02 09:23:12 crc kubenswrapper[4764]: I0202 09:23:12.527043 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="f495491c-57cd-4bc6-a813-8c3d28bc9bae" containerName="extract-utilities" Feb 02 09:23:12 crc kubenswrapper[4764]: E0202 09:23:12.527058 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f495491c-57cd-4bc6-a813-8c3d28bc9bae" containerName="registry-server" Feb 02 09:23:12 crc kubenswrapper[4764]: I0202 09:23:12.527065 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="f495491c-57cd-4bc6-a813-8c3d28bc9bae" containerName="registry-server" Feb 02 09:23:12 crc kubenswrapper[4764]: E0202 09:23:12.527075 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f495491c-57cd-4bc6-a813-8c3d28bc9bae" containerName="extract-content" Feb 02 09:23:12 crc kubenswrapper[4764]: I0202 09:23:12.527082 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="f495491c-57cd-4bc6-a813-8c3d28bc9bae" containerName="extract-content" Feb 02 09:23:12 crc kubenswrapper[4764]: I0202 09:23:12.527227 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="f495491c-57cd-4bc6-a813-8c3d28bc9bae" containerName="registry-server" Feb 02 09:23:12 crc kubenswrapper[4764]: I0202 09:23:12.527962 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f54874ffc-wmsqt" Feb 02 09:23:12 crc kubenswrapper[4764]: I0202 09:23:12.534574 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-f54874ffc-wmsqt"] Feb 02 09:23:12 crc kubenswrapper[4764]: I0202 09:23:12.597860 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/85525aa3-d70a-4d66-8bbf-536be10a0337-dns-svc\") pod \"dnsmasq-dns-f54874ffc-wmsqt\" (UID: \"85525aa3-d70a-4d66-8bbf-536be10a0337\") " pod="openstack/dnsmasq-dns-f54874ffc-wmsqt" Feb 02 09:23:12 crc kubenswrapper[4764]: I0202 09:23:12.597946 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/85525aa3-d70a-4d66-8bbf-536be10a0337-config\") pod \"dnsmasq-dns-f54874ffc-wmsqt\" (UID: \"85525aa3-d70a-4d66-8bbf-536be10a0337\") " pod="openstack/dnsmasq-dns-f54874ffc-wmsqt" Feb 02 09:23:12 crc kubenswrapper[4764]: I0202 09:23:12.597983 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jdr4z\" (UniqueName: \"kubernetes.io/projected/85525aa3-d70a-4d66-8bbf-536be10a0337-kube-api-access-jdr4z\") pod \"dnsmasq-dns-f54874ffc-wmsqt\" (UID: \"85525aa3-d70a-4d66-8bbf-536be10a0337\") " pod="openstack/dnsmasq-dns-f54874ffc-wmsqt" Feb 02 09:23:12 crc kubenswrapper[4764]: I0202 09:23:12.698711 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jdr4z\" (UniqueName: \"kubernetes.io/projected/85525aa3-d70a-4d66-8bbf-536be10a0337-kube-api-access-jdr4z\") pod \"dnsmasq-dns-f54874ffc-wmsqt\" (UID: \"85525aa3-d70a-4d66-8bbf-536be10a0337\") " pod="openstack/dnsmasq-dns-f54874ffc-wmsqt" Feb 02 09:23:12 crc kubenswrapper[4764]: I0202 09:23:12.698810 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/85525aa3-d70a-4d66-8bbf-536be10a0337-dns-svc\") pod \"dnsmasq-dns-f54874ffc-wmsqt\" (UID: \"85525aa3-d70a-4d66-8bbf-536be10a0337\") " pod="openstack/dnsmasq-dns-f54874ffc-wmsqt" Feb 02 09:23:12 crc kubenswrapper[4764]: I0202 09:23:12.698848 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/85525aa3-d70a-4d66-8bbf-536be10a0337-config\") pod \"dnsmasq-dns-f54874ffc-wmsqt\" (UID: \"85525aa3-d70a-4d66-8bbf-536be10a0337\") " pod="openstack/dnsmasq-dns-f54874ffc-wmsqt" Feb 02 09:23:12 crc kubenswrapper[4764]: I0202 09:23:12.699796 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/85525aa3-d70a-4d66-8bbf-536be10a0337-config\") pod \"dnsmasq-dns-f54874ffc-wmsqt\" (UID: \"85525aa3-d70a-4d66-8bbf-536be10a0337\") " pod="openstack/dnsmasq-dns-f54874ffc-wmsqt" Feb 02 09:23:12 crc kubenswrapper[4764]: I0202 09:23:12.699802 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/85525aa3-d70a-4d66-8bbf-536be10a0337-dns-svc\") pod \"dnsmasq-dns-f54874ffc-wmsqt\" (UID: \"85525aa3-d70a-4d66-8bbf-536be10a0337\") " pod="openstack/dnsmasq-dns-f54874ffc-wmsqt" Feb 02 09:23:12 crc kubenswrapper[4764]: I0202 09:23:12.728662 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jdr4z\" (UniqueName: \"kubernetes.io/projected/85525aa3-d70a-4d66-8bbf-536be10a0337-kube-api-access-jdr4z\") pod \"dnsmasq-dns-f54874ffc-wmsqt\" (UID: \"85525aa3-d70a-4d66-8bbf-536be10a0337\") " pod="openstack/dnsmasq-dns-f54874ffc-wmsqt" Feb 02 09:23:12 crc kubenswrapper[4764]: I0202 09:23:12.905537 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f54874ffc-wmsqt" Feb 02 09:23:13 crc kubenswrapper[4764]: I0202 09:23:13.040892 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-855cbc58c5-zcszj"] Feb 02 09:23:13 crc kubenswrapper[4764]: I0202 09:23:13.083040 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-67ff45466c-v6vn7"] Feb 02 09:23:13 crc kubenswrapper[4764]: I0202 09:23:13.084552 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67ff45466c-v6vn7" Feb 02 09:23:13 crc kubenswrapper[4764]: I0202 09:23:13.105686 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-67ff45466c-v6vn7"] Feb 02 09:23:13 crc kubenswrapper[4764]: I0202 09:23:13.206683 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210b0439-a7e6-4f9f-b847-0a04cf044f87-config\") pod \"dnsmasq-dns-67ff45466c-v6vn7\" (UID: \"210b0439-a7e6-4f9f-b847-0a04cf044f87\") " pod="openstack/dnsmasq-dns-67ff45466c-v6vn7" Feb 02 09:23:13 crc kubenswrapper[4764]: I0202 09:23:13.206774 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-54w4t\" (UniqueName: \"kubernetes.io/projected/210b0439-a7e6-4f9f-b847-0a04cf044f87-kube-api-access-54w4t\") pod \"dnsmasq-dns-67ff45466c-v6vn7\" (UID: \"210b0439-a7e6-4f9f-b847-0a04cf044f87\") " pod="openstack/dnsmasq-dns-67ff45466c-v6vn7" Feb 02 09:23:13 crc kubenswrapper[4764]: I0202 09:23:13.206819 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/210b0439-a7e6-4f9f-b847-0a04cf044f87-dns-svc\") pod \"dnsmasq-dns-67ff45466c-v6vn7\" (UID: \"210b0439-a7e6-4f9f-b847-0a04cf044f87\") " pod="openstack/dnsmasq-dns-67ff45466c-v6vn7" Feb 02 09:23:13 crc kubenswrapper[4764]: I0202 09:23:13.309633 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-54w4t\" (UniqueName: \"kubernetes.io/projected/210b0439-a7e6-4f9f-b847-0a04cf044f87-kube-api-access-54w4t\") pod \"dnsmasq-dns-67ff45466c-v6vn7\" (UID: \"210b0439-a7e6-4f9f-b847-0a04cf044f87\") " pod="openstack/dnsmasq-dns-67ff45466c-v6vn7" Feb 02 09:23:13 crc kubenswrapper[4764]: I0202 09:23:13.309725 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/210b0439-a7e6-4f9f-b847-0a04cf044f87-dns-svc\") pod \"dnsmasq-dns-67ff45466c-v6vn7\" (UID: \"210b0439-a7e6-4f9f-b847-0a04cf044f87\") " pod="openstack/dnsmasq-dns-67ff45466c-v6vn7" Feb 02 09:23:13 crc kubenswrapper[4764]: I0202 09:23:13.309807 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210b0439-a7e6-4f9f-b847-0a04cf044f87-config\") pod \"dnsmasq-dns-67ff45466c-v6vn7\" (UID: \"210b0439-a7e6-4f9f-b847-0a04cf044f87\") " pod="openstack/dnsmasq-dns-67ff45466c-v6vn7" Feb 02 09:23:13 crc kubenswrapper[4764]: I0202 09:23:13.311012 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210b0439-a7e6-4f9f-b847-0a04cf044f87-config\") pod \"dnsmasq-dns-67ff45466c-v6vn7\" (UID: \"210b0439-a7e6-4f9f-b847-0a04cf044f87\") " pod="openstack/dnsmasq-dns-67ff45466c-v6vn7" Feb 02 09:23:13 crc kubenswrapper[4764]: I0202 09:23:13.311205 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/210b0439-a7e6-4f9f-b847-0a04cf044f87-dns-svc\") pod \"dnsmasq-dns-67ff45466c-v6vn7\" (UID: \"210b0439-a7e6-4f9f-b847-0a04cf044f87\") " pod="openstack/dnsmasq-dns-67ff45466c-v6vn7" Feb 02 09:23:13 crc kubenswrapper[4764]: I0202 09:23:13.339324 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-54w4t\" (UniqueName: \"kubernetes.io/projected/210b0439-a7e6-4f9f-b847-0a04cf044f87-kube-api-access-54w4t\") pod \"dnsmasq-dns-67ff45466c-v6vn7\" (UID: \"210b0439-a7e6-4f9f-b847-0a04cf044f87\") " pod="openstack/dnsmasq-dns-67ff45466c-v6vn7" Feb 02 09:23:13 crc kubenswrapper[4764]: I0202 09:23:13.424368 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67ff45466c-v6vn7" Feb 02 09:23:13 crc kubenswrapper[4764]: I0202 09:23:13.523327 4764 patch_prober.go:28] interesting pod/machine-config-daemon-4ndm4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 09:23:13 crc kubenswrapper[4764]: I0202 09:23:13.523704 4764 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 09:23:13 crc kubenswrapper[4764]: I0202 09:23:13.523807 4764 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" Feb 02 09:23:13 crc kubenswrapper[4764]: I0202 09:23:13.524702 4764 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a4c68bd9cbe36bd80883876d850d402edf3b17e0660281a9f5527447a8167a49"} pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 02 09:23:13 crc kubenswrapper[4764]: I0202 09:23:13.524772 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" containerName="machine-config-daemon" containerID="cri-o://a4c68bd9cbe36bd80883876d850d402edf3b17e0660281a9f5527447a8167a49" gracePeriod=600 Feb 02 09:23:13 crc kubenswrapper[4764]: I0202 09:23:13.590417 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-f54874ffc-wmsqt"] Feb 02 09:23:13 crc kubenswrapper[4764]: I0202 09:23:13.687657 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Feb 02 09:23:13 crc kubenswrapper[4764]: I0202 09:23:13.690004 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Feb 02 09:23:13 crc kubenswrapper[4764]: I0202 09:23:13.692675 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Feb 02 09:23:13 crc kubenswrapper[4764]: I0202 09:23:13.692829 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Feb 02 09:23:13 crc kubenswrapper[4764]: I0202 09:23:13.693083 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-9q4h7" Feb 02 09:23:13 crc kubenswrapper[4764]: I0202 09:23:13.694096 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Feb 02 09:23:13 crc kubenswrapper[4764]: I0202 09:23:13.699542 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Feb 02 09:23:13 crc kubenswrapper[4764]: I0202 09:23:13.699777 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Feb 02 09:23:13 crc kubenswrapper[4764]: I0202 09:23:13.702511 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Feb 02 09:23:13 crc kubenswrapper[4764]: I0202 09:23:13.706662 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 02 09:23:13 crc kubenswrapper[4764]: I0202 09:23:13.816332 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/53d5d300-11ae-408d-8bb8-347defb93991-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"53d5d300-11ae-408d-8bb8-347defb93991\") " pod="openstack/rabbitmq-server-0" Feb 02 09:23:13 crc kubenswrapper[4764]: I0202 09:23:13.816795 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/53d5d300-11ae-408d-8bb8-347defb93991-server-conf\") pod \"rabbitmq-server-0\" (UID: \"53d5d300-11ae-408d-8bb8-347defb93991\") " pod="openstack/rabbitmq-server-0" Feb 02 09:23:13 crc kubenswrapper[4764]: I0202 09:23:13.816823 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-server-0\" (UID: \"53d5d300-11ae-408d-8bb8-347defb93991\") " pod="openstack/rabbitmq-server-0" Feb 02 09:23:13 crc kubenswrapper[4764]: I0202 09:23:13.816876 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/53d5d300-11ae-408d-8bb8-347defb93991-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"53d5d300-11ae-408d-8bb8-347defb93991\") " pod="openstack/rabbitmq-server-0" Feb 02 09:23:13 crc kubenswrapper[4764]: I0202 09:23:13.816898 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/53d5d300-11ae-408d-8bb8-347defb93991-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"53d5d300-11ae-408d-8bb8-347defb93991\") " pod="openstack/rabbitmq-server-0" Feb 02 09:23:13 crc kubenswrapper[4764]: I0202 09:23:13.816951 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vs58s\" (UniqueName: \"kubernetes.io/projected/53d5d300-11ae-408d-8bb8-347defb93991-kube-api-access-vs58s\") pod \"rabbitmq-server-0\" (UID: \"53d5d300-11ae-408d-8bb8-347defb93991\") " pod="openstack/rabbitmq-server-0" Feb 02 09:23:13 crc kubenswrapper[4764]: I0202 09:23:13.816976 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/53d5d300-11ae-408d-8bb8-347defb93991-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"53d5d300-11ae-408d-8bb8-347defb93991\") " pod="openstack/rabbitmq-server-0" Feb 02 09:23:13 crc kubenswrapper[4764]: I0202 09:23:13.816994 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/53d5d300-11ae-408d-8bb8-347defb93991-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"53d5d300-11ae-408d-8bb8-347defb93991\") " pod="openstack/rabbitmq-server-0" Feb 02 09:23:13 crc kubenswrapper[4764]: I0202 09:23:13.817018 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/53d5d300-11ae-408d-8bb8-347defb93991-config-data\") pod \"rabbitmq-server-0\" (UID: \"53d5d300-11ae-408d-8bb8-347defb93991\") " pod="openstack/rabbitmq-server-0" Feb 02 09:23:13 crc kubenswrapper[4764]: I0202 09:23:13.817037 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/53d5d300-11ae-408d-8bb8-347defb93991-pod-info\") pod \"rabbitmq-server-0\" (UID: \"53d5d300-11ae-408d-8bb8-347defb93991\") " pod="openstack/rabbitmq-server-0" Feb 02 09:23:13 crc kubenswrapper[4764]: I0202 09:23:13.817072 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/53d5d300-11ae-408d-8bb8-347defb93991-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"53d5d300-11ae-408d-8bb8-347defb93991\") " pod="openstack/rabbitmq-server-0" Feb 02 09:23:13 crc kubenswrapper[4764]: I0202 09:23:13.918382 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vs58s\" (UniqueName: \"kubernetes.io/projected/53d5d300-11ae-408d-8bb8-347defb93991-kube-api-access-vs58s\") pod \"rabbitmq-server-0\" (UID: \"53d5d300-11ae-408d-8bb8-347defb93991\") " pod="openstack/rabbitmq-server-0" Feb 02 09:23:13 crc kubenswrapper[4764]: I0202 09:23:13.918449 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/53d5d300-11ae-408d-8bb8-347defb93991-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"53d5d300-11ae-408d-8bb8-347defb93991\") " pod="openstack/rabbitmq-server-0" Feb 02 09:23:13 crc kubenswrapper[4764]: I0202 09:23:13.918472 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/53d5d300-11ae-408d-8bb8-347defb93991-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"53d5d300-11ae-408d-8bb8-347defb93991\") " pod="openstack/rabbitmq-server-0" Feb 02 09:23:13 crc kubenswrapper[4764]: I0202 09:23:13.918498 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/53d5d300-11ae-408d-8bb8-347defb93991-config-data\") pod \"rabbitmq-server-0\" (UID: \"53d5d300-11ae-408d-8bb8-347defb93991\") " pod="openstack/rabbitmq-server-0" Feb 02 09:23:13 crc kubenswrapper[4764]: I0202 09:23:13.918522 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/53d5d300-11ae-408d-8bb8-347defb93991-pod-info\") pod \"rabbitmq-server-0\" (UID: \"53d5d300-11ae-408d-8bb8-347defb93991\") " pod="openstack/rabbitmq-server-0" Feb 02 09:23:13 crc kubenswrapper[4764]: I0202 09:23:13.918559 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/53d5d300-11ae-408d-8bb8-347defb93991-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"53d5d300-11ae-408d-8bb8-347defb93991\") " pod="openstack/rabbitmq-server-0" Feb 02 09:23:13 crc kubenswrapper[4764]: I0202 09:23:13.918598 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/53d5d300-11ae-408d-8bb8-347defb93991-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"53d5d300-11ae-408d-8bb8-347defb93991\") " pod="openstack/rabbitmq-server-0" Feb 02 09:23:13 crc kubenswrapper[4764]: I0202 09:23:13.918621 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/53d5d300-11ae-408d-8bb8-347defb93991-server-conf\") pod \"rabbitmq-server-0\" (UID: \"53d5d300-11ae-408d-8bb8-347defb93991\") " pod="openstack/rabbitmq-server-0" Feb 02 09:23:13 crc kubenswrapper[4764]: I0202 09:23:13.918650 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-server-0\" (UID: \"53d5d300-11ae-408d-8bb8-347defb93991\") " pod="openstack/rabbitmq-server-0" Feb 02 09:23:13 crc kubenswrapper[4764]: I0202 09:23:13.918710 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/53d5d300-11ae-408d-8bb8-347defb93991-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"53d5d300-11ae-408d-8bb8-347defb93991\") " pod="openstack/rabbitmq-server-0" Feb 02 09:23:13 crc kubenswrapper[4764]: I0202 09:23:13.918733 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/53d5d300-11ae-408d-8bb8-347defb93991-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"53d5d300-11ae-408d-8bb8-347defb93991\") " pod="openstack/rabbitmq-server-0" Feb 02 09:23:13 crc kubenswrapper[4764]: I0202 09:23:13.919493 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/53d5d300-11ae-408d-8bb8-347defb93991-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"53d5d300-11ae-408d-8bb8-347defb93991\") " pod="openstack/rabbitmq-server-0" Feb 02 09:23:13 crc kubenswrapper[4764]: I0202 09:23:13.920046 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/53d5d300-11ae-408d-8bb8-347defb93991-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"53d5d300-11ae-408d-8bb8-347defb93991\") " pod="openstack/rabbitmq-server-0" Feb 02 09:23:13 crc kubenswrapper[4764]: I0202 09:23:13.920613 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/53d5d300-11ae-408d-8bb8-347defb93991-config-data\") pod \"rabbitmq-server-0\" (UID: \"53d5d300-11ae-408d-8bb8-347defb93991\") " pod="openstack/rabbitmq-server-0" Feb 02 09:23:13 crc kubenswrapper[4764]: I0202 09:23:13.921152 4764 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-server-0\" (UID: \"53d5d300-11ae-408d-8bb8-347defb93991\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/rabbitmq-server-0" Feb 02 09:23:13 crc kubenswrapper[4764]: I0202 09:23:13.921702 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/53d5d300-11ae-408d-8bb8-347defb93991-server-conf\") pod \"rabbitmq-server-0\" (UID: \"53d5d300-11ae-408d-8bb8-347defb93991\") " pod="openstack/rabbitmq-server-0" Feb 02 09:23:13 crc kubenswrapper[4764]: I0202 09:23:13.921816 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/53d5d300-11ae-408d-8bb8-347defb93991-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"53d5d300-11ae-408d-8bb8-347defb93991\") " pod="openstack/rabbitmq-server-0" Feb 02 09:23:13 crc kubenswrapper[4764]: I0202 09:23:13.928381 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/53d5d300-11ae-408d-8bb8-347defb93991-pod-info\") pod \"rabbitmq-server-0\" (UID: \"53d5d300-11ae-408d-8bb8-347defb93991\") " pod="openstack/rabbitmq-server-0" Feb 02 09:23:13 crc kubenswrapper[4764]: I0202 09:23:13.928770 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/53d5d300-11ae-408d-8bb8-347defb93991-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"53d5d300-11ae-408d-8bb8-347defb93991\") " pod="openstack/rabbitmq-server-0" Feb 02 09:23:13 crc kubenswrapper[4764]: I0202 09:23:13.928899 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/53d5d300-11ae-408d-8bb8-347defb93991-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"53d5d300-11ae-408d-8bb8-347defb93991\") " pod="openstack/rabbitmq-server-0" Feb 02 09:23:13 crc kubenswrapper[4764]: I0202 09:23:13.929087 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/53d5d300-11ae-408d-8bb8-347defb93991-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"53d5d300-11ae-408d-8bb8-347defb93991\") " pod="openstack/rabbitmq-server-0" Feb 02 09:23:13 crc kubenswrapper[4764]: I0202 09:23:13.944326 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vs58s\" (UniqueName: \"kubernetes.io/projected/53d5d300-11ae-408d-8bb8-347defb93991-kube-api-access-vs58s\") pod \"rabbitmq-server-0\" (UID: \"53d5d300-11ae-408d-8bb8-347defb93991\") " pod="openstack/rabbitmq-server-0" Feb 02 09:23:13 crc kubenswrapper[4764]: I0202 09:23:13.951473 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-server-0\" (UID: \"53d5d300-11ae-408d-8bb8-347defb93991\") " pod="openstack/rabbitmq-server-0" Feb 02 09:23:14 crc kubenswrapper[4764]: I0202 09:23:14.002876 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-67ff45466c-v6vn7"] Feb 02 09:23:14 crc kubenswrapper[4764]: W0202 09:23:14.021193 4764 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod210b0439_a7e6_4f9f_b847_0a04cf044f87.slice/crio-e9b82bb8d627c82badf1dccbdb307ef887aa920c8713078506a96369375f3e21 WatchSource:0}: Error finding container e9b82bb8d627c82badf1dccbdb307ef887aa920c8713078506a96369375f3e21: Status 404 returned error can't find the container with id e9b82bb8d627c82badf1dccbdb307ef887aa920c8713078506a96369375f3e21 Feb 02 09:23:14 crc kubenswrapper[4764]: I0202 09:23:14.030731 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Feb 02 09:23:14 crc kubenswrapper[4764]: I0202 09:23:14.208028 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 02 09:23:14 crc kubenswrapper[4764]: I0202 09:23:14.210120 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Feb 02 09:23:14 crc kubenswrapper[4764]: I0202 09:23:14.212577 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Feb 02 09:23:14 crc kubenswrapper[4764]: I0202 09:23:14.213479 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 02 09:23:14 crc kubenswrapper[4764]: I0202 09:23:14.214212 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Feb 02 09:23:14 crc kubenswrapper[4764]: I0202 09:23:14.215859 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Feb 02 09:23:14 crc kubenswrapper[4764]: I0202 09:23:14.215892 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Feb 02 09:23:14 crc kubenswrapper[4764]: I0202 09:23:14.216099 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Feb 02 09:23:14 crc kubenswrapper[4764]: I0202 09:23:14.216206 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-4mcsh" Feb 02 09:23:14 crc kubenswrapper[4764]: I0202 09:23:14.223295 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Feb 02 09:23:14 crc kubenswrapper[4764]: I0202 09:23:14.325750 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/03d9f090-4a93-4226-8fce-6fd0a425d929-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"03d9f090-4a93-4226-8fce-6fd0a425d929\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 09:23:14 crc kubenswrapper[4764]: I0202 09:23:14.325792 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/03d9f090-4a93-4226-8fce-6fd0a425d929-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"03d9f090-4a93-4226-8fce-6fd0a425d929\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 09:23:14 crc kubenswrapper[4764]: I0202 09:23:14.325826 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/03d9f090-4a93-4226-8fce-6fd0a425d929-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"03d9f090-4a93-4226-8fce-6fd0a425d929\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 09:23:14 crc kubenswrapper[4764]: I0202 09:23:14.325843 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/03d9f090-4a93-4226-8fce-6fd0a425d929-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"03d9f090-4a93-4226-8fce-6fd0a425d929\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 09:23:14 crc kubenswrapper[4764]: I0202 09:23:14.325996 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/03d9f090-4a93-4226-8fce-6fd0a425d929-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"03d9f090-4a93-4226-8fce-6fd0a425d929\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 09:23:14 crc kubenswrapper[4764]: I0202 09:23:14.326110 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"03d9f090-4a93-4226-8fce-6fd0a425d929\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 09:23:14 crc kubenswrapper[4764]: I0202 09:23:14.326169 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/03d9f090-4a93-4226-8fce-6fd0a425d929-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"03d9f090-4a93-4226-8fce-6fd0a425d929\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 09:23:14 crc kubenswrapper[4764]: I0202 09:23:14.326189 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/03d9f090-4a93-4226-8fce-6fd0a425d929-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"03d9f090-4a93-4226-8fce-6fd0a425d929\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 09:23:14 crc kubenswrapper[4764]: I0202 09:23:14.326220 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/03d9f090-4a93-4226-8fce-6fd0a425d929-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"03d9f090-4a93-4226-8fce-6fd0a425d929\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 09:23:14 crc kubenswrapper[4764]: I0202 09:23:14.326456 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/03d9f090-4a93-4226-8fce-6fd0a425d929-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"03d9f090-4a93-4226-8fce-6fd0a425d929\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 09:23:14 crc kubenswrapper[4764]: I0202 09:23:14.326578 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-27wk5\" (UniqueName: \"kubernetes.io/projected/03d9f090-4a93-4226-8fce-6fd0a425d929-kube-api-access-27wk5\") pod \"rabbitmq-cell1-server-0\" (UID: \"03d9f090-4a93-4226-8fce-6fd0a425d929\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 09:23:14 crc kubenswrapper[4764]: I0202 09:23:14.428280 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/03d9f090-4a93-4226-8fce-6fd0a425d929-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"03d9f090-4a93-4226-8fce-6fd0a425d929\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 09:23:14 crc kubenswrapper[4764]: I0202 09:23:14.428321 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/03d9f090-4a93-4226-8fce-6fd0a425d929-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"03d9f090-4a93-4226-8fce-6fd0a425d929\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 09:23:14 crc kubenswrapper[4764]: I0202 09:23:14.428338 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/03d9f090-4a93-4226-8fce-6fd0a425d929-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"03d9f090-4a93-4226-8fce-6fd0a425d929\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 09:23:14 crc kubenswrapper[4764]: I0202 09:23:14.428352 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/03d9f090-4a93-4226-8fce-6fd0a425d929-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"03d9f090-4a93-4226-8fce-6fd0a425d929\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 09:23:14 crc kubenswrapper[4764]: I0202 09:23:14.428395 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/03d9f090-4a93-4226-8fce-6fd0a425d929-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"03d9f090-4a93-4226-8fce-6fd0a425d929\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 09:23:14 crc kubenswrapper[4764]: I0202 09:23:14.428435 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"03d9f090-4a93-4226-8fce-6fd0a425d929\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 09:23:14 crc kubenswrapper[4764]: I0202 09:23:14.428466 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/03d9f090-4a93-4226-8fce-6fd0a425d929-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"03d9f090-4a93-4226-8fce-6fd0a425d929\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 09:23:14 crc kubenswrapper[4764]: I0202 09:23:14.428483 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/03d9f090-4a93-4226-8fce-6fd0a425d929-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"03d9f090-4a93-4226-8fce-6fd0a425d929\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 09:23:14 crc kubenswrapper[4764]: I0202 09:23:14.428502 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/03d9f090-4a93-4226-8fce-6fd0a425d929-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"03d9f090-4a93-4226-8fce-6fd0a425d929\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 09:23:14 crc kubenswrapper[4764]: I0202 09:23:14.428523 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/03d9f090-4a93-4226-8fce-6fd0a425d929-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"03d9f090-4a93-4226-8fce-6fd0a425d929\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 09:23:14 crc kubenswrapper[4764]: I0202 09:23:14.428544 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-27wk5\" (UniqueName: \"kubernetes.io/projected/03d9f090-4a93-4226-8fce-6fd0a425d929-kube-api-access-27wk5\") pod \"rabbitmq-cell1-server-0\" (UID: \"03d9f090-4a93-4226-8fce-6fd0a425d929\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 09:23:14 crc kubenswrapper[4764]: I0202 09:23:14.430198 4764 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"03d9f090-4a93-4226-8fce-6fd0a425d929\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/rabbitmq-cell1-server-0" Feb 02 09:23:14 crc kubenswrapper[4764]: I0202 09:23:14.436113 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/03d9f090-4a93-4226-8fce-6fd0a425d929-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"03d9f090-4a93-4226-8fce-6fd0a425d929\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 09:23:14 crc kubenswrapper[4764]: I0202 09:23:14.436115 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/03d9f090-4a93-4226-8fce-6fd0a425d929-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"03d9f090-4a93-4226-8fce-6fd0a425d929\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 09:23:14 crc kubenswrapper[4764]: I0202 09:23:14.436369 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/03d9f090-4a93-4226-8fce-6fd0a425d929-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"03d9f090-4a93-4226-8fce-6fd0a425d929\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 09:23:14 crc kubenswrapper[4764]: I0202 09:23:14.436655 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/03d9f090-4a93-4226-8fce-6fd0a425d929-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"03d9f090-4a93-4226-8fce-6fd0a425d929\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 09:23:14 crc kubenswrapper[4764]: I0202 09:23:14.444906 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/03d9f090-4a93-4226-8fce-6fd0a425d929-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"03d9f090-4a93-4226-8fce-6fd0a425d929\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 09:23:14 crc kubenswrapper[4764]: I0202 09:23:14.445266 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/03d9f090-4a93-4226-8fce-6fd0a425d929-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"03d9f090-4a93-4226-8fce-6fd0a425d929\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 09:23:14 crc kubenswrapper[4764]: I0202 09:23:14.446666 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/03d9f090-4a93-4226-8fce-6fd0a425d929-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"03d9f090-4a93-4226-8fce-6fd0a425d929\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 09:23:14 crc kubenswrapper[4764]: I0202 09:23:14.446794 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/03d9f090-4a93-4226-8fce-6fd0a425d929-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"03d9f090-4a93-4226-8fce-6fd0a425d929\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 09:23:14 crc kubenswrapper[4764]: I0202 09:23:14.456160 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/03d9f090-4a93-4226-8fce-6fd0a425d929-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"03d9f090-4a93-4226-8fce-6fd0a425d929\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 09:23:14 crc kubenswrapper[4764]: I0202 09:23:14.473244 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-27wk5\" (UniqueName: \"kubernetes.io/projected/03d9f090-4a93-4226-8fce-6fd0a425d929-kube-api-access-27wk5\") pod \"rabbitmq-cell1-server-0\" (UID: \"03d9f090-4a93-4226-8fce-6fd0a425d929\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 09:23:14 crc kubenswrapper[4764]: I0202 09:23:14.484358 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"03d9f090-4a93-4226-8fce-6fd0a425d929\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 09:23:14 crc kubenswrapper[4764]: I0202 09:23:14.521397 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67ff45466c-v6vn7" event={"ID":"210b0439-a7e6-4f9f-b847-0a04cf044f87","Type":"ContainerStarted","Data":"e9b82bb8d627c82badf1dccbdb307ef887aa920c8713078506a96369375f3e21"} Feb 02 09:23:14 crc kubenswrapper[4764]: I0202 09:23:14.535833 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Feb 02 09:23:14 crc kubenswrapper[4764]: I0202 09:23:14.559312 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f54874ffc-wmsqt" event={"ID":"85525aa3-d70a-4d66-8bbf-536be10a0337","Type":"ContainerStarted","Data":"574d06adc1e687e5f47f6871cdb34554c9d7875b54529338e583a5b3f80f5d33"} Feb 02 09:23:14 crc kubenswrapper[4764]: I0202 09:23:14.596684 4764 generic.go:334] "Generic (PLEG): container finished" podID="d192f670-9f9d-4539-9641-e4bed73acdd4" containerID="a4c68bd9cbe36bd80883876d850d402edf3b17e0660281a9f5527447a8167a49" exitCode=0 Feb 02 09:23:14 crc kubenswrapper[4764]: I0202 09:23:14.596783 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" event={"ID":"d192f670-9f9d-4539-9641-e4bed73acdd4","Type":"ContainerDied","Data":"a4c68bd9cbe36bd80883876d850d402edf3b17e0660281a9f5527447a8167a49"} Feb 02 09:23:14 crc kubenswrapper[4764]: I0202 09:23:14.596811 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" event={"ID":"d192f670-9f9d-4539-9641-e4bed73acdd4","Type":"ContainerStarted","Data":"bbb9c6cc86c729fedc84038f3af8b700a433083aa4652e832aca9d3d773be9fd"} Feb 02 09:23:14 crc kubenswrapper[4764]: I0202 09:23:14.596827 4764 scope.go:117] "RemoveContainer" containerID="942c4d2fe3179033f2460fe16951c41f44faf33abaa4e47531eeac47beeeae75" Feb 02 09:23:14 crc kubenswrapper[4764]: I0202 09:23:14.620168 4764 generic.go:334] "Generic (PLEG): container finished" podID="10c729f0-d54f-4b06-995b-ab35e4d583bb" containerID="ccb5dd565532e6e4b7d293886f44bad5362f5b8c79fc18074b57de20eb5cbdf0" exitCode=0 Feb 02 09:23:14 crc kubenswrapper[4764]: I0202 09:23:14.620209 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lx9rg" event={"ID":"10c729f0-d54f-4b06-995b-ab35e4d583bb","Type":"ContainerDied","Data":"ccb5dd565532e6e4b7d293886f44bad5362f5b8c79fc18074b57de20eb5cbdf0"} Feb 02 09:23:14 crc kubenswrapper[4764]: I0202 09:23:14.817065 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 02 09:23:15 crc kubenswrapper[4764]: I0202 09:23:15.187691 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 02 09:23:15 crc kubenswrapper[4764]: I0202 09:23:15.595651 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Feb 02 09:23:15 crc kubenswrapper[4764]: I0202 09:23:15.597659 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Feb 02 09:23:15 crc kubenswrapper[4764]: I0202 09:23:15.609775 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Feb 02 09:23:15 crc kubenswrapper[4764]: I0202 09:23:15.611458 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Feb 02 09:23:15 crc kubenswrapper[4764]: I0202 09:23:15.611611 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Feb 02 09:23:15 crc kubenswrapper[4764]: I0202 09:23:15.612005 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-wk9sx" Feb 02 09:23:15 crc kubenswrapper[4764]: I0202 09:23:15.621100 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Feb 02 09:23:15 crc kubenswrapper[4764]: I0202 09:23:15.630607 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Feb 02 09:23:15 crc kubenswrapper[4764]: I0202 09:23:15.651876 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"03d9f090-4a93-4226-8fce-6fd0a425d929","Type":"ContainerStarted","Data":"d923a6203fb8b53db25d87f66f30d098c92cff9a522b96722479fb01f94e1e95"} Feb 02 09:23:15 crc kubenswrapper[4764]: I0202 09:23:15.656774 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"53d5d300-11ae-408d-8bb8-347defb93991","Type":"ContainerStarted","Data":"5111e0090c4901a2d85fc188815016dd3aa41b8c15a1b826e3b6edbb5e86e2cc"} Feb 02 09:23:15 crc kubenswrapper[4764]: I0202 09:23:15.695273 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/1d6b6c31-819e-4098-b11b-6073cc751cad-config-data-default\") pod \"openstack-galera-0\" (UID: \"1d6b6c31-819e-4098-b11b-6073cc751cad\") " pod="openstack/openstack-galera-0" Feb 02 09:23:15 crc kubenswrapper[4764]: I0202 09:23:15.706160 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1d6b6c31-819e-4098-b11b-6073cc751cad-operator-scripts\") pod \"openstack-galera-0\" (UID: \"1d6b6c31-819e-4098-b11b-6073cc751cad\") " pod="openstack/openstack-galera-0" Feb 02 09:23:15 crc kubenswrapper[4764]: I0202 09:23:15.706240 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dm8vj\" (UniqueName: \"kubernetes.io/projected/1d6b6c31-819e-4098-b11b-6073cc751cad-kube-api-access-dm8vj\") pod \"openstack-galera-0\" (UID: \"1d6b6c31-819e-4098-b11b-6073cc751cad\") " pod="openstack/openstack-galera-0" Feb 02 09:23:15 crc kubenswrapper[4764]: I0202 09:23:15.706428 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/1d6b6c31-819e-4098-b11b-6073cc751cad-config-data-generated\") pod \"openstack-galera-0\" (UID: \"1d6b6c31-819e-4098-b11b-6073cc751cad\") " pod="openstack/openstack-galera-0" Feb 02 09:23:15 crc kubenswrapper[4764]: I0202 09:23:15.706479 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d6b6c31-819e-4098-b11b-6073cc751cad-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"1d6b6c31-819e-4098-b11b-6073cc751cad\") " pod="openstack/openstack-galera-0" Feb 02 09:23:15 crc kubenswrapper[4764]: I0202 09:23:15.706510 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-galera-0\" (UID: \"1d6b6c31-819e-4098-b11b-6073cc751cad\") " pod="openstack/openstack-galera-0" Feb 02 09:23:15 crc kubenswrapper[4764]: I0202 09:23:15.706715 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/1d6b6c31-819e-4098-b11b-6073cc751cad-kolla-config\") pod \"openstack-galera-0\" (UID: \"1d6b6c31-819e-4098-b11b-6073cc751cad\") " pod="openstack/openstack-galera-0" Feb 02 09:23:15 crc kubenswrapper[4764]: I0202 09:23:15.706741 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/1d6b6c31-819e-4098-b11b-6073cc751cad-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"1d6b6c31-819e-4098-b11b-6073cc751cad\") " pod="openstack/openstack-galera-0" Feb 02 09:23:15 crc kubenswrapper[4764]: I0202 09:23:15.816769 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/1d6b6c31-819e-4098-b11b-6073cc751cad-config-data-default\") pod \"openstack-galera-0\" (UID: \"1d6b6c31-819e-4098-b11b-6073cc751cad\") " pod="openstack/openstack-galera-0" Feb 02 09:23:15 crc kubenswrapper[4764]: I0202 09:23:15.816825 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1d6b6c31-819e-4098-b11b-6073cc751cad-operator-scripts\") pod \"openstack-galera-0\" (UID: \"1d6b6c31-819e-4098-b11b-6073cc751cad\") " pod="openstack/openstack-galera-0" Feb 02 09:23:15 crc kubenswrapper[4764]: I0202 09:23:15.816846 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dm8vj\" (UniqueName: \"kubernetes.io/projected/1d6b6c31-819e-4098-b11b-6073cc751cad-kube-api-access-dm8vj\") pod \"openstack-galera-0\" (UID: \"1d6b6c31-819e-4098-b11b-6073cc751cad\") " pod="openstack/openstack-galera-0" Feb 02 09:23:15 crc kubenswrapper[4764]: I0202 09:23:15.816897 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/1d6b6c31-819e-4098-b11b-6073cc751cad-config-data-generated\") pod \"openstack-galera-0\" (UID: \"1d6b6c31-819e-4098-b11b-6073cc751cad\") " pod="openstack/openstack-galera-0" Feb 02 09:23:15 crc kubenswrapper[4764]: I0202 09:23:15.816913 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d6b6c31-819e-4098-b11b-6073cc751cad-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"1d6b6c31-819e-4098-b11b-6073cc751cad\") " pod="openstack/openstack-galera-0" Feb 02 09:23:15 crc kubenswrapper[4764]: I0202 09:23:15.816985 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-galera-0\" (UID: \"1d6b6c31-819e-4098-b11b-6073cc751cad\") " pod="openstack/openstack-galera-0" Feb 02 09:23:15 crc kubenswrapper[4764]: I0202 09:23:15.817036 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/1d6b6c31-819e-4098-b11b-6073cc751cad-kolla-config\") pod \"openstack-galera-0\" (UID: \"1d6b6c31-819e-4098-b11b-6073cc751cad\") " pod="openstack/openstack-galera-0" Feb 02 09:23:15 crc kubenswrapper[4764]: I0202 09:23:15.817066 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/1d6b6c31-819e-4098-b11b-6073cc751cad-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"1d6b6c31-819e-4098-b11b-6073cc751cad\") " pod="openstack/openstack-galera-0" Feb 02 09:23:15 crc kubenswrapper[4764]: I0202 09:23:15.818556 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/1d6b6c31-819e-4098-b11b-6073cc751cad-config-data-generated\") pod \"openstack-galera-0\" (UID: \"1d6b6c31-819e-4098-b11b-6073cc751cad\") " pod="openstack/openstack-galera-0" Feb 02 09:23:15 crc kubenswrapper[4764]: I0202 09:23:15.820882 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/1d6b6c31-819e-4098-b11b-6073cc751cad-config-data-default\") pod \"openstack-galera-0\" (UID: \"1d6b6c31-819e-4098-b11b-6073cc751cad\") " pod="openstack/openstack-galera-0" Feb 02 09:23:15 crc kubenswrapper[4764]: I0202 09:23:15.821545 4764 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-galera-0\" (UID: \"1d6b6c31-819e-4098-b11b-6073cc751cad\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/openstack-galera-0" Feb 02 09:23:15 crc kubenswrapper[4764]: I0202 09:23:15.825910 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/1d6b6c31-819e-4098-b11b-6073cc751cad-kolla-config\") pod \"openstack-galera-0\" (UID: \"1d6b6c31-819e-4098-b11b-6073cc751cad\") " pod="openstack/openstack-galera-0" Feb 02 09:23:15 crc kubenswrapper[4764]: I0202 09:23:15.827111 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1d6b6c31-819e-4098-b11b-6073cc751cad-operator-scripts\") pod \"openstack-galera-0\" (UID: \"1d6b6c31-819e-4098-b11b-6073cc751cad\") " pod="openstack/openstack-galera-0" Feb 02 09:23:15 crc kubenswrapper[4764]: I0202 09:23:15.839546 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/1d6b6c31-819e-4098-b11b-6073cc751cad-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"1d6b6c31-819e-4098-b11b-6073cc751cad\") " pod="openstack/openstack-galera-0" Feb 02 09:23:15 crc kubenswrapper[4764]: I0202 09:23:15.848972 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dm8vj\" (UniqueName: \"kubernetes.io/projected/1d6b6c31-819e-4098-b11b-6073cc751cad-kube-api-access-dm8vj\") pod \"openstack-galera-0\" (UID: \"1d6b6c31-819e-4098-b11b-6073cc751cad\") " pod="openstack/openstack-galera-0" Feb 02 09:23:15 crc kubenswrapper[4764]: I0202 09:23:15.873721 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d6b6c31-819e-4098-b11b-6073cc751cad-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"1d6b6c31-819e-4098-b11b-6073cc751cad\") " pod="openstack/openstack-galera-0" Feb 02 09:23:15 crc kubenswrapper[4764]: I0202 09:23:15.952166 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-galera-0\" (UID: \"1d6b6c31-819e-4098-b11b-6073cc751cad\") " pod="openstack/openstack-galera-0" Feb 02 09:23:16 crc kubenswrapper[4764]: I0202 09:23:16.250534 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Feb 02 09:23:16 crc kubenswrapper[4764]: I0202 09:23:16.699589 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lx9rg" event={"ID":"10c729f0-d54f-4b06-995b-ab35e4d583bb","Type":"ContainerStarted","Data":"98a0ac0c06018d32adfbe9f8e8984832df81fc12c656ce5683faf6d86cd1d986"} Feb 02 09:23:16 crc kubenswrapper[4764]: I0202 09:23:16.730406 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-lx9rg" podStartSLOduration=3.553859813 podStartE2EDuration="6.730387299s" podCreationTimestamp="2026-02-02 09:23:10 +0000 UTC" firstStartedPulling="2026-02-02 09:23:12.452336384 +0000 UTC m=+955.386060462" lastFinishedPulling="2026-02-02 09:23:15.62886386 +0000 UTC m=+958.562587948" observedRunningTime="2026-02-02 09:23:16.72349714 +0000 UTC m=+959.657221228" watchObservedRunningTime="2026-02-02 09:23:16.730387299 +0000 UTC m=+959.664111397" Feb 02 09:23:16 crc kubenswrapper[4764]: I0202 09:23:16.974008 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Feb 02 09:23:16 crc kubenswrapper[4764]: I0202 09:23:16.983503 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Feb 02 09:23:16 crc kubenswrapper[4764]: I0202 09:23:16.984638 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Feb 02 09:23:16 crc kubenswrapper[4764]: I0202 09:23:16.992569 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-rq87l" Feb 02 09:23:16 crc kubenswrapper[4764]: I0202 09:23:16.992709 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Feb 02 09:23:16 crc kubenswrapper[4764]: I0202 09:23:16.992821 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Feb 02 09:23:17 crc kubenswrapper[4764]: I0202 09:23:17.005990 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Feb 02 09:23:17 crc kubenswrapper[4764]: I0202 09:23:17.008572 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Feb 02 09:23:17 crc kubenswrapper[4764]: I0202 09:23:17.156765 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/a9177511-b834-4e68-9f6f-8b1b4470b274-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"a9177511-b834-4e68-9f6f-8b1b4470b274\") " pod="openstack/openstack-cell1-galera-0" Feb 02 09:23:17 crc kubenswrapper[4764]: I0202 09:23:17.156815 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sltmn\" (UniqueName: \"kubernetes.io/projected/a9177511-b834-4e68-9f6f-8b1b4470b274-kube-api-access-sltmn\") pod \"openstack-cell1-galera-0\" (UID: \"a9177511-b834-4e68-9f6f-8b1b4470b274\") " pod="openstack/openstack-cell1-galera-0" Feb 02 09:23:17 crc kubenswrapper[4764]: I0202 09:23:17.156851 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/a9177511-b834-4e68-9f6f-8b1b4470b274-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"a9177511-b834-4e68-9f6f-8b1b4470b274\") " pod="openstack/openstack-cell1-galera-0" Feb 02 09:23:17 crc kubenswrapper[4764]: I0202 09:23:17.156892 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-cell1-galera-0\" (UID: \"a9177511-b834-4e68-9f6f-8b1b4470b274\") " pod="openstack/openstack-cell1-galera-0" Feb 02 09:23:17 crc kubenswrapper[4764]: I0202 09:23:17.156918 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/a9177511-b834-4e68-9f6f-8b1b4470b274-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"a9177511-b834-4e68-9f6f-8b1b4470b274\") " pod="openstack/openstack-cell1-galera-0" Feb 02 09:23:17 crc kubenswrapper[4764]: I0202 09:23:17.157060 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/a9177511-b834-4e68-9f6f-8b1b4470b274-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"a9177511-b834-4e68-9f6f-8b1b4470b274\") " pod="openstack/openstack-cell1-galera-0" Feb 02 09:23:17 crc kubenswrapper[4764]: I0202 09:23:17.157101 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9177511-b834-4e68-9f6f-8b1b4470b274-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"a9177511-b834-4e68-9f6f-8b1b4470b274\") " pod="openstack/openstack-cell1-galera-0" Feb 02 09:23:17 crc kubenswrapper[4764]: I0202 09:23:17.157130 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a9177511-b834-4e68-9f6f-8b1b4470b274-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"a9177511-b834-4e68-9f6f-8b1b4470b274\") " pod="openstack/openstack-cell1-galera-0" Feb 02 09:23:17 crc kubenswrapper[4764]: I0202 09:23:17.163715 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Feb 02 09:23:17 crc kubenswrapper[4764]: I0202 09:23:17.165085 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Feb 02 09:23:17 crc kubenswrapper[4764]: I0202 09:23:17.172367 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-8gljz" Feb 02 09:23:17 crc kubenswrapper[4764]: I0202 09:23:17.172573 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Feb 02 09:23:17 crc kubenswrapper[4764]: I0202 09:23:17.172701 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Feb 02 09:23:17 crc kubenswrapper[4764]: I0202 09:23:17.184877 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Feb 02 09:23:17 crc kubenswrapper[4764]: I0202 09:23:17.258054 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/a9177511-b834-4e68-9f6f-8b1b4470b274-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"a9177511-b834-4e68-9f6f-8b1b4470b274\") " pod="openstack/openstack-cell1-galera-0" Feb 02 09:23:17 crc kubenswrapper[4764]: I0202 09:23:17.258117 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd76e53a-13fc-4293-b5ca-2324224850cc-combined-ca-bundle\") pod \"memcached-0\" (UID: \"cd76e53a-13fc-4293-b5ca-2324224850cc\") " pod="openstack/memcached-0" Feb 02 09:23:17 crc kubenswrapper[4764]: I0202 09:23:17.258142 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9177511-b834-4e68-9f6f-8b1b4470b274-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"a9177511-b834-4e68-9f6f-8b1b4470b274\") " pod="openstack/openstack-cell1-galera-0" Feb 02 09:23:17 crc kubenswrapper[4764]: I0202 09:23:17.258164 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a9177511-b834-4e68-9f6f-8b1b4470b274-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"a9177511-b834-4e68-9f6f-8b1b4470b274\") " pod="openstack/openstack-cell1-galera-0" Feb 02 09:23:17 crc kubenswrapper[4764]: I0202 09:23:17.258189 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cd76e53a-13fc-4293-b5ca-2324224850cc-config-data\") pod \"memcached-0\" (UID: \"cd76e53a-13fc-4293-b5ca-2324224850cc\") " pod="openstack/memcached-0" Feb 02 09:23:17 crc kubenswrapper[4764]: I0202 09:23:17.258233 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/a9177511-b834-4e68-9f6f-8b1b4470b274-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"a9177511-b834-4e68-9f6f-8b1b4470b274\") " pod="openstack/openstack-cell1-galera-0" Feb 02 09:23:17 crc kubenswrapper[4764]: I0202 09:23:17.258251 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sltmn\" (UniqueName: \"kubernetes.io/projected/a9177511-b834-4e68-9f6f-8b1b4470b274-kube-api-access-sltmn\") pod \"openstack-cell1-galera-0\" (UID: \"a9177511-b834-4e68-9f6f-8b1b4470b274\") " pod="openstack/openstack-cell1-galera-0" Feb 02 09:23:17 crc kubenswrapper[4764]: I0202 09:23:17.258277 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/a9177511-b834-4e68-9f6f-8b1b4470b274-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"a9177511-b834-4e68-9f6f-8b1b4470b274\") " pod="openstack/openstack-cell1-galera-0" Feb 02 09:23:17 crc kubenswrapper[4764]: I0202 09:23:17.258292 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/cd76e53a-13fc-4293-b5ca-2324224850cc-memcached-tls-certs\") pod \"memcached-0\" (UID: \"cd76e53a-13fc-4293-b5ca-2324224850cc\") " pod="openstack/memcached-0" Feb 02 09:23:17 crc kubenswrapper[4764]: I0202 09:23:17.258315 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6v68h\" (UniqueName: \"kubernetes.io/projected/cd76e53a-13fc-4293-b5ca-2324224850cc-kube-api-access-6v68h\") pod \"memcached-0\" (UID: \"cd76e53a-13fc-4293-b5ca-2324224850cc\") " pod="openstack/memcached-0" Feb 02 09:23:17 crc kubenswrapper[4764]: I0202 09:23:17.258331 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/cd76e53a-13fc-4293-b5ca-2324224850cc-kolla-config\") pod \"memcached-0\" (UID: \"cd76e53a-13fc-4293-b5ca-2324224850cc\") " pod="openstack/memcached-0" Feb 02 09:23:17 crc kubenswrapper[4764]: I0202 09:23:17.258357 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-cell1-galera-0\" (UID: \"a9177511-b834-4e68-9f6f-8b1b4470b274\") " pod="openstack/openstack-cell1-galera-0" Feb 02 09:23:17 crc kubenswrapper[4764]: I0202 09:23:17.258375 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/a9177511-b834-4e68-9f6f-8b1b4470b274-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"a9177511-b834-4e68-9f6f-8b1b4470b274\") " pod="openstack/openstack-cell1-galera-0" Feb 02 09:23:17 crc kubenswrapper[4764]: I0202 09:23:17.259077 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/a9177511-b834-4e68-9f6f-8b1b4470b274-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"a9177511-b834-4e68-9f6f-8b1b4470b274\") " pod="openstack/openstack-cell1-galera-0" Feb 02 09:23:17 crc kubenswrapper[4764]: I0202 09:23:17.259664 4764 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-cell1-galera-0\" (UID: \"a9177511-b834-4e68-9f6f-8b1b4470b274\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/openstack-cell1-galera-0" Feb 02 09:23:17 crc kubenswrapper[4764]: I0202 09:23:17.263407 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/a9177511-b834-4e68-9f6f-8b1b4470b274-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"a9177511-b834-4e68-9f6f-8b1b4470b274\") " pod="openstack/openstack-cell1-galera-0" Feb 02 09:23:17 crc kubenswrapper[4764]: I0202 09:23:17.265237 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a9177511-b834-4e68-9f6f-8b1b4470b274-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"a9177511-b834-4e68-9f6f-8b1b4470b274\") " pod="openstack/openstack-cell1-galera-0" Feb 02 09:23:17 crc kubenswrapper[4764]: I0202 09:23:17.269548 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/a9177511-b834-4e68-9f6f-8b1b4470b274-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"a9177511-b834-4e68-9f6f-8b1b4470b274\") " pod="openstack/openstack-cell1-galera-0" Feb 02 09:23:17 crc kubenswrapper[4764]: I0202 09:23:17.287314 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/a9177511-b834-4e68-9f6f-8b1b4470b274-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"a9177511-b834-4e68-9f6f-8b1b4470b274\") " pod="openstack/openstack-cell1-galera-0" Feb 02 09:23:17 crc kubenswrapper[4764]: I0202 09:23:17.296308 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sltmn\" (UniqueName: \"kubernetes.io/projected/a9177511-b834-4e68-9f6f-8b1b4470b274-kube-api-access-sltmn\") pod \"openstack-cell1-galera-0\" (UID: \"a9177511-b834-4e68-9f6f-8b1b4470b274\") " pod="openstack/openstack-cell1-galera-0" Feb 02 09:23:17 crc kubenswrapper[4764]: I0202 09:23:17.299386 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9177511-b834-4e68-9f6f-8b1b4470b274-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"a9177511-b834-4e68-9f6f-8b1b4470b274\") " pod="openstack/openstack-cell1-galera-0" Feb 02 09:23:17 crc kubenswrapper[4764]: I0202 09:23:17.311946 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-cell1-galera-0\" (UID: \"a9177511-b834-4e68-9f6f-8b1b4470b274\") " pod="openstack/openstack-cell1-galera-0" Feb 02 09:23:17 crc kubenswrapper[4764]: I0202 09:23:17.368828 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cd76e53a-13fc-4293-b5ca-2324224850cc-config-data\") pod \"memcached-0\" (UID: \"cd76e53a-13fc-4293-b5ca-2324224850cc\") " pod="openstack/memcached-0" Feb 02 09:23:17 crc kubenswrapper[4764]: I0202 09:23:17.368928 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/cd76e53a-13fc-4293-b5ca-2324224850cc-memcached-tls-certs\") pod \"memcached-0\" (UID: \"cd76e53a-13fc-4293-b5ca-2324224850cc\") " pod="openstack/memcached-0" Feb 02 09:23:17 crc kubenswrapper[4764]: I0202 09:23:17.368971 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6v68h\" (UniqueName: \"kubernetes.io/projected/cd76e53a-13fc-4293-b5ca-2324224850cc-kube-api-access-6v68h\") pod \"memcached-0\" (UID: \"cd76e53a-13fc-4293-b5ca-2324224850cc\") " pod="openstack/memcached-0" Feb 02 09:23:17 crc kubenswrapper[4764]: I0202 09:23:17.368997 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/cd76e53a-13fc-4293-b5ca-2324224850cc-kolla-config\") pod \"memcached-0\" (UID: \"cd76e53a-13fc-4293-b5ca-2324224850cc\") " pod="openstack/memcached-0" Feb 02 09:23:17 crc kubenswrapper[4764]: I0202 09:23:17.369066 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd76e53a-13fc-4293-b5ca-2324224850cc-combined-ca-bundle\") pod \"memcached-0\" (UID: \"cd76e53a-13fc-4293-b5ca-2324224850cc\") " pod="openstack/memcached-0" Feb 02 09:23:17 crc kubenswrapper[4764]: I0202 09:23:17.372027 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cd76e53a-13fc-4293-b5ca-2324224850cc-config-data\") pod \"memcached-0\" (UID: \"cd76e53a-13fc-4293-b5ca-2324224850cc\") " pod="openstack/memcached-0" Feb 02 09:23:17 crc kubenswrapper[4764]: I0202 09:23:17.372953 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/cd76e53a-13fc-4293-b5ca-2324224850cc-kolla-config\") pod \"memcached-0\" (UID: \"cd76e53a-13fc-4293-b5ca-2324224850cc\") " pod="openstack/memcached-0" Feb 02 09:23:17 crc kubenswrapper[4764]: I0202 09:23:17.374659 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd76e53a-13fc-4293-b5ca-2324224850cc-combined-ca-bundle\") pod \"memcached-0\" (UID: \"cd76e53a-13fc-4293-b5ca-2324224850cc\") " pod="openstack/memcached-0" Feb 02 09:23:17 crc kubenswrapper[4764]: I0202 09:23:17.379379 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/cd76e53a-13fc-4293-b5ca-2324224850cc-memcached-tls-certs\") pod \"memcached-0\" (UID: \"cd76e53a-13fc-4293-b5ca-2324224850cc\") " pod="openstack/memcached-0" Feb 02 09:23:17 crc kubenswrapper[4764]: I0202 09:23:17.418720 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6v68h\" (UniqueName: \"kubernetes.io/projected/cd76e53a-13fc-4293-b5ca-2324224850cc-kube-api-access-6v68h\") pod \"memcached-0\" (UID: \"cd76e53a-13fc-4293-b5ca-2324224850cc\") " pod="openstack/memcached-0" Feb 02 09:23:17 crc kubenswrapper[4764]: I0202 09:23:17.553482 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Feb 02 09:23:17 crc kubenswrapper[4764]: I0202 09:23:17.615723 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Feb 02 09:23:17 crc kubenswrapper[4764]: I0202 09:23:17.787663 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"1d6b6c31-819e-4098-b11b-6073cc751cad","Type":"ContainerStarted","Data":"a62588e7e4c6b1ef51ca5c2f809fcf86e0b55ff455c3dcf16b29a3c2230208de"} Feb 02 09:23:18 crc kubenswrapper[4764]: I0202 09:23:18.068968 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Feb 02 09:23:18 crc kubenswrapper[4764]: I0202 09:23:18.076872 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Feb 02 09:23:18 crc kubenswrapper[4764]: W0202 09:23:18.160146 4764 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcd76e53a_13fc_4293_b5ca_2324224850cc.slice/crio-17480052cd2acb11c0838ec17febbf7af09c03d377b2b907039fb65585eb0c96 WatchSource:0}: Error finding container 17480052cd2acb11c0838ec17febbf7af09c03d377b2b907039fb65585eb0c96: Status 404 returned error can't find the container with id 17480052cd2acb11c0838ec17febbf7af09c03d377b2b907039fb65585eb0c96 Feb 02 09:23:18 crc kubenswrapper[4764]: W0202 09:23:18.187865 4764 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda9177511_b834_4e68_9f6f_8b1b4470b274.slice/crio-c826508594f3cd6b658c3b1bea9b8a00efd837336e26321931f0c29823592843 WatchSource:0}: Error finding container c826508594f3cd6b658c3b1bea9b8a00efd837336e26321931f0c29823592843: Status 404 returned error can't find the container with id c826508594f3cd6b658c3b1bea9b8a00efd837336e26321931f0c29823592843 Feb 02 09:23:18 crc kubenswrapper[4764]: I0202 09:23:18.844152 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Feb 02 09:23:18 crc kubenswrapper[4764]: I0202 09:23:18.845451 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Feb 02 09:23:18 crc kubenswrapper[4764]: I0202 09:23:18.855617 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-qkpvx" Feb 02 09:23:18 crc kubenswrapper[4764]: I0202 09:23:18.866136 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"a9177511-b834-4e68-9f6f-8b1b4470b274","Type":"ContainerStarted","Data":"c826508594f3cd6b658c3b1bea9b8a00efd837336e26321931f0c29823592843"} Feb 02 09:23:18 crc kubenswrapper[4764]: I0202 09:23:18.873593 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"cd76e53a-13fc-4293-b5ca-2324224850cc","Type":"ContainerStarted","Data":"17480052cd2acb11c0838ec17febbf7af09c03d377b2b907039fb65585eb0c96"} Feb 02 09:23:18 crc kubenswrapper[4764]: I0202 09:23:18.879971 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 02 09:23:18 crc kubenswrapper[4764]: I0202 09:23:18.920783 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-98v7s\" (UniqueName: \"kubernetes.io/projected/26d7a0b3-ea1e-417b-b417-69ae84bd5d24-kube-api-access-98v7s\") pod \"kube-state-metrics-0\" (UID: \"26d7a0b3-ea1e-417b-b417-69ae84bd5d24\") " pod="openstack/kube-state-metrics-0" Feb 02 09:23:19 crc kubenswrapper[4764]: I0202 09:23:19.026624 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-98v7s\" (UniqueName: \"kubernetes.io/projected/26d7a0b3-ea1e-417b-b417-69ae84bd5d24-kube-api-access-98v7s\") pod \"kube-state-metrics-0\" (UID: \"26d7a0b3-ea1e-417b-b417-69ae84bd5d24\") " pod="openstack/kube-state-metrics-0" Feb 02 09:23:19 crc kubenswrapper[4764]: I0202 09:23:19.072062 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-98v7s\" (UniqueName: \"kubernetes.io/projected/26d7a0b3-ea1e-417b-b417-69ae84bd5d24-kube-api-access-98v7s\") pod \"kube-state-metrics-0\" (UID: \"26d7a0b3-ea1e-417b-b417-69ae84bd5d24\") " pod="openstack/kube-state-metrics-0" Feb 02 09:23:19 crc kubenswrapper[4764]: I0202 09:23:19.169440 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Feb 02 09:23:19 crc kubenswrapper[4764]: I0202 09:23:19.826433 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 02 09:23:19 crc kubenswrapper[4764]: I0202 09:23:19.886095 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"26d7a0b3-ea1e-417b-b417-69ae84bd5d24","Type":"ContainerStarted","Data":"d8af482c5008789457daff07c4f5a0cf5cfc3e09895ed5dd3c018b07a566b275"} Feb 02 09:23:21 crc kubenswrapper[4764]: I0202 09:23:21.348039 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-lx9rg" Feb 02 09:23:21 crc kubenswrapper[4764]: I0202 09:23:21.348545 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-lx9rg" Feb 02 09:23:22 crc kubenswrapper[4764]: I0202 09:23:22.572671 4764 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-lx9rg" podUID="10c729f0-d54f-4b06-995b-ab35e4d583bb" containerName="registry-server" probeResult="failure" output=< Feb 02 09:23:22 crc kubenswrapper[4764]: timeout: failed to connect service ":50051" within 1s Feb 02 09:23:22 crc kubenswrapper[4764]: > Feb 02 09:23:22 crc kubenswrapper[4764]: I0202 09:23:22.921579 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Feb 02 09:23:22 crc kubenswrapper[4764]: I0202 09:23:22.923038 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Feb 02 09:23:22 crc kubenswrapper[4764]: I0202 09:23:22.928675 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Feb 02 09:23:22 crc kubenswrapper[4764]: I0202 09:23:22.928892 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Feb 02 09:23:22 crc kubenswrapper[4764]: I0202 09:23:22.929096 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Feb 02 09:23:22 crc kubenswrapper[4764]: I0202 09:23:22.941521 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Feb 02 09:23:22 crc kubenswrapper[4764]: I0202 09:23:22.948138 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-rbvsz" Feb 02 09:23:22 crc kubenswrapper[4764]: I0202 09:23:22.961439 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Feb 02 09:23:22 crc kubenswrapper[4764]: I0202 09:23:22.968451 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-6v7f6"] Feb 02 09:23:22 crc kubenswrapper[4764]: I0202 09:23:22.970585 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-6v7f6" Feb 02 09:23:22 crc kubenswrapper[4764]: I0202 09:23:22.975448 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Feb 02 09:23:22 crc kubenswrapper[4764]: I0202 09:23:22.977061 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Feb 02 09:23:22 crc kubenswrapper[4764]: I0202 09:23:22.978253 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-8rb5w" Feb 02 09:23:23 crc kubenswrapper[4764]: I0202 09:23:23.006981 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cba91a53-7d3d-481b-bd55-3f0bbe7ccbb6-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"cba91a53-7d3d-481b-bd55-3f0bbe7ccbb6\") " pod="openstack/ovsdbserver-nb-0" Feb 02 09:23:23 crc kubenswrapper[4764]: I0202 09:23:23.007039 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cba91a53-7d3d-481b-bd55-3f0bbe7ccbb6-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"cba91a53-7d3d-481b-bd55-3f0bbe7ccbb6\") " pod="openstack/ovsdbserver-nb-0" Feb 02 09:23:23 crc kubenswrapper[4764]: I0202 09:23:23.007069 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ovsdbserver-nb-0\" (UID: \"cba91a53-7d3d-481b-bd55-3f0bbe7ccbb6\") " pod="openstack/ovsdbserver-nb-0" Feb 02 09:23:23 crc kubenswrapper[4764]: I0202 09:23:23.007095 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/cba91a53-7d3d-481b-bd55-3f0bbe7ccbb6-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"cba91a53-7d3d-481b-bd55-3f0bbe7ccbb6\") " pod="openstack/ovsdbserver-nb-0" Feb 02 09:23:23 crc kubenswrapper[4764]: I0202 09:23:23.007127 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/cba91a53-7d3d-481b-bd55-3f0bbe7ccbb6-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"cba91a53-7d3d-481b-bd55-3f0bbe7ccbb6\") " pod="openstack/ovsdbserver-nb-0" Feb 02 09:23:23 crc kubenswrapper[4764]: I0202 09:23:23.007160 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cba91a53-7d3d-481b-bd55-3f0bbe7ccbb6-config\") pod \"ovsdbserver-nb-0\" (UID: \"cba91a53-7d3d-481b-bd55-3f0bbe7ccbb6\") " pod="openstack/ovsdbserver-nb-0" Feb 02 09:23:23 crc kubenswrapper[4764]: I0202 09:23:23.007196 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/cba91a53-7d3d-481b-bd55-3f0bbe7ccbb6-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"cba91a53-7d3d-481b-bd55-3f0bbe7ccbb6\") " pod="openstack/ovsdbserver-nb-0" Feb 02 09:23:23 crc kubenswrapper[4764]: I0202 09:23:23.007217 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8kstr\" (UniqueName: \"kubernetes.io/projected/cba91a53-7d3d-481b-bd55-3f0bbe7ccbb6-kube-api-access-8kstr\") pod \"ovsdbserver-nb-0\" (UID: \"cba91a53-7d3d-481b-bd55-3f0bbe7ccbb6\") " pod="openstack/ovsdbserver-nb-0" Feb 02 09:23:23 crc kubenswrapper[4764]: I0202 09:23:23.009260 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-6v7f6"] Feb 02 09:23:23 crc kubenswrapper[4764]: I0202 09:23:23.092055 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-2p2bk"] Feb 02 09:23:23 crc kubenswrapper[4764]: I0202 09:23:23.110811 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-2p2bk" Feb 02 09:23:23 crc kubenswrapper[4764]: I0202 09:23:23.140092 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8kstr\" (UniqueName: \"kubernetes.io/projected/cba91a53-7d3d-481b-bd55-3f0bbe7ccbb6-kube-api-access-8kstr\") pod \"ovsdbserver-nb-0\" (UID: \"cba91a53-7d3d-481b-bd55-3f0bbe7ccbb6\") " pod="openstack/ovsdbserver-nb-0" Feb 02 09:23:23 crc kubenswrapper[4764]: I0202 09:23:23.140173 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tb4mn\" (UniqueName: \"kubernetes.io/projected/66662206-d732-404d-a81d-6303d9947d60-kube-api-access-tb4mn\") pod \"ovn-controller-6v7f6\" (UID: \"66662206-d732-404d-a81d-6303d9947d60\") " pod="openstack/ovn-controller-6v7f6" Feb 02 09:23:23 crc kubenswrapper[4764]: I0202 09:23:23.140244 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cba91a53-7d3d-481b-bd55-3f0bbe7ccbb6-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"cba91a53-7d3d-481b-bd55-3f0bbe7ccbb6\") " pod="openstack/ovsdbserver-nb-0" Feb 02 09:23:23 crc kubenswrapper[4764]: I0202 09:23:23.140269 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/66662206-d732-404d-a81d-6303d9947d60-scripts\") pod \"ovn-controller-6v7f6\" (UID: \"66662206-d732-404d-a81d-6303d9947d60\") " pod="openstack/ovn-controller-6v7f6" Feb 02 09:23:23 crc kubenswrapper[4764]: I0202 09:23:23.140295 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66662206-d732-404d-a81d-6303d9947d60-combined-ca-bundle\") pod \"ovn-controller-6v7f6\" (UID: \"66662206-d732-404d-a81d-6303d9947d60\") " pod="openstack/ovn-controller-6v7f6" Feb 02 09:23:23 crc kubenswrapper[4764]: I0202 09:23:23.140347 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/66662206-d732-404d-a81d-6303d9947d60-var-run\") pod \"ovn-controller-6v7f6\" (UID: \"66662206-d732-404d-a81d-6303d9947d60\") " pod="openstack/ovn-controller-6v7f6" Feb 02 09:23:23 crc kubenswrapper[4764]: I0202 09:23:23.140371 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/66662206-d732-404d-a81d-6303d9947d60-ovn-controller-tls-certs\") pod \"ovn-controller-6v7f6\" (UID: \"66662206-d732-404d-a81d-6303d9947d60\") " pod="openstack/ovn-controller-6v7f6" Feb 02 09:23:23 crc kubenswrapper[4764]: I0202 09:23:23.140400 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/66662206-d732-404d-a81d-6303d9947d60-var-log-ovn\") pod \"ovn-controller-6v7f6\" (UID: \"66662206-d732-404d-a81d-6303d9947d60\") " pod="openstack/ovn-controller-6v7f6" Feb 02 09:23:23 crc kubenswrapper[4764]: I0202 09:23:23.140434 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cba91a53-7d3d-481b-bd55-3f0bbe7ccbb6-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"cba91a53-7d3d-481b-bd55-3f0bbe7ccbb6\") " pod="openstack/ovsdbserver-nb-0" Feb 02 09:23:23 crc kubenswrapper[4764]: I0202 09:23:23.140487 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ovsdbserver-nb-0\" (UID: \"cba91a53-7d3d-481b-bd55-3f0bbe7ccbb6\") " pod="openstack/ovsdbserver-nb-0" Feb 02 09:23:23 crc kubenswrapper[4764]: I0202 09:23:23.140514 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/cba91a53-7d3d-481b-bd55-3f0bbe7ccbb6-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"cba91a53-7d3d-481b-bd55-3f0bbe7ccbb6\") " pod="openstack/ovsdbserver-nb-0" Feb 02 09:23:23 crc kubenswrapper[4764]: I0202 09:23:23.140555 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/cba91a53-7d3d-481b-bd55-3f0bbe7ccbb6-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"cba91a53-7d3d-481b-bd55-3f0bbe7ccbb6\") " pod="openstack/ovsdbserver-nb-0" Feb 02 09:23:23 crc kubenswrapper[4764]: I0202 09:23:23.140630 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cba91a53-7d3d-481b-bd55-3f0bbe7ccbb6-config\") pod \"ovsdbserver-nb-0\" (UID: \"cba91a53-7d3d-481b-bd55-3f0bbe7ccbb6\") " pod="openstack/ovsdbserver-nb-0" Feb 02 09:23:23 crc kubenswrapper[4764]: I0202 09:23:23.140651 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/66662206-d732-404d-a81d-6303d9947d60-var-run-ovn\") pod \"ovn-controller-6v7f6\" (UID: \"66662206-d732-404d-a81d-6303d9947d60\") " pod="openstack/ovn-controller-6v7f6" Feb 02 09:23:23 crc kubenswrapper[4764]: I0202 09:23:23.140715 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/cba91a53-7d3d-481b-bd55-3f0bbe7ccbb6-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"cba91a53-7d3d-481b-bd55-3f0bbe7ccbb6\") " pod="openstack/ovsdbserver-nb-0" Feb 02 09:23:23 crc kubenswrapper[4764]: I0202 09:23:23.156299 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/cba91a53-7d3d-481b-bd55-3f0bbe7ccbb6-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"cba91a53-7d3d-481b-bd55-3f0bbe7ccbb6\") " pod="openstack/ovsdbserver-nb-0" Feb 02 09:23:23 crc kubenswrapper[4764]: I0202 09:23:23.156399 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cba91a53-7d3d-481b-bd55-3f0bbe7ccbb6-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"cba91a53-7d3d-481b-bd55-3f0bbe7ccbb6\") " pod="openstack/ovsdbserver-nb-0" Feb 02 09:23:23 crc kubenswrapper[4764]: I0202 09:23:23.158039 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/cba91a53-7d3d-481b-bd55-3f0bbe7ccbb6-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"cba91a53-7d3d-481b-bd55-3f0bbe7ccbb6\") " pod="openstack/ovsdbserver-nb-0" Feb 02 09:23:23 crc kubenswrapper[4764]: I0202 09:23:23.159810 4764 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ovsdbserver-nb-0\" (UID: \"cba91a53-7d3d-481b-bd55-3f0bbe7ccbb6\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/ovsdbserver-nb-0" Feb 02 09:23:23 crc kubenswrapper[4764]: I0202 09:23:23.160163 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cba91a53-7d3d-481b-bd55-3f0bbe7ccbb6-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"cba91a53-7d3d-481b-bd55-3f0bbe7ccbb6\") " pod="openstack/ovsdbserver-nb-0" Feb 02 09:23:23 crc kubenswrapper[4764]: I0202 09:23:23.160642 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cba91a53-7d3d-481b-bd55-3f0bbe7ccbb6-config\") pod \"ovsdbserver-nb-0\" (UID: \"cba91a53-7d3d-481b-bd55-3f0bbe7ccbb6\") " pod="openstack/ovsdbserver-nb-0" Feb 02 09:23:23 crc kubenswrapper[4764]: I0202 09:23:23.164325 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/cba91a53-7d3d-481b-bd55-3f0bbe7ccbb6-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"cba91a53-7d3d-481b-bd55-3f0bbe7ccbb6\") " pod="openstack/ovsdbserver-nb-0" Feb 02 09:23:23 crc kubenswrapper[4764]: I0202 09:23:23.170839 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-2p2bk"] Feb 02 09:23:23 crc kubenswrapper[4764]: I0202 09:23:23.203245 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8kstr\" (UniqueName: \"kubernetes.io/projected/cba91a53-7d3d-481b-bd55-3f0bbe7ccbb6-kube-api-access-8kstr\") pod \"ovsdbserver-nb-0\" (UID: \"cba91a53-7d3d-481b-bd55-3f0bbe7ccbb6\") " pod="openstack/ovsdbserver-nb-0" Feb 02 09:23:23 crc kubenswrapper[4764]: I0202 09:23:23.240653 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ovsdbserver-nb-0\" (UID: \"cba91a53-7d3d-481b-bd55-3f0bbe7ccbb6\") " pod="openstack/ovsdbserver-nb-0" Feb 02 09:23:23 crc kubenswrapper[4764]: I0202 09:23:23.249419 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tb4mn\" (UniqueName: \"kubernetes.io/projected/66662206-d732-404d-a81d-6303d9947d60-kube-api-access-tb4mn\") pod \"ovn-controller-6v7f6\" (UID: \"66662206-d732-404d-a81d-6303d9947d60\") " pod="openstack/ovn-controller-6v7f6" Feb 02 09:23:23 crc kubenswrapper[4764]: I0202 09:23:23.249570 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/66662206-d732-404d-a81d-6303d9947d60-scripts\") pod \"ovn-controller-6v7f6\" (UID: \"66662206-d732-404d-a81d-6303d9947d60\") " pod="openstack/ovn-controller-6v7f6" Feb 02 09:23:23 crc kubenswrapper[4764]: I0202 09:23:23.249628 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66662206-d732-404d-a81d-6303d9947d60-combined-ca-bundle\") pod \"ovn-controller-6v7f6\" (UID: \"66662206-d732-404d-a81d-6303d9947d60\") " pod="openstack/ovn-controller-6v7f6" Feb 02 09:23:23 crc kubenswrapper[4764]: I0202 09:23:23.249716 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/66662206-d732-404d-a81d-6303d9947d60-var-run\") pod \"ovn-controller-6v7f6\" (UID: \"66662206-d732-404d-a81d-6303d9947d60\") " pod="openstack/ovn-controller-6v7f6" Feb 02 09:23:23 crc kubenswrapper[4764]: I0202 09:23:23.249762 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/66662206-d732-404d-a81d-6303d9947d60-ovn-controller-tls-certs\") pod \"ovn-controller-6v7f6\" (UID: \"66662206-d732-404d-a81d-6303d9947d60\") " pod="openstack/ovn-controller-6v7f6" Feb 02 09:23:23 crc kubenswrapper[4764]: I0202 09:23:23.249801 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/29a433e1-2651-43cd-99d1-95694063c980-var-lib\") pod \"ovn-controller-ovs-2p2bk\" (UID: \"29a433e1-2651-43cd-99d1-95694063c980\") " pod="openstack/ovn-controller-ovs-2p2bk" Feb 02 09:23:23 crc kubenswrapper[4764]: I0202 09:23:23.249829 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/66662206-d732-404d-a81d-6303d9947d60-var-log-ovn\") pod \"ovn-controller-6v7f6\" (UID: \"66662206-d732-404d-a81d-6303d9947d60\") " pod="openstack/ovn-controller-6v7f6" Feb 02 09:23:23 crc kubenswrapper[4764]: I0202 09:23:23.250049 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/29a433e1-2651-43cd-99d1-95694063c980-etc-ovs\") pod \"ovn-controller-ovs-2p2bk\" (UID: \"29a433e1-2651-43cd-99d1-95694063c980\") " pod="openstack/ovn-controller-ovs-2p2bk" Feb 02 09:23:23 crc kubenswrapper[4764]: I0202 09:23:23.250092 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wphkm\" (UniqueName: \"kubernetes.io/projected/29a433e1-2651-43cd-99d1-95694063c980-kube-api-access-wphkm\") pod \"ovn-controller-ovs-2p2bk\" (UID: \"29a433e1-2651-43cd-99d1-95694063c980\") " pod="openstack/ovn-controller-ovs-2p2bk" Feb 02 09:23:23 crc kubenswrapper[4764]: I0202 09:23:23.250132 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/29a433e1-2651-43cd-99d1-95694063c980-var-log\") pod \"ovn-controller-ovs-2p2bk\" (UID: \"29a433e1-2651-43cd-99d1-95694063c980\") " pod="openstack/ovn-controller-ovs-2p2bk" Feb 02 09:23:23 crc kubenswrapper[4764]: I0202 09:23:23.250171 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/29a433e1-2651-43cd-99d1-95694063c980-scripts\") pod \"ovn-controller-ovs-2p2bk\" (UID: \"29a433e1-2651-43cd-99d1-95694063c980\") " pod="openstack/ovn-controller-ovs-2p2bk" Feb 02 09:23:23 crc kubenswrapper[4764]: I0202 09:23:23.250240 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/66662206-d732-404d-a81d-6303d9947d60-var-run-ovn\") pod \"ovn-controller-6v7f6\" (UID: \"66662206-d732-404d-a81d-6303d9947d60\") " pod="openstack/ovn-controller-6v7f6" Feb 02 09:23:23 crc kubenswrapper[4764]: I0202 09:23:23.250272 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/29a433e1-2651-43cd-99d1-95694063c980-var-run\") pod \"ovn-controller-ovs-2p2bk\" (UID: \"29a433e1-2651-43cd-99d1-95694063c980\") " pod="openstack/ovn-controller-ovs-2p2bk" Feb 02 09:23:23 crc kubenswrapper[4764]: I0202 09:23:23.251809 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/66662206-d732-404d-a81d-6303d9947d60-var-run\") pod \"ovn-controller-6v7f6\" (UID: \"66662206-d732-404d-a81d-6303d9947d60\") " pod="openstack/ovn-controller-6v7f6" Feb 02 09:23:23 crc kubenswrapper[4764]: I0202 09:23:23.253484 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/66662206-d732-404d-a81d-6303d9947d60-var-log-ovn\") pod \"ovn-controller-6v7f6\" (UID: \"66662206-d732-404d-a81d-6303d9947d60\") " pod="openstack/ovn-controller-6v7f6" Feb 02 09:23:23 crc kubenswrapper[4764]: I0202 09:23:23.255353 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/66662206-d732-404d-a81d-6303d9947d60-var-run-ovn\") pod \"ovn-controller-6v7f6\" (UID: \"66662206-d732-404d-a81d-6303d9947d60\") " pod="openstack/ovn-controller-6v7f6" Feb 02 09:23:23 crc kubenswrapper[4764]: I0202 09:23:23.255870 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/66662206-d732-404d-a81d-6303d9947d60-ovn-controller-tls-certs\") pod \"ovn-controller-6v7f6\" (UID: \"66662206-d732-404d-a81d-6303d9947d60\") " pod="openstack/ovn-controller-6v7f6" Feb 02 09:23:23 crc kubenswrapper[4764]: I0202 09:23:23.255874 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/66662206-d732-404d-a81d-6303d9947d60-scripts\") pod \"ovn-controller-6v7f6\" (UID: \"66662206-d732-404d-a81d-6303d9947d60\") " pod="openstack/ovn-controller-6v7f6" Feb 02 09:23:23 crc kubenswrapper[4764]: I0202 09:23:23.258464 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Feb 02 09:23:23 crc kubenswrapper[4764]: I0202 09:23:23.260232 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66662206-d732-404d-a81d-6303d9947d60-combined-ca-bundle\") pod \"ovn-controller-6v7f6\" (UID: \"66662206-d732-404d-a81d-6303d9947d60\") " pod="openstack/ovn-controller-6v7f6" Feb 02 09:23:23 crc kubenswrapper[4764]: I0202 09:23:23.287859 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tb4mn\" (UniqueName: \"kubernetes.io/projected/66662206-d732-404d-a81d-6303d9947d60-kube-api-access-tb4mn\") pod \"ovn-controller-6v7f6\" (UID: \"66662206-d732-404d-a81d-6303d9947d60\") " pod="openstack/ovn-controller-6v7f6" Feb 02 09:23:23 crc kubenswrapper[4764]: I0202 09:23:23.302746 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-6v7f6" Feb 02 09:23:23 crc kubenswrapper[4764]: I0202 09:23:23.352135 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/29a433e1-2651-43cd-99d1-95694063c980-var-lib\") pod \"ovn-controller-ovs-2p2bk\" (UID: \"29a433e1-2651-43cd-99d1-95694063c980\") " pod="openstack/ovn-controller-ovs-2p2bk" Feb 02 09:23:23 crc kubenswrapper[4764]: I0202 09:23:23.352219 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/29a433e1-2651-43cd-99d1-95694063c980-etc-ovs\") pod \"ovn-controller-ovs-2p2bk\" (UID: \"29a433e1-2651-43cd-99d1-95694063c980\") " pod="openstack/ovn-controller-ovs-2p2bk" Feb 02 09:23:23 crc kubenswrapper[4764]: I0202 09:23:23.352241 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wphkm\" (UniqueName: \"kubernetes.io/projected/29a433e1-2651-43cd-99d1-95694063c980-kube-api-access-wphkm\") pod \"ovn-controller-ovs-2p2bk\" (UID: \"29a433e1-2651-43cd-99d1-95694063c980\") " pod="openstack/ovn-controller-ovs-2p2bk" Feb 02 09:23:23 crc kubenswrapper[4764]: I0202 09:23:23.352262 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/29a433e1-2651-43cd-99d1-95694063c980-var-log\") pod \"ovn-controller-ovs-2p2bk\" (UID: \"29a433e1-2651-43cd-99d1-95694063c980\") " pod="openstack/ovn-controller-ovs-2p2bk" Feb 02 09:23:23 crc kubenswrapper[4764]: I0202 09:23:23.352280 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/29a433e1-2651-43cd-99d1-95694063c980-scripts\") pod \"ovn-controller-ovs-2p2bk\" (UID: \"29a433e1-2651-43cd-99d1-95694063c980\") " pod="openstack/ovn-controller-ovs-2p2bk" Feb 02 09:23:23 crc kubenswrapper[4764]: I0202 09:23:23.353602 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/29a433e1-2651-43cd-99d1-95694063c980-var-run\") pod \"ovn-controller-ovs-2p2bk\" (UID: \"29a433e1-2651-43cd-99d1-95694063c980\") " pod="openstack/ovn-controller-ovs-2p2bk" Feb 02 09:23:23 crc kubenswrapper[4764]: I0202 09:23:23.354143 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/29a433e1-2651-43cd-99d1-95694063c980-var-run\") pod \"ovn-controller-ovs-2p2bk\" (UID: \"29a433e1-2651-43cd-99d1-95694063c980\") " pod="openstack/ovn-controller-ovs-2p2bk" Feb 02 09:23:23 crc kubenswrapper[4764]: I0202 09:23:23.355782 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/29a433e1-2651-43cd-99d1-95694063c980-var-log\") pod \"ovn-controller-ovs-2p2bk\" (UID: \"29a433e1-2651-43cd-99d1-95694063c980\") " pod="openstack/ovn-controller-ovs-2p2bk" Feb 02 09:23:23 crc kubenswrapper[4764]: I0202 09:23:23.352685 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/29a433e1-2651-43cd-99d1-95694063c980-etc-ovs\") pod \"ovn-controller-ovs-2p2bk\" (UID: \"29a433e1-2651-43cd-99d1-95694063c980\") " pod="openstack/ovn-controller-ovs-2p2bk" Feb 02 09:23:23 crc kubenswrapper[4764]: I0202 09:23:23.357083 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/29a433e1-2651-43cd-99d1-95694063c980-scripts\") pod \"ovn-controller-ovs-2p2bk\" (UID: \"29a433e1-2651-43cd-99d1-95694063c980\") " pod="openstack/ovn-controller-ovs-2p2bk" Feb 02 09:23:23 crc kubenswrapper[4764]: I0202 09:23:23.357371 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/29a433e1-2651-43cd-99d1-95694063c980-var-lib\") pod \"ovn-controller-ovs-2p2bk\" (UID: \"29a433e1-2651-43cd-99d1-95694063c980\") " pod="openstack/ovn-controller-ovs-2p2bk" Feb 02 09:23:23 crc kubenswrapper[4764]: I0202 09:23:23.378424 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wphkm\" (UniqueName: \"kubernetes.io/projected/29a433e1-2651-43cd-99d1-95694063c980-kube-api-access-wphkm\") pod \"ovn-controller-ovs-2p2bk\" (UID: \"29a433e1-2651-43cd-99d1-95694063c980\") " pod="openstack/ovn-controller-ovs-2p2bk" Feb 02 09:23:23 crc kubenswrapper[4764]: I0202 09:23:23.456227 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-2p2bk" Feb 02 09:23:26 crc kubenswrapper[4764]: I0202 09:23:26.601466 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Feb 02 09:23:26 crc kubenswrapper[4764]: I0202 09:23:26.603885 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Feb 02 09:23:26 crc kubenswrapper[4764]: I0202 09:23:26.606134 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-lj7xv" Feb 02 09:23:26 crc kubenswrapper[4764]: I0202 09:23:26.606324 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Feb 02 09:23:26 crc kubenswrapper[4764]: I0202 09:23:26.607815 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Feb 02 09:23:26 crc kubenswrapper[4764]: I0202 09:23:26.608866 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Feb 02 09:23:26 crc kubenswrapper[4764]: I0202 09:23:26.614205 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Feb 02 09:23:26 crc kubenswrapper[4764]: I0202 09:23:26.653856 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/89c39976-ff91-4c6b-b369-5cba94e02db8-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"89c39976-ff91-4c6b-b369-5cba94e02db8\") " pod="openstack/ovsdbserver-sb-0" Feb 02 09:23:26 crc kubenswrapper[4764]: I0202 09:23:26.653921 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89c39976-ff91-4c6b-b369-5cba94e02db8-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"89c39976-ff91-4c6b-b369-5cba94e02db8\") " pod="openstack/ovsdbserver-sb-0" Feb 02 09:23:26 crc kubenswrapper[4764]: I0202 09:23:26.653966 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/89c39976-ff91-4c6b-b369-5cba94e02db8-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"89c39976-ff91-4c6b-b369-5cba94e02db8\") " pod="openstack/ovsdbserver-sb-0" Feb 02 09:23:26 crc kubenswrapper[4764]: I0202 09:23:26.653994 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"ovsdbserver-sb-0\" (UID: \"89c39976-ff91-4c6b-b369-5cba94e02db8\") " pod="openstack/ovsdbserver-sb-0" Feb 02 09:23:26 crc kubenswrapper[4764]: I0202 09:23:26.654046 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/89c39976-ff91-4c6b-b369-5cba94e02db8-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"89c39976-ff91-4c6b-b369-5cba94e02db8\") " pod="openstack/ovsdbserver-sb-0" Feb 02 09:23:26 crc kubenswrapper[4764]: I0202 09:23:26.654075 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/89c39976-ff91-4c6b-b369-5cba94e02db8-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"89c39976-ff91-4c6b-b369-5cba94e02db8\") " pod="openstack/ovsdbserver-sb-0" Feb 02 09:23:26 crc kubenswrapper[4764]: I0202 09:23:26.654102 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l8nzn\" (UniqueName: \"kubernetes.io/projected/89c39976-ff91-4c6b-b369-5cba94e02db8-kube-api-access-l8nzn\") pod \"ovsdbserver-sb-0\" (UID: \"89c39976-ff91-4c6b-b369-5cba94e02db8\") " pod="openstack/ovsdbserver-sb-0" Feb 02 09:23:26 crc kubenswrapper[4764]: I0202 09:23:26.654132 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/89c39976-ff91-4c6b-b369-5cba94e02db8-config\") pod \"ovsdbserver-sb-0\" (UID: \"89c39976-ff91-4c6b-b369-5cba94e02db8\") " pod="openstack/ovsdbserver-sb-0" Feb 02 09:23:26 crc kubenswrapper[4764]: I0202 09:23:26.755752 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/89c39976-ff91-4c6b-b369-5cba94e02db8-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"89c39976-ff91-4c6b-b369-5cba94e02db8\") " pod="openstack/ovsdbserver-sb-0" Feb 02 09:23:26 crc kubenswrapper[4764]: I0202 09:23:26.755798 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89c39976-ff91-4c6b-b369-5cba94e02db8-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"89c39976-ff91-4c6b-b369-5cba94e02db8\") " pod="openstack/ovsdbserver-sb-0" Feb 02 09:23:26 crc kubenswrapper[4764]: I0202 09:23:26.755823 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/89c39976-ff91-4c6b-b369-5cba94e02db8-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"89c39976-ff91-4c6b-b369-5cba94e02db8\") " pod="openstack/ovsdbserver-sb-0" Feb 02 09:23:26 crc kubenswrapper[4764]: I0202 09:23:26.755847 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"ovsdbserver-sb-0\" (UID: \"89c39976-ff91-4c6b-b369-5cba94e02db8\") " pod="openstack/ovsdbserver-sb-0" Feb 02 09:23:26 crc kubenswrapper[4764]: I0202 09:23:26.755902 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/89c39976-ff91-4c6b-b369-5cba94e02db8-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"89c39976-ff91-4c6b-b369-5cba94e02db8\") " pod="openstack/ovsdbserver-sb-0" Feb 02 09:23:26 crc kubenswrapper[4764]: I0202 09:23:26.755967 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/89c39976-ff91-4c6b-b369-5cba94e02db8-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"89c39976-ff91-4c6b-b369-5cba94e02db8\") " pod="openstack/ovsdbserver-sb-0" Feb 02 09:23:26 crc kubenswrapper[4764]: I0202 09:23:26.755999 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l8nzn\" (UniqueName: \"kubernetes.io/projected/89c39976-ff91-4c6b-b369-5cba94e02db8-kube-api-access-l8nzn\") pod \"ovsdbserver-sb-0\" (UID: \"89c39976-ff91-4c6b-b369-5cba94e02db8\") " pod="openstack/ovsdbserver-sb-0" Feb 02 09:23:26 crc kubenswrapper[4764]: I0202 09:23:26.756024 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/89c39976-ff91-4c6b-b369-5cba94e02db8-config\") pod \"ovsdbserver-sb-0\" (UID: \"89c39976-ff91-4c6b-b369-5cba94e02db8\") " pod="openstack/ovsdbserver-sb-0" Feb 02 09:23:26 crc kubenswrapper[4764]: I0202 09:23:26.756478 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/89c39976-ff91-4c6b-b369-5cba94e02db8-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"89c39976-ff91-4c6b-b369-5cba94e02db8\") " pod="openstack/ovsdbserver-sb-0" Feb 02 09:23:26 crc kubenswrapper[4764]: I0202 09:23:26.756490 4764 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"ovsdbserver-sb-0\" (UID: \"89c39976-ff91-4c6b-b369-5cba94e02db8\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/ovsdbserver-sb-0" Feb 02 09:23:26 crc kubenswrapper[4764]: I0202 09:23:26.758670 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/89c39976-ff91-4c6b-b369-5cba94e02db8-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"89c39976-ff91-4c6b-b369-5cba94e02db8\") " pod="openstack/ovsdbserver-sb-0" Feb 02 09:23:26 crc kubenswrapper[4764]: I0202 09:23:26.759416 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/89c39976-ff91-4c6b-b369-5cba94e02db8-config\") pod \"ovsdbserver-sb-0\" (UID: \"89c39976-ff91-4c6b-b369-5cba94e02db8\") " pod="openstack/ovsdbserver-sb-0" Feb 02 09:23:26 crc kubenswrapper[4764]: I0202 09:23:26.776816 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/89c39976-ff91-4c6b-b369-5cba94e02db8-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"89c39976-ff91-4c6b-b369-5cba94e02db8\") " pod="openstack/ovsdbserver-sb-0" Feb 02 09:23:26 crc kubenswrapper[4764]: I0202 09:23:26.777523 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89c39976-ff91-4c6b-b369-5cba94e02db8-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"89c39976-ff91-4c6b-b369-5cba94e02db8\") " pod="openstack/ovsdbserver-sb-0" Feb 02 09:23:26 crc kubenswrapper[4764]: I0202 09:23:26.779007 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/89c39976-ff91-4c6b-b369-5cba94e02db8-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"89c39976-ff91-4c6b-b369-5cba94e02db8\") " pod="openstack/ovsdbserver-sb-0" Feb 02 09:23:26 crc kubenswrapper[4764]: I0202 09:23:26.782883 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l8nzn\" (UniqueName: \"kubernetes.io/projected/89c39976-ff91-4c6b-b369-5cba94e02db8-kube-api-access-l8nzn\") pod \"ovsdbserver-sb-0\" (UID: \"89c39976-ff91-4c6b-b369-5cba94e02db8\") " pod="openstack/ovsdbserver-sb-0" Feb 02 09:23:26 crc kubenswrapper[4764]: I0202 09:23:26.791590 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"ovsdbserver-sb-0\" (UID: \"89c39976-ff91-4c6b-b369-5cba94e02db8\") " pod="openstack/ovsdbserver-sb-0" Feb 02 09:23:26 crc kubenswrapper[4764]: I0202 09:23:26.925912 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Feb 02 09:23:30 crc kubenswrapper[4764]: I0202 09:23:30.977612 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Feb 02 09:23:31 crc kubenswrapper[4764]: I0202 09:23:31.681205 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-2p2bk"] Feb 02 09:23:32 crc kubenswrapper[4764]: I0202 09:23:32.392638 4764 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-lx9rg" podUID="10c729f0-d54f-4b06-995b-ab35e4d583bb" containerName="registry-server" probeResult="failure" output=< Feb 02 09:23:32 crc kubenswrapper[4764]: timeout: failed to connect service ":50051" within 1s Feb 02 09:23:32 crc kubenswrapper[4764]: > Feb 02 09:23:34 crc kubenswrapper[4764]: I0202 09:23:34.748343 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-6v7f6"] Feb 02 09:23:35 crc kubenswrapper[4764]: I0202 09:23:35.928736 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-kgc8j"] Feb 02 09:23:35 crc kubenswrapper[4764]: I0202 09:23:35.929879 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-kgc8j" Feb 02 09:23:35 crc kubenswrapper[4764]: I0202 09:23:35.941498 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Feb 02 09:23:35 crc kubenswrapper[4764]: I0202 09:23:35.964017 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-kgc8j"] Feb 02 09:23:36 crc kubenswrapper[4764]: I0202 09:23:36.040830 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/b3becd85-0167-4ffe-802d-67988c486908-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-kgc8j\" (UID: \"b3becd85-0167-4ffe-802d-67988c486908\") " pod="openstack/ovn-controller-metrics-kgc8j" Feb 02 09:23:36 crc kubenswrapper[4764]: I0202 09:23:36.040897 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b3becd85-0167-4ffe-802d-67988c486908-config\") pod \"ovn-controller-metrics-kgc8j\" (UID: \"b3becd85-0167-4ffe-802d-67988c486908\") " pod="openstack/ovn-controller-metrics-kgc8j" Feb 02 09:23:36 crc kubenswrapper[4764]: I0202 09:23:36.040973 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/b3becd85-0167-4ffe-802d-67988c486908-ovs-rundir\") pod \"ovn-controller-metrics-kgc8j\" (UID: \"b3becd85-0167-4ffe-802d-67988c486908\") " pod="openstack/ovn-controller-metrics-kgc8j" Feb 02 09:23:36 crc kubenswrapper[4764]: I0202 09:23:36.041709 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3becd85-0167-4ffe-802d-67988c486908-combined-ca-bundle\") pod \"ovn-controller-metrics-kgc8j\" (UID: \"b3becd85-0167-4ffe-802d-67988c486908\") " pod="openstack/ovn-controller-metrics-kgc8j" Feb 02 09:23:36 crc kubenswrapper[4764]: I0202 09:23:36.041801 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8pdsr\" (UniqueName: \"kubernetes.io/projected/b3becd85-0167-4ffe-802d-67988c486908-kube-api-access-8pdsr\") pod \"ovn-controller-metrics-kgc8j\" (UID: \"b3becd85-0167-4ffe-802d-67988c486908\") " pod="openstack/ovn-controller-metrics-kgc8j" Feb 02 09:23:36 crc kubenswrapper[4764]: I0202 09:23:36.041899 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/b3becd85-0167-4ffe-802d-67988c486908-ovn-rundir\") pod \"ovn-controller-metrics-kgc8j\" (UID: \"b3becd85-0167-4ffe-802d-67988c486908\") " pod="openstack/ovn-controller-metrics-kgc8j" Feb 02 09:23:36 crc kubenswrapper[4764]: I0202 09:23:36.101506 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-67ff45466c-v6vn7"] Feb 02 09:23:36 crc kubenswrapper[4764]: I0202 09:23:36.139005 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-64f7f48db9-rz8h7"] Feb 02 09:23:36 crc kubenswrapper[4764]: I0202 09:23:36.140391 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64f7f48db9-rz8h7" Feb 02 09:23:36 crc kubenswrapper[4764]: I0202 09:23:36.143405 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Feb 02 09:23:36 crc kubenswrapper[4764]: I0202 09:23:36.144693 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/b3becd85-0167-4ffe-802d-67988c486908-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-kgc8j\" (UID: \"b3becd85-0167-4ffe-802d-67988c486908\") " pod="openstack/ovn-controller-metrics-kgc8j" Feb 02 09:23:36 crc kubenswrapper[4764]: I0202 09:23:36.144742 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b3becd85-0167-4ffe-802d-67988c486908-config\") pod \"ovn-controller-metrics-kgc8j\" (UID: \"b3becd85-0167-4ffe-802d-67988c486908\") " pod="openstack/ovn-controller-metrics-kgc8j" Feb 02 09:23:36 crc kubenswrapper[4764]: I0202 09:23:36.144785 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/b3becd85-0167-4ffe-802d-67988c486908-ovs-rundir\") pod \"ovn-controller-metrics-kgc8j\" (UID: \"b3becd85-0167-4ffe-802d-67988c486908\") " pod="openstack/ovn-controller-metrics-kgc8j" Feb 02 09:23:36 crc kubenswrapper[4764]: I0202 09:23:36.144832 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3becd85-0167-4ffe-802d-67988c486908-combined-ca-bundle\") pod \"ovn-controller-metrics-kgc8j\" (UID: \"b3becd85-0167-4ffe-802d-67988c486908\") " pod="openstack/ovn-controller-metrics-kgc8j" Feb 02 09:23:36 crc kubenswrapper[4764]: I0202 09:23:36.144879 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8pdsr\" (UniqueName: \"kubernetes.io/projected/b3becd85-0167-4ffe-802d-67988c486908-kube-api-access-8pdsr\") pod \"ovn-controller-metrics-kgc8j\" (UID: \"b3becd85-0167-4ffe-802d-67988c486908\") " pod="openstack/ovn-controller-metrics-kgc8j" Feb 02 09:23:36 crc kubenswrapper[4764]: I0202 09:23:36.144965 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/b3becd85-0167-4ffe-802d-67988c486908-ovn-rundir\") pod \"ovn-controller-metrics-kgc8j\" (UID: \"b3becd85-0167-4ffe-802d-67988c486908\") " pod="openstack/ovn-controller-metrics-kgc8j" Feb 02 09:23:36 crc kubenswrapper[4764]: I0202 09:23:36.145497 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/b3becd85-0167-4ffe-802d-67988c486908-ovs-rundir\") pod \"ovn-controller-metrics-kgc8j\" (UID: \"b3becd85-0167-4ffe-802d-67988c486908\") " pod="openstack/ovn-controller-metrics-kgc8j" Feb 02 09:23:36 crc kubenswrapper[4764]: I0202 09:23:36.145700 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/b3becd85-0167-4ffe-802d-67988c486908-ovn-rundir\") pod \"ovn-controller-metrics-kgc8j\" (UID: \"b3becd85-0167-4ffe-802d-67988c486908\") " pod="openstack/ovn-controller-metrics-kgc8j" Feb 02 09:23:36 crc kubenswrapper[4764]: I0202 09:23:36.146399 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b3becd85-0167-4ffe-802d-67988c486908-config\") pod \"ovn-controller-metrics-kgc8j\" (UID: \"b3becd85-0167-4ffe-802d-67988c486908\") " pod="openstack/ovn-controller-metrics-kgc8j" Feb 02 09:23:36 crc kubenswrapper[4764]: I0202 09:23:36.153414 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-64f7f48db9-rz8h7"] Feb 02 09:23:36 crc kubenswrapper[4764]: I0202 09:23:36.154275 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/b3becd85-0167-4ffe-802d-67988c486908-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-kgc8j\" (UID: \"b3becd85-0167-4ffe-802d-67988c486908\") " pod="openstack/ovn-controller-metrics-kgc8j" Feb 02 09:23:36 crc kubenswrapper[4764]: I0202 09:23:36.173619 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3becd85-0167-4ffe-802d-67988c486908-combined-ca-bundle\") pod \"ovn-controller-metrics-kgc8j\" (UID: \"b3becd85-0167-4ffe-802d-67988c486908\") " pod="openstack/ovn-controller-metrics-kgc8j" Feb 02 09:23:36 crc kubenswrapper[4764]: I0202 09:23:36.194817 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8pdsr\" (UniqueName: \"kubernetes.io/projected/b3becd85-0167-4ffe-802d-67988c486908-kube-api-access-8pdsr\") pod \"ovn-controller-metrics-kgc8j\" (UID: \"b3becd85-0167-4ffe-802d-67988c486908\") " pod="openstack/ovn-controller-metrics-kgc8j" Feb 02 09:23:36 crc kubenswrapper[4764]: I0202 09:23:36.246997 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/456b6ada-1b28-4df6-b74b-c470910598e5-config\") pod \"dnsmasq-dns-64f7f48db9-rz8h7\" (UID: \"456b6ada-1b28-4df6-b74b-c470910598e5\") " pod="openstack/dnsmasq-dns-64f7f48db9-rz8h7" Feb 02 09:23:36 crc kubenswrapper[4764]: I0202 09:23:36.247389 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/456b6ada-1b28-4df6-b74b-c470910598e5-ovsdbserver-nb\") pod \"dnsmasq-dns-64f7f48db9-rz8h7\" (UID: \"456b6ada-1b28-4df6-b74b-c470910598e5\") " pod="openstack/dnsmasq-dns-64f7f48db9-rz8h7" Feb 02 09:23:36 crc kubenswrapper[4764]: I0202 09:23:36.247515 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/456b6ada-1b28-4df6-b74b-c470910598e5-dns-svc\") pod \"dnsmasq-dns-64f7f48db9-rz8h7\" (UID: \"456b6ada-1b28-4df6-b74b-c470910598e5\") " pod="openstack/dnsmasq-dns-64f7f48db9-rz8h7" Feb 02 09:23:36 crc kubenswrapper[4764]: I0202 09:23:36.247638 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4pqz6\" (UniqueName: \"kubernetes.io/projected/456b6ada-1b28-4df6-b74b-c470910598e5-kube-api-access-4pqz6\") pod \"dnsmasq-dns-64f7f48db9-rz8h7\" (UID: \"456b6ada-1b28-4df6-b74b-c470910598e5\") " pod="openstack/dnsmasq-dns-64f7f48db9-rz8h7" Feb 02 09:23:36 crc kubenswrapper[4764]: I0202 09:23:36.257617 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-kgc8j" Feb 02 09:23:36 crc kubenswrapper[4764]: I0202 09:23:36.348999 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/456b6ada-1b28-4df6-b74b-c470910598e5-dns-svc\") pod \"dnsmasq-dns-64f7f48db9-rz8h7\" (UID: \"456b6ada-1b28-4df6-b74b-c470910598e5\") " pod="openstack/dnsmasq-dns-64f7f48db9-rz8h7" Feb 02 09:23:36 crc kubenswrapper[4764]: I0202 09:23:36.349078 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4pqz6\" (UniqueName: \"kubernetes.io/projected/456b6ada-1b28-4df6-b74b-c470910598e5-kube-api-access-4pqz6\") pod \"dnsmasq-dns-64f7f48db9-rz8h7\" (UID: \"456b6ada-1b28-4df6-b74b-c470910598e5\") " pod="openstack/dnsmasq-dns-64f7f48db9-rz8h7" Feb 02 09:23:36 crc kubenswrapper[4764]: I0202 09:23:36.349176 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/456b6ada-1b28-4df6-b74b-c470910598e5-config\") pod \"dnsmasq-dns-64f7f48db9-rz8h7\" (UID: \"456b6ada-1b28-4df6-b74b-c470910598e5\") " pod="openstack/dnsmasq-dns-64f7f48db9-rz8h7" Feb 02 09:23:36 crc kubenswrapper[4764]: I0202 09:23:36.349235 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/456b6ada-1b28-4df6-b74b-c470910598e5-ovsdbserver-nb\") pod \"dnsmasq-dns-64f7f48db9-rz8h7\" (UID: \"456b6ada-1b28-4df6-b74b-c470910598e5\") " pod="openstack/dnsmasq-dns-64f7f48db9-rz8h7" Feb 02 09:23:36 crc kubenswrapper[4764]: I0202 09:23:36.349839 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/456b6ada-1b28-4df6-b74b-c470910598e5-dns-svc\") pod \"dnsmasq-dns-64f7f48db9-rz8h7\" (UID: \"456b6ada-1b28-4df6-b74b-c470910598e5\") " pod="openstack/dnsmasq-dns-64f7f48db9-rz8h7" Feb 02 09:23:36 crc kubenswrapper[4764]: I0202 09:23:36.350679 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/456b6ada-1b28-4df6-b74b-c470910598e5-config\") pod \"dnsmasq-dns-64f7f48db9-rz8h7\" (UID: \"456b6ada-1b28-4df6-b74b-c470910598e5\") " pod="openstack/dnsmasq-dns-64f7f48db9-rz8h7" Feb 02 09:23:36 crc kubenswrapper[4764]: I0202 09:23:36.351725 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/456b6ada-1b28-4df6-b74b-c470910598e5-ovsdbserver-nb\") pod \"dnsmasq-dns-64f7f48db9-rz8h7\" (UID: \"456b6ada-1b28-4df6-b74b-c470910598e5\") " pod="openstack/dnsmasq-dns-64f7f48db9-rz8h7" Feb 02 09:23:36 crc kubenswrapper[4764]: I0202 09:23:36.370169 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4pqz6\" (UniqueName: \"kubernetes.io/projected/456b6ada-1b28-4df6-b74b-c470910598e5-kube-api-access-4pqz6\") pod \"dnsmasq-dns-64f7f48db9-rz8h7\" (UID: \"456b6ada-1b28-4df6-b74b-c470910598e5\") " pod="openstack/dnsmasq-dns-64f7f48db9-rz8h7" Feb 02 09:23:36 crc kubenswrapper[4764]: I0202 09:23:36.523148 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64f7f48db9-rz8h7" Feb 02 09:23:38 crc kubenswrapper[4764]: E0202 09:23:38.507347 4764 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-memcached@sha256:3c3b6a71bc3205fc3cf7616172526846dac02edd188be775b358a604448e5a66" Feb 02 09:23:38 crc kubenswrapper[4764]: E0202 09:23:38.507535 4764 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:memcached,Image:quay.io/podified-antelope-centos9/openstack-memcached@sha256:3c3b6a71bc3205fc3cf7616172526846dac02edd188be775b358a604448e5a66,Command:[/usr/bin/dumb-init -- /usr/local/bin/kolla_start],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:memcached,HostPort:0,ContainerPort:11211,Protocol:TCP,HostIP:,},ContainerPort{Name:memcached-tls,HostPort:0,ContainerPort:11212,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:POD_IPS,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIPs,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:CONFIG_HASH,Value:n7ch585h5b4h5d8h8dh78h79h56h5ddh5dbh86h668h586h58hbchfh87h89h88hcch64h669h597h58bh559h5ch9bh56fh594h94hddhd7q,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/src,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kolla-config,ReadOnly:true,MountPath:/var/lib/kolla/config_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:memcached-tls-certs,ReadOnly:true,MountPath:/var/lib/config-data/tls/certs/memcached.crt,SubPath:tls.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:memcached-tls-certs,ReadOnly:true,MountPath:/var/lib/config-data/tls/private/memcached.key,SubPath:tls.key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6v68h,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 11211 },Host:,},GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 11211 },Host:,},GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42457,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42457,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod memcached-0_openstack(cd76e53a-13fc-4293-b5ca-2324224850cc): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 02 09:23:38 crc kubenswrapper[4764]: E0202 09:23:38.508721 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"memcached\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/memcached-0" podUID="cd76e53a-13fc-4293-b5ca-2324224850cc" Feb 02 09:23:38 crc kubenswrapper[4764]: E0202 09:23:38.517871 4764 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-mariadb@sha256:0f7943e02fbdd3daec1d3db72fa9396bf37ad3fdd6b0f3119c90e29629e095ed" Feb 02 09:23:38 crc kubenswrapper[4764]: E0202 09:23:38.518087 4764 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:mysql-bootstrap,Image:quay.io/podified-antelope-centos9/openstack-mariadb@sha256:0f7943e02fbdd3daec1d3db72fa9396bf37ad3fdd6b0f3119c90e29629e095ed,Command:[bash /var/lib/operator-scripts/mysql_bootstrap.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:True,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:mysql-db,ReadOnly:false,MountPath:/var/lib/mysql,SubPath:mysql,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-default,ReadOnly:true,MountPath:/var/lib/config-data/default,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-generated,ReadOnly:false,MountPath:/var/lib/config-data/generated,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:operator-scripts,ReadOnly:true,MountPath:/var/lib/operator-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kolla-config,ReadOnly:true,MountPath:/var/lib/kolla/config_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-dm8vj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-galera-0_openstack(1d6b6c31-819e-4098-b11b-6073cc751cad): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 02 09:23:38 crc kubenswrapper[4764]: E0202 09:23:38.519237 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/openstack-galera-0" podUID="1d6b6c31-819e-4098-b11b-6073cc751cad" Feb 02 09:23:39 crc kubenswrapper[4764]: E0202 09:23:39.255917 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-mariadb@sha256:0f7943e02fbdd3daec1d3db72fa9396bf37ad3fdd6b0f3119c90e29629e095ed\\\"\"" pod="openstack/openstack-galera-0" podUID="1d6b6c31-819e-4098-b11b-6073cc751cad" Feb 02 09:23:39 crc kubenswrapper[4764]: E0202 09:23:39.256177 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"memcached\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-memcached@sha256:3c3b6a71bc3205fc3cf7616172526846dac02edd188be775b358a604448e5a66\\\"\"" pod="openstack/memcached-0" podUID="cd76e53a-13fc-4293-b5ca-2324224850cc" Feb 02 09:23:39 crc kubenswrapper[4764]: E0202 09:23:39.829138 4764 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-rabbitmq@sha256:221c84e162c46ac7454de6fb84343d0a605f2ea1d7d5647a34a66569e0a8fd76" Feb 02 09:23:39 crc kubenswrapper[4764]: E0202 09:23:39.829398 4764 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.io/podified-antelope-centos9/openstack-rabbitmq@sha256:221c84e162c46ac7454de6fb84343d0a605f2ea1d7d5647a34a66569e0a8fd76,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vs58s,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-server-0_openstack(53d5d300-11ae-408d-8bb8-347defb93991): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 02 09:23:39 crc kubenswrapper[4764]: E0202 09:23:39.831427 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-server-0" podUID="53d5d300-11ae-408d-8bb8-347defb93991" Feb 02 09:23:39 crc kubenswrapper[4764]: E0202 09:23:39.908968 4764 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-rabbitmq@sha256:221c84e162c46ac7454de6fb84343d0a605f2ea1d7d5647a34a66569e0a8fd76" Feb 02 09:23:39 crc kubenswrapper[4764]: E0202 09:23:39.909090 4764 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.io/podified-antelope-centos9/openstack-rabbitmq@sha256:221c84e162c46ac7454de6fb84343d0a605f2ea1d7d5647a34a66569e0a8fd76,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-27wk5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cell1-server-0_openstack(03d9f090-4a93-4226-8fce-6fd0a425d929): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 02 09:23:39 crc kubenswrapper[4764]: E0202 09:23:39.910324 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-cell1-server-0" podUID="03d9f090-4a93-4226-8fce-6fd0a425d929" Feb 02 09:23:40 crc kubenswrapper[4764]: I0202 09:23:40.261448 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"cba91a53-7d3d-481b-bd55-3f0bbe7ccbb6","Type":"ContainerStarted","Data":"c4f3398f9580a8c3bfe6f529bd852e3572c50a9592fc934241cf1499b9f42732"} Feb 02 09:23:40 crc kubenswrapper[4764]: I0202 09:23:40.263129 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-2p2bk" event={"ID":"29a433e1-2651-43cd-99d1-95694063c980","Type":"ContainerStarted","Data":"ea3d787fd0ece59a746baf586b26cdc993770fbb411e43ffb494f1c77f55ef26"} Feb 02 09:23:40 crc kubenswrapper[4764]: E0202 09:23:40.269407 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-rabbitmq@sha256:221c84e162c46ac7454de6fb84343d0a605f2ea1d7d5647a34a66569e0a8fd76\\\"\"" pod="openstack/rabbitmq-server-0" podUID="53d5d300-11ae-408d-8bb8-347defb93991" Feb 02 09:23:40 crc kubenswrapper[4764]: E0202 09:23:40.269431 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-rabbitmq@sha256:221c84e162c46ac7454de6fb84343d0a605f2ea1d7d5647a34a66569e0a8fd76\\\"\"" pod="openstack/rabbitmq-cell1-server-0" podUID="03d9f090-4a93-4226-8fce-6fd0a425d929" Feb 02 09:23:41 crc kubenswrapper[4764]: I0202 09:23:41.385556 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-lx9rg" Feb 02 09:23:41 crc kubenswrapper[4764]: I0202 09:23:41.427405 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-lx9rg" Feb 02 09:23:41 crc kubenswrapper[4764]: I0202 09:23:41.616206 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-lx9rg"] Feb 02 09:23:43 crc kubenswrapper[4764]: I0202 09:23:43.283839 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-lx9rg" podUID="10c729f0-d54f-4b06-995b-ab35e4d583bb" containerName="registry-server" containerID="cri-o://98a0ac0c06018d32adfbe9f8e8984832df81fc12c656ce5683faf6d86cd1d986" gracePeriod=2 Feb 02 09:23:44 crc kubenswrapper[4764]: I0202 09:23:44.300510 4764 generic.go:334] "Generic (PLEG): container finished" podID="10c729f0-d54f-4b06-995b-ab35e4d583bb" containerID="98a0ac0c06018d32adfbe9f8e8984832df81fc12c656ce5683faf6d86cd1d986" exitCode=0 Feb 02 09:23:44 crc kubenswrapper[4764]: I0202 09:23:44.300548 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lx9rg" event={"ID":"10c729f0-d54f-4b06-995b-ab35e4d583bb","Type":"ContainerDied","Data":"98a0ac0c06018d32adfbe9f8e8984832df81fc12c656ce5683faf6d86cd1d986"} Feb 02 09:23:45 crc kubenswrapper[4764]: I0202 09:23:45.310237 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-6v7f6" event={"ID":"66662206-d732-404d-a81d-6303d9947d60","Type":"ContainerStarted","Data":"a0b23fab94f0ced6bbf8453e263f6ffed9247a37453d3658424c398bfd04e775"} Feb 02 09:23:45 crc kubenswrapper[4764]: E0202 09:23:45.888984 4764 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:f391b842000dadaeb692eb6b5e845c2aa3125ef24679fbb4af2c8b98252de4b2" Feb 02 09:23:45 crc kubenswrapper[4764]: E0202 09:23:45.889158 4764 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:f391b842000dadaeb692eb6b5e845c2aa3125ef24679fbb4af2c8b98252de4b2,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-m62f5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-855cbc58c5-zcszj_openstack(985e7a8f-2a38-4ee8-a541-f2c8b7ff9fde): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 02 09:23:45 crc kubenswrapper[4764]: E0202 09:23:45.890518 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-855cbc58c5-zcszj" podUID="985e7a8f-2a38-4ee8-a541-f2c8b7ff9fde" Feb 02 09:23:45 crc kubenswrapper[4764]: E0202 09:23:45.899145 4764 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:f391b842000dadaeb692eb6b5e845c2aa3125ef24679fbb4af2c8b98252de4b2" Feb 02 09:23:45 crc kubenswrapper[4764]: E0202 09:23:45.899329 4764 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:f391b842000dadaeb692eb6b5e845c2aa3125ef24679fbb4af2c8b98252de4b2,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rkqmf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-6fcf94d689-4qtb8_openstack(58c314a7-0090-49dd-afbd-88826c8733b0): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 02 09:23:45 crc kubenswrapper[4764]: E0202 09:23:45.900594 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-6fcf94d689-4qtb8" podUID="58c314a7-0090-49dd-afbd-88826c8733b0" Feb 02 09:23:45 crc kubenswrapper[4764]: E0202 09:23:45.900999 4764 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:f391b842000dadaeb692eb6b5e845c2aa3125ef24679fbb4af2c8b98252de4b2" Feb 02 09:23:45 crc kubenswrapper[4764]: E0202 09:23:45.901239 4764 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:f391b842000dadaeb692eb6b5e845c2aa3125ef24679fbb4af2c8b98252de4b2,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n659h4h664hbh658h587h67ch89h587h8fh679hc6hf9h55fh644h5d5h698h68dh5cdh5ffh669h54ch9h689hb8hd4h5bfhd8h5d7h5fh665h574q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-54w4t,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-67ff45466c-v6vn7_openstack(210b0439-a7e6-4f9f-b847-0a04cf044f87): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 02 09:23:45 crc kubenswrapper[4764]: E0202 09:23:45.902624 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-67ff45466c-v6vn7" podUID="210b0439-a7e6-4f9f-b847-0a04cf044f87" Feb 02 09:23:46 crc kubenswrapper[4764]: E0202 09:23:46.034993 4764 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:f391b842000dadaeb692eb6b5e845c2aa3125ef24679fbb4af2c8b98252de4b2" Feb 02 09:23:46 crc kubenswrapper[4764]: E0202 09:23:46.035162 4764 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:f391b842000dadaeb692eb6b5e845c2aa3125ef24679fbb4af2c8b98252de4b2,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n68chd6h679hbfh55fhc6h5ffh5d8h94h56ch589hb4hc5h57bh677hcdh655h8dh667h675h654h66ch567h8fh659h5b4h675h566h55bh54h67dh6dq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-jdr4z,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-f54874ffc-wmsqt_openstack(85525aa3-d70a-4d66-8bbf-536be10a0337): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 02 09:23:46 crc kubenswrapper[4764]: E0202 09:23:46.037252 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-f54874ffc-wmsqt" podUID="85525aa3-d70a-4d66-8bbf-536be10a0337" Feb 02 09:23:46 crc kubenswrapper[4764]: E0202 09:23:46.335824 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:f391b842000dadaeb692eb6b5e845c2aa3125ef24679fbb4af2c8b98252de4b2\\\"\"" pod="openstack/dnsmasq-dns-f54874ffc-wmsqt" podUID="85525aa3-d70a-4d66-8bbf-536be10a0337" Feb 02 09:23:46 crc kubenswrapper[4764]: I0202 09:23:46.753954 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Feb 02 09:23:47 crc kubenswrapper[4764]: E0202 09:23:47.387696 4764 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.k8s.io/kube-state-metrics/kube-state-metrics@sha256:db384bf43222b066c378e77027a675d4cd9911107adba46c2922b3a55e10d6fb" Feb 02 09:23:47 crc kubenswrapper[4764]: E0202 09:23:47.387749 4764 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.k8s.io/kube-state-metrics/kube-state-metrics@sha256:db384bf43222b066c378e77027a675d4cd9911107adba46c2922b3a55e10d6fb" Feb 02 09:23:47 crc kubenswrapper[4764]: E0202 09:23:47.387870 4764 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-state-metrics,Image:registry.k8s.io/kube-state-metrics/kube-state-metrics@sha256:db384bf43222b066c378e77027a675d4cd9911107adba46c2922b3a55e10d6fb,Command:[],Args:[--resources=pods --namespaces=openstack],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:http-metrics,HostPort:0,ContainerPort:8080,Protocol:TCP,HostIP:,},ContainerPort{Name:telemetry,HostPort:0,ContainerPort:8081,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-98v7s,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/livez,Port:{0 8080 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:*true,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod kube-state-metrics-0_openstack(26d7a0b3-ea1e-417b-b417-69ae84bd5d24): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Feb 02 09:23:47 crc kubenswrapper[4764]: E0202 09:23:47.389149 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-state-metrics\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openstack/kube-state-metrics-0" podUID="26d7a0b3-ea1e-417b-b417-69ae84bd5d24" Feb 02 09:23:47 crc kubenswrapper[4764]: I0202 09:23:47.536995 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-855cbc58c5-zcszj" Feb 02 09:23:47 crc kubenswrapper[4764]: I0202 09:23:47.566604 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lx9rg" Feb 02 09:23:47 crc kubenswrapper[4764]: I0202 09:23:47.595036 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/10c729f0-d54f-4b06-995b-ab35e4d583bb-utilities\") pod \"10c729f0-d54f-4b06-995b-ab35e4d583bb\" (UID: \"10c729f0-d54f-4b06-995b-ab35e4d583bb\") " Feb 02 09:23:47 crc kubenswrapper[4764]: I0202 09:23:47.595087 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m62f5\" (UniqueName: \"kubernetes.io/projected/985e7a8f-2a38-4ee8-a541-f2c8b7ff9fde-kube-api-access-m62f5\") pod \"985e7a8f-2a38-4ee8-a541-f2c8b7ff9fde\" (UID: \"985e7a8f-2a38-4ee8-a541-f2c8b7ff9fde\") " Feb 02 09:23:47 crc kubenswrapper[4764]: I0202 09:23:47.595460 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xztz7\" (UniqueName: \"kubernetes.io/projected/10c729f0-d54f-4b06-995b-ab35e4d583bb-kube-api-access-xztz7\") pod \"10c729f0-d54f-4b06-995b-ab35e4d583bb\" (UID: \"10c729f0-d54f-4b06-995b-ab35e4d583bb\") " Feb 02 09:23:47 crc kubenswrapper[4764]: I0202 09:23:47.595618 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/985e7a8f-2a38-4ee8-a541-f2c8b7ff9fde-config\") pod \"985e7a8f-2a38-4ee8-a541-f2c8b7ff9fde\" (UID: \"985e7a8f-2a38-4ee8-a541-f2c8b7ff9fde\") " Feb 02 09:23:47 crc kubenswrapper[4764]: I0202 09:23:47.595767 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/10c729f0-d54f-4b06-995b-ab35e4d583bb-catalog-content\") pod \"10c729f0-d54f-4b06-995b-ab35e4d583bb\" (UID: \"10c729f0-d54f-4b06-995b-ab35e4d583bb\") " Feb 02 09:23:47 crc kubenswrapper[4764]: I0202 09:23:47.599712 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67ff45466c-v6vn7" Feb 02 09:23:47 crc kubenswrapper[4764]: I0202 09:23:47.600648 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/10c729f0-d54f-4b06-995b-ab35e4d583bb-utilities" (OuterVolumeSpecName: "utilities") pod "10c729f0-d54f-4b06-995b-ab35e4d583bb" (UID: "10c729f0-d54f-4b06-995b-ab35e4d583bb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 09:23:47 crc kubenswrapper[4764]: I0202 09:23:47.602381 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/985e7a8f-2a38-4ee8-a541-f2c8b7ff9fde-config" (OuterVolumeSpecName: "config") pod "985e7a8f-2a38-4ee8-a541-f2c8b7ff9fde" (UID: "985e7a8f-2a38-4ee8-a541-f2c8b7ff9fde"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:23:47 crc kubenswrapper[4764]: I0202 09:23:47.625573 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/985e7a8f-2a38-4ee8-a541-f2c8b7ff9fde-kube-api-access-m62f5" (OuterVolumeSpecName: "kube-api-access-m62f5") pod "985e7a8f-2a38-4ee8-a541-f2c8b7ff9fde" (UID: "985e7a8f-2a38-4ee8-a541-f2c8b7ff9fde"). InnerVolumeSpecName "kube-api-access-m62f5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:23:47 crc kubenswrapper[4764]: I0202 09:23:47.625715 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/10c729f0-d54f-4b06-995b-ab35e4d583bb-kube-api-access-xztz7" (OuterVolumeSpecName: "kube-api-access-xztz7") pod "10c729f0-d54f-4b06-995b-ab35e4d583bb" (UID: "10c729f0-d54f-4b06-995b-ab35e4d583bb"). InnerVolumeSpecName "kube-api-access-xztz7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:23:47 crc kubenswrapper[4764]: I0202 09:23:47.625759 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6fcf94d689-4qtb8" Feb 02 09:23:47 crc kubenswrapper[4764]: I0202 09:23:47.662728 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/10c729f0-d54f-4b06-995b-ab35e4d583bb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "10c729f0-d54f-4b06-995b-ab35e4d583bb" (UID: "10c729f0-d54f-4b06-995b-ab35e4d583bb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 09:23:47 crc kubenswrapper[4764]: I0202 09:23:47.698646 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-54w4t\" (UniqueName: \"kubernetes.io/projected/210b0439-a7e6-4f9f-b847-0a04cf044f87-kube-api-access-54w4t\") pod \"210b0439-a7e6-4f9f-b847-0a04cf044f87\" (UID: \"210b0439-a7e6-4f9f-b847-0a04cf044f87\") " Feb 02 09:23:47 crc kubenswrapper[4764]: I0202 09:23:47.699144 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/58c314a7-0090-49dd-afbd-88826c8733b0-config\") pod \"58c314a7-0090-49dd-afbd-88826c8733b0\" (UID: \"58c314a7-0090-49dd-afbd-88826c8733b0\") " Feb 02 09:23:47 crc kubenswrapper[4764]: I0202 09:23:47.699213 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rkqmf\" (UniqueName: \"kubernetes.io/projected/58c314a7-0090-49dd-afbd-88826c8733b0-kube-api-access-rkqmf\") pod \"58c314a7-0090-49dd-afbd-88826c8733b0\" (UID: \"58c314a7-0090-49dd-afbd-88826c8733b0\") " Feb 02 09:23:47 crc kubenswrapper[4764]: I0202 09:23:47.699259 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210b0439-a7e6-4f9f-b847-0a04cf044f87-config\") pod \"210b0439-a7e6-4f9f-b847-0a04cf044f87\" (UID: \"210b0439-a7e6-4f9f-b847-0a04cf044f87\") " Feb 02 09:23:47 crc kubenswrapper[4764]: I0202 09:23:47.699384 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/210b0439-a7e6-4f9f-b847-0a04cf044f87-dns-svc\") pod \"210b0439-a7e6-4f9f-b847-0a04cf044f87\" (UID: \"210b0439-a7e6-4f9f-b847-0a04cf044f87\") " Feb 02 09:23:47 crc kubenswrapper[4764]: I0202 09:23:47.699617 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/58c314a7-0090-49dd-afbd-88826c8733b0-dns-svc\") pod \"58c314a7-0090-49dd-afbd-88826c8733b0\" (UID: \"58c314a7-0090-49dd-afbd-88826c8733b0\") " Feb 02 09:23:47 crc kubenswrapper[4764]: I0202 09:23:47.699711 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/58c314a7-0090-49dd-afbd-88826c8733b0-config" (OuterVolumeSpecName: "config") pod "58c314a7-0090-49dd-afbd-88826c8733b0" (UID: "58c314a7-0090-49dd-afbd-88826c8733b0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:23:47 crc kubenswrapper[4764]: I0202 09:23:47.700070 4764 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/10c729f0-d54f-4b06-995b-ab35e4d583bb-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 09:23:47 crc kubenswrapper[4764]: I0202 09:23:47.700088 4764 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/58c314a7-0090-49dd-afbd-88826c8733b0-config\") on node \"crc\" DevicePath \"\"" Feb 02 09:23:47 crc kubenswrapper[4764]: I0202 09:23:47.700100 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m62f5\" (UniqueName: \"kubernetes.io/projected/985e7a8f-2a38-4ee8-a541-f2c8b7ff9fde-kube-api-access-m62f5\") on node \"crc\" DevicePath \"\"" Feb 02 09:23:47 crc kubenswrapper[4764]: I0202 09:23:47.700113 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xztz7\" (UniqueName: \"kubernetes.io/projected/10c729f0-d54f-4b06-995b-ab35e4d583bb-kube-api-access-xztz7\") on node \"crc\" DevicePath \"\"" Feb 02 09:23:47 crc kubenswrapper[4764]: I0202 09:23:47.700124 4764 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/985e7a8f-2a38-4ee8-a541-f2c8b7ff9fde-config\") on node \"crc\" DevicePath \"\"" Feb 02 09:23:47 crc kubenswrapper[4764]: I0202 09:23:47.700137 4764 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/10c729f0-d54f-4b06-995b-ab35e4d583bb-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 09:23:47 crc kubenswrapper[4764]: I0202 09:23:47.700168 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210b0439-a7e6-4f9f-b847-0a04cf044f87-config" (OuterVolumeSpecName: "config") pod "210b0439-a7e6-4f9f-b847-0a04cf044f87" (UID: "210b0439-a7e6-4f9f-b847-0a04cf044f87"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:23:47 crc kubenswrapper[4764]: I0202 09:23:47.700706 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210b0439-a7e6-4f9f-b847-0a04cf044f87-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "210b0439-a7e6-4f9f-b847-0a04cf044f87" (UID: "210b0439-a7e6-4f9f-b847-0a04cf044f87"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:23:47 crc kubenswrapper[4764]: I0202 09:23:47.702012 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/58c314a7-0090-49dd-afbd-88826c8733b0-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "58c314a7-0090-49dd-afbd-88826c8733b0" (UID: "58c314a7-0090-49dd-afbd-88826c8733b0"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:23:47 crc kubenswrapper[4764]: I0202 09:23:47.707287 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/58c314a7-0090-49dd-afbd-88826c8733b0-kube-api-access-rkqmf" (OuterVolumeSpecName: "kube-api-access-rkqmf") pod "58c314a7-0090-49dd-afbd-88826c8733b0" (UID: "58c314a7-0090-49dd-afbd-88826c8733b0"). InnerVolumeSpecName "kube-api-access-rkqmf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:23:47 crc kubenswrapper[4764]: I0202 09:23:47.710475 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210b0439-a7e6-4f9f-b847-0a04cf044f87-kube-api-access-54w4t" (OuterVolumeSpecName: "kube-api-access-54w4t") pod "210b0439-a7e6-4f9f-b847-0a04cf044f87" (UID: "210b0439-a7e6-4f9f-b847-0a04cf044f87"). InnerVolumeSpecName "kube-api-access-54w4t". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:23:47 crc kubenswrapper[4764]: I0202 09:23:47.801668 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rkqmf\" (UniqueName: \"kubernetes.io/projected/58c314a7-0090-49dd-afbd-88826c8733b0-kube-api-access-rkqmf\") on node \"crc\" DevicePath \"\"" Feb 02 09:23:47 crc kubenswrapper[4764]: I0202 09:23:47.802038 4764 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210b0439-a7e6-4f9f-b847-0a04cf044f87-config\") on node \"crc\" DevicePath \"\"" Feb 02 09:23:47 crc kubenswrapper[4764]: I0202 09:23:47.802049 4764 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/210b0439-a7e6-4f9f-b847-0a04cf044f87-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 02 09:23:47 crc kubenswrapper[4764]: I0202 09:23:47.802058 4764 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/58c314a7-0090-49dd-afbd-88826c8733b0-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 02 09:23:47 crc kubenswrapper[4764]: I0202 09:23:47.802066 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-54w4t\" (UniqueName: \"kubernetes.io/projected/210b0439-a7e6-4f9f-b847-0a04cf044f87-kube-api-access-54w4t\") on node \"crc\" DevicePath \"\"" Feb 02 09:23:47 crc kubenswrapper[4764]: I0202 09:23:47.996878 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-kgc8j"] Feb 02 09:23:48 crc kubenswrapper[4764]: I0202 09:23:48.004092 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-64f7f48db9-rz8h7"] Feb 02 09:23:48 crc kubenswrapper[4764]: W0202 09:23:48.008086 4764 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod456b6ada_1b28_4df6_b74b_c470910598e5.slice/crio-a2bfab2555ecacd859966ea06c0813ed79e2d4a2de6dbbf04bc852e597d9366e WatchSource:0}: Error finding container a2bfab2555ecacd859966ea06c0813ed79e2d4a2de6dbbf04bc852e597d9366e: Status 404 returned error can't find the container with id a2bfab2555ecacd859966ea06c0813ed79e2d4a2de6dbbf04bc852e597d9366e Feb 02 09:23:48 crc kubenswrapper[4764]: W0202 09:23:48.010363 4764 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb3becd85_0167_4ffe_802d_67988c486908.slice/crio-c9298bfb323d8f76e66d73510ea6d2e836bd6f7584307ed9b04a1ec480174f1a WatchSource:0}: Error finding container c9298bfb323d8f76e66d73510ea6d2e836bd6f7584307ed9b04a1ec480174f1a: Status 404 returned error can't find the container with id c9298bfb323d8f76e66d73510ea6d2e836bd6f7584307ed9b04a1ec480174f1a Feb 02 09:23:48 crc kubenswrapper[4764]: I0202 09:23:48.359352 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lx9rg" event={"ID":"10c729f0-d54f-4b06-995b-ab35e4d583bb","Type":"ContainerDied","Data":"44ce85ed6b7ce24f7e9e665a4856d6eb03422a9ddeb88c2f6d4222de281cf391"} Feb 02 09:23:48 crc kubenswrapper[4764]: I0202 09:23:48.359462 4764 scope.go:117] "RemoveContainer" containerID="98a0ac0c06018d32adfbe9f8e8984832df81fc12c656ce5683faf6d86cd1d986" Feb 02 09:23:48 crc kubenswrapper[4764]: I0202 09:23:48.359504 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lx9rg" Feb 02 09:23:48 crc kubenswrapper[4764]: I0202 09:23:48.368998 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6fcf94d689-4qtb8" Feb 02 09:23:48 crc kubenswrapper[4764]: I0202 09:23:48.369005 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6fcf94d689-4qtb8" event={"ID":"58c314a7-0090-49dd-afbd-88826c8733b0","Type":"ContainerDied","Data":"3685d9d71b7ceae49d109efc666e0e6c4a7e2c69318c9c0e08c1dc333747764e"} Feb 02 09:23:48 crc kubenswrapper[4764]: I0202 09:23:48.376272 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"a9177511-b834-4e68-9f6f-8b1b4470b274","Type":"ContainerStarted","Data":"94a6fc0b53a82ff5c22cf77f3512697e901d154e1d1d5ab233c5c7afa2a9b98a"} Feb 02 09:23:48 crc kubenswrapper[4764]: I0202 09:23:48.404414 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64f7f48db9-rz8h7" event={"ID":"456b6ada-1b28-4df6-b74b-c470910598e5","Type":"ContainerStarted","Data":"a2bfab2555ecacd859966ea06c0813ed79e2d4a2de6dbbf04bc852e597d9366e"} Feb 02 09:23:48 crc kubenswrapper[4764]: I0202 09:23:48.405819 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-kgc8j" event={"ID":"b3becd85-0167-4ffe-802d-67988c486908","Type":"ContainerStarted","Data":"c9298bfb323d8f76e66d73510ea6d2e836bd6f7584307ed9b04a1ec480174f1a"} Feb 02 09:23:48 crc kubenswrapper[4764]: I0202 09:23:48.409057 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"89c39976-ff91-4c6b-b369-5cba94e02db8","Type":"ContainerStarted","Data":"eeb333a4ff527304593596374d7c7c0b165b0912ebc0a8f7d8d0a2af3d733a8d"} Feb 02 09:23:48 crc kubenswrapper[4764]: I0202 09:23:48.412423 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-855cbc58c5-zcszj" event={"ID":"985e7a8f-2a38-4ee8-a541-f2c8b7ff9fde","Type":"ContainerDied","Data":"e707d1b26af371fcf8f95e95d6de95b9112fbf47773ac827fb844ae58376f42f"} Feb 02 09:23:48 crc kubenswrapper[4764]: I0202 09:23:48.412514 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-855cbc58c5-zcszj" Feb 02 09:23:48 crc kubenswrapper[4764]: I0202 09:23:48.417872 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-lx9rg"] Feb 02 09:23:48 crc kubenswrapper[4764]: I0202 09:23:48.426211 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67ff45466c-v6vn7" Feb 02 09:23:48 crc kubenswrapper[4764]: I0202 09:23:48.426837 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67ff45466c-v6vn7" event={"ID":"210b0439-a7e6-4f9f-b847-0a04cf044f87","Type":"ContainerDied","Data":"e9b82bb8d627c82badf1dccbdb307ef887aa920c8713078506a96369375f3e21"} Feb 02 09:23:48 crc kubenswrapper[4764]: I0202 09:23:48.430087 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-lx9rg"] Feb 02 09:23:48 crc kubenswrapper[4764]: E0202 09:23:48.431242 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-state-metrics\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.k8s.io/kube-state-metrics/kube-state-metrics@sha256:db384bf43222b066c378e77027a675d4cd9911107adba46c2922b3a55e10d6fb\\\"\"" pod="openstack/kube-state-metrics-0" podUID="26d7a0b3-ea1e-417b-b417-69ae84bd5d24" Feb 02 09:23:48 crc kubenswrapper[4764]: I0202 09:23:48.485028 4764 scope.go:117] "RemoveContainer" containerID="ccb5dd565532e6e4b7d293886f44bad5362f5b8c79fc18074b57de20eb5cbdf0" Feb 02 09:23:48 crc kubenswrapper[4764]: I0202 09:23:48.507210 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6fcf94d689-4qtb8"] Feb 02 09:23:48 crc kubenswrapper[4764]: I0202 09:23:48.514243 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6fcf94d689-4qtb8"] Feb 02 09:23:48 crc kubenswrapper[4764]: I0202 09:23:48.563683 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-67ff45466c-v6vn7"] Feb 02 09:23:48 crc kubenswrapper[4764]: I0202 09:23:48.569070 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-67ff45466c-v6vn7"] Feb 02 09:23:48 crc kubenswrapper[4764]: I0202 09:23:48.571878 4764 scope.go:117] "RemoveContainer" containerID="16b97d45425301e0814fee1ea9af4f1211a46ca037f1c773b9c55603eb0a3de5" Feb 02 09:23:48 crc kubenswrapper[4764]: I0202 09:23:48.619735 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-855cbc58c5-zcszj"] Feb 02 09:23:48 crc kubenswrapper[4764]: I0202 09:23:48.627587 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-855cbc58c5-zcszj"] Feb 02 09:23:49 crc kubenswrapper[4764]: I0202 09:23:49.841592 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="10c729f0-d54f-4b06-995b-ab35e4d583bb" path="/var/lib/kubelet/pods/10c729f0-d54f-4b06-995b-ab35e4d583bb/volumes" Feb 02 09:23:49 crc kubenswrapper[4764]: I0202 09:23:49.842571 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210b0439-a7e6-4f9f-b847-0a04cf044f87" path="/var/lib/kubelet/pods/210b0439-a7e6-4f9f-b847-0a04cf044f87/volumes" Feb 02 09:23:49 crc kubenswrapper[4764]: I0202 09:23:49.842984 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="58c314a7-0090-49dd-afbd-88826c8733b0" path="/var/lib/kubelet/pods/58c314a7-0090-49dd-afbd-88826c8733b0/volumes" Feb 02 09:23:49 crc kubenswrapper[4764]: I0202 09:23:49.843372 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="985e7a8f-2a38-4ee8-a541-f2c8b7ff9fde" path="/var/lib/kubelet/pods/985e7a8f-2a38-4ee8-a541-f2c8b7ff9fde/volumes" Feb 02 09:23:54 crc kubenswrapper[4764]: I0202 09:23:54.498272 4764 generic.go:334] "Generic (PLEG): container finished" podID="a9177511-b834-4e68-9f6f-8b1b4470b274" containerID="94a6fc0b53a82ff5c22cf77f3512697e901d154e1d1d5ab233c5c7afa2a9b98a" exitCode=0 Feb 02 09:23:54 crc kubenswrapper[4764]: I0202 09:23:54.498722 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"a9177511-b834-4e68-9f6f-8b1b4470b274","Type":"ContainerDied","Data":"94a6fc0b53a82ff5c22cf77f3512697e901d154e1d1d5ab233c5c7afa2a9b98a"} Feb 02 09:23:55 crc kubenswrapper[4764]: I0202 09:23:55.509680 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"a9177511-b834-4e68-9f6f-8b1b4470b274","Type":"ContainerStarted","Data":"f42efb48ac8c5194111ccedb91ad0851fbb9313ed5b115f8c2ce58f0b3ab42c1"} Feb 02 09:23:55 crc kubenswrapper[4764]: I0202 09:23:55.512470 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"cba91a53-7d3d-481b-bd55-3f0bbe7ccbb6","Type":"ContainerStarted","Data":"fdea530a3208f5180926054084586e8d52208415ac582e4ea0ed73a148304808"} Feb 02 09:23:55 crc kubenswrapper[4764]: I0202 09:23:55.512501 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"cba91a53-7d3d-481b-bd55-3f0bbe7ccbb6","Type":"ContainerStarted","Data":"7954cd7ebbe827d249249be681fdc9717e53fa75aab3bc83d4023116fb889469"} Feb 02 09:23:55 crc kubenswrapper[4764]: I0202 09:23:55.524922 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"cd76e53a-13fc-4293-b5ca-2324224850cc","Type":"ContainerStarted","Data":"0eb7fa6b0a934d287b7b5ff4fbeb398648b73cc8d8c3f41eeb4562d9dbe1ad26"} Feb 02 09:23:55 crc kubenswrapper[4764]: I0202 09:23:55.526292 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Feb 02 09:23:55 crc kubenswrapper[4764]: I0202 09:23:55.529430 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-6v7f6" event={"ID":"66662206-d732-404d-a81d-6303d9947d60","Type":"ContainerStarted","Data":"442ee247649b155234a189e9abdd1b9bfc36d7ae3e6d0741aa263ec83311c3a0"} Feb 02 09:23:55 crc kubenswrapper[4764]: I0202 09:23:55.530299 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-6v7f6" Feb 02 09:23:55 crc kubenswrapper[4764]: I0202 09:23:55.531421 4764 generic.go:334] "Generic (PLEG): container finished" podID="456b6ada-1b28-4df6-b74b-c470910598e5" containerID="73e4bf881e097e1b2174d65b867ec1e41ed8159785d4fd89e35ecf81136dc73e" exitCode=0 Feb 02 09:23:55 crc kubenswrapper[4764]: I0202 09:23:55.531593 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64f7f48db9-rz8h7" event={"ID":"456b6ada-1b28-4df6-b74b-c470910598e5","Type":"ContainerDied","Data":"73e4bf881e097e1b2174d65b867ec1e41ed8159785d4fd89e35ecf81136dc73e"} Feb 02 09:23:55 crc kubenswrapper[4764]: I0202 09:23:55.535764 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"03d9f090-4a93-4226-8fce-6fd0a425d929","Type":"ContainerStarted","Data":"d08998db15d7c2f061c9aa78b783e5e53996e688590d41367d09011c4224893d"} Feb 02 09:23:55 crc kubenswrapper[4764]: I0202 09:23:55.539358 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"89c39976-ff91-4c6b-b369-5cba94e02db8","Type":"ContainerStarted","Data":"6e0c3075ce91af4501b26db7645bb82701621ffccb77db61090a2bfb221f44d0"} Feb 02 09:23:55 crc kubenswrapper[4764]: I0202 09:23:55.539396 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"89c39976-ff91-4c6b-b369-5cba94e02db8","Type":"ContainerStarted","Data":"dbba54fe4596ccf577573dc78bc02d6d69478a505844a6c1ade64fbc71b1d877"} Feb 02 09:23:55 crc kubenswrapper[4764]: I0202 09:23:55.541765 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-kgc8j" event={"ID":"b3becd85-0167-4ffe-802d-67988c486908","Type":"ContainerStarted","Data":"136042a5ad9ede5507e8434f2c8ed17e72d9e236b1aa48ab7a762d6c34a4ec76"} Feb 02 09:23:55 crc kubenswrapper[4764]: I0202 09:23:55.559413 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=12.628211546 podStartE2EDuration="40.559390047s" podCreationTimestamp="2026-02-02 09:23:15 +0000 UTC" firstStartedPulling="2026-02-02 09:23:18.26442739 +0000 UTC m=+961.198151478" lastFinishedPulling="2026-02-02 09:23:46.195605891 +0000 UTC m=+989.129329979" observedRunningTime="2026-02-02 09:23:55.551615762 +0000 UTC m=+998.485339850" watchObservedRunningTime="2026-02-02 09:23:55.559390047 +0000 UTC m=+998.493114135" Feb 02 09:23:55 crc kubenswrapper[4764]: I0202 09:23:55.561580 4764 generic.go:334] "Generic (PLEG): container finished" podID="29a433e1-2651-43cd-99d1-95694063c980" containerID="274abb42626875b4d918969418c0ccccfafff53fb4c5dc1537789493bf1337e0" exitCode=0 Feb 02 09:23:55 crc kubenswrapper[4764]: I0202 09:23:55.561630 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-2p2bk" event={"ID":"29a433e1-2651-43cd-99d1-95694063c980","Type":"ContainerDied","Data":"274abb42626875b4d918969418c0ccccfafff53fb4c5dc1537789493bf1337e0"} Feb 02 09:23:55 crc kubenswrapper[4764]: I0202 09:23:55.563463 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"1d6b6c31-819e-4098-b11b-6073cc751cad","Type":"ContainerStarted","Data":"f4c8ff42b010f8abeb59946f8d2ed057ea1d2e9e6a5b9452138169d2101d23e7"} Feb 02 09:23:55 crc kubenswrapper[4764]: I0202 09:23:55.600963 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=23.719613029 podStartE2EDuration="30.600923525s" podCreationTimestamp="2026-02-02 09:23:25 +0000 UTC" firstStartedPulling="2026-02-02 09:23:47.430407186 +0000 UTC m=+990.364131274" lastFinishedPulling="2026-02-02 09:23:54.311717682 +0000 UTC m=+997.245441770" observedRunningTime="2026-02-02 09:23:55.599789085 +0000 UTC m=+998.533513173" watchObservedRunningTime="2026-02-02 09:23:55.600923525 +0000 UTC m=+998.534647603" Feb 02 09:23:55 crc kubenswrapper[4764]: I0202 09:23:55.639484 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=20.306372389 podStartE2EDuration="34.639459553s" podCreationTimestamp="2026-02-02 09:23:21 +0000 UTC" firstStartedPulling="2026-02-02 09:23:39.89809131 +0000 UTC m=+982.831815398" lastFinishedPulling="2026-02-02 09:23:54.231178464 +0000 UTC m=+997.164902562" observedRunningTime="2026-02-02 09:23:55.626466969 +0000 UTC m=+998.560191077" watchObservedRunningTime="2026-02-02 09:23:55.639459553 +0000 UTC m=+998.573183641" Feb 02 09:23:55 crc kubenswrapper[4764]: I0202 09:23:55.646881 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=2.507390422 podStartE2EDuration="38.646862918s" podCreationTimestamp="2026-02-02 09:23:17 +0000 UTC" firstStartedPulling="2026-02-02 09:23:18.174135386 +0000 UTC m=+961.107859474" lastFinishedPulling="2026-02-02 09:23:54.313607882 +0000 UTC m=+997.247331970" observedRunningTime="2026-02-02 09:23:55.64390618 +0000 UTC m=+998.577630268" watchObservedRunningTime="2026-02-02 09:23:55.646862918 +0000 UTC m=+998.580587006" Feb 02 09:23:55 crc kubenswrapper[4764]: I0202 09:23:55.762496 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-kgc8j" podStartSLOduration=14.424205544 podStartE2EDuration="20.762475983s" podCreationTimestamp="2026-02-02 09:23:35 +0000 UTC" firstStartedPulling="2026-02-02 09:23:48.015052603 +0000 UTC m=+990.948776681" lastFinishedPulling="2026-02-02 09:23:54.353323032 +0000 UTC m=+997.287047120" observedRunningTime="2026-02-02 09:23:55.708365313 +0000 UTC m=+998.642089401" watchObservedRunningTime="2026-02-02 09:23:55.762475983 +0000 UTC m=+998.696200071" Feb 02 09:23:55 crc kubenswrapper[4764]: I0202 09:23:55.812744 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-6v7f6" podStartSLOduration=24.537853642 podStartE2EDuration="33.812726881s" podCreationTimestamp="2026-02-02 09:23:22 +0000 UTC" firstStartedPulling="2026-02-02 09:23:45.036864764 +0000 UTC m=+987.970588862" lastFinishedPulling="2026-02-02 09:23:54.311738013 +0000 UTC m=+997.245462101" observedRunningTime="2026-02-02 09:23:55.808258213 +0000 UTC m=+998.741982301" watchObservedRunningTime="2026-02-02 09:23:55.812726881 +0000 UTC m=+998.746450969" Feb 02 09:23:56 crc kubenswrapper[4764]: I0202 09:23:56.033562 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-f54874ffc-wmsqt"] Feb 02 09:23:56 crc kubenswrapper[4764]: I0202 09:23:56.083967 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-56df986d9c-vkqnn"] Feb 02 09:23:56 crc kubenswrapper[4764]: E0202 09:23:56.084247 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10c729f0-d54f-4b06-995b-ab35e4d583bb" containerName="extract-utilities" Feb 02 09:23:56 crc kubenswrapper[4764]: I0202 09:23:56.084258 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="10c729f0-d54f-4b06-995b-ab35e4d583bb" containerName="extract-utilities" Feb 02 09:23:56 crc kubenswrapper[4764]: E0202 09:23:56.084298 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10c729f0-d54f-4b06-995b-ab35e4d583bb" containerName="registry-server" Feb 02 09:23:56 crc kubenswrapper[4764]: I0202 09:23:56.084304 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="10c729f0-d54f-4b06-995b-ab35e4d583bb" containerName="registry-server" Feb 02 09:23:56 crc kubenswrapper[4764]: E0202 09:23:56.084318 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10c729f0-d54f-4b06-995b-ab35e4d583bb" containerName="extract-content" Feb 02 09:23:56 crc kubenswrapper[4764]: I0202 09:23:56.084326 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="10c729f0-d54f-4b06-995b-ab35e4d583bb" containerName="extract-content" Feb 02 09:23:56 crc kubenswrapper[4764]: I0202 09:23:56.084471 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="10c729f0-d54f-4b06-995b-ab35e4d583bb" containerName="registry-server" Feb 02 09:23:56 crc kubenswrapper[4764]: I0202 09:23:56.085169 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df986d9c-vkqnn" Feb 02 09:23:56 crc kubenswrapper[4764]: I0202 09:23:56.089543 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Feb 02 09:23:56 crc kubenswrapper[4764]: I0202 09:23:56.103262 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/22c2fdff-3f57-4fc3-bbf4-f459784fa876-ovsdbserver-nb\") pod \"dnsmasq-dns-56df986d9c-vkqnn\" (UID: \"22c2fdff-3f57-4fc3-bbf4-f459784fa876\") " pod="openstack/dnsmasq-dns-56df986d9c-vkqnn" Feb 02 09:23:56 crc kubenswrapper[4764]: I0202 09:23:56.103298 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c2fdff-3f57-4fc3-bbf4-f459784fa876-config\") pod \"dnsmasq-dns-56df986d9c-vkqnn\" (UID: \"22c2fdff-3f57-4fc3-bbf4-f459784fa876\") " pod="openstack/dnsmasq-dns-56df986d9c-vkqnn" Feb 02 09:23:56 crc kubenswrapper[4764]: I0202 09:23:56.103326 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/22c2fdff-3f57-4fc3-bbf4-f459784fa876-ovsdbserver-sb\") pod \"dnsmasq-dns-56df986d9c-vkqnn\" (UID: \"22c2fdff-3f57-4fc3-bbf4-f459784fa876\") " pod="openstack/dnsmasq-dns-56df986d9c-vkqnn" Feb 02 09:23:56 crc kubenswrapper[4764]: I0202 09:23:56.103356 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/22c2fdff-3f57-4fc3-bbf4-f459784fa876-dns-svc\") pod \"dnsmasq-dns-56df986d9c-vkqnn\" (UID: \"22c2fdff-3f57-4fc3-bbf4-f459784fa876\") " pod="openstack/dnsmasq-dns-56df986d9c-vkqnn" Feb 02 09:23:56 crc kubenswrapper[4764]: I0202 09:23:56.103401 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gzgzn\" (UniqueName: \"kubernetes.io/projected/22c2fdff-3f57-4fc3-bbf4-f459784fa876-kube-api-access-gzgzn\") pod \"dnsmasq-dns-56df986d9c-vkqnn\" (UID: \"22c2fdff-3f57-4fc3-bbf4-f459784fa876\") " pod="openstack/dnsmasq-dns-56df986d9c-vkqnn" Feb 02 09:23:56 crc kubenswrapper[4764]: I0202 09:23:56.142573 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-56df986d9c-vkqnn"] Feb 02 09:23:56 crc kubenswrapper[4764]: I0202 09:23:56.204406 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/22c2fdff-3f57-4fc3-bbf4-f459784fa876-ovsdbserver-sb\") pod \"dnsmasq-dns-56df986d9c-vkqnn\" (UID: \"22c2fdff-3f57-4fc3-bbf4-f459784fa876\") " pod="openstack/dnsmasq-dns-56df986d9c-vkqnn" Feb 02 09:23:56 crc kubenswrapper[4764]: I0202 09:23:56.204459 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/22c2fdff-3f57-4fc3-bbf4-f459784fa876-dns-svc\") pod \"dnsmasq-dns-56df986d9c-vkqnn\" (UID: \"22c2fdff-3f57-4fc3-bbf4-f459784fa876\") " pod="openstack/dnsmasq-dns-56df986d9c-vkqnn" Feb 02 09:23:56 crc kubenswrapper[4764]: I0202 09:23:56.204504 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gzgzn\" (UniqueName: \"kubernetes.io/projected/22c2fdff-3f57-4fc3-bbf4-f459784fa876-kube-api-access-gzgzn\") pod \"dnsmasq-dns-56df986d9c-vkqnn\" (UID: \"22c2fdff-3f57-4fc3-bbf4-f459784fa876\") " pod="openstack/dnsmasq-dns-56df986d9c-vkqnn" Feb 02 09:23:56 crc kubenswrapper[4764]: I0202 09:23:56.204560 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/22c2fdff-3f57-4fc3-bbf4-f459784fa876-ovsdbserver-nb\") pod \"dnsmasq-dns-56df986d9c-vkqnn\" (UID: \"22c2fdff-3f57-4fc3-bbf4-f459784fa876\") " pod="openstack/dnsmasq-dns-56df986d9c-vkqnn" Feb 02 09:23:56 crc kubenswrapper[4764]: I0202 09:23:56.204582 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c2fdff-3f57-4fc3-bbf4-f459784fa876-config\") pod \"dnsmasq-dns-56df986d9c-vkqnn\" (UID: \"22c2fdff-3f57-4fc3-bbf4-f459784fa876\") " pod="openstack/dnsmasq-dns-56df986d9c-vkqnn" Feb 02 09:23:56 crc kubenswrapper[4764]: I0202 09:23:56.205899 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c2fdff-3f57-4fc3-bbf4-f459784fa876-config\") pod \"dnsmasq-dns-56df986d9c-vkqnn\" (UID: \"22c2fdff-3f57-4fc3-bbf4-f459784fa876\") " pod="openstack/dnsmasq-dns-56df986d9c-vkqnn" Feb 02 09:23:56 crc kubenswrapper[4764]: I0202 09:23:56.206419 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/22c2fdff-3f57-4fc3-bbf4-f459784fa876-ovsdbserver-sb\") pod \"dnsmasq-dns-56df986d9c-vkqnn\" (UID: \"22c2fdff-3f57-4fc3-bbf4-f459784fa876\") " pod="openstack/dnsmasq-dns-56df986d9c-vkqnn" Feb 02 09:23:56 crc kubenswrapper[4764]: I0202 09:23:56.206887 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/22c2fdff-3f57-4fc3-bbf4-f459784fa876-dns-svc\") pod \"dnsmasq-dns-56df986d9c-vkqnn\" (UID: \"22c2fdff-3f57-4fc3-bbf4-f459784fa876\") " pod="openstack/dnsmasq-dns-56df986d9c-vkqnn" Feb 02 09:23:56 crc kubenswrapper[4764]: I0202 09:23:56.207744 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/22c2fdff-3f57-4fc3-bbf4-f459784fa876-ovsdbserver-nb\") pod \"dnsmasq-dns-56df986d9c-vkqnn\" (UID: \"22c2fdff-3f57-4fc3-bbf4-f459784fa876\") " pod="openstack/dnsmasq-dns-56df986d9c-vkqnn" Feb 02 09:23:56 crc kubenswrapper[4764]: I0202 09:23:56.232483 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gzgzn\" (UniqueName: \"kubernetes.io/projected/22c2fdff-3f57-4fc3-bbf4-f459784fa876-kube-api-access-gzgzn\") pod \"dnsmasq-dns-56df986d9c-vkqnn\" (UID: \"22c2fdff-3f57-4fc3-bbf4-f459784fa876\") " pod="openstack/dnsmasq-dns-56df986d9c-vkqnn" Feb 02 09:23:56 crc kubenswrapper[4764]: I0202 09:23:56.259870 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Feb 02 09:23:56 crc kubenswrapper[4764]: I0202 09:23:56.403457 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df986d9c-vkqnn" Feb 02 09:23:56 crc kubenswrapper[4764]: I0202 09:23:56.509469 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f54874ffc-wmsqt" Feb 02 09:23:56 crc kubenswrapper[4764]: I0202 09:23:56.572603 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"53d5d300-11ae-408d-8bb8-347defb93991","Type":"ContainerStarted","Data":"f5641821f6b2d966369e040f6849f892c7244535560aa2d911c4599545382008"} Feb 02 09:23:56 crc kubenswrapper[4764]: I0202 09:23:56.576742 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f54874ffc-wmsqt" event={"ID":"85525aa3-d70a-4d66-8bbf-536be10a0337","Type":"ContainerDied","Data":"574d06adc1e687e5f47f6871cdb34554c9d7875b54529338e583a5b3f80f5d33"} Feb 02 09:23:56 crc kubenswrapper[4764]: I0202 09:23:56.576830 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f54874ffc-wmsqt" Feb 02 09:23:56 crc kubenswrapper[4764]: I0202 09:23:56.580035 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64f7f48db9-rz8h7" event={"ID":"456b6ada-1b28-4df6-b74b-c470910598e5","Type":"ContainerStarted","Data":"81ca5aae341ed7dd3dd43c8d9df176155909c2161d3978da1edd39d23566c0e1"} Feb 02 09:23:56 crc kubenswrapper[4764]: I0202 09:23:56.580159 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-64f7f48db9-rz8h7" Feb 02 09:23:56 crc kubenswrapper[4764]: I0202 09:23:56.589045 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-2p2bk" event={"ID":"29a433e1-2651-43cd-99d1-95694063c980","Type":"ContainerStarted","Data":"9b0a09e76f920bbc8ef2e5ed75c813af8449ef431d3b30608ae74aa2b5aa0d8f"} Feb 02 09:23:56 crc kubenswrapper[4764]: I0202 09:23:56.589076 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-2p2bk" Feb 02 09:23:56 crc kubenswrapper[4764]: I0202 09:23:56.589086 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-2p2bk" event={"ID":"29a433e1-2651-43cd-99d1-95694063c980","Type":"ContainerStarted","Data":"3a0743a8543a4d097461681c8407c441a64c69c40352785f185e0ece7aedd3f5"} Feb 02 09:23:56 crc kubenswrapper[4764]: I0202 09:23:56.589773 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-2p2bk" Feb 02 09:23:56 crc kubenswrapper[4764]: I0202 09:23:56.682030 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-2p2bk" podStartSLOduration=19.255950948 podStartE2EDuration="33.682006469s" podCreationTimestamp="2026-02-02 09:23:23 +0000 UTC" firstStartedPulling="2026-02-02 09:23:39.889747379 +0000 UTC m=+982.823471467" lastFinishedPulling="2026-02-02 09:23:54.31580289 +0000 UTC m=+997.249526988" observedRunningTime="2026-02-02 09:23:56.673250868 +0000 UTC m=+999.606974956" watchObservedRunningTime="2026-02-02 09:23:56.682006469 +0000 UTC m=+999.615730557" Feb 02 09:23:56 crc kubenswrapper[4764]: I0202 09:23:56.684314 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-64f7f48db9-rz8h7" podStartSLOduration=14.340253928 podStartE2EDuration="20.684298059s" podCreationTimestamp="2026-02-02 09:23:36 +0000 UTC" firstStartedPulling="2026-02-02 09:23:48.017037615 +0000 UTC m=+990.950761703" lastFinishedPulling="2026-02-02 09:23:54.361081746 +0000 UTC m=+997.294805834" observedRunningTime="2026-02-02 09:23:56.647956969 +0000 UTC m=+999.581681077" watchObservedRunningTime="2026-02-02 09:23:56.684298059 +0000 UTC m=+999.618022157" Feb 02 09:23:56 crc kubenswrapper[4764]: I0202 09:23:56.717657 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/85525aa3-d70a-4d66-8bbf-536be10a0337-dns-svc\") pod \"85525aa3-d70a-4d66-8bbf-536be10a0337\" (UID: \"85525aa3-d70a-4d66-8bbf-536be10a0337\") " Feb 02 09:23:56 crc kubenswrapper[4764]: I0202 09:23:56.717740 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jdr4z\" (UniqueName: \"kubernetes.io/projected/85525aa3-d70a-4d66-8bbf-536be10a0337-kube-api-access-jdr4z\") pod \"85525aa3-d70a-4d66-8bbf-536be10a0337\" (UID: \"85525aa3-d70a-4d66-8bbf-536be10a0337\") " Feb 02 09:23:56 crc kubenswrapper[4764]: I0202 09:23:56.717814 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/85525aa3-d70a-4d66-8bbf-536be10a0337-config\") pod \"85525aa3-d70a-4d66-8bbf-536be10a0337\" (UID: \"85525aa3-d70a-4d66-8bbf-536be10a0337\") " Feb 02 09:23:56 crc kubenswrapper[4764]: I0202 09:23:56.718126 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/85525aa3-d70a-4d66-8bbf-536be10a0337-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "85525aa3-d70a-4d66-8bbf-536be10a0337" (UID: "85525aa3-d70a-4d66-8bbf-536be10a0337"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:23:56 crc kubenswrapper[4764]: I0202 09:23:56.718683 4764 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/85525aa3-d70a-4d66-8bbf-536be10a0337-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 02 09:23:56 crc kubenswrapper[4764]: I0202 09:23:56.719148 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/85525aa3-d70a-4d66-8bbf-536be10a0337-config" (OuterVolumeSpecName: "config") pod "85525aa3-d70a-4d66-8bbf-536be10a0337" (UID: "85525aa3-d70a-4d66-8bbf-536be10a0337"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:23:56 crc kubenswrapper[4764]: I0202 09:23:56.722297 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/85525aa3-d70a-4d66-8bbf-536be10a0337-kube-api-access-jdr4z" (OuterVolumeSpecName: "kube-api-access-jdr4z") pod "85525aa3-d70a-4d66-8bbf-536be10a0337" (UID: "85525aa3-d70a-4d66-8bbf-536be10a0337"). InnerVolumeSpecName "kube-api-access-jdr4z". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:23:56 crc kubenswrapper[4764]: I0202 09:23:56.821469 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jdr4z\" (UniqueName: \"kubernetes.io/projected/85525aa3-d70a-4d66-8bbf-536be10a0337-kube-api-access-jdr4z\") on node \"crc\" DevicePath \"\"" Feb 02 09:23:56 crc kubenswrapper[4764]: I0202 09:23:56.821538 4764 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/85525aa3-d70a-4d66-8bbf-536be10a0337-config\") on node \"crc\" DevicePath \"\"" Feb 02 09:23:56 crc kubenswrapper[4764]: I0202 09:23:56.918629 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-f54874ffc-wmsqt"] Feb 02 09:23:56 crc kubenswrapper[4764]: I0202 09:23:56.926462 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Feb 02 09:23:56 crc kubenswrapper[4764]: I0202 09:23:56.926514 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Feb 02 09:23:56 crc kubenswrapper[4764]: I0202 09:23:56.939820 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-f54874ffc-wmsqt"] Feb 02 09:23:56 crc kubenswrapper[4764]: I0202 09:23:56.953047 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-56df986d9c-vkqnn"] Feb 02 09:23:56 crc kubenswrapper[4764]: W0202 09:23:56.953512 4764 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod22c2fdff_3f57_4fc3_bbf4_f459784fa876.slice/crio-1de7784d8dbb29eef5cbc116714bb69dee0f3a82be15a2859d2d9910b2822da4 WatchSource:0}: Error finding container 1de7784d8dbb29eef5cbc116714bb69dee0f3a82be15a2859d2d9910b2822da4: Status 404 returned error can't find the container with id 1de7784d8dbb29eef5cbc116714bb69dee0f3a82be15a2859d2d9910b2822da4 Feb 02 09:23:57 crc kubenswrapper[4764]: I0202 09:23:57.594728 4764 generic.go:334] "Generic (PLEG): container finished" podID="22c2fdff-3f57-4fc3-bbf4-f459784fa876" containerID="a92d9770b278125ed99e3f4e33511ef8b270f5a22578dabe0ff27e3c86793d5c" exitCode=0 Feb 02 09:23:57 crc kubenswrapper[4764]: I0202 09:23:57.594837 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df986d9c-vkqnn" event={"ID":"22c2fdff-3f57-4fc3-bbf4-f459784fa876","Type":"ContainerDied","Data":"a92d9770b278125ed99e3f4e33511ef8b270f5a22578dabe0ff27e3c86793d5c"} Feb 02 09:23:57 crc kubenswrapper[4764]: I0202 09:23:57.595717 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df986d9c-vkqnn" event={"ID":"22c2fdff-3f57-4fc3-bbf4-f459784fa876","Type":"ContainerStarted","Data":"1de7784d8dbb29eef5cbc116714bb69dee0f3a82be15a2859d2d9910b2822da4"} Feb 02 09:23:57 crc kubenswrapper[4764]: I0202 09:23:57.619353 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Feb 02 09:23:57 crc kubenswrapper[4764]: I0202 09:23:57.619386 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Feb 02 09:23:57 crc kubenswrapper[4764]: I0202 09:23:57.835668 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="85525aa3-d70a-4d66-8bbf-536be10a0337" path="/var/lib/kubelet/pods/85525aa3-d70a-4d66-8bbf-536be10a0337/volumes" Feb 02 09:23:58 crc kubenswrapper[4764]: I0202 09:23:58.260414 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Feb 02 09:23:58 crc kubenswrapper[4764]: I0202 09:23:58.604827 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df986d9c-vkqnn" event={"ID":"22c2fdff-3f57-4fc3-bbf4-f459784fa876","Type":"ContainerStarted","Data":"e4c97ac9d116f802f33a0f092f6268eaea37dd2a894d084f76cf34a1525465ec"} Feb 02 09:23:58 crc kubenswrapper[4764]: I0202 09:23:58.605315 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-56df986d9c-vkqnn" Feb 02 09:23:58 crc kubenswrapper[4764]: I0202 09:23:58.607009 4764 generic.go:334] "Generic (PLEG): container finished" podID="1d6b6c31-819e-4098-b11b-6073cc751cad" containerID="f4c8ff42b010f8abeb59946f8d2ed057ea1d2e9e6a5b9452138169d2101d23e7" exitCode=0 Feb 02 09:23:58 crc kubenswrapper[4764]: I0202 09:23:58.607093 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"1d6b6c31-819e-4098-b11b-6073cc751cad","Type":"ContainerDied","Data":"f4c8ff42b010f8abeb59946f8d2ed057ea1d2e9e6a5b9452138169d2101d23e7"} Feb 02 09:23:58 crc kubenswrapper[4764]: I0202 09:23:58.625636 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-56df986d9c-vkqnn" podStartSLOduration=2.625580342 podStartE2EDuration="2.625580342s" podCreationTimestamp="2026-02-02 09:23:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:23:58.62435412 +0000 UTC m=+1001.558078228" watchObservedRunningTime="2026-02-02 09:23:58.625580342 +0000 UTC m=+1001.559304430" Feb 02 09:23:59 crc kubenswrapper[4764]: I0202 09:23:59.301649 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Feb 02 09:23:59 crc kubenswrapper[4764]: I0202 09:23:59.615411 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"1d6b6c31-819e-4098-b11b-6073cc751cad","Type":"ContainerStarted","Data":"4314eae38acb44b64073c3bc40e655c227904d19437e3899976ebbf6804a236a"} Feb 02 09:23:59 crc kubenswrapper[4764]: I0202 09:23:59.650418 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=-9223371991.204391 podStartE2EDuration="45.650384738s" podCreationTimestamp="2026-02-02 09:23:14 +0000 UTC" firstStartedPulling="2026-02-02 09:23:17.082980281 +0000 UTC m=+960.016704369" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:23:59.636570493 +0000 UTC m=+1002.570294621" watchObservedRunningTime="2026-02-02 09:23:59.650384738 +0000 UTC m=+1002.584108866" Feb 02 09:23:59 crc kubenswrapper[4764]: I0202 09:23:59.676621 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Feb 02 09:23:59 crc kubenswrapper[4764]: I0202 09:23:59.966758 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Feb 02 09:24:00 crc kubenswrapper[4764]: I0202 09:24:00.013361 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Feb 02 09:24:00 crc kubenswrapper[4764]: I0202 09:24:00.431781 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Feb 02 09:24:00 crc kubenswrapper[4764]: I0202 09:24:00.432988 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Feb 02 09:24:00 crc kubenswrapper[4764]: I0202 09:24:00.438873 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Feb 02 09:24:00 crc kubenswrapper[4764]: I0202 09:24:00.439169 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-4dmc4" Feb 02 09:24:00 crc kubenswrapper[4764]: I0202 09:24:00.439293 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Feb 02 09:24:00 crc kubenswrapper[4764]: I0202 09:24:00.449860 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Feb 02 09:24:00 crc kubenswrapper[4764]: I0202 09:24:00.482277 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Feb 02 09:24:00 crc kubenswrapper[4764]: I0202 09:24:00.483354 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f56696ad-9508-4f7b-89fb-28b59af303a0-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"f56696ad-9508-4f7b-89fb-28b59af303a0\") " pod="openstack/ovn-northd-0" Feb 02 09:24:00 crc kubenswrapper[4764]: I0202 09:24:00.483421 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f56696ad-9508-4f7b-89fb-28b59af303a0-scripts\") pod \"ovn-northd-0\" (UID: \"f56696ad-9508-4f7b-89fb-28b59af303a0\") " pod="openstack/ovn-northd-0" Feb 02 09:24:00 crc kubenswrapper[4764]: I0202 09:24:00.483443 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/f56696ad-9508-4f7b-89fb-28b59af303a0-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"f56696ad-9508-4f7b-89fb-28b59af303a0\") " pod="openstack/ovn-northd-0" Feb 02 09:24:00 crc kubenswrapper[4764]: I0202 09:24:00.483509 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/f56696ad-9508-4f7b-89fb-28b59af303a0-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"f56696ad-9508-4f7b-89fb-28b59af303a0\") " pod="openstack/ovn-northd-0" Feb 02 09:24:00 crc kubenswrapper[4764]: I0202 09:24:00.483526 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f56696ad-9508-4f7b-89fb-28b59af303a0-config\") pod \"ovn-northd-0\" (UID: \"f56696ad-9508-4f7b-89fb-28b59af303a0\") " pod="openstack/ovn-northd-0" Feb 02 09:24:00 crc kubenswrapper[4764]: I0202 09:24:00.483560 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/f56696ad-9508-4f7b-89fb-28b59af303a0-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"f56696ad-9508-4f7b-89fb-28b59af303a0\") " pod="openstack/ovn-northd-0" Feb 02 09:24:00 crc kubenswrapper[4764]: I0202 09:24:00.483613 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n9l7v\" (UniqueName: \"kubernetes.io/projected/f56696ad-9508-4f7b-89fb-28b59af303a0-kube-api-access-n9l7v\") pod \"ovn-northd-0\" (UID: \"f56696ad-9508-4f7b-89fb-28b59af303a0\") " pod="openstack/ovn-northd-0" Feb 02 09:24:00 crc kubenswrapper[4764]: I0202 09:24:00.585188 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n9l7v\" (UniqueName: \"kubernetes.io/projected/f56696ad-9508-4f7b-89fb-28b59af303a0-kube-api-access-n9l7v\") pod \"ovn-northd-0\" (UID: \"f56696ad-9508-4f7b-89fb-28b59af303a0\") " pod="openstack/ovn-northd-0" Feb 02 09:24:00 crc kubenswrapper[4764]: I0202 09:24:00.585251 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f56696ad-9508-4f7b-89fb-28b59af303a0-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"f56696ad-9508-4f7b-89fb-28b59af303a0\") " pod="openstack/ovn-northd-0" Feb 02 09:24:00 crc kubenswrapper[4764]: I0202 09:24:00.585283 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f56696ad-9508-4f7b-89fb-28b59af303a0-scripts\") pod \"ovn-northd-0\" (UID: \"f56696ad-9508-4f7b-89fb-28b59af303a0\") " pod="openstack/ovn-northd-0" Feb 02 09:24:00 crc kubenswrapper[4764]: I0202 09:24:00.586075 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f56696ad-9508-4f7b-89fb-28b59af303a0-scripts\") pod \"ovn-northd-0\" (UID: \"f56696ad-9508-4f7b-89fb-28b59af303a0\") " pod="openstack/ovn-northd-0" Feb 02 09:24:00 crc kubenswrapper[4764]: I0202 09:24:00.586171 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/f56696ad-9508-4f7b-89fb-28b59af303a0-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"f56696ad-9508-4f7b-89fb-28b59af303a0\") " pod="openstack/ovn-northd-0" Feb 02 09:24:00 crc kubenswrapper[4764]: I0202 09:24:00.586242 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/f56696ad-9508-4f7b-89fb-28b59af303a0-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"f56696ad-9508-4f7b-89fb-28b59af303a0\") " pod="openstack/ovn-northd-0" Feb 02 09:24:00 crc kubenswrapper[4764]: I0202 09:24:00.586261 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f56696ad-9508-4f7b-89fb-28b59af303a0-config\") pod \"ovn-northd-0\" (UID: \"f56696ad-9508-4f7b-89fb-28b59af303a0\") " pod="openstack/ovn-northd-0" Feb 02 09:24:00 crc kubenswrapper[4764]: I0202 09:24:00.586281 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/f56696ad-9508-4f7b-89fb-28b59af303a0-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"f56696ad-9508-4f7b-89fb-28b59af303a0\") " pod="openstack/ovn-northd-0" Feb 02 09:24:00 crc kubenswrapper[4764]: I0202 09:24:00.586959 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/f56696ad-9508-4f7b-89fb-28b59af303a0-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"f56696ad-9508-4f7b-89fb-28b59af303a0\") " pod="openstack/ovn-northd-0" Feb 02 09:24:00 crc kubenswrapper[4764]: I0202 09:24:00.587224 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f56696ad-9508-4f7b-89fb-28b59af303a0-config\") pod \"ovn-northd-0\" (UID: \"f56696ad-9508-4f7b-89fb-28b59af303a0\") " pod="openstack/ovn-northd-0" Feb 02 09:24:00 crc kubenswrapper[4764]: I0202 09:24:00.590669 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f56696ad-9508-4f7b-89fb-28b59af303a0-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"f56696ad-9508-4f7b-89fb-28b59af303a0\") " pod="openstack/ovn-northd-0" Feb 02 09:24:00 crc kubenswrapper[4764]: I0202 09:24:00.590917 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/f56696ad-9508-4f7b-89fb-28b59af303a0-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"f56696ad-9508-4f7b-89fb-28b59af303a0\") " pod="openstack/ovn-northd-0" Feb 02 09:24:00 crc kubenswrapper[4764]: I0202 09:24:00.595412 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/f56696ad-9508-4f7b-89fb-28b59af303a0-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"f56696ad-9508-4f7b-89fb-28b59af303a0\") " pod="openstack/ovn-northd-0" Feb 02 09:24:00 crc kubenswrapper[4764]: I0202 09:24:00.604309 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n9l7v\" (UniqueName: \"kubernetes.io/projected/f56696ad-9508-4f7b-89fb-28b59af303a0-kube-api-access-n9l7v\") pod \"ovn-northd-0\" (UID: \"f56696ad-9508-4f7b-89fb-28b59af303a0\") " pod="openstack/ovn-northd-0" Feb 02 09:24:00 crc kubenswrapper[4764]: I0202 09:24:00.748906 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Feb 02 09:24:01 crc kubenswrapper[4764]: I0202 09:24:01.192904 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Feb 02 09:24:01 crc kubenswrapper[4764]: I0202 09:24:01.524108 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-64f7f48db9-rz8h7" Feb 02 09:24:01 crc kubenswrapper[4764]: I0202 09:24:01.633958 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"26d7a0b3-ea1e-417b-b417-69ae84bd5d24","Type":"ContainerStarted","Data":"1cb1bd29c445477024eea879da1cceda1244dd5e3b7e54856e664ed0cfe5ec68"} Feb 02 09:24:01 crc kubenswrapper[4764]: I0202 09:24:01.634262 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Feb 02 09:24:01 crc kubenswrapper[4764]: I0202 09:24:01.639352 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"f56696ad-9508-4f7b-89fb-28b59af303a0","Type":"ContainerStarted","Data":"e66a6c1e0c881a222f10aa42844542c37e19121dd8dee1727d8400f27ab1e4fc"} Feb 02 09:24:01 crc kubenswrapper[4764]: I0202 09:24:01.658925 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.304460556 podStartE2EDuration="43.658905537s" podCreationTimestamp="2026-02-02 09:23:18 +0000 UTC" firstStartedPulling="2026-02-02 09:23:19.859805028 +0000 UTC m=+962.793529116" lastFinishedPulling="2026-02-02 09:24:01.214250009 +0000 UTC m=+1004.147974097" observedRunningTime="2026-02-02 09:24:01.656703709 +0000 UTC m=+1004.590427787" watchObservedRunningTime="2026-02-02 09:24:01.658905537 +0000 UTC m=+1004.592629625" Feb 02 09:24:01 crc kubenswrapper[4764]: I0202 09:24:01.868611 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Feb 02 09:24:01 crc kubenswrapper[4764]: I0202 09:24:01.963210 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Feb 02 09:24:02 crc kubenswrapper[4764]: I0202 09:24:02.560293 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Feb 02 09:24:04 crc kubenswrapper[4764]: I0202 09:24:04.663069 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"f56696ad-9508-4f7b-89fb-28b59af303a0","Type":"ContainerStarted","Data":"d0341795bb50d081da80fa088bd92de3e4e661042ebdd35b486d4f9c851bc0a4"} Feb 02 09:24:05 crc kubenswrapper[4764]: I0202 09:24:05.673206 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"f56696ad-9508-4f7b-89fb-28b59af303a0","Type":"ContainerStarted","Data":"112620f8fe2c98977d3ebd91a4ad14423766040f6722d152d0eaf080b9b51c1d"} Feb 02 09:24:05 crc kubenswrapper[4764]: I0202 09:24:05.673559 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Feb 02 09:24:05 crc kubenswrapper[4764]: I0202 09:24:05.699777 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=2.486467933 podStartE2EDuration="5.699744581s" podCreationTimestamp="2026-02-02 09:24:00 +0000 UTC" firstStartedPulling="2026-02-02 09:24:01.210155291 +0000 UTC m=+1004.143879379" lastFinishedPulling="2026-02-02 09:24:04.423431939 +0000 UTC m=+1007.357156027" observedRunningTime="2026-02-02 09:24:05.69250357 +0000 UTC m=+1008.626227728" watchObservedRunningTime="2026-02-02 09:24:05.699744581 +0000 UTC m=+1008.633468709" Feb 02 09:24:06 crc kubenswrapper[4764]: I0202 09:24:06.019876 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/root-account-create-update-4rrf5"] Feb 02 09:24:06 crc kubenswrapper[4764]: I0202 09:24:06.021418 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-4rrf5" Feb 02 09:24:06 crc kubenswrapper[4764]: I0202 09:24:06.028751 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-mariadb-root-db-secret" Feb 02 09:24:06 crc kubenswrapper[4764]: I0202 09:24:06.029448 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-4rrf5"] Feb 02 09:24:06 crc kubenswrapper[4764]: I0202 09:24:06.187082 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/82d1fbc3-b839-4f16-ad55-bb8a8ce288e1-operator-scripts\") pod \"root-account-create-update-4rrf5\" (UID: \"82d1fbc3-b839-4f16-ad55-bb8a8ce288e1\") " pod="openstack/root-account-create-update-4rrf5" Feb 02 09:24:06 crc kubenswrapper[4764]: I0202 09:24:06.187133 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4b4wn\" (UniqueName: \"kubernetes.io/projected/82d1fbc3-b839-4f16-ad55-bb8a8ce288e1-kube-api-access-4b4wn\") pod \"root-account-create-update-4rrf5\" (UID: \"82d1fbc3-b839-4f16-ad55-bb8a8ce288e1\") " pod="openstack/root-account-create-update-4rrf5" Feb 02 09:24:06 crc kubenswrapper[4764]: I0202 09:24:06.251638 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Feb 02 09:24:06 crc kubenswrapper[4764]: I0202 09:24:06.251712 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Feb 02 09:24:06 crc kubenswrapper[4764]: I0202 09:24:06.288693 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/82d1fbc3-b839-4f16-ad55-bb8a8ce288e1-operator-scripts\") pod \"root-account-create-update-4rrf5\" (UID: \"82d1fbc3-b839-4f16-ad55-bb8a8ce288e1\") " pod="openstack/root-account-create-update-4rrf5" Feb 02 09:24:06 crc kubenswrapper[4764]: I0202 09:24:06.288760 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4b4wn\" (UniqueName: \"kubernetes.io/projected/82d1fbc3-b839-4f16-ad55-bb8a8ce288e1-kube-api-access-4b4wn\") pod \"root-account-create-update-4rrf5\" (UID: \"82d1fbc3-b839-4f16-ad55-bb8a8ce288e1\") " pod="openstack/root-account-create-update-4rrf5" Feb 02 09:24:06 crc kubenswrapper[4764]: I0202 09:24:06.289472 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/82d1fbc3-b839-4f16-ad55-bb8a8ce288e1-operator-scripts\") pod \"root-account-create-update-4rrf5\" (UID: \"82d1fbc3-b839-4f16-ad55-bb8a8ce288e1\") " pod="openstack/root-account-create-update-4rrf5" Feb 02 09:24:06 crc kubenswrapper[4764]: I0202 09:24:06.317412 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4b4wn\" (UniqueName: \"kubernetes.io/projected/82d1fbc3-b839-4f16-ad55-bb8a8ce288e1-kube-api-access-4b4wn\") pod \"root-account-create-update-4rrf5\" (UID: \"82d1fbc3-b839-4f16-ad55-bb8a8ce288e1\") " pod="openstack/root-account-create-update-4rrf5" Feb 02 09:24:06 crc kubenswrapper[4764]: I0202 09:24:06.340398 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-4rrf5" Feb 02 09:24:06 crc kubenswrapper[4764]: I0202 09:24:06.400624 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Feb 02 09:24:06 crc kubenswrapper[4764]: I0202 09:24:06.405090 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-56df986d9c-vkqnn" Feb 02 09:24:06 crc kubenswrapper[4764]: I0202 09:24:06.507227 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-64f7f48db9-rz8h7"] Feb 02 09:24:06 crc kubenswrapper[4764]: I0202 09:24:06.510953 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-64f7f48db9-rz8h7" podUID="456b6ada-1b28-4df6-b74b-c470910598e5" containerName="dnsmasq-dns" containerID="cri-o://81ca5aae341ed7dd3dd43c8d9df176155909c2161d3978da1edd39d23566c0e1" gracePeriod=10 Feb 02 09:24:06 crc kubenswrapper[4764]: I0202 09:24:06.690450 4764 generic.go:334] "Generic (PLEG): container finished" podID="456b6ada-1b28-4df6-b74b-c470910598e5" containerID="81ca5aae341ed7dd3dd43c8d9df176155909c2161d3978da1edd39d23566c0e1" exitCode=0 Feb 02 09:24:06 crc kubenswrapper[4764]: I0202 09:24:06.691956 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64f7f48db9-rz8h7" event={"ID":"456b6ada-1b28-4df6-b74b-c470910598e5","Type":"ContainerDied","Data":"81ca5aae341ed7dd3dd43c8d9df176155909c2161d3978da1edd39d23566c0e1"} Feb 02 09:24:06 crc kubenswrapper[4764]: I0202 09:24:06.795661 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Feb 02 09:24:06 crc kubenswrapper[4764]: I0202 09:24:06.958050 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-4rrf5"] Feb 02 09:24:07 crc kubenswrapper[4764]: I0202 09:24:07.051908 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64f7f48db9-rz8h7" Feb 02 09:24:07 crc kubenswrapper[4764]: I0202 09:24:07.076281 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-wqk4n"] Feb 02 09:24:07 crc kubenswrapper[4764]: E0202 09:24:07.076666 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="456b6ada-1b28-4df6-b74b-c470910598e5" containerName="dnsmasq-dns" Feb 02 09:24:07 crc kubenswrapper[4764]: I0202 09:24:07.076680 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="456b6ada-1b28-4df6-b74b-c470910598e5" containerName="dnsmasq-dns" Feb 02 09:24:07 crc kubenswrapper[4764]: E0202 09:24:07.076703 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="456b6ada-1b28-4df6-b74b-c470910598e5" containerName="init" Feb 02 09:24:07 crc kubenswrapper[4764]: I0202 09:24:07.076709 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="456b6ada-1b28-4df6-b74b-c470910598e5" containerName="init" Feb 02 09:24:07 crc kubenswrapper[4764]: I0202 09:24:07.076864 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="456b6ada-1b28-4df6-b74b-c470910598e5" containerName="dnsmasq-dns" Feb 02 09:24:07 crc kubenswrapper[4764]: I0202 09:24:07.077380 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-wqk4n" Feb 02 09:24:07 crc kubenswrapper[4764]: I0202 09:24:07.083887 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-wqk4n"] Feb 02 09:24:07 crc kubenswrapper[4764]: I0202 09:24:07.207598 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4pqz6\" (UniqueName: \"kubernetes.io/projected/456b6ada-1b28-4df6-b74b-c470910598e5-kube-api-access-4pqz6\") pod \"456b6ada-1b28-4df6-b74b-c470910598e5\" (UID: \"456b6ada-1b28-4df6-b74b-c470910598e5\") " Feb 02 09:24:07 crc kubenswrapper[4764]: I0202 09:24:07.207968 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/456b6ada-1b28-4df6-b74b-c470910598e5-config\") pod \"456b6ada-1b28-4df6-b74b-c470910598e5\" (UID: \"456b6ada-1b28-4df6-b74b-c470910598e5\") " Feb 02 09:24:07 crc kubenswrapper[4764]: I0202 09:24:07.208006 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/456b6ada-1b28-4df6-b74b-c470910598e5-ovsdbserver-nb\") pod \"456b6ada-1b28-4df6-b74b-c470910598e5\" (UID: \"456b6ada-1b28-4df6-b74b-c470910598e5\") " Feb 02 09:24:07 crc kubenswrapper[4764]: I0202 09:24:07.208077 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/456b6ada-1b28-4df6-b74b-c470910598e5-dns-svc\") pod \"456b6ada-1b28-4df6-b74b-c470910598e5\" (UID: \"456b6ada-1b28-4df6-b74b-c470910598e5\") " Feb 02 09:24:07 crc kubenswrapper[4764]: I0202 09:24:07.208982 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7d99k\" (UniqueName: \"kubernetes.io/projected/a3a49c31-c4c9-415b-89d9-bee375775e38-kube-api-access-7d99k\") pod \"keystone-db-create-wqk4n\" (UID: \"a3a49c31-c4c9-415b-89d9-bee375775e38\") " pod="openstack/keystone-db-create-wqk4n" Feb 02 09:24:07 crc kubenswrapper[4764]: I0202 09:24:07.209034 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a3a49c31-c4c9-415b-89d9-bee375775e38-operator-scripts\") pod \"keystone-db-create-wqk4n\" (UID: \"a3a49c31-c4c9-415b-89d9-bee375775e38\") " pod="openstack/keystone-db-create-wqk4n" Feb 02 09:24:07 crc kubenswrapper[4764]: I0202 09:24:07.213720 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/456b6ada-1b28-4df6-b74b-c470910598e5-kube-api-access-4pqz6" (OuterVolumeSpecName: "kube-api-access-4pqz6") pod "456b6ada-1b28-4df6-b74b-c470910598e5" (UID: "456b6ada-1b28-4df6-b74b-c470910598e5"). InnerVolumeSpecName "kube-api-access-4pqz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:24:07 crc kubenswrapper[4764]: I0202 09:24:07.257914 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/456b6ada-1b28-4df6-b74b-c470910598e5-config" (OuterVolumeSpecName: "config") pod "456b6ada-1b28-4df6-b74b-c470910598e5" (UID: "456b6ada-1b28-4df6-b74b-c470910598e5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:24:07 crc kubenswrapper[4764]: I0202 09:24:07.258359 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/456b6ada-1b28-4df6-b74b-c470910598e5-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "456b6ada-1b28-4df6-b74b-c470910598e5" (UID: "456b6ada-1b28-4df6-b74b-c470910598e5"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:24:07 crc kubenswrapper[4764]: I0202 09:24:07.260175 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/456b6ada-1b28-4df6-b74b-c470910598e5-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "456b6ada-1b28-4df6-b74b-c470910598e5" (UID: "456b6ada-1b28-4df6-b74b-c470910598e5"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:24:07 crc kubenswrapper[4764]: I0202 09:24:07.300071 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-48c4-account-create-update-lvzbb"] Feb 02 09:24:07 crc kubenswrapper[4764]: I0202 09:24:07.301210 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-48c4-account-create-update-lvzbb" Feb 02 09:24:07 crc kubenswrapper[4764]: I0202 09:24:07.307362 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Feb 02 09:24:07 crc kubenswrapper[4764]: I0202 09:24:07.310427 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-48c4-account-create-update-lvzbb"] Feb 02 09:24:07 crc kubenswrapper[4764]: I0202 09:24:07.310575 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a3a49c31-c4c9-415b-89d9-bee375775e38-operator-scripts\") pod \"keystone-db-create-wqk4n\" (UID: \"a3a49c31-c4c9-415b-89d9-bee375775e38\") " pod="openstack/keystone-db-create-wqk4n" Feb 02 09:24:07 crc kubenswrapper[4764]: I0202 09:24:07.310709 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7d99k\" (UniqueName: \"kubernetes.io/projected/a3a49c31-c4c9-415b-89d9-bee375775e38-kube-api-access-7d99k\") pod \"keystone-db-create-wqk4n\" (UID: \"a3a49c31-c4c9-415b-89d9-bee375775e38\") " pod="openstack/keystone-db-create-wqk4n" Feb 02 09:24:07 crc kubenswrapper[4764]: I0202 09:24:07.310772 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4pqz6\" (UniqueName: \"kubernetes.io/projected/456b6ada-1b28-4df6-b74b-c470910598e5-kube-api-access-4pqz6\") on node \"crc\" DevicePath \"\"" Feb 02 09:24:07 crc kubenswrapper[4764]: I0202 09:24:07.310783 4764 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/456b6ada-1b28-4df6-b74b-c470910598e5-config\") on node \"crc\" DevicePath \"\"" Feb 02 09:24:07 crc kubenswrapper[4764]: I0202 09:24:07.310791 4764 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/456b6ada-1b28-4df6-b74b-c470910598e5-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 02 09:24:07 crc kubenswrapper[4764]: I0202 09:24:07.310820 4764 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/456b6ada-1b28-4df6-b74b-c470910598e5-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 02 09:24:07 crc kubenswrapper[4764]: I0202 09:24:07.311298 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a3a49c31-c4c9-415b-89d9-bee375775e38-operator-scripts\") pod \"keystone-db-create-wqk4n\" (UID: \"a3a49c31-c4c9-415b-89d9-bee375775e38\") " pod="openstack/keystone-db-create-wqk4n" Feb 02 09:24:07 crc kubenswrapper[4764]: I0202 09:24:07.328194 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7d99k\" (UniqueName: \"kubernetes.io/projected/a3a49c31-c4c9-415b-89d9-bee375775e38-kube-api-access-7d99k\") pod \"keystone-db-create-wqk4n\" (UID: \"a3a49c31-c4c9-415b-89d9-bee375775e38\") " pod="openstack/keystone-db-create-wqk4n" Feb 02 09:24:07 crc kubenswrapper[4764]: I0202 09:24:07.399696 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-wqk4n" Feb 02 09:24:07 crc kubenswrapper[4764]: I0202 09:24:07.412454 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zggfh\" (UniqueName: \"kubernetes.io/projected/9c4c3400-5c5d-4d8b-82a9-c8ca34abe7c0-kube-api-access-zggfh\") pod \"keystone-48c4-account-create-update-lvzbb\" (UID: \"9c4c3400-5c5d-4d8b-82a9-c8ca34abe7c0\") " pod="openstack/keystone-48c4-account-create-update-lvzbb" Feb 02 09:24:07 crc kubenswrapper[4764]: I0202 09:24:07.412507 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9c4c3400-5c5d-4d8b-82a9-c8ca34abe7c0-operator-scripts\") pod \"keystone-48c4-account-create-update-lvzbb\" (UID: \"9c4c3400-5c5d-4d8b-82a9-c8ca34abe7c0\") " pod="openstack/keystone-48c4-account-create-update-lvzbb" Feb 02 09:24:07 crc kubenswrapper[4764]: I0202 09:24:07.445364 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-2d92n"] Feb 02 09:24:07 crc kubenswrapper[4764]: I0202 09:24:07.467501 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-2d92n" Feb 02 09:24:07 crc kubenswrapper[4764]: I0202 09:24:07.487011 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-2d92n"] Feb 02 09:24:07 crc kubenswrapper[4764]: I0202 09:24:07.515899 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rf9th\" (UniqueName: \"kubernetes.io/projected/b0e0b60e-0289-4700-b747-a6442ca6e34a-kube-api-access-rf9th\") pod \"placement-db-create-2d92n\" (UID: \"b0e0b60e-0289-4700-b747-a6442ca6e34a\") " pod="openstack/placement-db-create-2d92n" Feb 02 09:24:07 crc kubenswrapper[4764]: I0202 09:24:07.516133 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b0e0b60e-0289-4700-b747-a6442ca6e34a-operator-scripts\") pod \"placement-db-create-2d92n\" (UID: \"b0e0b60e-0289-4700-b747-a6442ca6e34a\") " pod="openstack/placement-db-create-2d92n" Feb 02 09:24:07 crc kubenswrapper[4764]: I0202 09:24:07.516214 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zggfh\" (UniqueName: \"kubernetes.io/projected/9c4c3400-5c5d-4d8b-82a9-c8ca34abe7c0-kube-api-access-zggfh\") pod \"keystone-48c4-account-create-update-lvzbb\" (UID: \"9c4c3400-5c5d-4d8b-82a9-c8ca34abe7c0\") " pod="openstack/keystone-48c4-account-create-update-lvzbb" Feb 02 09:24:07 crc kubenswrapper[4764]: I0202 09:24:07.516278 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9c4c3400-5c5d-4d8b-82a9-c8ca34abe7c0-operator-scripts\") pod \"keystone-48c4-account-create-update-lvzbb\" (UID: \"9c4c3400-5c5d-4d8b-82a9-c8ca34abe7c0\") " pod="openstack/keystone-48c4-account-create-update-lvzbb" Feb 02 09:24:07 crc kubenswrapper[4764]: I0202 09:24:07.517420 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9c4c3400-5c5d-4d8b-82a9-c8ca34abe7c0-operator-scripts\") pod \"keystone-48c4-account-create-update-lvzbb\" (UID: \"9c4c3400-5c5d-4d8b-82a9-c8ca34abe7c0\") " pod="openstack/keystone-48c4-account-create-update-lvzbb" Feb 02 09:24:07 crc kubenswrapper[4764]: I0202 09:24:07.550994 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zggfh\" (UniqueName: \"kubernetes.io/projected/9c4c3400-5c5d-4d8b-82a9-c8ca34abe7c0-kube-api-access-zggfh\") pod \"keystone-48c4-account-create-update-lvzbb\" (UID: \"9c4c3400-5c5d-4d8b-82a9-c8ca34abe7c0\") " pod="openstack/keystone-48c4-account-create-update-lvzbb" Feb 02 09:24:07 crc kubenswrapper[4764]: I0202 09:24:07.615634 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-48c4-account-create-update-lvzbb" Feb 02 09:24:07 crc kubenswrapper[4764]: I0202 09:24:07.617645 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rf9th\" (UniqueName: \"kubernetes.io/projected/b0e0b60e-0289-4700-b747-a6442ca6e34a-kube-api-access-rf9th\") pod \"placement-db-create-2d92n\" (UID: \"b0e0b60e-0289-4700-b747-a6442ca6e34a\") " pod="openstack/placement-db-create-2d92n" Feb 02 09:24:07 crc kubenswrapper[4764]: I0202 09:24:07.617692 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b0e0b60e-0289-4700-b747-a6442ca6e34a-operator-scripts\") pod \"placement-db-create-2d92n\" (UID: \"b0e0b60e-0289-4700-b747-a6442ca6e34a\") " pod="openstack/placement-db-create-2d92n" Feb 02 09:24:07 crc kubenswrapper[4764]: I0202 09:24:07.619259 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b0e0b60e-0289-4700-b747-a6442ca6e34a-operator-scripts\") pod \"placement-db-create-2d92n\" (UID: \"b0e0b60e-0289-4700-b747-a6442ca6e34a\") " pod="openstack/placement-db-create-2d92n" Feb 02 09:24:07 crc kubenswrapper[4764]: I0202 09:24:07.677203 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-2a71-account-create-update-6bfns"] Feb 02 09:24:07 crc kubenswrapper[4764]: I0202 09:24:07.681742 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-2a71-account-create-update-6bfns" Feb 02 09:24:07 crc kubenswrapper[4764]: I0202 09:24:07.688003 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rf9th\" (UniqueName: \"kubernetes.io/projected/b0e0b60e-0289-4700-b747-a6442ca6e34a-kube-api-access-rf9th\") pod \"placement-db-create-2d92n\" (UID: \"b0e0b60e-0289-4700-b747-a6442ca6e34a\") " pod="openstack/placement-db-create-2d92n" Feb 02 09:24:07 crc kubenswrapper[4764]: I0202 09:24:07.689888 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Feb 02 09:24:07 crc kubenswrapper[4764]: I0202 09:24:07.693914 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-2a71-account-create-update-6bfns"] Feb 02 09:24:07 crc kubenswrapper[4764]: I0202 09:24:07.720071 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1a85da62-5c86-4f54-9d14-b80c3974189a-operator-scripts\") pod \"placement-2a71-account-create-update-6bfns\" (UID: \"1a85da62-5c86-4f54-9d14-b80c3974189a\") " pod="openstack/placement-2a71-account-create-update-6bfns" Feb 02 09:24:07 crc kubenswrapper[4764]: I0202 09:24:07.720190 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6l24x\" (UniqueName: \"kubernetes.io/projected/1a85da62-5c86-4f54-9d14-b80c3974189a-kube-api-access-6l24x\") pod \"placement-2a71-account-create-update-6bfns\" (UID: \"1a85da62-5c86-4f54-9d14-b80c3974189a\") " pod="openstack/placement-2a71-account-create-update-6bfns" Feb 02 09:24:07 crc kubenswrapper[4764]: I0202 09:24:07.732293 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64f7f48db9-rz8h7" Feb 02 09:24:07 crc kubenswrapper[4764]: I0202 09:24:07.732474 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64f7f48db9-rz8h7" event={"ID":"456b6ada-1b28-4df6-b74b-c470910598e5","Type":"ContainerDied","Data":"a2bfab2555ecacd859966ea06c0813ed79e2d4a2de6dbbf04bc852e597d9366e"} Feb 02 09:24:07 crc kubenswrapper[4764]: I0202 09:24:07.732539 4764 scope.go:117] "RemoveContainer" containerID="81ca5aae341ed7dd3dd43c8d9df176155909c2161d3978da1edd39d23566c0e1" Feb 02 09:24:07 crc kubenswrapper[4764]: I0202 09:24:07.744899 4764 generic.go:334] "Generic (PLEG): container finished" podID="82d1fbc3-b839-4f16-ad55-bb8a8ce288e1" containerID="588e10f44de4046cd270174efdbfc218e7cd319ddead99865bd4a949dab2c722" exitCode=0 Feb 02 09:24:07 crc kubenswrapper[4764]: I0202 09:24:07.746366 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-4rrf5" event={"ID":"82d1fbc3-b839-4f16-ad55-bb8a8ce288e1","Type":"ContainerDied","Data":"588e10f44de4046cd270174efdbfc218e7cd319ddead99865bd4a949dab2c722"} Feb 02 09:24:07 crc kubenswrapper[4764]: I0202 09:24:07.746520 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-4rrf5" event={"ID":"82d1fbc3-b839-4f16-ad55-bb8a8ce288e1","Type":"ContainerStarted","Data":"4b45018ab95716f45eee3c9426df2d50bd04b6019589bcd8bc53aea9136cf886"} Feb 02 09:24:07 crc kubenswrapper[4764]: I0202 09:24:07.761527 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-wqk4n"] Feb 02 09:24:07 crc kubenswrapper[4764]: I0202 09:24:07.767180 4764 scope.go:117] "RemoveContainer" containerID="73e4bf881e097e1b2174d65b867ec1e41ed8159785d4fd89e35ecf81136dc73e" Feb 02 09:24:07 crc kubenswrapper[4764]: I0202 09:24:07.796765 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-64f7f48db9-rz8h7"] Feb 02 09:24:07 crc kubenswrapper[4764]: I0202 09:24:07.812697 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-64f7f48db9-rz8h7"] Feb 02 09:24:07 crc kubenswrapper[4764]: I0202 09:24:07.817289 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-2d92n" Feb 02 09:24:07 crc kubenswrapper[4764]: I0202 09:24:07.827028 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1a85da62-5c86-4f54-9d14-b80c3974189a-operator-scripts\") pod \"placement-2a71-account-create-update-6bfns\" (UID: \"1a85da62-5c86-4f54-9d14-b80c3974189a\") " pod="openstack/placement-2a71-account-create-update-6bfns" Feb 02 09:24:07 crc kubenswrapper[4764]: I0202 09:24:07.827158 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6l24x\" (UniqueName: \"kubernetes.io/projected/1a85da62-5c86-4f54-9d14-b80c3974189a-kube-api-access-6l24x\") pod \"placement-2a71-account-create-update-6bfns\" (UID: \"1a85da62-5c86-4f54-9d14-b80c3974189a\") " pod="openstack/placement-2a71-account-create-update-6bfns" Feb 02 09:24:07 crc kubenswrapper[4764]: I0202 09:24:07.829681 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1a85da62-5c86-4f54-9d14-b80c3974189a-operator-scripts\") pod \"placement-2a71-account-create-update-6bfns\" (UID: \"1a85da62-5c86-4f54-9d14-b80c3974189a\") " pod="openstack/placement-2a71-account-create-update-6bfns" Feb 02 09:24:07 crc kubenswrapper[4764]: I0202 09:24:07.868582 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="456b6ada-1b28-4df6-b74b-c470910598e5" path="/var/lib/kubelet/pods/456b6ada-1b28-4df6-b74b-c470910598e5/volumes" Feb 02 09:24:07 crc kubenswrapper[4764]: I0202 09:24:07.874850 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6l24x\" (UniqueName: \"kubernetes.io/projected/1a85da62-5c86-4f54-9d14-b80c3974189a-kube-api-access-6l24x\") pod \"placement-2a71-account-create-update-6bfns\" (UID: \"1a85da62-5c86-4f54-9d14-b80c3974189a\") " pod="openstack/placement-2a71-account-create-update-6bfns" Feb 02 09:24:08 crc kubenswrapper[4764]: I0202 09:24:08.017235 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-2a71-account-create-update-6bfns" Feb 02 09:24:08 crc kubenswrapper[4764]: I0202 09:24:08.200622 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-48c4-account-create-update-lvzbb"] Feb 02 09:24:08 crc kubenswrapper[4764]: W0202 09:24:08.218814 4764 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9c4c3400_5c5d_4d8b_82a9_c8ca34abe7c0.slice/crio-12eb78775474df2bcd9f7e74647a009c5efd052d78eae394ffe87fff61360771 WatchSource:0}: Error finding container 12eb78775474df2bcd9f7e74647a009c5efd052d78eae394ffe87fff61360771: Status 404 returned error can't find the container with id 12eb78775474df2bcd9f7e74647a009c5efd052d78eae394ffe87fff61360771 Feb 02 09:24:08 crc kubenswrapper[4764]: I0202 09:24:08.302976 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-2d92n"] Feb 02 09:24:08 crc kubenswrapper[4764]: I0202 09:24:08.477291 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-2a71-account-create-update-6bfns"] Feb 02 09:24:08 crc kubenswrapper[4764]: I0202 09:24:08.754671 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-2a71-account-create-update-6bfns" event={"ID":"1a85da62-5c86-4f54-9d14-b80c3974189a","Type":"ContainerStarted","Data":"d62d6a087221f0d57aacf9b33844eb86a5a5c129fd54d5dd9da5ebf675b4c402"} Feb 02 09:24:08 crc kubenswrapper[4764]: I0202 09:24:08.754726 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-2a71-account-create-update-6bfns" event={"ID":"1a85da62-5c86-4f54-9d14-b80c3974189a","Type":"ContainerStarted","Data":"d04235a6d9a0b8b0b7f989d59c81216046bf07d18a123399fb9d2a3e1c838a0f"} Feb 02 09:24:08 crc kubenswrapper[4764]: I0202 09:24:08.759173 4764 generic.go:334] "Generic (PLEG): container finished" podID="a3a49c31-c4c9-415b-89d9-bee375775e38" containerID="884c3cc99e06d5ffc876d00aabdecbb9303f422c3cafc35b2045e7f55876c9c3" exitCode=0 Feb 02 09:24:08 crc kubenswrapper[4764]: I0202 09:24:08.759267 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-wqk4n" event={"ID":"a3a49c31-c4c9-415b-89d9-bee375775e38","Type":"ContainerDied","Data":"884c3cc99e06d5ffc876d00aabdecbb9303f422c3cafc35b2045e7f55876c9c3"} Feb 02 09:24:08 crc kubenswrapper[4764]: I0202 09:24:08.759292 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-wqk4n" event={"ID":"a3a49c31-c4c9-415b-89d9-bee375775e38","Type":"ContainerStarted","Data":"ef15e32c0c09f8b8bf041754d1ea709d12bc18f74445463b052d8de8cfeb2a69"} Feb 02 09:24:08 crc kubenswrapper[4764]: I0202 09:24:08.762734 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-48c4-account-create-update-lvzbb" event={"ID":"9c4c3400-5c5d-4d8b-82a9-c8ca34abe7c0","Type":"ContainerStarted","Data":"7e4be9945a21765d2e816b5d7ef51e7f14627efe60631690e6d27f83860286f3"} Feb 02 09:24:08 crc kubenswrapper[4764]: I0202 09:24:08.762775 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-48c4-account-create-update-lvzbb" event={"ID":"9c4c3400-5c5d-4d8b-82a9-c8ca34abe7c0","Type":"ContainerStarted","Data":"12eb78775474df2bcd9f7e74647a009c5efd052d78eae394ffe87fff61360771"} Feb 02 09:24:08 crc kubenswrapper[4764]: I0202 09:24:08.764471 4764 generic.go:334] "Generic (PLEG): container finished" podID="b0e0b60e-0289-4700-b747-a6442ca6e34a" containerID="15a867576bfef1a3a727c912d74f23f8a035f461961a78d5ad32e4e7c8833fc0" exitCode=0 Feb 02 09:24:08 crc kubenswrapper[4764]: I0202 09:24:08.765418 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-2d92n" event={"ID":"b0e0b60e-0289-4700-b747-a6442ca6e34a","Type":"ContainerDied","Data":"15a867576bfef1a3a727c912d74f23f8a035f461961a78d5ad32e4e7c8833fc0"} Feb 02 09:24:08 crc kubenswrapper[4764]: I0202 09:24:08.765441 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-2d92n" event={"ID":"b0e0b60e-0289-4700-b747-a6442ca6e34a","Type":"ContainerStarted","Data":"e9bd7d579598242363a720861f2448e473f5579d221b01f61293a761ee817906"} Feb 02 09:24:08 crc kubenswrapper[4764]: I0202 09:24:08.779277 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-2a71-account-create-update-6bfns" podStartSLOduration=1.779255477 podStartE2EDuration="1.779255477s" podCreationTimestamp="2026-02-02 09:24:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:24:08.769333044 +0000 UTC m=+1011.703057122" watchObservedRunningTime="2026-02-02 09:24:08.779255477 +0000 UTC m=+1011.712979565" Feb 02 09:24:09 crc kubenswrapper[4764]: I0202 09:24:09.179175 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Feb 02 09:24:09 crc kubenswrapper[4764]: I0202 09:24:09.369617 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-4rrf5" Feb 02 09:24:09 crc kubenswrapper[4764]: I0202 09:24:09.472459 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4b4wn\" (UniqueName: \"kubernetes.io/projected/82d1fbc3-b839-4f16-ad55-bb8a8ce288e1-kube-api-access-4b4wn\") pod \"82d1fbc3-b839-4f16-ad55-bb8a8ce288e1\" (UID: \"82d1fbc3-b839-4f16-ad55-bb8a8ce288e1\") " Feb 02 09:24:09 crc kubenswrapper[4764]: I0202 09:24:09.472577 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/82d1fbc3-b839-4f16-ad55-bb8a8ce288e1-operator-scripts\") pod \"82d1fbc3-b839-4f16-ad55-bb8a8ce288e1\" (UID: \"82d1fbc3-b839-4f16-ad55-bb8a8ce288e1\") " Feb 02 09:24:09 crc kubenswrapper[4764]: I0202 09:24:09.473303 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/82d1fbc3-b839-4f16-ad55-bb8a8ce288e1-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "82d1fbc3-b839-4f16-ad55-bb8a8ce288e1" (UID: "82d1fbc3-b839-4f16-ad55-bb8a8ce288e1"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:24:09 crc kubenswrapper[4764]: I0202 09:24:09.482525 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/82d1fbc3-b839-4f16-ad55-bb8a8ce288e1-kube-api-access-4b4wn" (OuterVolumeSpecName: "kube-api-access-4b4wn") pod "82d1fbc3-b839-4f16-ad55-bb8a8ce288e1" (UID: "82d1fbc3-b839-4f16-ad55-bb8a8ce288e1"). InnerVolumeSpecName "kube-api-access-4b4wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:24:09 crc kubenswrapper[4764]: I0202 09:24:09.576050 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4b4wn\" (UniqueName: \"kubernetes.io/projected/82d1fbc3-b839-4f16-ad55-bb8a8ce288e1-kube-api-access-4b4wn\") on node \"crc\" DevicePath \"\"" Feb 02 09:24:09 crc kubenswrapper[4764]: I0202 09:24:09.576097 4764 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/82d1fbc3-b839-4f16-ad55-bb8a8ce288e1-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 09:24:09 crc kubenswrapper[4764]: I0202 09:24:09.774339 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-4rrf5" event={"ID":"82d1fbc3-b839-4f16-ad55-bb8a8ce288e1","Type":"ContainerDied","Data":"4b45018ab95716f45eee3c9426df2d50bd04b6019589bcd8bc53aea9136cf886"} Feb 02 09:24:09 crc kubenswrapper[4764]: I0202 09:24:09.774459 4764 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4b45018ab95716f45eee3c9426df2d50bd04b6019589bcd8bc53aea9136cf886" Feb 02 09:24:09 crc kubenswrapper[4764]: I0202 09:24:09.774409 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-4rrf5" Feb 02 09:24:09 crc kubenswrapper[4764]: I0202 09:24:09.775907 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-48c4-account-create-update-lvzbb" event={"ID":"9c4c3400-5c5d-4d8b-82a9-c8ca34abe7c0","Type":"ContainerDied","Data":"7e4be9945a21765d2e816b5d7ef51e7f14627efe60631690e6d27f83860286f3"} Feb 02 09:24:09 crc kubenswrapper[4764]: I0202 09:24:09.775874 4764 generic.go:334] "Generic (PLEG): container finished" podID="9c4c3400-5c5d-4d8b-82a9-c8ca34abe7c0" containerID="7e4be9945a21765d2e816b5d7ef51e7f14627efe60631690e6d27f83860286f3" exitCode=0 Feb 02 09:24:09 crc kubenswrapper[4764]: I0202 09:24:09.777670 4764 generic.go:334] "Generic (PLEG): container finished" podID="1a85da62-5c86-4f54-9d14-b80c3974189a" containerID="d62d6a087221f0d57aacf9b33844eb86a5a5c129fd54d5dd9da5ebf675b4c402" exitCode=0 Feb 02 09:24:09 crc kubenswrapper[4764]: I0202 09:24:09.777748 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-2a71-account-create-update-6bfns" event={"ID":"1a85da62-5c86-4f54-9d14-b80c3974189a","Type":"ContainerDied","Data":"d62d6a087221f0d57aacf9b33844eb86a5a5c129fd54d5dd9da5ebf675b4c402"} Feb 02 09:24:10 crc kubenswrapper[4764]: I0202 09:24:10.213792 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-wqk4n" Feb 02 09:24:10 crc kubenswrapper[4764]: I0202 09:24:10.220628 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-48c4-account-create-update-lvzbb" Feb 02 09:24:10 crc kubenswrapper[4764]: I0202 09:24:10.233590 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-2d92n" Feb 02 09:24:10 crc kubenswrapper[4764]: I0202 09:24:10.287690 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b0e0b60e-0289-4700-b747-a6442ca6e34a-operator-scripts\") pod \"b0e0b60e-0289-4700-b747-a6442ca6e34a\" (UID: \"b0e0b60e-0289-4700-b747-a6442ca6e34a\") " Feb 02 09:24:10 crc kubenswrapper[4764]: I0202 09:24:10.287846 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9c4c3400-5c5d-4d8b-82a9-c8ca34abe7c0-operator-scripts\") pod \"9c4c3400-5c5d-4d8b-82a9-c8ca34abe7c0\" (UID: \"9c4c3400-5c5d-4d8b-82a9-c8ca34abe7c0\") " Feb 02 09:24:10 crc kubenswrapper[4764]: I0202 09:24:10.287899 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zggfh\" (UniqueName: \"kubernetes.io/projected/9c4c3400-5c5d-4d8b-82a9-c8ca34abe7c0-kube-api-access-zggfh\") pod \"9c4c3400-5c5d-4d8b-82a9-c8ca34abe7c0\" (UID: \"9c4c3400-5c5d-4d8b-82a9-c8ca34abe7c0\") " Feb 02 09:24:10 crc kubenswrapper[4764]: I0202 09:24:10.287950 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7d99k\" (UniqueName: \"kubernetes.io/projected/a3a49c31-c4c9-415b-89d9-bee375775e38-kube-api-access-7d99k\") pod \"a3a49c31-c4c9-415b-89d9-bee375775e38\" (UID: \"a3a49c31-c4c9-415b-89d9-bee375775e38\") " Feb 02 09:24:10 crc kubenswrapper[4764]: I0202 09:24:10.287976 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a3a49c31-c4c9-415b-89d9-bee375775e38-operator-scripts\") pod \"a3a49c31-c4c9-415b-89d9-bee375775e38\" (UID: \"a3a49c31-c4c9-415b-89d9-bee375775e38\") " Feb 02 09:24:10 crc kubenswrapper[4764]: I0202 09:24:10.288031 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rf9th\" (UniqueName: \"kubernetes.io/projected/b0e0b60e-0289-4700-b747-a6442ca6e34a-kube-api-access-rf9th\") pod \"b0e0b60e-0289-4700-b747-a6442ca6e34a\" (UID: \"b0e0b60e-0289-4700-b747-a6442ca6e34a\") " Feb 02 09:24:10 crc kubenswrapper[4764]: I0202 09:24:10.293723 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b0e0b60e-0289-4700-b747-a6442ca6e34a-kube-api-access-rf9th" (OuterVolumeSpecName: "kube-api-access-rf9th") pod "b0e0b60e-0289-4700-b747-a6442ca6e34a" (UID: "b0e0b60e-0289-4700-b747-a6442ca6e34a"). InnerVolumeSpecName "kube-api-access-rf9th". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:24:10 crc kubenswrapper[4764]: I0202 09:24:10.294163 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c4c3400-5c5d-4d8b-82a9-c8ca34abe7c0-kube-api-access-zggfh" (OuterVolumeSpecName: "kube-api-access-zggfh") pod "9c4c3400-5c5d-4d8b-82a9-c8ca34abe7c0" (UID: "9c4c3400-5c5d-4d8b-82a9-c8ca34abe7c0"). InnerVolumeSpecName "kube-api-access-zggfh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:24:10 crc kubenswrapper[4764]: I0202 09:24:10.294308 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a3a49c31-c4c9-415b-89d9-bee375775e38-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a3a49c31-c4c9-415b-89d9-bee375775e38" (UID: "a3a49c31-c4c9-415b-89d9-bee375775e38"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:24:10 crc kubenswrapper[4764]: I0202 09:24:10.294618 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b0e0b60e-0289-4700-b747-a6442ca6e34a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b0e0b60e-0289-4700-b747-a6442ca6e34a" (UID: "b0e0b60e-0289-4700-b747-a6442ca6e34a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:24:10 crc kubenswrapper[4764]: I0202 09:24:10.294733 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9c4c3400-5c5d-4d8b-82a9-c8ca34abe7c0-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "9c4c3400-5c5d-4d8b-82a9-c8ca34abe7c0" (UID: "9c4c3400-5c5d-4d8b-82a9-c8ca34abe7c0"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:24:10 crc kubenswrapper[4764]: I0202 09:24:10.297207 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a3a49c31-c4c9-415b-89d9-bee375775e38-kube-api-access-7d99k" (OuterVolumeSpecName: "kube-api-access-7d99k") pod "a3a49c31-c4c9-415b-89d9-bee375775e38" (UID: "a3a49c31-c4c9-415b-89d9-bee375775e38"). InnerVolumeSpecName "kube-api-access-7d99k". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:24:10 crc kubenswrapper[4764]: I0202 09:24:10.391720 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zggfh\" (UniqueName: \"kubernetes.io/projected/9c4c3400-5c5d-4d8b-82a9-c8ca34abe7c0-kube-api-access-zggfh\") on node \"crc\" DevicePath \"\"" Feb 02 09:24:10 crc kubenswrapper[4764]: I0202 09:24:10.391749 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7d99k\" (UniqueName: \"kubernetes.io/projected/a3a49c31-c4c9-415b-89d9-bee375775e38-kube-api-access-7d99k\") on node \"crc\" DevicePath \"\"" Feb 02 09:24:10 crc kubenswrapper[4764]: I0202 09:24:10.391759 4764 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a3a49c31-c4c9-415b-89d9-bee375775e38-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 09:24:10 crc kubenswrapper[4764]: I0202 09:24:10.391768 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rf9th\" (UniqueName: \"kubernetes.io/projected/b0e0b60e-0289-4700-b747-a6442ca6e34a-kube-api-access-rf9th\") on node \"crc\" DevicePath \"\"" Feb 02 09:24:10 crc kubenswrapper[4764]: I0202 09:24:10.391780 4764 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b0e0b60e-0289-4700-b747-a6442ca6e34a-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 09:24:10 crc kubenswrapper[4764]: I0202 09:24:10.391788 4764 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9c4c3400-5c5d-4d8b-82a9-c8ca34abe7c0-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 09:24:10 crc kubenswrapper[4764]: I0202 09:24:10.793714 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-2d92n" Feb 02 09:24:10 crc kubenswrapper[4764]: I0202 09:24:10.793821 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-2d92n" event={"ID":"b0e0b60e-0289-4700-b747-a6442ca6e34a","Type":"ContainerDied","Data":"e9bd7d579598242363a720861f2448e473f5579d221b01f61293a761ee817906"} Feb 02 09:24:10 crc kubenswrapper[4764]: I0202 09:24:10.794240 4764 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e9bd7d579598242363a720861f2448e473f5579d221b01f61293a761ee817906" Feb 02 09:24:10 crc kubenswrapper[4764]: I0202 09:24:10.795971 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-wqk4n" event={"ID":"a3a49c31-c4c9-415b-89d9-bee375775e38","Type":"ContainerDied","Data":"ef15e32c0c09f8b8bf041754d1ea709d12bc18f74445463b052d8de8cfeb2a69"} Feb 02 09:24:10 crc kubenswrapper[4764]: I0202 09:24:10.796040 4764 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ef15e32c0c09f8b8bf041754d1ea709d12bc18f74445463b052d8de8cfeb2a69" Feb 02 09:24:10 crc kubenswrapper[4764]: I0202 09:24:10.796117 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-wqk4n" Feb 02 09:24:10 crc kubenswrapper[4764]: I0202 09:24:10.799525 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-48c4-account-create-update-lvzbb" event={"ID":"9c4c3400-5c5d-4d8b-82a9-c8ca34abe7c0","Type":"ContainerDied","Data":"12eb78775474df2bcd9f7e74647a009c5efd052d78eae394ffe87fff61360771"} Feb 02 09:24:10 crc kubenswrapper[4764]: I0202 09:24:10.799555 4764 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="12eb78775474df2bcd9f7e74647a009c5efd052d78eae394ffe87fff61360771" Feb 02 09:24:10 crc kubenswrapper[4764]: I0202 09:24:10.799570 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-48c4-account-create-update-lvzbb" Feb 02 09:24:11 crc kubenswrapper[4764]: I0202 09:24:11.535064 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-2a71-account-create-update-6bfns" Feb 02 09:24:11 crc kubenswrapper[4764]: I0202 09:24:11.616013 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6l24x\" (UniqueName: \"kubernetes.io/projected/1a85da62-5c86-4f54-9d14-b80c3974189a-kube-api-access-6l24x\") pod \"1a85da62-5c86-4f54-9d14-b80c3974189a\" (UID: \"1a85da62-5c86-4f54-9d14-b80c3974189a\") " Feb 02 09:24:11 crc kubenswrapper[4764]: I0202 09:24:11.616703 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1a85da62-5c86-4f54-9d14-b80c3974189a-operator-scripts\") pod \"1a85da62-5c86-4f54-9d14-b80c3974189a\" (UID: \"1a85da62-5c86-4f54-9d14-b80c3974189a\") " Feb 02 09:24:11 crc kubenswrapper[4764]: I0202 09:24:11.617620 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1a85da62-5c86-4f54-9d14-b80c3974189a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "1a85da62-5c86-4f54-9d14-b80c3974189a" (UID: "1a85da62-5c86-4f54-9d14-b80c3974189a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:24:11 crc kubenswrapper[4764]: I0202 09:24:11.625210 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a85da62-5c86-4f54-9d14-b80c3974189a-kube-api-access-6l24x" (OuterVolumeSpecName: "kube-api-access-6l24x") pod "1a85da62-5c86-4f54-9d14-b80c3974189a" (UID: "1a85da62-5c86-4f54-9d14-b80c3974189a"). InnerVolumeSpecName "kube-api-access-6l24x". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:24:11 crc kubenswrapper[4764]: I0202 09:24:11.718992 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6l24x\" (UniqueName: \"kubernetes.io/projected/1a85da62-5c86-4f54-9d14-b80c3974189a-kube-api-access-6l24x\") on node \"crc\" DevicePath \"\"" Feb 02 09:24:11 crc kubenswrapper[4764]: I0202 09:24:11.719061 4764 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1a85da62-5c86-4f54-9d14-b80c3974189a-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 09:24:11 crc kubenswrapper[4764]: I0202 09:24:11.813784 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-2a71-account-create-update-6bfns" event={"ID":"1a85da62-5c86-4f54-9d14-b80c3974189a","Type":"ContainerDied","Data":"d04235a6d9a0b8b0b7f989d59c81216046bf07d18a123399fb9d2a3e1c838a0f"} Feb 02 09:24:11 crc kubenswrapper[4764]: I0202 09:24:11.813845 4764 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d04235a6d9a0b8b0b7f989d59c81216046bf07d18a123399fb9d2a3e1c838a0f" Feb 02 09:24:11 crc kubenswrapper[4764]: I0202 09:24:11.813927 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-2a71-account-create-update-6bfns" Feb 02 09:24:12 crc kubenswrapper[4764]: I0202 09:24:12.739688 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-z55kp"] Feb 02 09:24:12 crc kubenswrapper[4764]: E0202 09:24:12.740742 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82d1fbc3-b839-4f16-ad55-bb8a8ce288e1" containerName="mariadb-account-create-update" Feb 02 09:24:12 crc kubenswrapper[4764]: I0202 09:24:12.740771 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="82d1fbc3-b839-4f16-ad55-bb8a8ce288e1" containerName="mariadb-account-create-update" Feb 02 09:24:12 crc kubenswrapper[4764]: E0202 09:24:12.740812 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a85da62-5c86-4f54-9d14-b80c3974189a" containerName="mariadb-account-create-update" Feb 02 09:24:12 crc kubenswrapper[4764]: I0202 09:24:12.740822 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a85da62-5c86-4f54-9d14-b80c3974189a" containerName="mariadb-account-create-update" Feb 02 09:24:12 crc kubenswrapper[4764]: E0202 09:24:12.740838 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c4c3400-5c5d-4d8b-82a9-c8ca34abe7c0" containerName="mariadb-account-create-update" Feb 02 09:24:12 crc kubenswrapper[4764]: I0202 09:24:12.740847 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c4c3400-5c5d-4d8b-82a9-c8ca34abe7c0" containerName="mariadb-account-create-update" Feb 02 09:24:12 crc kubenswrapper[4764]: E0202 09:24:12.740869 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0e0b60e-0289-4700-b747-a6442ca6e34a" containerName="mariadb-database-create" Feb 02 09:24:12 crc kubenswrapper[4764]: I0202 09:24:12.740880 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0e0b60e-0289-4700-b747-a6442ca6e34a" containerName="mariadb-database-create" Feb 02 09:24:12 crc kubenswrapper[4764]: E0202 09:24:12.740920 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3a49c31-c4c9-415b-89d9-bee375775e38" containerName="mariadb-database-create" Feb 02 09:24:12 crc kubenswrapper[4764]: I0202 09:24:12.740927 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3a49c31-c4c9-415b-89d9-bee375775e38" containerName="mariadb-database-create" Feb 02 09:24:12 crc kubenswrapper[4764]: I0202 09:24:12.741173 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="a3a49c31-c4c9-415b-89d9-bee375775e38" containerName="mariadb-database-create" Feb 02 09:24:12 crc kubenswrapper[4764]: I0202 09:24:12.741199 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="b0e0b60e-0289-4700-b747-a6442ca6e34a" containerName="mariadb-database-create" Feb 02 09:24:12 crc kubenswrapper[4764]: I0202 09:24:12.741223 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a85da62-5c86-4f54-9d14-b80c3974189a" containerName="mariadb-account-create-update" Feb 02 09:24:12 crc kubenswrapper[4764]: I0202 09:24:12.741235 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="82d1fbc3-b839-4f16-ad55-bb8a8ce288e1" containerName="mariadb-account-create-update" Feb 02 09:24:12 crc kubenswrapper[4764]: I0202 09:24:12.741245 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c4c3400-5c5d-4d8b-82a9-c8ca34abe7c0" containerName="mariadb-account-create-update" Feb 02 09:24:12 crc kubenswrapper[4764]: I0202 09:24:12.742089 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-z55kp" Feb 02 09:24:12 crc kubenswrapper[4764]: I0202 09:24:12.751962 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-z55kp"] Feb 02 09:24:12 crc kubenswrapper[4764]: I0202 09:24:12.861304 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d2kwp\" (UniqueName: \"kubernetes.io/projected/52dfc490-3882-4fb5-9c6f-93253a7f794b-kube-api-access-d2kwp\") pod \"glance-db-create-z55kp\" (UID: \"52dfc490-3882-4fb5-9c6f-93253a7f794b\") " pod="openstack/glance-db-create-z55kp" Feb 02 09:24:12 crc kubenswrapper[4764]: I0202 09:24:12.861994 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/52dfc490-3882-4fb5-9c6f-93253a7f794b-operator-scripts\") pod \"glance-db-create-z55kp\" (UID: \"52dfc490-3882-4fb5-9c6f-93253a7f794b\") " pod="openstack/glance-db-create-z55kp" Feb 02 09:24:12 crc kubenswrapper[4764]: I0202 09:24:12.864234 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-a0fd-account-create-update-qgrmb"] Feb 02 09:24:12 crc kubenswrapper[4764]: I0202 09:24:12.868340 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-a0fd-account-create-update-qgrmb" Feb 02 09:24:12 crc kubenswrapper[4764]: I0202 09:24:12.870669 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Feb 02 09:24:12 crc kubenswrapper[4764]: I0202 09:24:12.873554 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-a0fd-account-create-update-qgrmb"] Feb 02 09:24:12 crc kubenswrapper[4764]: I0202 09:24:12.963898 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d2kwp\" (UniqueName: \"kubernetes.io/projected/52dfc490-3882-4fb5-9c6f-93253a7f794b-kube-api-access-d2kwp\") pod \"glance-db-create-z55kp\" (UID: \"52dfc490-3882-4fb5-9c6f-93253a7f794b\") " pod="openstack/glance-db-create-z55kp" Feb 02 09:24:12 crc kubenswrapper[4764]: I0202 09:24:12.964192 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xxm6v\" (UniqueName: \"kubernetes.io/projected/80593eb5-7787-484f-b3a5-8cd856f0a6b9-kube-api-access-xxm6v\") pod \"glance-a0fd-account-create-update-qgrmb\" (UID: \"80593eb5-7787-484f-b3a5-8cd856f0a6b9\") " pod="openstack/glance-a0fd-account-create-update-qgrmb" Feb 02 09:24:12 crc kubenswrapper[4764]: I0202 09:24:12.964387 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/80593eb5-7787-484f-b3a5-8cd856f0a6b9-operator-scripts\") pod \"glance-a0fd-account-create-update-qgrmb\" (UID: \"80593eb5-7787-484f-b3a5-8cd856f0a6b9\") " pod="openstack/glance-a0fd-account-create-update-qgrmb" Feb 02 09:24:12 crc kubenswrapper[4764]: I0202 09:24:12.964518 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/52dfc490-3882-4fb5-9c6f-93253a7f794b-operator-scripts\") pod \"glance-db-create-z55kp\" (UID: \"52dfc490-3882-4fb5-9c6f-93253a7f794b\") " pod="openstack/glance-db-create-z55kp" Feb 02 09:24:12 crc kubenswrapper[4764]: I0202 09:24:12.965258 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/52dfc490-3882-4fb5-9c6f-93253a7f794b-operator-scripts\") pod \"glance-db-create-z55kp\" (UID: \"52dfc490-3882-4fb5-9c6f-93253a7f794b\") " pod="openstack/glance-db-create-z55kp" Feb 02 09:24:12 crc kubenswrapper[4764]: I0202 09:24:12.984523 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d2kwp\" (UniqueName: \"kubernetes.io/projected/52dfc490-3882-4fb5-9c6f-93253a7f794b-kube-api-access-d2kwp\") pod \"glance-db-create-z55kp\" (UID: \"52dfc490-3882-4fb5-9c6f-93253a7f794b\") " pod="openstack/glance-db-create-z55kp" Feb 02 09:24:13 crc kubenswrapper[4764]: I0202 09:24:13.066393 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/80593eb5-7787-484f-b3a5-8cd856f0a6b9-operator-scripts\") pod \"glance-a0fd-account-create-update-qgrmb\" (UID: \"80593eb5-7787-484f-b3a5-8cd856f0a6b9\") " pod="openstack/glance-a0fd-account-create-update-qgrmb" Feb 02 09:24:13 crc kubenswrapper[4764]: I0202 09:24:13.066540 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xxm6v\" (UniqueName: \"kubernetes.io/projected/80593eb5-7787-484f-b3a5-8cd856f0a6b9-kube-api-access-xxm6v\") pod \"glance-a0fd-account-create-update-qgrmb\" (UID: \"80593eb5-7787-484f-b3a5-8cd856f0a6b9\") " pod="openstack/glance-a0fd-account-create-update-qgrmb" Feb 02 09:24:13 crc kubenswrapper[4764]: I0202 09:24:13.067143 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/80593eb5-7787-484f-b3a5-8cd856f0a6b9-operator-scripts\") pod \"glance-a0fd-account-create-update-qgrmb\" (UID: \"80593eb5-7787-484f-b3a5-8cd856f0a6b9\") " pod="openstack/glance-a0fd-account-create-update-qgrmb" Feb 02 09:24:13 crc kubenswrapper[4764]: I0202 09:24:13.083053 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xxm6v\" (UniqueName: \"kubernetes.io/projected/80593eb5-7787-484f-b3a5-8cd856f0a6b9-kube-api-access-xxm6v\") pod \"glance-a0fd-account-create-update-qgrmb\" (UID: \"80593eb5-7787-484f-b3a5-8cd856f0a6b9\") " pod="openstack/glance-a0fd-account-create-update-qgrmb" Feb 02 09:24:13 crc kubenswrapper[4764]: I0202 09:24:13.102652 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-z55kp" Feb 02 09:24:13 crc kubenswrapper[4764]: I0202 09:24:13.183914 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-a0fd-account-create-update-qgrmb" Feb 02 09:24:13 crc kubenswrapper[4764]: I0202 09:24:13.531437 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-z55kp"] Feb 02 09:24:13 crc kubenswrapper[4764]: W0202 09:24:13.537205 4764 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod52dfc490_3882_4fb5_9c6f_93253a7f794b.slice/crio-8f49736691b34d75dc24a506434861a9884b491100d178bfae7c7865877768fd WatchSource:0}: Error finding container 8f49736691b34d75dc24a506434861a9884b491100d178bfae7c7865877768fd: Status 404 returned error can't find the container with id 8f49736691b34d75dc24a506434861a9884b491100d178bfae7c7865877768fd Feb 02 09:24:13 crc kubenswrapper[4764]: I0202 09:24:13.682382 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-a0fd-account-create-update-qgrmb"] Feb 02 09:24:13 crc kubenswrapper[4764]: W0202 09:24:13.687131 4764 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod80593eb5_7787_484f_b3a5_8cd856f0a6b9.slice/crio-f7319ceec1413bff895567066dfed9814978ad418baa5cf3c3a082d14296184c WatchSource:0}: Error finding container f7319ceec1413bff895567066dfed9814978ad418baa5cf3c3a082d14296184c: Status 404 returned error can't find the container with id f7319ceec1413bff895567066dfed9814978ad418baa5cf3c3a082d14296184c Feb 02 09:24:13 crc kubenswrapper[4764]: I0202 09:24:13.846965 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-z55kp" event={"ID":"52dfc490-3882-4fb5-9c6f-93253a7f794b","Type":"ContainerStarted","Data":"eeac11c8826e4af748480f8b6e34d466eccf10b88248d850ef138bc625c07bec"} Feb 02 09:24:13 crc kubenswrapper[4764]: I0202 09:24:13.847006 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-z55kp" event={"ID":"52dfc490-3882-4fb5-9c6f-93253a7f794b","Type":"ContainerStarted","Data":"8f49736691b34d75dc24a506434861a9884b491100d178bfae7c7865877768fd"} Feb 02 09:24:13 crc kubenswrapper[4764]: I0202 09:24:13.848534 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-a0fd-account-create-update-qgrmb" event={"ID":"80593eb5-7787-484f-b3a5-8cd856f0a6b9","Type":"ContainerStarted","Data":"8b7fb4bd5dbf9110d437990b0859eb46aa316b79e92b75be85900e53ef6305d1"} Feb 02 09:24:13 crc kubenswrapper[4764]: I0202 09:24:13.848561 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-a0fd-account-create-update-qgrmb" event={"ID":"80593eb5-7787-484f-b3a5-8cd856f0a6b9","Type":"ContainerStarted","Data":"f7319ceec1413bff895567066dfed9814978ad418baa5cf3c3a082d14296184c"} Feb 02 09:24:13 crc kubenswrapper[4764]: I0202 09:24:13.864783 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-create-z55kp" podStartSLOduration=1.864767783 podStartE2EDuration="1.864767783s" podCreationTimestamp="2026-02-02 09:24:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:24:13.859374171 +0000 UTC m=+1016.793098259" watchObservedRunningTime="2026-02-02 09:24:13.864767783 +0000 UTC m=+1016.798491871" Feb 02 09:24:13 crc kubenswrapper[4764]: I0202 09:24:13.881464 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-a0fd-account-create-update-qgrmb" podStartSLOduration=1.8814515840000001 podStartE2EDuration="1.881451584s" podCreationTimestamp="2026-02-02 09:24:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:24:13.87752081 +0000 UTC m=+1016.811244898" watchObservedRunningTime="2026-02-02 09:24:13.881451584 +0000 UTC m=+1016.815175672" Feb 02 09:24:14 crc kubenswrapper[4764]: I0202 09:24:14.529321 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/root-account-create-update-4rrf5"] Feb 02 09:24:14 crc kubenswrapper[4764]: I0202 09:24:14.535827 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/root-account-create-update-4rrf5"] Feb 02 09:24:14 crc kubenswrapper[4764]: I0202 09:24:14.542276 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/root-account-create-update-rd2c2"] Feb 02 09:24:14 crc kubenswrapper[4764]: I0202 09:24:14.543423 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-rd2c2" Feb 02 09:24:14 crc kubenswrapper[4764]: I0202 09:24:14.545388 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-mariadb-root-db-secret" Feb 02 09:24:14 crc kubenswrapper[4764]: I0202 09:24:14.555840 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-rd2c2"] Feb 02 09:24:14 crc kubenswrapper[4764]: I0202 09:24:14.599331 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q68nk\" (UniqueName: \"kubernetes.io/projected/9a6a24ad-01bc-4766-bb16-b40605a07ae6-kube-api-access-q68nk\") pod \"root-account-create-update-rd2c2\" (UID: \"9a6a24ad-01bc-4766-bb16-b40605a07ae6\") " pod="openstack/root-account-create-update-rd2c2" Feb 02 09:24:14 crc kubenswrapper[4764]: I0202 09:24:14.599684 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9a6a24ad-01bc-4766-bb16-b40605a07ae6-operator-scripts\") pod \"root-account-create-update-rd2c2\" (UID: \"9a6a24ad-01bc-4766-bb16-b40605a07ae6\") " pod="openstack/root-account-create-update-rd2c2" Feb 02 09:24:14 crc kubenswrapper[4764]: I0202 09:24:14.701297 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q68nk\" (UniqueName: \"kubernetes.io/projected/9a6a24ad-01bc-4766-bb16-b40605a07ae6-kube-api-access-q68nk\") pod \"root-account-create-update-rd2c2\" (UID: \"9a6a24ad-01bc-4766-bb16-b40605a07ae6\") " pod="openstack/root-account-create-update-rd2c2" Feb 02 09:24:14 crc kubenswrapper[4764]: I0202 09:24:14.701616 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9a6a24ad-01bc-4766-bb16-b40605a07ae6-operator-scripts\") pod \"root-account-create-update-rd2c2\" (UID: \"9a6a24ad-01bc-4766-bb16-b40605a07ae6\") " pod="openstack/root-account-create-update-rd2c2" Feb 02 09:24:14 crc kubenswrapper[4764]: I0202 09:24:14.702371 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9a6a24ad-01bc-4766-bb16-b40605a07ae6-operator-scripts\") pod \"root-account-create-update-rd2c2\" (UID: \"9a6a24ad-01bc-4766-bb16-b40605a07ae6\") " pod="openstack/root-account-create-update-rd2c2" Feb 02 09:24:14 crc kubenswrapper[4764]: I0202 09:24:14.725786 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q68nk\" (UniqueName: \"kubernetes.io/projected/9a6a24ad-01bc-4766-bb16-b40605a07ae6-kube-api-access-q68nk\") pod \"root-account-create-update-rd2c2\" (UID: \"9a6a24ad-01bc-4766-bb16-b40605a07ae6\") " pod="openstack/root-account-create-update-rd2c2" Feb 02 09:24:14 crc kubenswrapper[4764]: I0202 09:24:14.875700 4764 generic.go:334] "Generic (PLEG): container finished" podID="52dfc490-3882-4fb5-9c6f-93253a7f794b" containerID="eeac11c8826e4af748480f8b6e34d466eccf10b88248d850ef138bc625c07bec" exitCode=0 Feb 02 09:24:14 crc kubenswrapper[4764]: I0202 09:24:14.875803 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-z55kp" event={"ID":"52dfc490-3882-4fb5-9c6f-93253a7f794b","Type":"ContainerDied","Data":"eeac11c8826e4af748480f8b6e34d466eccf10b88248d850ef138bc625c07bec"} Feb 02 09:24:14 crc kubenswrapper[4764]: I0202 09:24:14.881361 4764 generic.go:334] "Generic (PLEG): container finished" podID="80593eb5-7787-484f-b3a5-8cd856f0a6b9" containerID="8b7fb4bd5dbf9110d437990b0859eb46aa316b79e92b75be85900e53ef6305d1" exitCode=0 Feb 02 09:24:14 crc kubenswrapper[4764]: I0202 09:24:14.881450 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-a0fd-account-create-update-qgrmb" event={"ID":"80593eb5-7787-484f-b3a5-8cd856f0a6b9","Type":"ContainerDied","Data":"8b7fb4bd5dbf9110d437990b0859eb46aa316b79e92b75be85900e53ef6305d1"} Feb 02 09:24:14 crc kubenswrapper[4764]: I0202 09:24:14.902261 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-rd2c2" Feb 02 09:24:15 crc kubenswrapper[4764]: I0202 09:24:15.394882 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-rd2c2"] Feb 02 09:24:15 crc kubenswrapper[4764]: I0202 09:24:15.834290 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="82d1fbc3-b839-4f16-ad55-bb8a8ce288e1" path="/var/lib/kubelet/pods/82d1fbc3-b839-4f16-ad55-bb8a8ce288e1/volumes" Feb 02 09:24:15 crc kubenswrapper[4764]: I0202 09:24:15.889203 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-rd2c2" event={"ID":"9a6a24ad-01bc-4766-bb16-b40605a07ae6","Type":"ContainerStarted","Data":"00e343bf0be6a97406e0c63cd0a3dc73dccc3aa1062c62938cb7d13b3623dd5b"} Feb 02 09:24:15 crc kubenswrapper[4764]: I0202 09:24:15.890183 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-rd2c2" event={"ID":"9a6a24ad-01bc-4766-bb16-b40605a07ae6","Type":"ContainerStarted","Data":"a27335e0e1b4a31acb8d7d920e726300d4c81ceeb75f9187fbe782344a89a9bd"} Feb 02 09:24:15 crc kubenswrapper[4764]: I0202 09:24:15.914651 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/root-account-create-update-rd2c2" podStartSLOduration=1.914630675 podStartE2EDuration="1.914630675s" podCreationTimestamp="2026-02-02 09:24:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:24:15.907914067 +0000 UTC m=+1018.841638155" watchObservedRunningTime="2026-02-02 09:24:15.914630675 +0000 UTC m=+1018.848354763" Feb 02 09:24:16 crc kubenswrapper[4764]: I0202 09:24:16.182097 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-a0fd-account-create-update-qgrmb" Feb 02 09:24:16 crc kubenswrapper[4764]: I0202 09:24:16.228521 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/80593eb5-7787-484f-b3a5-8cd856f0a6b9-operator-scripts\") pod \"80593eb5-7787-484f-b3a5-8cd856f0a6b9\" (UID: \"80593eb5-7787-484f-b3a5-8cd856f0a6b9\") " Feb 02 09:24:16 crc kubenswrapper[4764]: I0202 09:24:16.228718 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xxm6v\" (UniqueName: \"kubernetes.io/projected/80593eb5-7787-484f-b3a5-8cd856f0a6b9-kube-api-access-xxm6v\") pod \"80593eb5-7787-484f-b3a5-8cd856f0a6b9\" (UID: \"80593eb5-7787-484f-b3a5-8cd856f0a6b9\") " Feb 02 09:24:16 crc kubenswrapper[4764]: I0202 09:24:16.230852 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/80593eb5-7787-484f-b3a5-8cd856f0a6b9-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "80593eb5-7787-484f-b3a5-8cd856f0a6b9" (UID: "80593eb5-7787-484f-b3a5-8cd856f0a6b9"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:24:16 crc kubenswrapper[4764]: I0202 09:24:16.235376 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/80593eb5-7787-484f-b3a5-8cd856f0a6b9-kube-api-access-xxm6v" (OuterVolumeSpecName: "kube-api-access-xxm6v") pod "80593eb5-7787-484f-b3a5-8cd856f0a6b9" (UID: "80593eb5-7787-484f-b3a5-8cd856f0a6b9"). InnerVolumeSpecName "kube-api-access-xxm6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:24:16 crc kubenswrapper[4764]: I0202 09:24:16.267906 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-z55kp" Feb 02 09:24:16 crc kubenswrapper[4764]: I0202 09:24:16.330033 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d2kwp\" (UniqueName: \"kubernetes.io/projected/52dfc490-3882-4fb5-9c6f-93253a7f794b-kube-api-access-d2kwp\") pod \"52dfc490-3882-4fb5-9c6f-93253a7f794b\" (UID: \"52dfc490-3882-4fb5-9c6f-93253a7f794b\") " Feb 02 09:24:16 crc kubenswrapper[4764]: I0202 09:24:16.330145 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/52dfc490-3882-4fb5-9c6f-93253a7f794b-operator-scripts\") pod \"52dfc490-3882-4fb5-9c6f-93253a7f794b\" (UID: \"52dfc490-3882-4fb5-9c6f-93253a7f794b\") " Feb 02 09:24:16 crc kubenswrapper[4764]: I0202 09:24:16.330547 4764 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/80593eb5-7787-484f-b3a5-8cd856f0a6b9-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 09:24:16 crc kubenswrapper[4764]: I0202 09:24:16.330570 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xxm6v\" (UniqueName: \"kubernetes.io/projected/80593eb5-7787-484f-b3a5-8cd856f0a6b9-kube-api-access-xxm6v\") on node \"crc\" DevicePath \"\"" Feb 02 09:24:16 crc kubenswrapper[4764]: I0202 09:24:16.331060 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/52dfc490-3882-4fb5-9c6f-93253a7f794b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "52dfc490-3882-4fb5-9c6f-93253a7f794b" (UID: "52dfc490-3882-4fb5-9c6f-93253a7f794b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:24:16 crc kubenswrapper[4764]: I0202 09:24:16.334113 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/52dfc490-3882-4fb5-9c6f-93253a7f794b-kube-api-access-d2kwp" (OuterVolumeSpecName: "kube-api-access-d2kwp") pod "52dfc490-3882-4fb5-9c6f-93253a7f794b" (UID: "52dfc490-3882-4fb5-9c6f-93253a7f794b"). InnerVolumeSpecName "kube-api-access-d2kwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:24:16 crc kubenswrapper[4764]: I0202 09:24:16.432364 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d2kwp\" (UniqueName: \"kubernetes.io/projected/52dfc490-3882-4fb5-9c6f-93253a7f794b-kube-api-access-d2kwp\") on node \"crc\" DevicePath \"\"" Feb 02 09:24:16 crc kubenswrapper[4764]: I0202 09:24:16.432391 4764 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/52dfc490-3882-4fb5-9c6f-93253a7f794b-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 09:24:16 crc kubenswrapper[4764]: I0202 09:24:16.900453 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-z55kp" event={"ID":"52dfc490-3882-4fb5-9c6f-93253a7f794b","Type":"ContainerDied","Data":"8f49736691b34d75dc24a506434861a9884b491100d178bfae7c7865877768fd"} Feb 02 09:24:16 crc kubenswrapper[4764]: I0202 09:24:16.900515 4764 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8f49736691b34d75dc24a506434861a9884b491100d178bfae7c7865877768fd" Feb 02 09:24:16 crc kubenswrapper[4764]: I0202 09:24:16.900634 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-z55kp" Feb 02 09:24:16 crc kubenswrapper[4764]: I0202 09:24:16.917686 4764 generic.go:334] "Generic (PLEG): container finished" podID="9a6a24ad-01bc-4766-bb16-b40605a07ae6" containerID="00e343bf0be6a97406e0c63cd0a3dc73dccc3aa1062c62938cb7d13b3623dd5b" exitCode=0 Feb 02 09:24:16 crc kubenswrapper[4764]: I0202 09:24:16.917761 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-rd2c2" event={"ID":"9a6a24ad-01bc-4766-bb16-b40605a07ae6","Type":"ContainerDied","Data":"00e343bf0be6a97406e0c63cd0a3dc73dccc3aa1062c62938cb7d13b3623dd5b"} Feb 02 09:24:16 crc kubenswrapper[4764]: I0202 09:24:16.920831 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-a0fd-account-create-update-qgrmb" event={"ID":"80593eb5-7787-484f-b3a5-8cd856f0a6b9","Type":"ContainerDied","Data":"f7319ceec1413bff895567066dfed9814978ad418baa5cf3c3a082d14296184c"} Feb 02 09:24:16 crc kubenswrapper[4764]: I0202 09:24:16.920898 4764 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f7319ceec1413bff895567066dfed9814978ad418baa5cf3c3a082d14296184c" Feb 02 09:24:16 crc kubenswrapper[4764]: I0202 09:24:16.921027 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-a0fd-account-create-update-qgrmb" Feb 02 09:24:18 crc kubenswrapper[4764]: I0202 09:24:18.126721 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-wzb8q"] Feb 02 09:24:18 crc kubenswrapper[4764]: E0202 09:24:18.127031 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80593eb5-7787-484f-b3a5-8cd856f0a6b9" containerName="mariadb-account-create-update" Feb 02 09:24:18 crc kubenswrapper[4764]: I0202 09:24:18.127043 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="80593eb5-7787-484f-b3a5-8cd856f0a6b9" containerName="mariadb-account-create-update" Feb 02 09:24:18 crc kubenswrapper[4764]: E0202 09:24:18.127060 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52dfc490-3882-4fb5-9c6f-93253a7f794b" containerName="mariadb-database-create" Feb 02 09:24:18 crc kubenswrapper[4764]: I0202 09:24:18.127067 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="52dfc490-3882-4fb5-9c6f-93253a7f794b" containerName="mariadb-database-create" Feb 02 09:24:18 crc kubenswrapper[4764]: I0202 09:24:18.127216 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="52dfc490-3882-4fb5-9c6f-93253a7f794b" containerName="mariadb-database-create" Feb 02 09:24:18 crc kubenswrapper[4764]: I0202 09:24:18.127230 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="80593eb5-7787-484f-b3a5-8cd856f0a6b9" containerName="mariadb-account-create-update" Feb 02 09:24:18 crc kubenswrapper[4764]: I0202 09:24:18.127691 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-wzb8q" Feb 02 09:24:18 crc kubenswrapper[4764]: I0202 09:24:18.133014 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Feb 02 09:24:18 crc kubenswrapper[4764]: I0202 09:24:18.133431 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-snxv4" Feb 02 09:24:18 crc kubenswrapper[4764]: I0202 09:24:18.140822 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-wzb8q"] Feb 02 09:24:18 crc kubenswrapper[4764]: I0202 09:24:18.158904 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nzr8l\" (UniqueName: \"kubernetes.io/projected/b81f3f76-ea90-43bb-a084-c4631433b0da-kube-api-access-nzr8l\") pod \"glance-db-sync-wzb8q\" (UID: \"b81f3f76-ea90-43bb-a084-c4631433b0da\") " pod="openstack/glance-db-sync-wzb8q" Feb 02 09:24:18 crc kubenswrapper[4764]: I0202 09:24:18.158991 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b81f3f76-ea90-43bb-a084-c4631433b0da-combined-ca-bundle\") pod \"glance-db-sync-wzb8q\" (UID: \"b81f3f76-ea90-43bb-a084-c4631433b0da\") " pod="openstack/glance-db-sync-wzb8q" Feb 02 09:24:18 crc kubenswrapper[4764]: I0202 09:24:18.159068 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b81f3f76-ea90-43bb-a084-c4631433b0da-config-data\") pod \"glance-db-sync-wzb8q\" (UID: \"b81f3f76-ea90-43bb-a084-c4631433b0da\") " pod="openstack/glance-db-sync-wzb8q" Feb 02 09:24:18 crc kubenswrapper[4764]: I0202 09:24:18.159094 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b81f3f76-ea90-43bb-a084-c4631433b0da-db-sync-config-data\") pod \"glance-db-sync-wzb8q\" (UID: \"b81f3f76-ea90-43bb-a084-c4631433b0da\") " pod="openstack/glance-db-sync-wzb8q" Feb 02 09:24:18 crc kubenswrapper[4764]: I0202 09:24:18.260671 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nzr8l\" (UniqueName: \"kubernetes.io/projected/b81f3f76-ea90-43bb-a084-c4631433b0da-kube-api-access-nzr8l\") pod \"glance-db-sync-wzb8q\" (UID: \"b81f3f76-ea90-43bb-a084-c4631433b0da\") " pod="openstack/glance-db-sync-wzb8q" Feb 02 09:24:18 crc kubenswrapper[4764]: I0202 09:24:18.260718 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b81f3f76-ea90-43bb-a084-c4631433b0da-combined-ca-bundle\") pod \"glance-db-sync-wzb8q\" (UID: \"b81f3f76-ea90-43bb-a084-c4631433b0da\") " pod="openstack/glance-db-sync-wzb8q" Feb 02 09:24:18 crc kubenswrapper[4764]: I0202 09:24:18.260770 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b81f3f76-ea90-43bb-a084-c4631433b0da-config-data\") pod \"glance-db-sync-wzb8q\" (UID: \"b81f3f76-ea90-43bb-a084-c4631433b0da\") " pod="openstack/glance-db-sync-wzb8q" Feb 02 09:24:18 crc kubenswrapper[4764]: I0202 09:24:18.260791 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b81f3f76-ea90-43bb-a084-c4631433b0da-db-sync-config-data\") pod \"glance-db-sync-wzb8q\" (UID: \"b81f3f76-ea90-43bb-a084-c4631433b0da\") " pod="openstack/glance-db-sync-wzb8q" Feb 02 09:24:18 crc kubenswrapper[4764]: I0202 09:24:18.269952 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b81f3f76-ea90-43bb-a084-c4631433b0da-combined-ca-bundle\") pod \"glance-db-sync-wzb8q\" (UID: \"b81f3f76-ea90-43bb-a084-c4631433b0da\") " pod="openstack/glance-db-sync-wzb8q" Feb 02 09:24:18 crc kubenswrapper[4764]: I0202 09:24:18.269982 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b81f3f76-ea90-43bb-a084-c4631433b0da-db-sync-config-data\") pod \"glance-db-sync-wzb8q\" (UID: \"b81f3f76-ea90-43bb-a084-c4631433b0da\") " pod="openstack/glance-db-sync-wzb8q" Feb 02 09:24:18 crc kubenswrapper[4764]: I0202 09:24:18.271663 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b81f3f76-ea90-43bb-a084-c4631433b0da-config-data\") pod \"glance-db-sync-wzb8q\" (UID: \"b81f3f76-ea90-43bb-a084-c4631433b0da\") " pod="openstack/glance-db-sync-wzb8q" Feb 02 09:24:18 crc kubenswrapper[4764]: I0202 09:24:18.283677 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nzr8l\" (UniqueName: \"kubernetes.io/projected/b81f3f76-ea90-43bb-a084-c4631433b0da-kube-api-access-nzr8l\") pod \"glance-db-sync-wzb8q\" (UID: \"b81f3f76-ea90-43bb-a084-c4631433b0da\") " pod="openstack/glance-db-sync-wzb8q" Feb 02 09:24:18 crc kubenswrapper[4764]: I0202 09:24:18.341667 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-rd2c2" Feb 02 09:24:18 crc kubenswrapper[4764]: I0202 09:24:18.362830 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9a6a24ad-01bc-4766-bb16-b40605a07ae6-operator-scripts\") pod \"9a6a24ad-01bc-4766-bb16-b40605a07ae6\" (UID: \"9a6a24ad-01bc-4766-bb16-b40605a07ae6\") " Feb 02 09:24:18 crc kubenswrapper[4764]: I0202 09:24:18.362885 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q68nk\" (UniqueName: \"kubernetes.io/projected/9a6a24ad-01bc-4766-bb16-b40605a07ae6-kube-api-access-q68nk\") pod \"9a6a24ad-01bc-4766-bb16-b40605a07ae6\" (UID: \"9a6a24ad-01bc-4766-bb16-b40605a07ae6\") " Feb 02 09:24:18 crc kubenswrapper[4764]: I0202 09:24:18.363638 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9a6a24ad-01bc-4766-bb16-b40605a07ae6-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "9a6a24ad-01bc-4766-bb16-b40605a07ae6" (UID: "9a6a24ad-01bc-4766-bb16-b40605a07ae6"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:24:18 crc kubenswrapper[4764]: I0202 09:24:18.370837 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9a6a24ad-01bc-4766-bb16-b40605a07ae6-kube-api-access-q68nk" (OuterVolumeSpecName: "kube-api-access-q68nk") pod "9a6a24ad-01bc-4766-bb16-b40605a07ae6" (UID: "9a6a24ad-01bc-4766-bb16-b40605a07ae6"). InnerVolumeSpecName "kube-api-access-q68nk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:24:18 crc kubenswrapper[4764]: I0202 09:24:18.456305 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-wzb8q" Feb 02 09:24:18 crc kubenswrapper[4764]: I0202 09:24:18.463912 4764 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9a6a24ad-01bc-4766-bb16-b40605a07ae6-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 09:24:18 crc kubenswrapper[4764]: I0202 09:24:18.463948 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q68nk\" (UniqueName: \"kubernetes.io/projected/9a6a24ad-01bc-4766-bb16-b40605a07ae6-kube-api-access-q68nk\") on node \"crc\" DevicePath \"\"" Feb 02 09:24:18 crc kubenswrapper[4764]: I0202 09:24:18.949661 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-rd2c2" event={"ID":"9a6a24ad-01bc-4766-bb16-b40605a07ae6","Type":"ContainerDied","Data":"a27335e0e1b4a31acb8d7d920e726300d4c81ceeb75f9187fbe782344a89a9bd"} Feb 02 09:24:18 crc kubenswrapper[4764]: I0202 09:24:18.949695 4764 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a27335e0e1b4a31acb8d7d920e726300d4c81ceeb75f9187fbe782344a89a9bd" Feb 02 09:24:18 crc kubenswrapper[4764]: I0202 09:24:18.949712 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-rd2c2" Feb 02 09:24:18 crc kubenswrapper[4764]: I0202 09:24:18.992957 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-wzb8q"] Feb 02 09:24:18 crc kubenswrapper[4764]: W0202 09:24:18.995854 4764 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb81f3f76_ea90_43bb_a084_c4631433b0da.slice/crio-f0d44e8f7440499940cc75a57adb0d14cd94cb4b96c47a7bb998363a14532669 WatchSource:0}: Error finding container f0d44e8f7440499940cc75a57adb0d14cd94cb4b96c47a7bb998363a14532669: Status 404 returned error can't find the container with id f0d44e8f7440499940cc75a57adb0d14cd94cb4b96c47a7bb998363a14532669 Feb 02 09:24:19 crc kubenswrapper[4764]: I0202 09:24:19.962498 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-wzb8q" event={"ID":"b81f3f76-ea90-43bb-a084-c4631433b0da","Type":"ContainerStarted","Data":"f0d44e8f7440499940cc75a57adb0d14cd94cb4b96c47a7bb998363a14532669"} Feb 02 09:24:20 crc kubenswrapper[4764]: I0202 09:24:20.842686 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Feb 02 09:24:28 crc kubenswrapper[4764]: I0202 09:24:28.025139 4764 generic.go:334] "Generic (PLEG): container finished" podID="03d9f090-4a93-4226-8fce-6fd0a425d929" containerID="d08998db15d7c2f061c9aa78b783e5e53996e688590d41367d09011c4224893d" exitCode=0 Feb 02 09:24:28 crc kubenswrapper[4764]: I0202 09:24:28.025232 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"03d9f090-4a93-4226-8fce-6fd0a425d929","Type":"ContainerDied","Data":"d08998db15d7c2f061c9aa78b783e5e53996e688590d41367d09011c4224893d"} Feb 02 09:24:28 crc kubenswrapper[4764]: I0202 09:24:28.030277 4764 generic.go:334] "Generic (PLEG): container finished" podID="53d5d300-11ae-408d-8bb8-347defb93991" containerID="f5641821f6b2d966369e040f6849f892c7244535560aa2d911c4599545382008" exitCode=0 Feb 02 09:24:28 crc kubenswrapper[4764]: I0202 09:24:28.030315 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"53d5d300-11ae-408d-8bb8-347defb93991","Type":"ContainerDied","Data":"f5641821f6b2d966369e040f6849f892c7244535560aa2d911c4599545382008"} Feb 02 09:24:28 crc kubenswrapper[4764]: I0202 09:24:28.344554 4764 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-6v7f6" podUID="66662206-d732-404d-a81d-6303d9947d60" containerName="ovn-controller" probeResult="failure" output=< Feb 02 09:24:28 crc kubenswrapper[4764]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Feb 02 09:24:28 crc kubenswrapper[4764]: > Feb 02 09:24:28 crc kubenswrapper[4764]: I0202 09:24:28.494829 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-2p2bk" Feb 02 09:24:28 crc kubenswrapper[4764]: I0202 09:24:28.502661 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-2p2bk" Feb 02 09:24:28 crc kubenswrapper[4764]: I0202 09:24:28.706759 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-6v7f6-config-ftnmq"] Feb 02 09:24:28 crc kubenswrapper[4764]: E0202 09:24:28.708432 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a6a24ad-01bc-4766-bb16-b40605a07ae6" containerName="mariadb-account-create-update" Feb 02 09:24:28 crc kubenswrapper[4764]: I0202 09:24:28.708488 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a6a24ad-01bc-4766-bb16-b40605a07ae6" containerName="mariadb-account-create-update" Feb 02 09:24:28 crc kubenswrapper[4764]: I0202 09:24:28.708639 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a6a24ad-01bc-4766-bb16-b40605a07ae6" containerName="mariadb-account-create-update" Feb 02 09:24:28 crc kubenswrapper[4764]: I0202 09:24:28.735389 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-6v7f6-config-ftnmq"] Feb 02 09:24:28 crc kubenswrapper[4764]: I0202 09:24:28.735497 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-6v7f6-config-ftnmq" Feb 02 09:24:28 crc kubenswrapper[4764]: I0202 09:24:28.737582 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Feb 02 09:24:28 crc kubenswrapper[4764]: I0202 09:24:28.858551 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/459dcb22-63e0-487d-bcf5-ece077d0b977-scripts\") pod \"ovn-controller-6v7f6-config-ftnmq\" (UID: \"459dcb22-63e0-487d-bcf5-ece077d0b977\") " pod="openstack/ovn-controller-6v7f6-config-ftnmq" Feb 02 09:24:28 crc kubenswrapper[4764]: I0202 09:24:28.858594 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/459dcb22-63e0-487d-bcf5-ece077d0b977-additional-scripts\") pod \"ovn-controller-6v7f6-config-ftnmq\" (UID: \"459dcb22-63e0-487d-bcf5-ece077d0b977\") " pod="openstack/ovn-controller-6v7f6-config-ftnmq" Feb 02 09:24:28 crc kubenswrapper[4764]: I0202 09:24:28.858622 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/459dcb22-63e0-487d-bcf5-ece077d0b977-var-run\") pod \"ovn-controller-6v7f6-config-ftnmq\" (UID: \"459dcb22-63e0-487d-bcf5-ece077d0b977\") " pod="openstack/ovn-controller-6v7f6-config-ftnmq" Feb 02 09:24:28 crc kubenswrapper[4764]: I0202 09:24:28.858660 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/459dcb22-63e0-487d-bcf5-ece077d0b977-var-log-ovn\") pod \"ovn-controller-6v7f6-config-ftnmq\" (UID: \"459dcb22-63e0-487d-bcf5-ece077d0b977\") " pod="openstack/ovn-controller-6v7f6-config-ftnmq" Feb 02 09:24:28 crc kubenswrapper[4764]: I0202 09:24:28.858686 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-msj9q\" (UniqueName: \"kubernetes.io/projected/459dcb22-63e0-487d-bcf5-ece077d0b977-kube-api-access-msj9q\") pod \"ovn-controller-6v7f6-config-ftnmq\" (UID: \"459dcb22-63e0-487d-bcf5-ece077d0b977\") " pod="openstack/ovn-controller-6v7f6-config-ftnmq" Feb 02 09:24:28 crc kubenswrapper[4764]: I0202 09:24:28.858765 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/459dcb22-63e0-487d-bcf5-ece077d0b977-var-run-ovn\") pod \"ovn-controller-6v7f6-config-ftnmq\" (UID: \"459dcb22-63e0-487d-bcf5-ece077d0b977\") " pod="openstack/ovn-controller-6v7f6-config-ftnmq" Feb 02 09:24:28 crc kubenswrapper[4764]: I0202 09:24:28.959728 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/459dcb22-63e0-487d-bcf5-ece077d0b977-var-run\") pod \"ovn-controller-6v7f6-config-ftnmq\" (UID: \"459dcb22-63e0-487d-bcf5-ece077d0b977\") " pod="openstack/ovn-controller-6v7f6-config-ftnmq" Feb 02 09:24:28 crc kubenswrapper[4764]: I0202 09:24:28.959805 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/459dcb22-63e0-487d-bcf5-ece077d0b977-var-log-ovn\") pod \"ovn-controller-6v7f6-config-ftnmq\" (UID: \"459dcb22-63e0-487d-bcf5-ece077d0b977\") " pod="openstack/ovn-controller-6v7f6-config-ftnmq" Feb 02 09:24:28 crc kubenswrapper[4764]: I0202 09:24:28.959845 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-msj9q\" (UniqueName: \"kubernetes.io/projected/459dcb22-63e0-487d-bcf5-ece077d0b977-kube-api-access-msj9q\") pod \"ovn-controller-6v7f6-config-ftnmq\" (UID: \"459dcb22-63e0-487d-bcf5-ece077d0b977\") " pod="openstack/ovn-controller-6v7f6-config-ftnmq" Feb 02 09:24:28 crc kubenswrapper[4764]: I0202 09:24:28.959893 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/459dcb22-63e0-487d-bcf5-ece077d0b977-var-run-ovn\") pod \"ovn-controller-6v7f6-config-ftnmq\" (UID: \"459dcb22-63e0-487d-bcf5-ece077d0b977\") " pod="openstack/ovn-controller-6v7f6-config-ftnmq" Feb 02 09:24:28 crc kubenswrapper[4764]: I0202 09:24:28.960020 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/459dcb22-63e0-487d-bcf5-ece077d0b977-scripts\") pod \"ovn-controller-6v7f6-config-ftnmq\" (UID: \"459dcb22-63e0-487d-bcf5-ece077d0b977\") " pod="openstack/ovn-controller-6v7f6-config-ftnmq" Feb 02 09:24:28 crc kubenswrapper[4764]: I0202 09:24:28.960051 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/459dcb22-63e0-487d-bcf5-ece077d0b977-additional-scripts\") pod \"ovn-controller-6v7f6-config-ftnmq\" (UID: \"459dcb22-63e0-487d-bcf5-ece077d0b977\") " pod="openstack/ovn-controller-6v7f6-config-ftnmq" Feb 02 09:24:28 crc kubenswrapper[4764]: I0202 09:24:28.961169 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/459dcb22-63e0-487d-bcf5-ece077d0b977-additional-scripts\") pod \"ovn-controller-6v7f6-config-ftnmq\" (UID: \"459dcb22-63e0-487d-bcf5-ece077d0b977\") " pod="openstack/ovn-controller-6v7f6-config-ftnmq" Feb 02 09:24:28 crc kubenswrapper[4764]: I0202 09:24:28.961525 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/459dcb22-63e0-487d-bcf5-ece077d0b977-var-log-ovn\") pod \"ovn-controller-6v7f6-config-ftnmq\" (UID: \"459dcb22-63e0-487d-bcf5-ece077d0b977\") " pod="openstack/ovn-controller-6v7f6-config-ftnmq" Feb 02 09:24:28 crc kubenswrapper[4764]: I0202 09:24:28.962074 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/459dcb22-63e0-487d-bcf5-ece077d0b977-var-run\") pod \"ovn-controller-6v7f6-config-ftnmq\" (UID: \"459dcb22-63e0-487d-bcf5-ece077d0b977\") " pod="openstack/ovn-controller-6v7f6-config-ftnmq" Feb 02 09:24:28 crc kubenswrapper[4764]: I0202 09:24:28.963677 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/459dcb22-63e0-487d-bcf5-ece077d0b977-scripts\") pod \"ovn-controller-6v7f6-config-ftnmq\" (UID: \"459dcb22-63e0-487d-bcf5-ece077d0b977\") " pod="openstack/ovn-controller-6v7f6-config-ftnmq" Feb 02 09:24:28 crc kubenswrapper[4764]: I0202 09:24:28.963736 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/459dcb22-63e0-487d-bcf5-ece077d0b977-var-run-ovn\") pod \"ovn-controller-6v7f6-config-ftnmq\" (UID: \"459dcb22-63e0-487d-bcf5-ece077d0b977\") " pod="openstack/ovn-controller-6v7f6-config-ftnmq" Feb 02 09:24:29 crc kubenswrapper[4764]: I0202 09:24:29.001239 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-msj9q\" (UniqueName: \"kubernetes.io/projected/459dcb22-63e0-487d-bcf5-ece077d0b977-kube-api-access-msj9q\") pod \"ovn-controller-6v7f6-config-ftnmq\" (UID: \"459dcb22-63e0-487d-bcf5-ece077d0b977\") " pod="openstack/ovn-controller-6v7f6-config-ftnmq" Feb 02 09:24:29 crc kubenswrapper[4764]: I0202 09:24:29.056453 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-6v7f6-config-ftnmq" Feb 02 09:24:31 crc kubenswrapper[4764]: I0202 09:24:31.446006 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-6v7f6-config-ftnmq"] Feb 02 09:24:32 crc kubenswrapper[4764]: I0202 09:24:32.086149 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-wzb8q" event={"ID":"b81f3f76-ea90-43bb-a084-c4631433b0da","Type":"ContainerStarted","Data":"5543885f98fc6fc86034790d4953d84f4d1c24207aeca5e8f52d8e0eb431f891"} Feb 02 09:24:32 crc kubenswrapper[4764]: I0202 09:24:32.089420 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"03d9f090-4a93-4226-8fce-6fd0a425d929","Type":"ContainerStarted","Data":"b0556eef8b622b39f414f040258f94501b1ff75d65d0e71032e333b6b15e8b4f"} Feb 02 09:24:32 crc kubenswrapper[4764]: I0202 09:24:32.089664 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Feb 02 09:24:32 crc kubenswrapper[4764]: I0202 09:24:32.090990 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"53d5d300-11ae-408d-8bb8-347defb93991","Type":"ContainerStarted","Data":"352eac9c43f64896394cf8c78ba1f8a36be490958f2022459a08cd79e7b607ea"} Feb 02 09:24:32 crc kubenswrapper[4764]: I0202 09:24:32.091442 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Feb 02 09:24:32 crc kubenswrapper[4764]: I0202 09:24:32.092160 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-6v7f6-config-ftnmq" event={"ID":"459dcb22-63e0-487d-bcf5-ece077d0b977","Type":"ContainerStarted","Data":"ceb8422566de0c1e2db3ceb17cfb3eb09bd59d5db0c9758695b7c739686570c0"} Feb 02 09:24:32 crc kubenswrapper[4764]: I0202 09:24:32.092187 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-6v7f6-config-ftnmq" event={"ID":"459dcb22-63e0-487d-bcf5-ece077d0b977","Type":"ContainerStarted","Data":"905cc9297238842b49aecd1420263ef11080c0b4071217a0e43f5b20cc4cb741"} Feb 02 09:24:32 crc kubenswrapper[4764]: I0202 09:24:32.110478 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-wzb8q" podStartSLOduration=2.020308325 podStartE2EDuration="14.110454956s" podCreationTimestamp="2026-02-02 09:24:18 +0000 UTC" firstStartedPulling="2026-02-02 09:24:18.998463525 +0000 UTC m=+1021.932187623" lastFinishedPulling="2026-02-02 09:24:31.088610166 +0000 UTC m=+1034.022334254" observedRunningTime="2026-02-02 09:24:32.101124969 +0000 UTC m=+1035.034849077" watchObservedRunningTime="2026-02-02 09:24:32.110454956 +0000 UTC m=+1035.044179044" Feb 02 09:24:32 crc kubenswrapper[4764]: I0202 09:24:32.128256 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=40.59044395 podStartE2EDuration="1m20.128235995s" podCreationTimestamp="2026-02-02 09:23:12 +0000 UTC" firstStartedPulling="2026-02-02 09:23:14.830608715 +0000 UTC m=+957.764332803" lastFinishedPulling="2026-02-02 09:23:54.36840076 +0000 UTC m=+997.302124848" observedRunningTime="2026-02-02 09:24:32.126544201 +0000 UTC m=+1035.060268309" watchObservedRunningTime="2026-02-02 09:24:32.128235995 +0000 UTC m=+1035.061960083" Feb 02 09:24:32 crc kubenswrapper[4764]: I0202 09:24:32.158256 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=40.003263838 podStartE2EDuration="1m19.158241638s" podCreationTimestamp="2026-02-02 09:23:13 +0000 UTC" firstStartedPulling="2026-02-02 09:23:15.204547785 +0000 UTC m=+958.138271873" lastFinishedPulling="2026-02-02 09:23:54.359525585 +0000 UTC m=+997.293249673" observedRunningTime="2026-02-02 09:24:32.153869853 +0000 UTC m=+1035.087593941" watchObservedRunningTime="2026-02-02 09:24:32.158241638 +0000 UTC m=+1035.091965726" Feb 02 09:24:33 crc kubenswrapper[4764]: I0202 09:24:33.100972 4764 generic.go:334] "Generic (PLEG): container finished" podID="459dcb22-63e0-487d-bcf5-ece077d0b977" containerID="ceb8422566de0c1e2db3ceb17cfb3eb09bd59d5db0c9758695b7c739686570c0" exitCode=0 Feb 02 09:24:33 crc kubenswrapper[4764]: I0202 09:24:33.101074 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-6v7f6-config-ftnmq" event={"ID":"459dcb22-63e0-487d-bcf5-ece077d0b977","Type":"ContainerDied","Data":"ceb8422566de0c1e2db3ceb17cfb3eb09bd59d5db0c9758695b7c739686570c0"} Feb 02 09:24:33 crc kubenswrapper[4764]: I0202 09:24:33.395789 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-6v7f6" Feb 02 09:24:33 crc kubenswrapper[4764]: I0202 09:24:33.489767 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-6v7f6-config-ftnmq" Feb 02 09:24:33 crc kubenswrapper[4764]: I0202 09:24:33.566999 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/459dcb22-63e0-487d-bcf5-ece077d0b977-var-log-ovn\") pod \"459dcb22-63e0-487d-bcf5-ece077d0b977\" (UID: \"459dcb22-63e0-487d-bcf5-ece077d0b977\") " Feb 02 09:24:33 crc kubenswrapper[4764]: I0202 09:24:33.567038 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/459dcb22-63e0-487d-bcf5-ece077d0b977-var-run\") pod \"459dcb22-63e0-487d-bcf5-ece077d0b977\" (UID: \"459dcb22-63e0-487d-bcf5-ece077d0b977\") " Feb 02 09:24:33 crc kubenswrapper[4764]: I0202 09:24:33.567071 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/459dcb22-63e0-487d-bcf5-ece077d0b977-scripts\") pod \"459dcb22-63e0-487d-bcf5-ece077d0b977\" (UID: \"459dcb22-63e0-487d-bcf5-ece077d0b977\") " Feb 02 09:24:33 crc kubenswrapper[4764]: I0202 09:24:33.567106 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-msj9q\" (UniqueName: \"kubernetes.io/projected/459dcb22-63e0-487d-bcf5-ece077d0b977-kube-api-access-msj9q\") pod \"459dcb22-63e0-487d-bcf5-ece077d0b977\" (UID: \"459dcb22-63e0-487d-bcf5-ece077d0b977\") " Feb 02 09:24:33 crc kubenswrapper[4764]: I0202 09:24:33.567165 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/459dcb22-63e0-487d-bcf5-ece077d0b977-var-run-ovn\") pod \"459dcb22-63e0-487d-bcf5-ece077d0b977\" (UID: \"459dcb22-63e0-487d-bcf5-ece077d0b977\") " Feb 02 09:24:33 crc kubenswrapper[4764]: I0202 09:24:33.567254 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/459dcb22-63e0-487d-bcf5-ece077d0b977-additional-scripts\") pod \"459dcb22-63e0-487d-bcf5-ece077d0b977\" (UID: \"459dcb22-63e0-487d-bcf5-ece077d0b977\") " Feb 02 09:24:33 crc kubenswrapper[4764]: I0202 09:24:33.567843 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/459dcb22-63e0-487d-bcf5-ece077d0b977-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "459dcb22-63e0-487d-bcf5-ece077d0b977" (UID: "459dcb22-63e0-487d-bcf5-ece077d0b977"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 09:24:33 crc kubenswrapper[4764]: I0202 09:24:33.567904 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/459dcb22-63e0-487d-bcf5-ece077d0b977-var-run" (OuterVolumeSpecName: "var-run") pod "459dcb22-63e0-487d-bcf5-ece077d0b977" (UID: "459dcb22-63e0-487d-bcf5-ece077d0b977"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 09:24:33 crc kubenswrapper[4764]: I0202 09:24:33.568194 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/459dcb22-63e0-487d-bcf5-ece077d0b977-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "459dcb22-63e0-487d-bcf5-ece077d0b977" (UID: "459dcb22-63e0-487d-bcf5-ece077d0b977"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:24:33 crc kubenswrapper[4764]: I0202 09:24:33.568838 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/459dcb22-63e0-487d-bcf5-ece077d0b977-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "459dcb22-63e0-487d-bcf5-ece077d0b977" (UID: "459dcb22-63e0-487d-bcf5-ece077d0b977"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 09:24:33 crc kubenswrapper[4764]: I0202 09:24:33.568902 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/459dcb22-63e0-487d-bcf5-ece077d0b977-scripts" (OuterVolumeSpecName: "scripts") pod "459dcb22-63e0-487d-bcf5-ece077d0b977" (UID: "459dcb22-63e0-487d-bcf5-ece077d0b977"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:24:33 crc kubenswrapper[4764]: I0202 09:24:33.576807 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/459dcb22-63e0-487d-bcf5-ece077d0b977-kube-api-access-msj9q" (OuterVolumeSpecName: "kube-api-access-msj9q") pod "459dcb22-63e0-487d-bcf5-ece077d0b977" (UID: "459dcb22-63e0-487d-bcf5-ece077d0b977"). InnerVolumeSpecName "kube-api-access-msj9q". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:24:33 crc kubenswrapper[4764]: I0202 09:24:33.669078 4764 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/459dcb22-63e0-487d-bcf5-ece077d0b977-additional-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 09:24:33 crc kubenswrapper[4764]: I0202 09:24:33.669110 4764 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/459dcb22-63e0-487d-bcf5-ece077d0b977-var-log-ovn\") on node \"crc\" DevicePath \"\"" Feb 02 09:24:33 crc kubenswrapper[4764]: I0202 09:24:33.669120 4764 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/459dcb22-63e0-487d-bcf5-ece077d0b977-var-run\") on node \"crc\" DevicePath \"\"" Feb 02 09:24:33 crc kubenswrapper[4764]: I0202 09:24:33.669128 4764 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/459dcb22-63e0-487d-bcf5-ece077d0b977-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 09:24:33 crc kubenswrapper[4764]: I0202 09:24:33.669140 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-msj9q\" (UniqueName: \"kubernetes.io/projected/459dcb22-63e0-487d-bcf5-ece077d0b977-kube-api-access-msj9q\") on node \"crc\" DevicePath \"\"" Feb 02 09:24:33 crc kubenswrapper[4764]: I0202 09:24:33.669154 4764 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/459dcb22-63e0-487d-bcf5-ece077d0b977-var-run-ovn\") on node \"crc\" DevicePath \"\"" Feb 02 09:24:34 crc kubenswrapper[4764]: I0202 09:24:34.109833 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-6v7f6-config-ftnmq" event={"ID":"459dcb22-63e0-487d-bcf5-ece077d0b977","Type":"ContainerDied","Data":"905cc9297238842b49aecd1420263ef11080c0b4071217a0e43f5b20cc4cb741"} Feb 02 09:24:34 crc kubenswrapper[4764]: I0202 09:24:34.109877 4764 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="905cc9297238842b49aecd1420263ef11080c0b4071217a0e43f5b20cc4cb741" Feb 02 09:24:34 crc kubenswrapper[4764]: I0202 09:24:34.109950 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-6v7f6-config-ftnmq" Feb 02 09:24:34 crc kubenswrapper[4764]: I0202 09:24:34.643422 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-6v7f6-config-ftnmq"] Feb 02 09:24:34 crc kubenswrapper[4764]: I0202 09:24:34.651899 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-6v7f6-config-ftnmq"] Feb 02 09:24:35 crc kubenswrapper[4764]: I0202 09:24:35.838188 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="459dcb22-63e0-487d-bcf5-ece077d0b977" path="/var/lib/kubelet/pods/459dcb22-63e0-487d-bcf5-ece077d0b977/volumes" Feb 02 09:24:39 crc kubenswrapper[4764]: I0202 09:24:39.146158 4764 generic.go:334] "Generic (PLEG): container finished" podID="b81f3f76-ea90-43bb-a084-c4631433b0da" containerID="5543885f98fc6fc86034790d4953d84f4d1c24207aeca5e8f52d8e0eb431f891" exitCode=0 Feb 02 09:24:39 crc kubenswrapper[4764]: I0202 09:24:39.146232 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-wzb8q" event={"ID":"b81f3f76-ea90-43bb-a084-c4631433b0da","Type":"ContainerDied","Data":"5543885f98fc6fc86034790d4953d84f4d1c24207aeca5e8f52d8e0eb431f891"} Feb 02 09:24:40 crc kubenswrapper[4764]: I0202 09:24:40.507576 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-wzb8q" Feb 02 09:24:40 crc kubenswrapper[4764]: I0202 09:24:40.580160 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b81f3f76-ea90-43bb-a084-c4631433b0da-combined-ca-bundle\") pod \"b81f3f76-ea90-43bb-a084-c4631433b0da\" (UID: \"b81f3f76-ea90-43bb-a084-c4631433b0da\") " Feb 02 09:24:40 crc kubenswrapper[4764]: I0202 09:24:40.580260 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b81f3f76-ea90-43bb-a084-c4631433b0da-db-sync-config-data\") pod \"b81f3f76-ea90-43bb-a084-c4631433b0da\" (UID: \"b81f3f76-ea90-43bb-a084-c4631433b0da\") " Feb 02 09:24:40 crc kubenswrapper[4764]: I0202 09:24:40.580290 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzr8l\" (UniqueName: \"kubernetes.io/projected/b81f3f76-ea90-43bb-a084-c4631433b0da-kube-api-access-nzr8l\") pod \"b81f3f76-ea90-43bb-a084-c4631433b0da\" (UID: \"b81f3f76-ea90-43bb-a084-c4631433b0da\") " Feb 02 09:24:40 crc kubenswrapper[4764]: I0202 09:24:40.580377 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b81f3f76-ea90-43bb-a084-c4631433b0da-config-data\") pod \"b81f3f76-ea90-43bb-a084-c4631433b0da\" (UID: \"b81f3f76-ea90-43bb-a084-c4631433b0da\") " Feb 02 09:24:40 crc kubenswrapper[4764]: I0202 09:24:40.585706 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b81f3f76-ea90-43bb-a084-c4631433b0da-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "b81f3f76-ea90-43bb-a084-c4631433b0da" (UID: "b81f3f76-ea90-43bb-a084-c4631433b0da"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:24:40 crc kubenswrapper[4764]: I0202 09:24:40.591858 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b81f3f76-ea90-43bb-a084-c4631433b0da-kube-api-access-nzr8l" (OuterVolumeSpecName: "kube-api-access-nzr8l") pod "b81f3f76-ea90-43bb-a084-c4631433b0da" (UID: "b81f3f76-ea90-43bb-a084-c4631433b0da"). InnerVolumeSpecName "kube-api-access-nzr8l". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:24:40 crc kubenswrapper[4764]: I0202 09:24:40.602570 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b81f3f76-ea90-43bb-a084-c4631433b0da-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b81f3f76-ea90-43bb-a084-c4631433b0da" (UID: "b81f3f76-ea90-43bb-a084-c4631433b0da"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:24:40 crc kubenswrapper[4764]: I0202 09:24:40.618964 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b81f3f76-ea90-43bb-a084-c4631433b0da-config-data" (OuterVolumeSpecName: "config-data") pod "b81f3f76-ea90-43bb-a084-c4631433b0da" (UID: "b81f3f76-ea90-43bb-a084-c4631433b0da"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:24:40 crc kubenswrapper[4764]: I0202 09:24:40.685572 4764 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b81f3f76-ea90-43bb-a084-c4631433b0da-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 09:24:40 crc kubenswrapper[4764]: I0202 09:24:40.685641 4764 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b81f3f76-ea90-43bb-a084-c4631433b0da-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 09:24:40 crc kubenswrapper[4764]: I0202 09:24:40.685661 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzr8l\" (UniqueName: \"kubernetes.io/projected/b81f3f76-ea90-43bb-a084-c4631433b0da-kube-api-access-nzr8l\") on node \"crc\" DevicePath \"\"" Feb 02 09:24:40 crc kubenswrapper[4764]: I0202 09:24:40.685680 4764 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b81f3f76-ea90-43bb-a084-c4631433b0da-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 09:24:41 crc kubenswrapper[4764]: I0202 09:24:41.164480 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-wzb8q" event={"ID":"b81f3f76-ea90-43bb-a084-c4631433b0da","Type":"ContainerDied","Data":"f0d44e8f7440499940cc75a57adb0d14cd94cb4b96c47a7bb998363a14532669"} Feb 02 09:24:41 crc kubenswrapper[4764]: I0202 09:24:41.164524 4764 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f0d44e8f7440499940cc75a57adb0d14cd94cb4b96c47a7bb998363a14532669" Feb 02 09:24:41 crc kubenswrapper[4764]: I0202 09:24:41.164536 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-wzb8q" Feb 02 09:24:41 crc kubenswrapper[4764]: I0202 09:24:41.566067 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-58976fdf49-bb6rk"] Feb 02 09:24:41 crc kubenswrapper[4764]: E0202 09:24:41.566561 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="459dcb22-63e0-487d-bcf5-ece077d0b977" containerName="ovn-config" Feb 02 09:24:41 crc kubenswrapper[4764]: I0202 09:24:41.566572 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="459dcb22-63e0-487d-bcf5-ece077d0b977" containerName="ovn-config" Feb 02 09:24:41 crc kubenswrapper[4764]: E0202 09:24:41.566594 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b81f3f76-ea90-43bb-a084-c4631433b0da" containerName="glance-db-sync" Feb 02 09:24:41 crc kubenswrapper[4764]: I0202 09:24:41.566600 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="b81f3f76-ea90-43bb-a084-c4631433b0da" containerName="glance-db-sync" Feb 02 09:24:41 crc kubenswrapper[4764]: I0202 09:24:41.566741 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="b81f3f76-ea90-43bb-a084-c4631433b0da" containerName="glance-db-sync" Feb 02 09:24:41 crc kubenswrapper[4764]: I0202 09:24:41.566753 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="459dcb22-63e0-487d-bcf5-ece077d0b977" containerName="ovn-config" Feb 02 09:24:41 crc kubenswrapper[4764]: I0202 09:24:41.567490 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58976fdf49-bb6rk" Feb 02 09:24:41 crc kubenswrapper[4764]: I0202 09:24:41.587139 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-58976fdf49-bb6rk"] Feb 02 09:24:41 crc kubenswrapper[4764]: I0202 09:24:41.716735 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cbcg4\" (UniqueName: \"kubernetes.io/projected/30d22725-a706-485b-8cc2-964b86137ac1-kube-api-access-cbcg4\") pod \"dnsmasq-dns-58976fdf49-bb6rk\" (UID: \"30d22725-a706-485b-8cc2-964b86137ac1\") " pod="openstack/dnsmasq-dns-58976fdf49-bb6rk" Feb 02 09:24:41 crc kubenswrapper[4764]: I0202 09:24:41.716783 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/30d22725-a706-485b-8cc2-964b86137ac1-ovsdbserver-sb\") pod \"dnsmasq-dns-58976fdf49-bb6rk\" (UID: \"30d22725-a706-485b-8cc2-964b86137ac1\") " pod="openstack/dnsmasq-dns-58976fdf49-bb6rk" Feb 02 09:24:41 crc kubenswrapper[4764]: I0202 09:24:41.716814 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/30d22725-a706-485b-8cc2-964b86137ac1-ovsdbserver-nb\") pod \"dnsmasq-dns-58976fdf49-bb6rk\" (UID: \"30d22725-a706-485b-8cc2-964b86137ac1\") " pod="openstack/dnsmasq-dns-58976fdf49-bb6rk" Feb 02 09:24:41 crc kubenswrapper[4764]: I0202 09:24:41.716884 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/30d22725-a706-485b-8cc2-964b86137ac1-dns-svc\") pod \"dnsmasq-dns-58976fdf49-bb6rk\" (UID: \"30d22725-a706-485b-8cc2-964b86137ac1\") " pod="openstack/dnsmasq-dns-58976fdf49-bb6rk" Feb 02 09:24:41 crc kubenswrapper[4764]: I0202 09:24:41.717028 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/30d22725-a706-485b-8cc2-964b86137ac1-config\") pod \"dnsmasq-dns-58976fdf49-bb6rk\" (UID: \"30d22725-a706-485b-8cc2-964b86137ac1\") " pod="openstack/dnsmasq-dns-58976fdf49-bb6rk" Feb 02 09:24:41 crc kubenswrapper[4764]: I0202 09:24:41.818695 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/30d22725-a706-485b-8cc2-964b86137ac1-config\") pod \"dnsmasq-dns-58976fdf49-bb6rk\" (UID: \"30d22725-a706-485b-8cc2-964b86137ac1\") " pod="openstack/dnsmasq-dns-58976fdf49-bb6rk" Feb 02 09:24:41 crc kubenswrapper[4764]: I0202 09:24:41.818789 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cbcg4\" (UniqueName: \"kubernetes.io/projected/30d22725-a706-485b-8cc2-964b86137ac1-kube-api-access-cbcg4\") pod \"dnsmasq-dns-58976fdf49-bb6rk\" (UID: \"30d22725-a706-485b-8cc2-964b86137ac1\") " pod="openstack/dnsmasq-dns-58976fdf49-bb6rk" Feb 02 09:24:41 crc kubenswrapper[4764]: I0202 09:24:41.818814 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/30d22725-a706-485b-8cc2-964b86137ac1-ovsdbserver-sb\") pod \"dnsmasq-dns-58976fdf49-bb6rk\" (UID: \"30d22725-a706-485b-8cc2-964b86137ac1\") " pod="openstack/dnsmasq-dns-58976fdf49-bb6rk" Feb 02 09:24:41 crc kubenswrapper[4764]: I0202 09:24:41.818837 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/30d22725-a706-485b-8cc2-964b86137ac1-ovsdbserver-nb\") pod \"dnsmasq-dns-58976fdf49-bb6rk\" (UID: \"30d22725-a706-485b-8cc2-964b86137ac1\") " pod="openstack/dnsmasq-dns-58976fdf49-bb6rk" Feb 02 09:24:41 crc kubenswrapper[4764]: I0202 09:24:41.818865 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/30d22725-a706-485b-8cc2-964b86137ac1-dns-svc\") pod \"dnsmasq-dns-58976fdf49-bb6rk\" (UID: \"30d22725-a706-485b-8cc2-964b86137ac1\") " pod="openstack/dnsmasq-dns-58976fdf49-bb6rk" Feb 02 09:24:41 crc kubenswrapper[4764]: I0202 09:24:41.819577 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/30d22725-a706-485b-8cc2-964b86137ac1-config\") pod \"dnsmasq-dns-58976fdf49-bb6rk\" (UID: \"30d22725-a706-485b-8cc2-964b86137ac1\") " pod="openstack/dnsmasq-dns-58976fdf49-bb6rk" Feb 02 09:24:41 crc kubenswrapper[4764]: I0202 09:24:41.819638 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/30d22725-a706-485b-8cc2-964b86137ac1-dns-svc\") pod \"dnsmasq-dns-58976fdf49-bb6rk\" (UID: \"30d22725-a706-485b-8cc2-964b86137ac1\") " pod="openstack/dnsmasq-dns-58976fdf49-bb6rk" Feb 02 09:24:41 crc kubenswrapper[4764]: I0202 09:24:41.820156 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/30d22725-a706-485b-8cc2-964b86137ac1-ovsdbserver-sb\") pod \"dnsmasq-dns-58976fdf49-bb6rk\" (UID: \"30d22725-a706-485b-8cc2-964b86137ac1\") " pod="openstack/dnsmasq-dns-58976fdf49-bb6rk" Feb 02 09:24:41 crc kubenswrapper[4764]: I0202 09:24:41.820307 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/30d22725-a706-485b-8cc2-964b86137ac1-ovsdbserver-nb\") pod \"dnsmasq-dns-58976fdf49-bb6rk\" (UID: \"30d22725-a706-485b-8cc2-964b86137ac1\") " pod="openstack/dnsmasq-dns-58976fdf49-bb6rk" Feb 02 09:24:41 crc kubenswrapper[4764]: I0202 09:24:41.844021 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cbcg4\" (UniqueName: \"kubernetes.io/projected/30d22725-a706-485b-8cc2-964b86137ac1-kube-api-access-cbcg4\") pod \"dnsmasq-dns-58976fdf49-bb6rk\" (UID: \"30d22725-a706-485b-8cc2-964b86137ac1\") " pod="openstack/dnsmasq-dns-58976fdf49-bb6rk" Feb 02 09:24:41 crc kubenswrapper[4764]: I0202 09:24:41.887579 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58976fdf49-bb6rk" Feb 02 09:24:42 crc kubenswrapper[4764]: I0202 09:24:42.338228 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-58976fdf49-bb6rk"] Feb 02 09:24:43 crc kubenswrapper[4764]: I0202 09:24:43.184004 4764 generic.go:334] "Generic (PLEG): container finished" podID="30d22725-a706-485b-8cc2-964b86137ac1" containerID="12bd136d4e0414940957c9b807bf9d845d345d00ab87fc4e0292591dce89decb" exitCode=0 Feb 02 09:24:43 crc kubenswrapper[4764]: I0202 09:24:43.184623 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58976fdf49-bb6rk" event={"ID":"30d22725-a706-485b-8cc2-964b86137ac1","Type":"ContainerDied","Data":"12bd136d4e0414940957c9b807bf9d845d345d00ab87fc4e0292591dce89decb"} Feb 02 09:24:43 crc kubenswrapper[4764]: I0202 09:24:43.184652 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58976fdf49-bb6rk" event={"ID":"30d22725-a706-485b-8cc2-964b86137ac1","Type":"ContainerStarted","Data":"7a566970df23dee365ff4021a5e463f656969b25cc50ccec9520791c93fc4ba6"} Feb 02 09:24:44 crc kubenswrapper[4764]: I0202 09:24:44.035206 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Feb 02 09:24:44 crc kubenswrapper[4764]: I0202 09:24:44.191253 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58976fdf49-bb6rk" event={"ID":"30d22725-a706-485b-8cc2-964b86137ac1","Type":"ContainerStarted","Data":"3160f95e976a6f6e0632f2d7adc40045a33aab5f58543831305f2e7c08360f26"} Feb 02 09:24:44 crc kubenswrapper[4764]: I0202 09:24:44.191827 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-58976fdf49-bb6rk" Feb 02 09:24:44 crc kubenswrapper[4764]: I0202 09:24:44.235127 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-58976fdf49-bb6rk" podStartSLOduration=3.235107379 podStartE2EDuration="3.235107379s" podCreationTimestamp="2026-02-02 09:24:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:24:44.230961709 +0000 UTC m=+1047.164685807" watchObservedRunningTime="2026-02-02 09:24:44.235107379 +0000 UTC m=+1047.168831467" Feb 02 09:24:44 crc kubenswrapper[4764]: I0202 09:24:44.539155 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Feb 02 09:24:44 crc kubenswrapper[4764]: I0202 09:24:44.729099 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-nbgkq"] Feb 02 09:24:44 crc kubenswrapper[4764]: I0202 09:24:44.729968 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-nbgkq" Feb 02 09:24:44 crc kubenswrapper[4764]: I0202 09:24:44.741925 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-nbgkq"] Feb 02 09:24:44 crc kubenswrapper[4764]: I0202 09:24:44.769078 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rkqwh\" (UniqueName: \"kubernetes.io/projected/4bb792d3-0d3c-44fc-8d1f-50e9763773dc-kube-api-access-rkqwh\") pod \"barbican-db-create-nbgkq\" (UID: \"4bb792d3-0d3c-44fc-8d1f-50e9763773dc\") " pod="openstack/barbican-db-create-nbgkq" Feb 02 09:24:44 crc kubenswrapper[4764]: I0202 09:24:44.769152 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4bb792d3-0d3c-44fc-8d1f-50e9763773dc-operator-scripts\") pod \"barbican-db-create-nbgkq\" (UID: \"4bb792d3-0d3c-44fc-8d1f-50e9763773dc\") " pod="openstack/barbican-db-create-nbgkq" Feb 02 09:24:44 crc kubenswrapper[4764]: I0202 09:24:44.803198 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-0818-account-create-update-rkmtz"] Feb 02 09:24:44 crc kubenswrapper[4764]: I0202 09:24:44.804327 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-0818-account-create-update-rkmtz" Feb 02 09:24:44 crc kubenswrapper[4764]: I0202 09:24:44.809209 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Feb 02 09:24:44 crc kubenswrapper[4764]: I0202 09:24:44.820444 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-0818-account-create-update-rkmtz"] Feb 02 09:24:44 crc kubenswrapper[4764]: I0202 09:24:44.851322 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-4fxzb"] Feb 02 09:24:44 crc kubenswrapper[4764]: I0202 09:24:44.852456 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-4fxzb" Feb 02 09:24:44 crc kubenswrapper[4764]: I0202 09:24:44.867855 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-4fxzb"] Feb 02 09:24:44 crc kubenswrapper[4764]: I0202 09:24:44.871080 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4bb792d3-0d3c-44fc-8d1f-50e9763773dc-operator-scripts\") pod \"barbican-db-create-nbgkq\" (UID: \"4bb792d3-0d3c-44fc-8d1f-50e9763773dc\") " pod="openstack/barbican-db-create-nbgkq" Feb 02 09:24:44 crc kubenswrapper[4764]: I0202 09:24:44.871272 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-df9zx\" (UniqueName: \"kubernetes.io/projected/49eeb3bb-263f-452e-a2dc-9c24c0fe62b1-kube-api-access-df9zx\") pod \"barbican-0818-account-create-update-rkmtz\" (UID: \"49eeb3bb-263f-452e-a2dc-9c24c0fe62b1\") " pod="openstack/barbican-0818-account-create-update-rkmtz" Feb 02 09:24:44 crc kubenswrapper[4764]: I0202 09:24:44.871489 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/49eeb3bb-263f-452e-a2dc-9c24c0fe62b1-operator-scripts\") pod \"barbican-0818-account-create-update-rkmtz\" (UID: \"49eeb3bb-263f-452e-a2dc-9c24c0fe62b1\") " pod="openstack/barbican-0818-account-create-update-rkmtz" Feb 02 09:24:44 crc kubenswrapper[4764]: I0202 09:24:44.871750 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4bb792d3-0d3c-44fc-8d1f-50e9763773dc-operator-scripts\") pod \"barbican-db-create-nbgkq\" (UID: \"4bb792d3-0d3c-44fc-8d1f-50e9763773dc\") " pod="openstack/barbican-db-create-nbgkq" Feb 02 09:24:44 crc kubenswrapper[4764]: I0202 09:24:44.871772 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rkqwh\" (UniqueName: \"kubernetes.io/projected/4bb792d3-0d3c-44fc-8d1f-50e9763773dc-kube-api-access-rkqwh\") pod \"barbican-db-create-nbgkq\" (UID: \"4bb792d3-0d3c-44fc-8d1f-50e9763773dc\") " pod="openstack/barbican-db-create-nbgkq" Feb 02 09:24:44 crc kubenswrapper[4764]: I0202 09:24:44.926512 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rkqwh\" (UniqueName: \"kubernetes.io/projected/4bb792d3-0d3c-44fc-8d1f-50e9763773dc-kube-api-access-rkqwh\") pod \"barbican-db-create-nbgkq\" (UID: \"4bb792d3-0d3c-44fc-8d1f-50e9763773dc\") " pod="openstack/barbican-db-create-nbgkq" Feb 02 09:24:44 crc kubenswrapper[4764]: I0202 09:24:44.973200 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a3915739-a4ee-4cdd-a27c-58e518e0dade-operator-scripts\") pod \"cinder-db-create-4fxzb\" (UID: \"a3915739-a4ee-4cdd-a27c-58e518e0dade\") " pod="openstack/cinder-db-create-4fxzb" Feb 02 09:24:44 crc kubenswrapper[4764]: I0202 09:24:44.973316 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-df9zx\" (UniqueName: \"kubernetes.io/projected/49eeb3bb-263f-452e-a2dc-9c24c0fe62b1-kube-api-access-df9zx\") pod \"barbican-0818-account-create-update-rkmtz\" (UID: \"49eeb3bb-263f-452e-a2dc-9c24c0fe62b1\") " pod="openstack/barbican-0818-account-create-update-rkmtz" Feb 02 09:24:44 crc kubenswrapper[4764]: I0202 09:24:44.973363 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/49eeb3bb-263f-452e-a2dc-9c24c0fe62b1-operator-scripts\") pod \"barbican-0818-account-create-update-rkmtz\" (UID: \"49eeb3bb-263f-452e-a2dc-9c24c0fe62b1\") " pod="openstack/barbican-0818-account-create-update-rkmtz" Feb 02 09:24:44 crc kubenswrapper[4764]: I0202 09:24:44.973388 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-snjjk\" (UniqueName: \"kubernetes.io/projected/a3915739-a4ee-4cdd-a27c-58e518e0dade-kube-api-access-snjjk\") pod \"cinder-db-create-4fxzb\" (UID: \"a3915739-a4ee-4cdd-a27c-58e518e0dade\") " pod="openstack/cinder-db-create-4fxzb" Feb 02 09:24:44 crc kubenswrapper[4764]: I0202 09:24:44.974596 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/49eeb3bb-263f-452e-a2dc-9c24c0fe62b1-operator-scripts\") pod \"barbican-0818-account-create-update-rkmtz\" (UID: \"49eeb3bb-263f-452e-a2dc-9c24c0fe62b1\") " pod="openstack/barbican-0818-account-create-update-rkmtz" Feb 02 09:24:44 crc kubenswrapper[4764]: I0202 09:24:44.975103 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-6k5q7"] Feb 02 09:24:44 crc kubenswrapper[4764]: I0202 09:24:44.976285 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-6k5q7" Feb 02 09:24:44 crc kubenswrapper[4764]: I0202 09:24:44.980226 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-dds8z" Feb 02 09:24:44 crc kubenswrapper[4764]: I0202 09:24:44.982194 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Feb 02 09:24:44 crc kubenswrapper[4764]: I0202 09:24:44.982723 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Feb 02 09:24:44 crc kubenswrapper[4764]: I0202 09:24:44.992621 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Feb 02 09:24:44 crc kubenswrapper[4764]: I0202 09:24:44.997959 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-6k5q7"] Feb 02 09:24:45 crc kubenswrapper[4764]: I0202 09:24:45.015257 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-df9zx\" (UniqueName: \"kubernetes.io/projected/49eeb3bb-263f-452e-a2dc-9c24c0fe62b1-kube-api-access-df9zx\") pod \"barbican-0818-account-create-update-rkmtz\" (UID: \"49eeb3bb-263f-452e-a2dc-9c24c0fe62b1\") " pod="openstack/barbican-0818-account-create-update-rkmtz" Feb 02 09:24:45 crc kubenswrapper[4764]: I0202 09:24:45.046136 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-bqq2t"] Feb 02 09:24:45 crc kubenswrapper[4764]: I0202 09:24:45.047065 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-bqq2t" Feb 02 09:24:45 crc kubenswrapper[4764]: I0202 09:24:45.049375 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-nbgkq" Feb 02 09:24:45 crc kubenswrapper[4764]: I0202 09:24:45.074717 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd7663ce-b995-4503-99d9-d5d22952db2c-combined-ca-bundle\") pod \"keystone-db-sync-6k5q7\" (UID: \"fd7663ce-b995-4503-99d9-d5d22952db2c\") " pod="openstack/keystone-db-sync-6k5q7" Feb 02 09:24:45 crc kubenswrapper[4764]: I0202 09:24:45.075041 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-snjjk\" (UniqueName: \"kubernetes.io/projected/a3915739-a4ee-4cdd-a27c-58e518e0dade-kube-api-access-snjjk\") pod \"cinder-db-create-4fxzb\" (UID: \"a3915739-a4ee-4cdd-a27c-58e518e0dade\") " pod="openstack/cinder-db-create-4fxzb" Feb 02 09:24:45 crc kubenswrapper[4764]: I0202 09:24:45.075080 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gphw2\" (UniqueName: \"kubernetes.io/projected/fd7663ce-b995-4503-99d9-d5d22952db2c-kube-api-access-gphw2\") pod \"keystone-db-sync-6k5q7\" (UID: \"fd7663ce-b995-4503-99d9-d5d22952db2c\") " pod="openstack/keystone-db-sync-6k5q7" Feb 02 09:24:45 crc kubenswrapper[4764]: I0202 09:24:45.075114 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a3915739-a4ee-4cdd-a27c-58e518e0dade-operator-scripts\") pod \"cinder-db-create-4fxzb\" (UID: \"a3915739-a4ee-4cdd-a27c-58e518e0dade\") " pod="openstack/cinder-db-create-4fxzb" Feb 02 09:24:45 crc kubenswrapper[4764]: I0202 09:24:45.075156 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd7663ce-b995-4503-99d9-d5d22952db2c-config-data\") pod \"keystone-db-sync-6k5q7\" (UID: \"fd7663ce-b995-4503-99d9-d5d22952db2c\") " pod="openstack/keystone-db-sync-6k5q7" Feb 02 09:24:45 crc kubenswrapper[4764]: I0202 09:24:45.076076 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a3915739-a4ee-4cdd-a27c-58e518e0dade-operator-scripts\") pod \"cinder-db-create-4fxzb\" (UID: \"a3915739-a4ee-4cdd-a27c-58e518e0dade\") " pod="openstack/cinder-db-create-4fxzb" Feb 02 09:24:45 crc kubenswrapper[4764]: I0202 09:24:45.078999 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-bqq2t"] Feb 02 09:24:45 crc kubenswrapper[4764]: I0202 09:24:45.100883 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-66d2-account-create-update-6fznc"] Feb 02 09:24:45 crc kubenswrapper[4764]: I0202 09:24:45.101879 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-66d2-account-create-update-6fznc" Feb 02 09:24:45 crc kubenswrapper[4764]: I0202 09:24:45.107182 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Feb 02 09:24:45 crc kubenswrapper[4764]: I0202 09:24:45.115944 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-66d2-account-create-update-6fznc"] Feb 02 09:24:45 crc kubenswrapper[4764]: I0202 09:24:45.120968 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-0818-account-create-update-rkmtz" Feb 02 09:24:45 crc kubenswrapper[4764]: I0202 09:24:45.141689 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-snjjk\" (UniqueName: \"kubernetes.io/projected/a3915739-a4ee-4cdd-a27c-58e518e0dade-kube-api-access-snjjk\") pod \"cinder-db-create-4fxzb\" (UID: \"a3915739-a4ee-4cdd-a27c-58e518e0dade\") " pod="openstack/cinder-db-create-4fxzb" Feb 02 09:24:45 crc kubenswrapper[4764]: I0202 09:24:45.169762 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-4fxzb" Feb 02 09:24:45 crc kubenswrapper[4764]: I0202 09:24:45.179762 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gphw2\" (UniqueName: \"kubernetes.io/projected/fd7663ce-b995-4503-99d9-d5d22952db2c-kube-api-access-gphw2\") pod \"keystone-db-sync-6k5q7\" (UID: \"fd7663ce-b995-4503-99d9-d5d22952db2c\") " pod="openstack/keystone-db-sync-6k5q7" Feb 02 09:24:45 crc kubenswrapper[4764]: I0202 09:24:45.179815 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1709efed-3308-4801-a4b4-efd067c6ce90-operator-scripts\") pod \"neutron-db-create-bqq2t\" (UID: \"1709efed-3308-4801-a4b4-efd067c6ce90\") " pod="openstack/neutron-db-create-bqq2t" Feb 02 09:24:45 crc kubenswrapper[4764]: I0202 09:24:45.179848 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tlt8c\" (UniqueName: \"kubernetes.io/projected/1709efed-3308-4801-a4b4-efd067c6ce90-kube-api-access-tlt8c\") pod \"neutron-db-create-bqq2t\" (UID: \"1709efed-3308-4801-a4b4-efd067c6ce90\") " pod="openstack/neutron-db-create-bqq2t" Feb 02 09:24:45 crc kubenswrapper[4764]: I0202 09:24:45.179901 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd7663ce-b995-4503-99d9-d5d22952db2c-config-data\") pod \"keystone-db-sync-6k5q7\" (UID: \"fd7663ce-b995-4503-99d9-d5d22952db2c\") " pod="openstack/keystone-db-sync-6k5q7" Feb 02 09:24:45 crc kubenswrapper[4764]: I0202 09:24:45.179942 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-srvr5\" (UniqueName: \"kubernetes.io/projected/467a028d-ccfb-4d13-84d5-b77cdd1e96ac-kube-api-access-srvr5\") pod \"neutron-66d2-account-create-update-6fznc\" (UID: \"467a028d-ccfb-4d13-84d5-b77cdd1e96ac\") " pod="openstack/neutron-66d2-account-create-update-6fznc" Feb 02 09:24:45 crc kubenswrapper[4764]: I0202 09:24:45.179969 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd7663ce-b995-4503-99d9-d5d22952db2c-combined-ca-bundle\") pod \"keystone-db-sync-6k5q7\" (UID: \"fd7663ce-b995-4503-99d9-d5d22952db2c\") " pod="openstack/keystone-db-sync-6k5q7" Feb 02 09:24:45 crc kubenswrapper[4764]: I0202 09:24:45.180008 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/467a028d-ccfb-4d13-84d5-b77cdd1e96ac-operator-scripts\") pod \"neutron-66d2-account-create-update-6fznc\" (UID: \"467a028d-ccfb-4d13-84d5-b77cdd1e96ac\") " pod="openstack/neutron-66d2-account-create-update-6fznc" Feb 02 09:24:45 crc kubenswrapper[4764]: I0202 09:24:45.193227 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd7663ce-b995-4503-99d9-d5d22952db2c-config-data\") pod \"keystone-db-sync-6k5q7\" (UID: \"fd7663ce-b995-4503-99d9-d5d22952db2c\") " pod="openstack/keystone-db-sync-6k5q7" Feb 02 09:24:45 crc kubenswrapper[4764]: I0202 09:24:45.198106 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd7663ce-b995-4503-99d9-d5d22952db2c-combined-ca-bundle\") pod \"keystone-db-sync-6k5q7\" (UID: \"fd7663ce-b995-4503-99d9-d5d22952db2c\") " pod="openstack/keystone-db-sync-6k5q7" Feb 02 09:24:45 crc kubenswrapper[4764]: I0202 09:24:45.204035 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-873f-account-create-update-n74bq"] Feb 02 09:24:45 crc kubenswrapper[4764]: I0202 09:24:45.208138 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-873f-account-create-update-n74bq" Feb 02 09:24:45 crc kubenswrapper[4764]: I0202 09:24:45.209701 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-873f-account-create-update-n74bq"] Feb 02 09:24:45 crc kubenswrapper[4764]: I0202 09:24:45.215133 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Feb 02 09:24:45 crc kubenswrapper[4764]: I0202 09:24:45.229550 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gphw2\" (UniqueName: \"kubernetes.io/projected/fd7663ce-b995-4503-99d9-d5d22952db2c-kube-api-access-gphw2\") pod \"keystone-db-sync-6k5q7\" (UID: \"fd7663ce-b995-4503-99d9-d5d22952db2c\") " pod="openstack/keystone-db-sync-6k5q7" Feb 02 09:24:45 crc kubenswrapper[4764]: I0202 09:24:45.281154 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a2004c87-c998-4892-ba8f-adeeeb31e746-operator-scripts\") pod \"cinder-873f-account-create-update-n74bq\" (UID: \"a2004c87-c998-4892-ba8f-adeeeb31e746\") " pod="openstack/cinder-873f-account-create-update-n74bq" Feb 02 09:24:45 crc kubenswrapper[4764]: I0202 09:24:45.281215 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-srvr5\" (UniqueName: \"kubernetes.io/projected/467a028d-ccfb-4d13-84d5-b77cdd1e96ac-kube-api-access-srvr5\") pod \"neutron-66d2-account-create-update-6fznc\" (UID: \"467a028d-ccfb-4d13-84d5-b77cdd1e96ac\") " pod="openstack/neutron-66d2-account-create-update-6fznc" Feb 02 09:24:45 crc kubenswrapper[4764]: I0202 09:24:45.281267 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/467a028d-ccfb-4d13-84d5-b77cdd1e96ac-operator-scripts\") pod \"neutron-66d2-account-create-update-6fznc\" (UID: \"467a028d-ccfb-4d13-84d5-b77cdd1e96ac\") " pod="openstack/neutron-66d2-account-create-update-6fznc" Feb 02 09:24:45 crc kubenswrapper[4764]: I0202 09:24:45.281300 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rsscn\" (UniqueName: \"kubernetes.io/projected/a2004c87-c998-4892-ba8f-adeeeb31e746-kube-api-access-rsscn\") pod \"cinder-873f-account-create-update-n74bq\" (UID: \"a2004c87-c998-4892-ba8f-adeeeb31e746\") " pod="openstack/cinder-873f-account-create-update-n74bq" Feb 02 09:24:45 crc kubenswrapper[4764]: I0202 09:24:45.281334 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1709efed-3308-4801-a4b4-efd067c6ce90-operator-scripts\") pod \"neutron-db-create-bqq2t\" (UID: \"1709efed-3308-4801-a4b4-efd067c6ce90\") " pod="openstack/neutron-db-create-bqq2t" Feb 02 09:24:45 crc kubenswrapper[4764]: I0202 09:24:45.281361 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tlt8c\" (UniqueName: \"kubernetes.io/projected/1709efed-3308-4801-a4b4-efd067c6ce90-kube-api-access-tlt8c\") pod \"neutron-db-create-bqq2t\" (UID: \"1709efed-3308-4801-a4b4-efd067c6ce90\") " pod="openstack/neutron-db-create-bqq2t" Feb 02 09:24:45 crc kubenswrapper[4764]: I0202 09:24:45.283100 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/467a028d-ccfb-4d13-84d5-b77cdd1e96ac-operator-scripts\") pod \"neutron-66d2-account-create-update-6fznc\" (UID: \"467a028d-ccfb-4d13-84d5-b77cdd1e96ac\") " pod="openstack/neutron-66d2-account-create-update-6fznc" Feb 02 09:24:45 crc kubenswrapper[4764]: I0202 09:24:45.283546 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1709efed-3308-4801-a4b4-efd067c6ce90-operator-scripts\") pod \"neutron-db-create-bqq2t\" (UID: \"1709efed-3308-4801-a4b4-efd067c6ce90\") " pod="openstack/neutron-db-create-bqq2t" Feb 02 09:24:45 crc kubenswrapper[4764]: I0202 09:24:45.302572 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-srvr5\" (UniqueName: \"kubernetes.io/projected/467a028d-ccfb-4d13-84d5-b77cdd1e96ac-kube-api-access-srvr5\") pod \"neutron-66d2-account-create-update-6fznc\" (UID: \"467a028d-ccfb-4d13-84d5-b77cdd1e96ac\") " pod="openstack/neutron-66d2-account-create-update-6fznc" Feb 02 09:24:45 crc kubenswrapper[4764]: I0202 09:24:45.304500 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tlt8c\" (UniqueName: \"kubernetes.io/projected/1709efed-3308-4801-a4b4-efd067c6ce90-kube-api-access-tlt8c\") pod \"neutron-db-create-bqq2t\" (UID: \"1709efed-3308-4801-a4b4-efd067c6ce90\") " pod="openstack/neutron-db-create-bqq2t" Feb 02 09:24:45 crc kubenswrapper[4764]: I0202 09:24:45.305900 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-6k5q7" Feb 02 09:24:45 crc kubenswrapper[4764]: I0202 09:24:45.384371 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a2004c87-c998-4892-ba8f-adeeeb31e746-operator-scripts\") pod \"cinder-873f-account-create-update-n74bq\" (UID: \"a2004c87-c998-4892-ba8f-adeeeb31e746\") " pod="openstack/cinder-873f-account-create-update-n74bq" Feb 02 09:24:45 crc kubenswrapper[4764]: I0202 09:24:45.384776 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rsscn\" (UniqueName: \"kubernetes.io/projected/a2004c87-c998-4892-ba8f-adeeeb31e746-kube-api-access-rsscn\") pod \"cinder-873f-account-create-update-n74bq\" (UID: \"a2004c87-c998-4892-ba8f-adeeeb31e746\") " pod="openstack/cinder-873f-account-create-update-n74bq" Feb 02 09:24:45 crc kubenswrapper[4764]: I0202 09:24:45.385312 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a2004c87-c998-4892-ba8f-adeeeb31e746-operator-scripts\") pod \"cinder-873f-account-create-update-n74bq\" (UID: \"a2004c87-c998-4892-ba8f-adeeeb31e746\") " pod="openstack/cinder-873f-account-create-update-n74bq" Feb 02 09:24:45 crc kubenswrapper[4764]: I0202 09:24:45.404483 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rsscn\" (UniqueName: \"kubernetes.io/projected/a2004c87-c998-4892-ba8f-adeeeb31e746-kube-api-access-rsscn\") pod \"cinder-873f-account-create-update-n74bq\" (UID: \"a2004c87-c998-4892-ba8f-adeeeb31e746\") " pod="openstack/cinder-873f-account-create-update-n74bq" Feb 02 09:24:45 crc kubenswrapper[4764]: I0202 09:24:45.520739 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-bqq2t" Feb 02 09:24:45 crc kubenswrapper[4764]: I0202 09:24:45.547266 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-66d2-account-create-update-6fznc" Feb 02 09:24:45 crc kubenswrapper[4764]: I0202 09:24:45.556209 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-873f-account-create-update-n74bq" Feb 02 09:24:45 crc kubenswrapper[4764]: I0202 09:24:45.633597 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-0818-account-create-update-rkmtz"] Feb 02 09:24:45 crc kubenswrapper[4764]: I0202 09:24:45.704087 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-4fxzb"] Feb 02 09:24:45 crc kubenswrapper[4764]: I0202 09:24:45.729245 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-6k5q7"] Feb 02 09:24:45 crc kubenswrapper[4764]: I0202 09:24:45.850292 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-nbgkq"] Feb 02 09:24:45 crc kubenswrapper[4764]: W0202 09:24:45.894892 4764 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4bb792d3_0d3c_44fc_8d1f_50e9763773dc.slice/crio-796b83afd45f740fe95140d7d7d80989c4471d19862fe85205e7510bbde6d758 WatchSource:0}: Error finding container 796b83afd45f740fe95140d7d7d80989c4471d19862fe85205e7510bbde6d758: Status 404 returned error can't find the container with id 796b83afd45f740fe95140d7d7d80989c4471d19862fe85205e7510bbde6d758 Feb 02 09:24:46 crc kubenswrapper[4764]: I0202 09:24:46.217828 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-6k5q7" event={"ID":"fd7663ce-b995-4503-99d9-d5d22952db2c","Type":"ContainerStarted","Data":"8d86513c96b1463daf3678356279dfe8b542918e28d64a602fc52d54aaf12094"} Feb 02 09:24:46 crc kubenswrapper[4764]: I0202 09:24:46.231571 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-4fxzb" event={"ID":"a3915739-a4ee-4cdd-a27c-58e518e0dade","Type":"ContainerStarted","Data":"2bafaf1ec4086fb317564b3ab7647ca03464edcb1c64eb49e00c153bf3d7d09f"} Feb 02 09:24:46 crc kubenswrapper[4764]: I0202 09:24:46.231611 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-4fxzb" event={"ID":"a3915739-a4ee-4cdd-a27c-58e518e0dade","Type":"ContainerStarted","Data":"287ab5eaa3e23fbe987090153fe25dd643f5951e4b32b95e6578effd38855f24"} Feb 02 09:24:46 crc kubenswrapper[4764]: I0202 09:24:46.236610 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-bqq2t"] Feb 02 09:24:46 crc kubenswrapper[4764]: I0202 09:24:46.237216 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-nbgkq" event={"ID":"4bb792d3-0d3c-44fc-8d1f-50e9763773dc","Type":"ContainerStarted","Data":"3cfba45ffa4b9101660134caf144daceeb7817ff8fbd85821de2c11047da9cc9"} Feb 02 09:24:46 crc kubenswrapper[4764]: I0202 09:24:46.237238 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-nbgkq" event={"ID":"4bb792d3-0d3c-44fc-8d1f-50e9763773dc","Type":"ContainerStarted","Data":"796b83afd45f740fe95140d7d7d80989c4471d19862fe85205e7510bbde6d758"} Feb 02 09:24:46 crc kubenswrapper[4764]: I0202 09:24:46.238728 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-0818-account-create-update-rkmtz" event={"ID":"49eeb3bb-263f-452e-a2dc-9c24c0fe62b1","Type":"ContainerStarted","Data":"f10b287949161d05b18bc14b4c82c7334d4201f0215296bfcdbf79060570f7c7"} Feb 02 09:24:46 crc kubenswrapper[4764]: I0202 09:24:46.238749 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-0818-account-create-update-rkmtz" event={"ID":"49eeb3bb-263f-452e-a2dc-9c24c0fe62b1","Type":"ContainerStarted","Data":"f2ff8e1a30965d3fd10a918fc1d950b71ea18510f98140d28ec28b100ba7653f"} Feb 02 09:24:46 crc kubenswrapper[4764]: I0202 09:24:46.274602 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-create-4fxzb" podStartSLOduration=2.274586815 podStartE2EDuration="2.274586815s" podCreationTimestamp="2026-02-02 09:24:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:24:46.268026382 +0000 UTC m=+1049.201750470" watchObservedRunningTime="2026-02-02 09:24:46.274586815 +0000 UTC m=+1049.208310903" Feb 02 09:24:46 crc kubenswrapper[4764]: I0202 09:24:46.331175 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-0818-account-create-update-rkmtz" podStartSLOduration=2.33115278 podStartE2EDuration="2.33115278s" podCreationTimestamp="2026-02-02 09:24:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:24:46.320107038 +0000 UTC m=+1049.253831126" watchObservedRunningTime="2026-02-02 09:24:46.33115278 +0000 UTC m=+1049.264876868" Feb 02 09:24:46 crc kubenswrapper[4764]: I0202 09:24:46.335285 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-create-nbgkq" podStartSLOduration=2.335263058 podStartE2EDuration="2.335263058s" podCreationTimestamp="2026-02-02 09:24:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:24:46.29823912 +0000 UTC m=+1049.231963208" watchObservedRunningTime="2026-02-02 09:24:46.335263058 +0000 UTC m=+1049.268987136" Feb 02 09:24:46 crc kubenswrapper[4764]: I0202 09:24:46.382115 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-66d2-account-create-update-6fznc"] Feb 02 09:24:46 crc kubenswrapper[4764]: W0202 09:24:46.406871 4764 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod467a028d_ccfb_4d13_84d5_b77cdd1e96ac.slice/crio-e66319a8a21c2320ba0a2bb77cc75036a5cdba2c2519d1d98af11138182792ac WatchSource:0}: Error finding container e66319a8a21c2320ba0a2bb77cc75036a5cdba2c2519d1d98af11138182792ac: Status 404 returned error can't find the container with id e66319a8a21c2320ba0a2bb77cc75036a5cdba2c2519d1d98af11138182792ac Feb 02 09:24:46 crc kubenswrapper[4764]: I0202 09:24:46.492054 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-873f-account-create-update-n74bq"] Feb 02 09:24:47 crc kubenswrapper[4764]: I0202 09:24:47.246218 4764 generic.go:334] "Generic (PLEG): container finished" podID="467a028d-ccfb-4d13-84d5-b77cdd1e96ac" containerID="0a9aa80c6f0babc8f62ada7c7954125d50f6816eafcd5d7a65e1b84be7a050bd" exitCode=0 Feb 02 09:24:47 crc kubenswrapper[4764]: I0202 09:24:47.246268 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-66d2-account-create-update-6fznc" event={"ID":"467a028d-ccfb-4d13-84d5-b77cdd1e96ac","Type":"ContainerDied","Data":"0a9aa80c6f0babc8f62ada7c7954125d50f6816eafcd5d7a65e1b84be7a050bd"} Feb 02 09:24:47 crc kubenswrapper[4764]: I0202 09:24:47.246627 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-66d2-account-create-update-6fznc" event={"ID":"467a028d-ccfb-4d13-84d5-b77cdd1e96ac","Type":"ContainerStarted","Data":"e66319a8a21c2320ba0a2bb77cc75036a5cdba2c2519d1d98af11138182792ac"} Feb 02 09:24:47 crc kubenswrapper[4764]: I0202 09:24:47.248865 4764 generic.go:334] "Generic (PLEG): container finished" podID="1709efed-3308-4801-a4b4-efd067c6ce90" containerID="af83f130e1580a8eb2b2162ba0adf21c90b40709512ef5e33a6c9de59adaa512" exitCode=0 Feb 02 09:24:47 crc kubenswrapper[4764]: I0202 09:24:47.248909 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-bqq2t" event={"ID":"1709efed-3308-4801-a4b4-efd067c6ce90","Type":"ContainerDied","Data":"af83f130e1580a8eb2b2162ba0adf21c90b40709512ef5e33a6c9de59adaa512"} Feb 02 09:24:47 crc kubenswrapper[4764]: I0202 09:24:47.248924 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-bqq2t" event={"ID":"1709efed-3308-4801-a4b4-efd067c6ce90","Type":"ContainerStarted","Data":"2ad4f842624752f50da5a3bdbf862260f2a70025a9f5e3197b088970e7822d29"} Feb 02 09:24:47 crc kubenswrapper[4764]: I0202 09:24:47.250002 4764 generic.go:334] "Generic (PLEG): container finished" podID="49eeb3bb-263f-452e-a2dc-9c24c0fe62b1" containerID="f10b287949161d05b18bc14b4c82c7334d4201f0215296bfcdbf79060570f7c7" exitCode=0 Feb 02 09:24:47 crc kubenswrapper[4764]: I0202 09:24:47.250062 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-0818-account-create-update-rkmtz" event={"ID":"49eeb3bb-263f-452e-a2dc-9c24c0fe62b1","Type":"ContainerDied","Data":"f10b287949161d05b18bc14b4c82c7334d4201f0215296bfcdbf79060570f7c7"} Feb 02 09:24:47 crc kubenswrapper[4764]: I0202 09:24:47.251623 4764 generic.go:334] "Generic (PLEG): container finished" podID="a3915739-a4ee-4cdd-a27c-58e518e0dade" containerID="2bafaf1ec4086fb317564b3ab7647ca03464edcb1c64eb49e00c153bf3d7d09f" exitCode=0 Feb 02 09:24:47 crc kubenswrapper[4764]: I0202 09:24:47.251661 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-4fxzb" event={"ID":"a3915739-a4ee-4cdd-a27c-58e518e0dade","Type":"ContainerDied","Data":"2bafaf1ec4086fb317564b3ab7647ca03464edcb1c64eb49e00c153bf3d7d09f"} Feb 02 09:24:47 crc kubenswrapper[4764]: I0202 09:24:47.252955 4764 generic.go:334] "Generic (PLEG): container finished" podID="a2004c87-c998-4892-ba8f-adeeeb31e746" containerID="c626eaf56455d96ba04df17db01428f27600844910ce734771f5704cb5969915" exitCode=0 Feb 02 09:24:47 crc kubenswrapper[4764]: I0202 09:24:47.252991 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-873f-account-create-update-n74bq" event={"ID":"a2004c87-c998-4892-ba8f-adeeeb31e746","Type":"ContainerDied","Data":"c626eaf56455d96ba04df17db01428f27600844910ce734771f5704cb5969915"} Feb 02 09:24:47 crc kubenswrapper[4764]: I0202 09:24:47.253005 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-873f-account-create-update-n74bq" event={"ID":"a2004c87-c998-4892-ba8f-adeeeb31e746","Type":"ContainerStarted","Data":"c5d50e9ec698f8fd85ae03ed412aec5961884e3db6b5e1393b654243f93fb84c"} Feb 02 09:24:47 crc kubenswrapper[4764]: I0202 09:24:47.262641 4764 generic.go:334] "Generic (PLEG): container finished" podID="4bb792d3-0d3c-44fc-8d1f-50e9763773dc" containerID="3cfba45ffa4b9101660134caf144daceeb7817ff8fbd85821de2c11047da9cc9" exitCode=0 Feb 02 09:24:47 crc kubenswrapper[4764]: I0202 09:24:47.262699 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-nbgkq" event={"ID":"4bb792d3-0d3c-44fc-8d1f-50e9763773dc","Type":"ContainerDied","Data":"3cfba45ffa4b9101660134caf144daceeb7817ff8fbd85821de2c11047da9cc9"} Feb 02 09:24:51 crc kubenswrapper[4764]: I0202 09:24:51.303430 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-4fxzb" event={"ID":"a3915739-a4ee-4cdd-a27c-58e518e0dade","Type":"ContainerDied","Data":"287ab5eaa3e23fbe987090153fe25dd643f5951e4b32b95e6578effd38855f24"} Feb 02 09:24:51 crc kubenswrapper[4764]: I0202 09:24:51.303967 4764 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="287ab5eaa3e23fbe987090153fe25dd643f5951e4b32b95e6578effd38855f24" Feb 02 09:24:51 crc kubenswrapper[4764]: I0202 09:24:51.305811 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-873f-account-create-update-n74bq" event={"ID":"a2004c87-c998-4892-ba8f-adeeeb31e746","Type":"ContainerDied","Data":"c5d50e9ec698f8fd85ae03ed412aec5961884e3db6b5e1393b654243f93fb84c"} Feb 02 09:24:51 crc kubenswrapper[4764]: I0202 09:24:51.305850 4764 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c5d50e9ec698f8fd85ae03ed412aec5961884e3db6b5e1393b654243f93fb84c" Feb 02 09:24:51 crc kubenswrapper[4764]: I0202 09:24:51.308375 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-nbgkq" event={"ID":"4bb792d3-0d3c-44fc-8d1f-50e9763773dc","Type":"ContainerDied","Data":"796b83afd45f740fe95140d7d7d80989c4471d19862fe85205e7510bbde6d758"} Feb 02 09:24:51 crc kubenswrapper[4764]: I0202 09:24:51.308394 4764 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="796b83afd45f740fe95140d7d7d80989c4471d19862fe85205e7510bbde6d758" Feb 02 09:24:51 crc kubenswrapper[4764]: I0202 09:24:51.312077 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-66d2-account-create-update-6fznc" event={"ID":"467a028d-ccfb-4d13-84d5-b77cdd1e96ac","Type":"ContainerDied","Data":"e66319a8a21c2320ba0a2bb77cc75036a5cdba2c2519d1d98af11138182792ac"} Feb 02 09:24:51 crc kubenswrapper[4764]: I0202 09:24:51.312129 4764 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e66319a8a21c2320ba0a2bb77cc75036a5cdba2c2519d1d98af11138182792ac" Feb 02 09:24:51 crc kubenswrapper[4764]: I0202 09:24:51.318187 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-bqq2t" event={"ID":"1709efed-3308-4801-a4b4-efd067c6ce90","Type":"ContainerDied","Data":"2ad4f842624752f50da5a3bdbf862260f2a70025a9f5e3197b088970e7822d29"} Feb 02 09:24:51 crc kubenswrapper[4764]: I0202 09:24:51.318238 4764 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2ad4f842624752f50da5a3bdbf862260f2a70025a9f5e3197b088970e7822d29" Feb 02 09:24:51 crc kubenswrapper[4764]: I0202 09:24:51.320323 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-0818-account-create-update-rkmtz" event={"ID":"49eeb3bb-263f-452e-a2dc-9c24c0fe62b1","Type":"ContainerDied","Data":"f2ff8e1a30965d3fd10a918fc1d950b71ea18510f98140d28ec28b100ba7653f"} Feb 02 09:24:51 crc kubenswrapper[4764]: I0202 09:24:51.320350 4764 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f2ff8e1a30965d3fd10a918fc1d950b71ea18510f98140d28ec28b100ba7653f" Feb 02 09:24:51 crc kubenswrapper[4764]: I0202 09:24:51.432470 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-66d2-account-create-update-6fznc" Feb 02 09:24:51 crc kubenswrapper[4764]: I0202 09:24:51.442414 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-bqq2t" Feb 02 09:24:51 crc kubenswrapper[4764]: I0202 09:24:51.458163 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-873f-account-create-update-n74bq" Feb 02 09:24:51 crc kubenswrapper[4764]: I0202 09:24:51.470383 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-0818-account-create-update-rkmtz" Feb 02 09:24:51 crc kubenswrapper[4764]: I0202 09:24:51.487463 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-nbgkq" Feb 02 09:24:51 crc kubenswrapper[4764]: I0202 09:24:51.503438 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-4fxzb" Feb 02 09:24:51 crc kubenswrapper[4764]: I0202 09:24:51.525788 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rkqwh\" (UniqueName: \"kubernetes.io/projected/4bb792d3-0d3c-44fc-8d1f-50e9763773dc-kube-api-access-rkqwh\") pod \"4bb792d3-0d3c-44fc-8d1f-50e9763773dc\" (UID: \"4bb792d3-0d3c-44fc-8d1f-50e9763773dc\") " Feb 02 09:24:51 crc kubenswrapper[4764]: I0202 09:24:51.525849 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rsscn\" (UniqueName: \"kubernetes.io/projected/a2004c87-c998-4892-ba8f-adeeeb31e746-kube-api-access-rsscn\") pod \"a2004c87-c998-4892-ba8f-adeeeb31e746\" (UID: \"a2004c87-c998-4892-ba8f-adeeeb31e746\") " Feb 02 09:24:51 crc kubenswrapper[4764]: I0202 09:24:51.525880 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-df9zx\" (UniqueName: \"kubernetes.io/projected/49eeb3bb-263f-452e-a2dc-9c24c0fe62b1-kube-api-access-df9zx\") pod \"49eeb3bb-263f-452e-a2dc-9c24c0fe62b1\" (UID: \"49eeb3bb-263f-452e-a2dc-9c24c0fe62b1\") " Feb 02 09:24:51 crc kubenswrapper[4764]: I0202 09:24:51.525903 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a2004c87-c998-4892-ba8f-adeeeb31e746-operator-scripts\") pod \"a2004c87-c998-4892-ba8f-adeeeb31e746\" (UID: \"a2004c87-c998-4892-ba8f-adeeeb31e746\") " Feb 02 09:24:51 crc kubenswrapper[4764]: I0202 09:24:51.525945 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/49eeb3bb-263f-452e-a2dc-9c24c0fe62b1-operator-scripts\") pod \"49eeb3bb-263f-452e-a2dc-9c24c0fe62b1\" (UID: \"49eeb3bb-263f-452e-a2dc-9c24c0fe62b1\") " Feb 02 09:24:51 crc kubenswrapper[4764]: I0202 09:24:51.525978 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-srvr5\" (UniqueName: \"kubernetes.io/projected/467a028d-ccfb-4d13-84d5-b77cdd1e96ac-kube-api-access-srvr5\") pod \"467a028d-ccfb-4d13-84d5-b77cdd1e96ac\" (UID: \"467a028d-ccfb-4d13-84d5-b77cdd1e96ac\") " Feb 02 09:24:51 crc kubenswrapper[4764]: I0202 09:24:51.525999 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/467a028d-ccfb-4d13-84d5-b77cdd1e96ac-operator-scripts\") pod \"467a028d-ccfb-4d13-84d5-b77cdd1e96ac\" (UID: \"467a028d-ccfb-4d13-84d5-b77cdd1e96ac\") " Feb 02 09:24:51 crc kubenswrapper[4764]: I0202 09:24:51.526045 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4bb792d3-0d3c-44fc-8d1f-50e9763773dc-operator-scripts\") pod \"4bb792d3-0d3c-44fc-8d1f-50e9763773dc\" (UID: \"4bb792d3-0d3c-44fc-8d1f-50e9763773dc\") " Feb 02 09:24:51 crc kubenswrapper[4764]: I0202 09:24:51.526089 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tlt8c\" (UniqueName: \"kubernetes.io/projected/1709efed-3308-4801-a4b4-efd067c6ce90-kube-api-access-tlt8c\") pod \"1709efed-3308-4801-a4b4-efd067c6ce90\" (UID: \"1709efed-3308-4801-a4b4-efd067c6ce90\") " Feb 02 09:24:51 crc kubenswrapper[4764]: I0202 09:24:51.526107 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1709efed-3308-4801-a4b4-efd067c6ce90-operator-scripts\") pod \"1709efed-3308-4801-a4b4-efd067c6ce90\" (UID: \"1709efed-3308-4801-a4b4-efd067c6ce90\") " Feb 02 09:24:51 crc kubenswrapper[4764]: I0202 09:24:51.527008 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1709efed-3308-4801-a4b4-efd067c6ce90-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "1709efed-3308-4801-a4b4-efd067c6ce90" (UID: "1709efed-3308-4801-a4b4-efd067c6ce90"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:24:51 crc kubenswrapper[4764]: I0202 09:24:51.527273 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49eeb3bb-263f-452e-a2dc-9c24c0fe62b1-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "49eeb3bb-263f-452e-a2dc-9c24c0fe62b1" (UID: "49eeb3bb-263f-452e-a2dc-9c24c0fe62b1"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:24:51 crc kubenswrapper[4764]: I0202 09:24:51.528474 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb792d3-0d3c-44fc-8d1f-50e9763773dc-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "4bb792d3-0d3c-44fc-8d1f-50e9763773dc" (UID: "4bb792d3-0d3c-44fc-8d1f-50e9763773dc"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:24:51 crc kubenswrapper[4764]: I0202 09:24:51.528811 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/467a028d-ccfb-4d13-84d5-b77cdd1e96ac-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "467a028d-ccfb-4d13-84d5-b77cdd1e96ac" (UID: "467a028d-ccfb-4d13-84d5-b77cdd1e96ac"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:24:51 crc kubenswrapper[4764]: I0202 09:24:51.532125 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a2004c87-c998-4892-ba8f-adeeeb31e746-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a2004c87-c998-4892-ba8f-adeeeb31e746" (UID: "a2004c87-c998-4892-ba8f-adeeeb31e746"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:24:51 crc kubenswrapper[4764]: I0202 09:24:51.535582 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1709efed-3308-4801-a4b4-efd067c6ce90-kube-api-access-tlt8c" (OuterVolumeSpecName: "kube-api-access-tlt8c") pod "1709efed-3308-4801-a4b4-efd067c6ce90" (UID: "1709efed-3308-4801-a4b4-efd067c6ce90"). InnerVolumeSpecName "kube-api-access-tlt8c". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:24:51 crc kubenswrapper[4764]: I0202 09:24:51.535875 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb792d3-0d3c-44fc-8d1f-50e9763773dc-kube-api-access-rkqwh" (OuterVolumeSpecName: "kube-api-access-rkqwh") pod "4bb792d3-0d3c-44fc-8d1f-50e9763773dc" (UID: "4bb792d3-0d3c-44fc-8d1f-50e9763773dc"). InnerVolumeSpecName "kube-api-access-rkqwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:24:51 crc kubenswrapper[4764]: I0202 09:24:51.538773 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49eeb3bb-263f-452e-a2dc-9c24c0fe62b1-kube-api-access-df9zx" (OuterVolumeSpecName: "kube-api-access-df9zx") pod "49eeb3bb-263f-452e-a2dc-9c24c0fe62b1" (UID: "49eeb3bb-263f-452e-a2dc-9c24c0fe62b1"). InnerVolumeSpecName "kube-api-access-df9zx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:24:51 crc kubenswrapper[4764]: I0202 09:24:51.552575 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/467a028d-ccfb-4d13-84d5-b77cdd1e96ac-kube-api-access-srvr5" (OuterVolumeSpecName: "kube-api-access-srvr5") pod "467a028d-ccfb-4d13-84d5-b77cdd1e96ac" (UID: "467a028d-ccfb-4d13-84d5-b77cdd1e96ac"). InnerVolumeSpecName "kube-api-access-srvr5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:24:51 crc kubenswrapper[4764]: I0202 09:24:51.563437 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a2004c87-c998-4892-ba8f-adeeeb31e746-kube-api-access-rsscn" (OuterVolumeSpecName: "kube-api-access-rsscn") pod "a2004c87-c998-4892-ba8f-adeeeb31e746" (UID: "a2004c87-c998-4892-ba8f-adeeeb31e746"). InnerVolumeSpecName "kube-api-access-rsscn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:24:51 crc kubenswrapper[4764]: I0202 09:24:51.628230 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a3915739-a4ee-4cdd-a27c-58e518e0dade-operator-scripts\") pod \"a3915739-a4ee-4cdd-a27c-58e518e0dade\" (UID: \"a3915739-a4ee-4cdd-a27c-58e518e0dade\") " Feb 02 09:24:51 crc kubenswrapper[4764]: I0202 09:24:51.628360 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-snjjk\" (UniqueName: \"kubernetes.io/projected/a3915739-a4ee-4cdd-a27c-58e518e0dade-kube-api-access-snjjk\") pod \"a3915739-a4ee-4cdd-a27c-58e518e0dade\" (UID: \"a3915739-a4ee-4cdd-a27c-58e518e0dade\") " Feb 02 09:24:51 crc kubenswrapper[4764]: I0202 09:24:51.628717 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rkqwh\" (UniqueName: \"kubernetes.io/projected/4bb792d3-0d3c-44fc-8d1f-50e9763773dc-kube-api-access-rkqwh\") on node \"crc\" DevicePath \"\"" Feb 02 09:24:51 crc kubenswrapper[4764]: I0202 09:24:51.628754 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rsscn\" (UniqueName: \"kubernetes.io/projected/a2004c87-c998-4892-ba8f-adeeeb31e746-kube-api-access-rsscn\") on node \"crc\" DevicePath \"\"" Feb 02 09:24:51 crc kubenswrapper[4764]: I0202 09:24:51.628768 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-df9zx\" (UniqueName: \"kubernetes.io/projected/49eeb3bb-263f-452e-a2dc-9c24c0fe62b1-kube-api-access-df9zx\") on node \"crc\" DevicePath \"\"" Feb 02 09:24:51 crc kubenswrapper[4764]: I0202 09:24:51.628780 4764 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a2004c87-c998-4892-ba8f-adeeeb31e746-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 09:24:51 crc kubenswrapper[4764]: I0202 09:24:51.628791 4764 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/49eeb3bb-263f-452e-a2dc-9c24c0fe62b1-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 09:24:51 crc kubenswrapper[4764]: I0202 09:24:51.628803 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-srvr5\" (UniqueName: \"kubernetes.io/projected/467a028d-ccfb-4d13-84d5-b77cdd1e96ac-kube-api-access-srvr5\") on node \"crc\" DevicePath \"\"" Feb 02 09:24:51 crc kubenswrapper[4764]: I0202 09:24:51.628814 4764 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/467a028d-ccfb-4d13-84d5-b77cdd1e96ac-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 09:24:51 crc kubenswrapper[4764]: I0202 09:24:51.628826 4764 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4bb792d3-0d3c-44fc-8d1f-50e9763773dc-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 09:24:51 crc kubenswrapper[4764]: I0202 09:24:51.628836 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tlt8c\" (UniqueName: \"kubernetes.io/projected/1709efed-3308-4801-a4b4-efd067c6ce90-kube-api-access-tlt8c\") on node \"crc\" DevicePath \"\"" Feb 02 09:24:51 crc kubenswrapper[4764]: I0202 09:24:51.628847 4764 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1709efed-3308-4801-a4b4-efd067c6ce90-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 09:24:51 crc kubenswrapper[4764]: I0202 09:24:51.629058 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a3915739-a4ee-4cdd-a27c-58e518e0dade-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a3915739-a4ee-4cdd-a27c-58e518e0dade" (UID: "a3915739-a4ee-4cdd-a27c-58e518e0dade"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:24:51 crc kubenswrapper[4764]: I0202 09:24:51.632534 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a3915739-a4ee-4cdd-a27c-58e518e0dade-kube-api-access-snjjk" (OuterVolumeSpecName: "kube-api-access-snjjk") pod "a3915739-a4ee-4cdd-a27c-58e518e0dade" (UID: "a3915739-a4ee-4cdd-a27c-58e518e0dade"). InnerVolumeSpecName "kube-api-access-snjjk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:24:51 crc kubenswrapper[4764]: I0202 09:24:51.730478 4764 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a3915739-a4ee-4cdd-a27c-58e518e0dade-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 09:24:51 crc kubenswrapper[4764]: I0202 09:24:51.730730 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-snjjk\" (UniqueName: \"kubernetes.io/projected/a3915739-a4ee-4cdd-a27c-58e518e0dade-kube-api-access-snjjk\") on node \"crc\" DevicePath \"\"" Feb 02 09:24:51 crc kubenswrapper[4764]: I0202 09:24:51.896262 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-58976fdf49-bb6rk" Feb 02 09:24:51 crc kubenswrapper[4764]: I0202 09:24:51.991338 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-56df986d9c-vkqnn"] Feb 02 09:24:51 crc kubenswrapper[4764]: I0202 09:24:51.991613 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-56df986d9c-vkqnn" podUID="22c2fdff-3f57-4fc3-bbf4-f459784fa876" containerName="dnsmasq-dns" containerID="cri-o://e4c97ac9d116f802f33a0f092f6268eaea37dd2a894d084f76cf34a1525465ec" gracePeriod=10 Feb 02 09:24:52 crc kubenswrapper[4764]: I0202 09:24:52.327507 4764 generic.go:334] "Generic (PLEG): container finished" podID="22c2fdff-3f57-4fc3-bbf4-f459784fa876" containerID="e4c97ac9d116f802f33a0f092f6268eaea37dd2a894d084f76cf34a1525465ec" exitCode=0 Feb 02 09:24:52 crc kubenswrapper[4764]: I0202 09:24:52.328552 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df986d9c-vkqnn" event={"ID":"22c2fdff-3f57-4fc3-bbf4-f459784fa876","Type":"ContainerDied","Data":"e4c97ac9d116f802f33a0f092f6268eaea37dd2a894d084f76cf34a1525465ec"} Feb 02 09:24:52 crc kubenswrapper[4764]: I0202 09:24:52.330171 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-66d2-account-create-update-6fznc" Feb 02 09:24:52 crc kubenswrapper[4764]: I0202 09:24:52.331555 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-6k5q7" event={"ID":"fd7663ce-b995-4503-99d9-d5d22952db2c","Type":"ContainerStarted","Data":"c48c0eb8f08401abc4475fcbb3a0ef88bad3296f2474f680e07a7865ce799f47"} Feb 02 09:24:52 crc kubenswrapper[4764]: I0202 09:24:52.331702 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-4fxzb" Feb 02 09:24:52 crc kubenswrapper[4764]: I0202 09:24:52.332134 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-873f-account-create-update-n74bq" Feb 02 09:24:52 crc kubenswrapper[4764]: I0202 09:24:52.332613 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-0818-account-create-update-rkmtz" Feb 02 09:24:52 crc kubenswrapper[4764]: I0202 09:24:52.333250 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-nbgkq" Feb 02 09:24:52 crc kubenswrapper[4764]: I0202 09:24:52.334026 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-bqq2t" Feb 02 09:24:52 crc kubenswrapper[4764]: I0202 09:24:52.350051 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-6k5q7" podStartSLOduration=2.5743065229999997 podStartE2EDuration="8.350035508s" podCreationTimestamp="2026-02-02 09:24:44 +0000 UTC" firstStartedPulling="2026-02-02 09:24:45.788558833 +0000 UTC m=+1048.722282921" lastFinishedPulling="2026-02-02 09:24:51.564287818 +0000 UTC m=+1054.498011906" observedRunningTime="2026-02-02 09:24:52.344549523 +0000 UTC m=+1055.278273631" watchObservedRunningTime="2026-02-02 09:24:52.350035508 +0000 UTC m=+1055.283759596" Feb 02 09:24:53 crc kubenswrapper[4764]: I0202 09:24:52.917447 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df986d9c-vkqnn" Feb 02 09:24:53 crc kubenswrapper[4764]: I0202 09:24:52.959295 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/22c2fdff-3f57-4fc3-bbf4-f459784fa876-ovsdbserver-sb\") pod \"22c2fdff-3f57-4fc3-bbf4-f459784fa876\" (UID: \"22c2fdff-3f57-4fc3-bbf4-f459784fa876\") " Feb 02 09:24:53 crc kubenswrapper[4764]: I0202 09:24:52.959422 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c2fdff-3f57-4fc3-bbf4-f459784fa876-config\") pod \"22c2fdff-3f57-4fc3-bbf4-f459784fa876\" (UID: \"22c2fdff-3f57-4fc3-bbf4-f459784fa876\") " Feb 02 09:24:53 crc kubenswrapper[4764]: I0202 09:24:52.959526 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/22c2fdff-3f57-4fc3-bbf4-f459784fa876-dns-svc\") pod \"22c2fdff-3f57-4fc3-bbf4-f459784fa876\" (UID: \"22c2fdff-3f57-4fc3-bbf4-f459784fa876\") " Feb 02 09:24:53 crc kubenswrapper[4764]: I0202 09:24:52.959586 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/22c2fdff-3f57-4fc3-bbf4-f459784fa876-ovsdbserver-nb\") pod \"22c2fdff-3f57-4fc3-bbf4-f459784fa876\" (UID: \"22c2fdff-3f57-4fc3-bbf4-f459784fa876\") " Feb 02 09:24:53 crc kubenswrapper[4764]: I0202 09:24:52.959630 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gzgzn\" (UniqueName: \"kubernetes.io/projected/22c2fdff-3f57-4fc3-bbf4-f459784fa876-kube-api-access-gzgzn\") pod \"22c2fdff-3f57-4fc3-bbf4-f459784fa876\" (UID: \"22c2fdff-3f57-4fc3-bbf4-f459784fa876\") " Feb 02 09:24:53 crc kubenswrapper[4764]: I0202 09:24:52.971366 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c2fdff-3f57-4fc3-bbf4-f459784fa876-kube-api-access-gzgzn" (OuterVolumeSpecName: "kube-api-access-gzgzn") pod "22c2fdff-3f57-4fc3-bbf4-f459784fa876" (UID: "22c2fdff-3f57-4fc3-bbf4-f459784fa876"). InnerVolumeSpecName "kube-api-access-gzgzn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:24:53 crc kubenswrapper[4764]: I0202 09:24:53.023925 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c2fdff-3f57-4fc3-bbf4-f459784fa876-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "22c2fdff-3f57-4fc3-bbf4-f459784fa876" (UID: "22c2fdff-3f57-4fc3-bbf4-f459784fa876"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:24:53 crc kubenswrapper[4764]: I0202 09:24:53.029909 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c2fdff-3f57-4fc3-bbf4-f459784fa876-config" (OuterVolumeSpecName: "config") pod "22c2fdff-3f57-4fc3-bbf4-f459784fa876" (UID: "22c2fdff-3f57-4fc3-bbf4-f459784fa876"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:24:53 crc kubenswrapper[4764]: I0202 09:24:53.034534 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c2fdff-3f57-4fc3-bbf4-f459784fa876-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "22c2fdff-3f57-4fc3-bbf4-f459784fa876" (UID: "22c2fdff-3f57-4fc3-bbf4-f459784fa876"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:24:53 crc kubenswrapper[4764]: I0202 09:24:53.037308 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c2fdff-3f57-4fc3-bbf4-f459784fa876-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "22c2fdff-3f57-4fc3-bbf4-f459784fa876" (UID: "22c2fdff-3f57-4fc3-bbf4-f459784fa876"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:24:53 crc kubenswrapper[4764]: I0202 09:24:53.062021 4764 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c2fdff-3f57-4fc3-bbf4-f459784fa876-config\") on node \"crc\" DevicePath \"\"" Feb 02 09:24:53 crc kubenswrapper[4764]: I0202 09:24:53.062056 4764 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/22c2fdff-3f57-4fc3-bbf4-f459784fa876-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 02 09:24:53 crc kubenswrapper[4764]: I0202 09:24:53.062066 4764 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/22c2fdff-3f57-4fc3-bbf4-f459784fa876-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 02 09:24:53 crc kubenswrapper[4764]: I0202 09:24:53.062078 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gzgzn\" (UniqueName: \"kubernetes.io/projected/22c2fdff-3f57-4fc3-bbf4-f459784fa876-kube-api-access-gzgzn\") on node \"crc\" DevicePath \"\"" Feb 02 09:24:53 crc kubenswrapper[4764]: I0202 09:24:53.062087 4764 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/22c2fdff-3f57-4fc3-bbf4-f459784fa876-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 02 09:24:53 crc kubenswrapper[4764]: I0202 09:24:53.340040 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df986d9c-vkqnn" Feb 02 09:24:53 crc kubenswrapper[4764]: I0202 09:24:53.340412 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df986d9c-vkqnn" event={"ID":"22c2fdff-3f57-4fc3-bbf4-f459784fa876","Type":"ContainerDied","Data":"1de7784d8dbb29eef5cbc116714bb69dee0f3a82be15a2859d2d9910b2822da4"} Feb 02 09:24:53 crc kubenswrapper[4764]: I0202 09:24:53.340450 4764 scope.go:117] "RemoveContainer" containerID="e4c97ac9d116f802f33a0f092f6268eaea37dd2a894d084f76cf34a1525465ec" Feb 02 09:24:53 crc kubenswrapper[4764]: I0202 09:24:53.375184 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-56df986d9c-vkqnn"] Feb 02 09:24:53 crc kubenswrapper[4764]: I0202 09:24:53.378182 4764 scope.go:117] "RemoveContainer" containerID="a92d9770b278125ed99e3f4e33511ef8b270f5a22578dabe0ff27e3c86793d5c" Feb 02 09:24:53 crc kubenswrapper[4764]: I0202 09:24:53.380988 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-56df986d9c-vkqnn"] Feb 02 09:24:53 crc kubenswrapper[4764]: I0202 09:24:53.842998 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c2fdff-3f57-4fc3-bbf4-f459784fa876" path="/var/lib/kubelet/pods/22c2fdff-3f57-4fc3-bbf4-f459784fa876/volumes" Feb 02 09:24:55 crc kubenswrapper[4764]: I0202 09:24:55.364901 4764 generic.go:334] "Generic (PLEG): container finished" podID="fd7663ce-b995-4503-99d9-d5d22952db2c" containerID="c48c0eb8f08401abc4475fcbb3a0ef88bad3296f2474f680e07a7865ce799f47" exitCode=0 Feb 02 09:24:55 crc kubenswrapper[4764]: I0202 09:24:55.365122 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-6k5q7" event={"ID":"fd7663ce-b995-4503-99d9-d5d22952db2c","Type":"ContainerDied","Data":"c48c0eb8f08401abc4475fcbb3a0ef88bad3296f2474f680e07a7865ce799f47"} Feb 02 09:24:56 crc kubenswrapper[4764]: I0202 09:24:56.702002 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-6k5q7" Feb 02 09:24:56 crc kubenswrapper[4764]: I0202 09:24:56.738518 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd7663ce-b995-4503-99d9-d5d22952db2c-combined-ca-bundle\") pod \"fd7663ce-b995-4503-99d9-d5d22952db2c\" (UID: \"fd7663ce-b995-4503-99d9-d5d22952db2c\") " Feb 02 09:24:56 crc kubenswrapper[4764]: I0202 09:24:56.738806 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd7663ce-b995-4503-99d9-d5d22952db2c-config-data\") pod \"fd7663ce-b995-4503-99d9-d5d22952db2c\" (UID: \"fd7663ce-b995-4503-99d9-d5d22952db2c\") " Feb 02 09:24:56 crc kubenswrapper[4764]: I0202 09:24:56.739112 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gphw2\" (UniqueName: \"kubernetes.io/projected/fd7663ce-b995-4503-99d9-d5d22952db2c-kube-api-access-gphw2\") pod \"fd7663ce-b995-4503-99d9-d5d22952db2c\" (UID: \"fd7663ce-b995-4503-99d9-d5d22952db2c\") " Feb 02 09:24:56 crc kubenswrapper[4764]: I0202 09:24:56.744744 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fd7663ce-b995-4503-99d9-d5d22952db2c-kube-api-access-gphw2" (OuterVolumeSpecName: "kube-api-access-gphw2") pod "fd7663ce-b995-4503-99d9-d5d22952db2c" (UID: "fd7663ce-b995-4503-99d9-d5d22952db2c"). InnerVolumeSpecName "kube-api-access-gphw2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:24:56 crc kubenswrapper[4764]: I0202 09:24:56.765516 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd7663ce-b995-4503-99d9-d5d22952db2c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fd7663ce-b995-4503-99d9-d5d22952db2c" (UID: "fd7663ce-b995-4503-99d9-d5d22952db2c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:24:56 crc kubenswrapper[4764]: I0202 09:24:56.777007 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd7663ce-b995-4503-99d9-d5d22952db2c-config-data" (OuterVolumeSpecName: "config-data") pod "fd7663ce-b995-4503-99d9-d5d22952db2c" (UID: "fd7663ce-b995-4503-99d9-d5d22952db2c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:24:56 crc kubenswrapper[4764]: I0202 09:24:56.841571 4764 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd7663ce-b995-4503-99d9-d5d22952db2c-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 09:24:56 crc kubenswrapper[4764]: I0202 09:24:56.841620 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gphw2\" (UniqueName: \"kubernetes.io/projected/fd7663ce-b995-4503-99d9-d5d22952db2c-kube-api-access-gphw2\") on node \"crc\" DevicePath \"\"" Feb 02 09:24:56 crc kubenswrapper[4764]: I0202 09:24:56.841634 4764 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd7663ce-b995-4503-99d9-d5d22952db2c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 09:24:57 crc kubenswrapper[4764]: I0202 09:24:57.398406 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-6k5q7" event={"ID":"fd7663ce-b995-4503-99d9-d5d22952db2c","Type":"ContainerDied","Data":"8d86513c96b1463daf3678356279dfe8b542918e28d64a602fc52d54aaf12094"} Feb 02 09:24:57 crc kubenswrapper[4764]: I0202 09:24:57.398850 4764 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8d86513c96b1463daf3678356279dfe8b542918e28d64a602fc52d54aaf12094" Feb 02 09:24:57 crc kubenswrapper[4764]: I0202 09:24:57.398511 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-6k5q7" Feb 02 09:24:57 crc kubenswrapper[4764]: I0202 09:24:57.696740 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-nk659"] Feb 02 09:24:57 crc kubenswrapper[4764]: E0202 09:24:57.697057 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22c2fdff-3f57-4fc3-bbf4-f459784fa876" containerName="init" Feb 02 09:24:57 crc kubenswrapper[4764]: I0202 09:24:57.697073 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="22c2fdff-3f57-4fc3-bbf4-f459784fa876" containerName="init" Feb 02 09:24:57 crc kubenswrapper[4764]: E0202 09:24:57.697086 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2004c87-c998-4892-ba8f-adeeeb31e746" containerName="mariadb-account-create-update" Feb 02 09:24:57 crc kubenswrapper[4764]: I0202 09:24:57.697093 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2004c87-c998-4892-ba8f-adeeeb31e746" containerName="mariadb-account-create-update" Feb 02 09:24:57 crc kubenswrapper[4764]: E0202 09:24:57.697103 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49eeb3bb-263f-452e-a2dc-9c24c0fe62b1" containerName="mariadb-account-create-update" Feb 02 09:24:57 crc kubenswrapper[4764]: I0202 09:24:57.697109 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="49eeb3bb-263f-452e-a2dc-9c24c0fe62b1" containerName="mariadb-account-create-update" Feb 02 09:24:57 crc kubenswrapper[4764]: E0202 09:24:57.697117 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd7663ce-b995-4503-99d9-d5d22952db2c" containerName="keystone-db-sync" Feb 02 09:24:57 crc kubenswrapper[4764]: I0202 09:24:57.697123 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd7663ce-b995-4503-99d9-d5d22952db2c" containerName="keystone-db-sync" Feb 02 09:24:57 crc kubenswrapper[4764]: E0202 09:24:57.697132 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22c2fdff-3f57-4fc3-bbf4-f459784fa876" containerName="dnsmasq-dns" Feb 02 09:24:57 crc kubenswrapper[4764]: I0202 09:24:57.697137 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="22c2fdff-3f57-4fc3-bbf4-f459784fa876" containerName="dnsmasq-dns" Feb 02 09:24:57 crc kubenswrapper[4764]: E0202 09:24:57.697150 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1709efed-3308-4801-a4b4-efd067c6ce90" containerName="mariadb-database-create" Feb 02 09:24:57 crc kubenswrapper[4764]: I0202 09:24:57.697157 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="1709efed-3308-4801-a4b4-efd067c6ce90" containerName="mariadb-database-create" Feb 02 09:24:57 crc kubenswrapper[4764]: E0202 09:24:57.697167 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4bb792d3-0d3c-44fc-8d1f-50e9763773dc" containerName="mariadb-database-create" Feb 02 09:24:57 crc kubenswrapper[4764]: I0202 09:24:57.697173 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="4bb792d3-0d3c-44fc-8d1f-50e9763773dc" containerName="mariadb-database-create" Feb 02 09:24:57 crc kubenswrapper[4764]: E0202 09:24:57.697184 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3915739-a4ee-4cdd-a27c-58e518e0dade" containerName="mariadb-database-create" Feb 02 09:24:57 crc kubenswrapper[4764]: I0202 09:24:57.697190 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3915739-a4ee-4cdd-a27c-58e518e0dade" containerName="mariadb-database-create" Feb 02 09:24:57 crc kubenswrapper[4764]: E0202 09:24:57.697205 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="467a028d-ccfb-4d13-84d5-b77cdd1e96ac" containerName="mariadb-account-create-update" Feb 02 09:24:57 crc kubenswrapper[4764]: I0202 09:24:57.697212 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="467a028d-ccfb-4d13-84d5-b77cdd1e96ac" containerName="mariadb-account-create-update" Feb 02 09:24:57 crc kubenswrapper[4764]: I0202 09:24:57.697351 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="467a028d-ccfb-4d13-84d5-b77cdd1e96ac" containerName="mariadb-account-create-update" Feb 02 09:24:57 crc kubenswrapper[4764]: I0202 09:24:57.697360 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="49eeb3bb-263f-452e-a2dc-9c24c0fe62b1" containerName="mariadb-account-create-update" Feb 02 09:24:57 crc kubenswrapper[4764]: I0202 09:24:57.697371 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="1709efed-3308-4801-a4b4-efd067c6ce90" containerName="mariadb-database-create" Feb 02 09:24:57 crc kubenswrapper[4764]: I0202 09:24:57.697380 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="a3915739-a4ee-4cdd-a27c-58e518e0dade" containerName="mariadb-database-create" Feb 02 09:24:57 crc kubenswrapper[4764]: I0202 09:24:57.697388 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="22c2fdff-3f57-4fc3-bbf4-f459784fa876" containerName="dnsmasq-dns" Feb 02 09:24:57 crc kubenswrapper[4764]: I0202 09:24:57.697404 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="4bb792d3-0d3c-44fc-8d1f-50e9763773dc" containerName="mariadb-database-create" Feb 02 09:24:57 crc kubenswrapper[4764]: I0202 09:24:57.697411 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd7663ce-b995-4503-99d9-d5d22952db2c" containerName="keystone-db-sync" Feb 02 09:24:57 crc kubenswrapper[4764]: I0202 09:24:57.697419 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="a2004c87-c998-4892-ba8f-adeeeb31e746" containerName="mariadb-account-create-update" Feb 02 09:24:57 crc kubenswrapper[4764]: I0202 09:24:57.698971 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-nk659" Feb 02 09:24:57 crc kubenswrapper[4764]: I0202 09:24:57.706611 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Feb 02 09:24:57 crc kubenswrapper[4764]: I0202 09:24:57.706796 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Feb 02 09:24:57 crc kubenswrapper[4764]: I0202 09:24:57.707038 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Feb 02 09:24:57 crc kubenswrapper[4764]: I0202 09:24:57.707165 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Feb 02 09:24:57 crc kubenswrapper[4764]: I0202 09:24:57.707263 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-dds8z" Feb 02 09:24:57 crc kubenswrapper[4764]: I0202 09:24:57.717283 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-nk659"] Feb 02 09:24:57 crc kubenswrapper[4764]: I0202 09:24:57.760739 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9eb8eedc-a81b-4988-acc5-7a4ef9744aa2-scripts\") pod \"keystone-bootstrap-nk659\" (UID: \"9eb8eedc-a81b-4988-acc5-7a4ef9744aa2\") " pod="openstack/keystone-bootstrap-nk659" Feb 02 09:24:57 crc kubenswrapper[4764]: I0202 09:24:57.760829 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/9eb8eedc-a81b-4988-acc5-7a4ef9744aa2-credential-keys\") pod \"keystone-bootstrap-nk659\" (UID: \"9eb8eedc-a81b-4988-acc5-7a4ef9744aa2\") " pod="openstack/keystone-bootstrap-nk659" Feb 02 09:24:57 crc kubenswrapper[4764]: I0202 09:24:57.760853 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9eb8eedc-a81b-4988-acc5-7a4ef9744aa2-fernet-keys\") pod \"keystone-bootstrap-nk659\" (UID: \"9eb8eedc-a81b-4988-acc5-7a4ef9744aa2\") " pod="openstack/keystone-bootstrap-nk659" Feb 02 09:24:57 crc kubenswrapper[4764]: I0202 09:24:57.760869 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wprzq\" (UniqueName: \"kubernetes.io/projected/9eb8eedc-a81b-4988-acc5-7a4ef9744aa2-kube-api-access-wprzq\") pod \"keystone-bootstrap-nk659\" (UID: \"9eb8eedc-a81b-4988-acc5-7a4ef9744aa2\") " pod="openstack/keystone-bootstrap-nk659" Feb 02 09:24:57 crc kubenswrapper[4764]: I0202 09:24:57.760891 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9eb8eedc-a81b-4988-acc5-7a4ef9744aa2-combined-ca-bundle\") pod \"keystone-bootstrap-nk659\" (UID: \"9eb8eedc-a81b-4988-acc5-7a4ef9744aa2\") " pod="openstack/keystone-bootstrap-nk659" Feb 02 09:24:57 crc kubenswrapper[4764]: I0202 09:24:57.760939 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9eb8eedc-a81b-4988-acc5-7a4ef9744aa2-config-data\") pod \"keystone-bootstrap-nk659\" (UID: \"9eb8eedc-a81b-4988-acc5-7a4ef9744aa2\") " pod="openstack/keystone-bootstrap-nk659" Feb 02 09:24:57 crc kubenswrapper[4764]: I0202 09:24:57.762573 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7bc867d685-gw7hp"] Feb 02 09:24:57 crc kubenswrapper[4764]: I0202 09:24:57.763951 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bc867d685-gw7hp" Feb 02 09:24:57 crc kubenswrapper[4764]: I0202 09:24:57.796178 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7bc867d685-gw7hp"] Feb 02 09:24:57 crc kubenswrapper[4764]: I0202 09:24:57.862102 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8b367c30-2cb1-40c4-b259-d9abbfe4cd16-ovsdbserver-sb\") pod \"dnsmasq-dns-7bc867d685-gw7hp\" (UID: \"8b367c30-2cb1-40c4-b259-d9abbfe4cd16\") " pod="openstack/dnsmasq-dns-7bc867d685-gw7hp" Feb 02 09:24:57 crc kubenswrapper[4764]: I0202 09:24:57.862147 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9eb8eedc-a81b-4988-acc5-7a4ef9744aa2-config-data\") pod \"keystone-bootstrap-nk659\" (UID: \"9eb8eedc-a81b-4988-acc5-7a4ef9744aa2\") " pod="openstack/keystone-bootstrap-nk659" Feb 02 09:24:57 crc kubenswrapper[4764]: I0202 09:24:57.862192 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9eb8eedc-a81b-4988-acc5-7a4ef9744aa2-scripts\") pod \"keystone-bootstrap-nk659\" (UID: \"9eb8eedc-a81b-4988-acc5-7a4ef9744aa2\") " pod="openstack/keystone-bootstrap-nk659" Feb 02 09:24:57 crc kubenswrapper[4764]: I0202 09:24:57.862211 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8b367c30-2cb1-40c4-b259-d9abbfe4cd16-dns-svc\") pod \"dnsmasq-dns-7bc867d685-gw7hp\" (UID: \"8b367c30-2cb1-40c4-b259-d9abbfe4cd16\") " pod="openstack/dnsmasq-dns-7bc867d685-gw7hp" Feb 02 09:24:57 crc kubenswrapper[4764]: I0202 09:24:57.862253 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8b367c30-2cb1-40c4-b259-d9abbfe4cd16-config\") pod \"dnsmasq-dns-7bc867d685-gw7hp\" (UID: \"8b367c30-2cb1-40c4-b259-d9abbfe4cd16\") " pod="openstack/dnsmasq-dns-7bc867d685-gw7hp" Feb 02 09:24:57 crc kubenswrapper[4764]: I0202 09:24:57.862315 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pmghh\" (UniqueName: \"kubernetes.io/projected/8b367c30-2cb1-40c4-b259-d9abbfe4cd16-kube-api-access-pmghh\") pod \"dnsmasq-dns-7bc867d685-gw7hp\" (UID: \"8b367c30-2cb1-40c4-b259-d9abbfe4cd16\") " pod="openstack/dnsmasq-dns-7bc867d685-gw7hp" Feb 02 09:24:57 crc kubenswrapper[4764]: I0202 09:24:57.862333 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/9eb8eedc-a81b-4988-acc5-7a4ef9744aa2-credential-keys\") pod \"keystone-bootstrap-nk659\" (UID: \"9eb8eedc-a81b-4988-acc5-7a4ef9744aa2\") " pod="openstack/keystone-bootstrap-nk659" Feb 02 09:24:57 crc kubenswrapper[4764]: I0202 09:24:57.862362 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9eb8eedc-a81b-4988-acc5-7a4ef9744aa2-fernet-keys\") pod \"keystone-bootstrap-nk659\" (UID: \"9eb8eedc-a81b-4988-acc5-7a4ef9744aa2\") " pod="openstack/keystone-bootstrap-nk659" Feb 02 09:24:57 crc kubenswrapper[4764]: I0202 09:24:57.862377 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wprzq\" (UniqueName: \"kubernetes.io/projected/9eb8eedc-a81b-4988-acc5-7a4ef9744aa2-kube-api-access-wprzq\") pod \"keystone-bootstrap-nk659\" (UID: \"9eb8eedc-a81b-4988-acc5-7a4ef9744aa2\") " pod="openstack/keystone-bootstrap-nk659" Feb 02 09:24:57 crc kubenswrapper[4764]: I0202 09:24:57.862397 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9eb8eedc-a81b-4988-acc5-7a4ef9744aa2-combined-ca-bundle\") pod \"keystone-bootstrap-nk659\" (UID: \"9eb8eedc-a81b-4988-acc5-7a4ef9744aa2\") " pod="openstack/keystone-bootstrap-nk659" Feb 02 09:24:57 crc kubenswrapper[4764]: I0202 09:24:57.862434 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8b367c30-2cb1-40c4-b259-d9abbfe4cd16-ovsdbserver-nb\") pod \"dnsmasq-dns-7bc867d685-gw7hp\" (UID: \"8b367c30-2cb1-40c4-b259-d9abbfe4cd16\") " pod="openstack/dnsmasq-dns-7bc867d685-gw7hp" Feb 02 09:24:57 crc kubenswrapper[4764]: I0202 09:24:57.876837 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9eb8eedc-a81b-4988-acc5-7a4ef9744aa2-combined-ca-bundle\") pod \"keystone-bootstrap-nk659\" (UID: \"9eb8eedc-a81b-4988-acc5-7a4ef9744aa2\") " pod="openstack/keystone-bootstrap-nk659" Feb 02 09:24:57 crc kubenswrapper[4764]: I0202 09:24:57.879121 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9eb8eedc-a81b-4988-acc5-7a4ef9744aa2-fernet-keys\") pod \"keystone-bootstrap-nk659\" (UID: \"9eb8eedc-a81b-4988-acc5-7a4ef9744aa2\") " pod="openstack/keystone-bootstrap-nk659" Feb 02 09:24:57 crc kubenswrapper[4764]: I0202 09:24:57.880230 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9eb8eedc-a81b-4988-acc5-7a4ef9744aa2-config-data\") pod \"keystone-bootstrap-nk659\" (UID: \"9eb8eedc-a81b-4988-acc5-7a4ef9744aa2\") " pod="openstack/keystone-bootstrap-nk659" Feb 02 09:24:57 crc kubenswrapper[4764]: I0202 09:24:57.882530 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/9eb8eedc-a81b-4988-acc5-7a4ef9744aa2-credential-keys\") pod \"keystone-bootstrap-nk659\" (UID: \"9eb8eedc-a81b-4988-acc5-7a4ef9744aa2\") " pod="openstack/keystone-bootstrap-nk659" Feb 02 09:24:57 crc kubenswrapper[4764]: I0202 09:24:57.891256 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9eb8eedc-a81b-4988-acc5-7a4ef9744aa2-scripts\") pod \"keystone-bootstrap-nk659\" (UID: \"9eb8eedc-a81b-4988-acc5-7a4ef9744aa2\") " pod="openstack/keystone-bootstrap-nk659" Feb 02 09:24:57 crc kubenswrapper[4764]: I0202 09:24:57.897989 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wprzq\" (UniqueName: \"kubernetes.io/projected/9eb8eedc-a81b-4988-acc5-7a4ef9744aa2-kube-api-access-wprzq\") pod \"keystone-bootstrap-nk659\" (UID: \"9eb8eedc-a81b-4988-acc5-7a4ef9744aa2\") " pod="openstack/keystone-bootstrap-nk659" Feb 02 09:24:57 crc kubenswrapper[4764]: I0202 09:24:57.963360 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pmghh\" (UniqueName: \"kubernetes.io/projected/8b367c30-2cb1-40c4-b259-d9abbfe4cd16-kube-api-access-pmghh\") pod \"dnsmasq-dns-7bc867d685-gw7hp\" (UID: \"8b367c30-2cb1-40c4-b259-d9abbfe4cd16\") " pod="openstack/dnsmasq-dns-7bc867d685-gw7hp" Feb 02 09:24:57 crc kubenswrapper[4764]: I0202 09:24:57.963442 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8b367c30-2cb1-40c4-b259-d9abbfe4cd16-ovsdbserver-nb\") pod \"dnsmasq-dns-7bc867d685-gw7hp\" (UID: \"8b367c30-2cb1-40c4-b259-d9abbfe4cd16\") " pod="openstack/dnsmasq-dns-7bc867d685-gw7hp" Feb 02 09:24:57 crc kubenswrapper[4764]: I0202 09:24:57.963487 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8b367c30-2cb1-40c4-b259-d9abbfe4cd16-ovsdbserver-sb\") pod \"dnsmasq-dns-7bc867d685-gw7hp\" (UID: \"8b367c30-2cb1-40c4-b259-d9abbfe4cd16\") " pod="openstack/dnsmasq-dns-7bc867d685-gw7hp" Feb 02 09:24:57 crc kubenswrapper[4764]: I0202 09:24:57.963534 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8b367c30-2cb1-40c4-b259-d9abbfe4cd16-dns-svc\") pod \"dnsmasq-dns-7bc867d685-gw7hp\" (UID: \"8b367c30-2cb1-40c4-b259-d9abbfe4cd16\") " pod="openstack/dnsmasq-dns-7bc867d685-gw7hp" Feb 02 09:24:57 crc kubenswrapper[4764]: I0202 09:24:57.963582 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8b367c30-2cb1-40c4-b259-d9abbfe4cd16-config\") pod \"dnsmasq-dns-7bc867d685-gw7hp\" (UID: \"8b367c30-2cb1-40c4-b259-d9abbfe4cd16\") " pod="openstack/dnsmasq-dns-7bc867d685-gw7hp" Feb 02 09:24:57 crc kubenswrapper[4764]: I0202 09:24:57.967901 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8b367c30-2cb1-40c4-b259-d9abbfe4cd16-ovsdbserver-sb\") pod \"dnsmasq-dns-7bc867d685-gw7hp\" (UID: \"8b367c30-2cb1-40c4-b259-d9abbfe4cd16\") " pod="openstack/dnsmasq-dns-7bc867d685-gw7hp" Feb 02 09:24:57 crc kubenswrapper[4764]: I0202 09:24:57.968092 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8b367c30-2cb1-40c4-b259-d9abbfe4cd16-dns-svc\") pod \"dnsmasq-dns-7bc867d685-gw7hp\" (UID: \"8b367c30-2cb1-40c4-b259-d9abbfe4cd16\") " pod="openstack/dnsmasq-dns-7bc867d685-gw7hp" Feb 02 09:24:57 crc kubenswrapper[4764]: I0202 09:24:57.968182 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8b367c30-2cb1-40c4-b259-d9abbfe4cd16-config\") pod \"dnsmasq-dns-7bc867d685-gw7hp\" (UID: \"8b367c30-2cb1-40c4-b259-d9abbfe4cd16\") " pod="openstack/dnsmasq-dns-7bc867d685-gw7hp" Feb 02 09:24:57 crc kubenswrapper[4764]: I0202 09:24:57.968221 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8b367c30-2cb1-40c4-b259-d9abbfe4cd16-ovsdbserver-nb\") pod \"dnsmasq-dns-7bc867d685-gw7hp\" (UID: \"8b367c30-2cb1-40c4-b259-d9abbfe4cd16\") " pod="openstack/dnsmasq-dns-7bc867d685-gw7hp" Feb 02 09:24:57 crc kubenswrapper[4764]: I0202 09:24:57.998776 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pmghh\" (UniqueName: \"kubernetes.io/projected/8b367c30-2cb1-40c4-b259-d9abbfe4cd16-kube-api-access-pmghh\") pod \"dnsmasq-dns-7bc867d685-gw7hp\" (UID: \"8b367c30-2cb1-40c4-b259-d9abbfe4cd16\") " pod="openstack/dnsmasq-dns-7bc867d685-gw7hp" Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.033370 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-nk659" Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.039087 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-dkrnx"] Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.040105 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-dkrnx" Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.043752 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-4ftqb" Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.066418 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.068744 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.088536 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-dkrnx"] Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.112497 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bc867d685-gw7hp" Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.123763 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b3d3314-2f16-4dc3-aa52-5f7f8576813a-combined-ca-bundle\") pod \"cinder-db-sync-dkrnx\" (UID: \"6b3d3314-2f16-4dc3-aa52-5f7f8576813a\") " pod="openstack/cinder-db-sync-dkrnx" Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.123828 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b3d3314-2f16-4dc3-aa52-5f7f8576813a-scripts\") pod \"cinder-db-sync-dkrnx\" (UID: \"6b3d3314-2f16-4dc3-aa52-5f7f8576813a\") " pod="openstack/cinder-db-sync-dkrnx" Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.123932 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hgkdq\" (UniqueName: \"kubernetes.io/projected/6b3d3314-2f16-4dc3-aa52-5f7f8576813a-kube-api-access-hgkdq\") pod \"cinder-db-sync-dkrnx\" (UID: \"6b3d3314-2f16-4dc3-aa52-5f7f8576813a\") " pod="openstack/cinder-db-sync-dkrnx" Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.124021 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/6b3d3314-2f16-4dc3-aa52-5f7f8576813a-db-sync-config-data\") pod \"cinder-db-sync-dkrnx\" (UID: \"6b3d3314-2f16-4dc3-aa52-5f7f8576813a\") " pod="openstack/cinder-db-sync-dkrnx" Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.124096 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b3d3314-2f16-4dc3-aa52-5f7f8576813a-config-data\") pod \"cinder-db-sync-dkrnx\" (UID: \"6b3d3314-2f16-4dc3-aa52-5f7f8576813a\") " pod="openstack/cinder-db-sync-dkrnx" Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.124183 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6b3d3314-2f16-4dc3-aa52-5f7f8576813a-etc-machine-id\") pod \"cinder-db-sync-dkrnx\" (UID: \"6b3d3314-2f16-4dc3-aa52-5f7f8576813a\") " pod="openstack/cinder-db-sync-dkrnx" Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.176104 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-sm8gf"] Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.179891 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-sm8gf" Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.187369 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-g8sjd" Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.187635 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.203828 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-sm8gf"] Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.220076 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-p7g2s"] Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.221007 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-p7g2s" Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.225182 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b3d3314-2f16-4dc3-aa52-5f7f8576813a-combined-ca-bundle\") pod \"cinder-db-sync-dkrnx\" (UID: \"6b3d3314-2f16-4dc3-aa52-5f7f8576813a\") " pod="openstack/cinder-db-sync-dkrnx" Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.225221 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b3d3314-2f16-4dc3-aa52-5f7f8576813a-scripts\") pod \"cinder-db-sync-dkrnx\" (UID: \"6b3d3314-2f16-4dc3-aa52-5f7f8576813a\") " pod="openstack/cinder-db-sync-dkrnx" Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.225246 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/e61fa54a-f23a-409c-923f-522e04a8099f-db-sync-config-data\") pod \"barbican-db-sync-sm8gf\" (UID: \"e61fa54a-f23a-409c-923f-522e04a8099f\") " pod="openstack/barbican-db-sync-sm8gf" Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.225291 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hgkdq\" (UniqueName: \"kubernetes.io/projected/6b3d3314-2f16-4dc3-aa52-5f7f8576813a-kube-api-access-hgkdq\") pod \"cinder-db-sync-dkrnx\" (UID: \"6b3d3314-2f16-4dc3-aa52-5f7f8576813a\") " pod="openstack/cinder-db-sync-dkrnx" Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.225316 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/6b3d3314-2f16-4dc3-aa52-5f7f8576813a-db-sync-config-data\") pod \"cinder-db-sync-dkrnx\" (UID: \"6b3d3314-2f16-4dc3-aa52-5f7f8576813a\") " pod="openstack/cinder-db-sync-dkrnx" Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.225343 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b3d3314-2f16-4dc3-aa52-5f7f8576813a-config-data\") pod \"cinder-db-sync-dkrnx\" (UID: \"6b3d3314-2f16-4dc3-aa52-5f7f8576813a\") " pod="openstack/cinder-db-sync-dkrnx" Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.225362 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b8pms\" (UniqueName: \"kubernetes.io/projected/e61fa54a-f23a-409c-923f-522e04a8099f-kube-api-access-b8pms\") pod \"barbican-db-sync-sm8gf\" (UID: \"e61fa54a-f23a-409c-923f-522e04a8099f\") " pod="openstack/barbican-db-sync-sm8gf" Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.225381 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e61fa54a-f23a-409c-923f-522e04a8099f-combined-ca-bundle\") pod \"barbican-db-sync-sm8gf\" (UID: \"e61fa54a-f23a-409c-923f-522e04a8099f\") " pod="openstack/barbican-db-sync-sm8gf" Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.225416 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6b3d3314-2f16-4dc3-aa52-5f7f8576813a-etc-machine-id\") pod \"cinder-db-sync-dkrnx\" (UID: \"6b3d3314-2f16-4dc3-aa52-5f7f8576813a\") " pod="openstack/cinder-db-sync-dkrnx" Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.225540 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6b3d3314-2f16-4dc3-aa52-5f7f8576813a-etc-machine-id\") pod \"cinder-db-sync-dkrnx\" (UID: \"6b3d3314-2f16-4dc3-aa52-5f7f8576813a\") " pod="openstack/cinder-db-sync-dkrnx" Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.232384 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-hpwwg" Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.232569 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.232687 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.239175 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b3d3314-2f16-4dc3-aa52-5f7f8576813a-combined-ca-bundle\") pod \"cinder-db-sync-dkrnx\" (UID: \"6b3d3314-2f16-4dc3-aa52-5f7f8576813a\") " pod="openstack/cinder-db-sync-dkrnx" Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.240242 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/6b3d3314-2f16-4dc3-aa52-5f7f8576813a-db-sync-config-data\") pod \"cinder-db-sync-dkrnx\" (UID: \"6b3d3314-2f16-4dc3-aa52-5f7f8576813a\") " pod="openstack/cinder-db-sync-dkrnx" Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.241065 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b3d3314-2f16-4dc3-aa52-5f7f8576813a-config-data\") pod \"cinder-db-sync-dkrnx\" (UID: \"6b3d3314-2f16-4dc3-aa52-5f7f8576813a\") " pod="openstack/cinder-db-sync-dkrnx" Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.247052 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b3d3314-2f16-4dc3-aa52-5f7f8576813a-scripts\") pod \"cinder-db-sync-dkrnx\" (UID: \"6b3d3314-2f16-4dc3-aa52-5f7f8576813a\") " pod="openstack/cinder-db-sync-dkrnx" Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.256458 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7bc867d685-gw7hp"] Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.256507 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-229dq"] Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.257602 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-229dq" Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.276057 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-w8vss" Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.276218 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.276416 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.282495 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hgkdq\" (UniqueName: \"kubernetes.io/projected/6b3d3314-2f16-4dc3-aa52-5f7f8576813a-kube-api-access-hgkdq\") pod \"cinder-db-sync-dkrnx\" (UID: \"6b3d3314-2f16-4dc3-aa52-5f7f8576813a\") " pod="openstack/cinder-db-sync-dkrnx" Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.295174 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5bfb76cfc-8k8x2"] Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.296383 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bfb76cfc-8k8x2" Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.334472 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b8pms\" (UniqueName: \"kubernetes.io/projected/e61fa54a-f23a-409c-923f-522e04a8099f-kube-api-access-b8pms\") pod \"barbican-db-sync-sm8gf\" (UID: \"e61fa54a-f23a-409c-923f-522e04a8099f\") " pod="openstack/barbican-db-sync-sm8gf" Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.336626 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e61fa54a-f23a-409c-923f-522e04a8099f-combined-ca-bundle\") pod \"barbican-db-sync-sm8gf\" (UID: \"e61fa54a-f23a-409c-923f-522e04a8099f\") " pod="openstack/barbican-db-sync-sm8gf" Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.336769 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/e61fa54a-f23a-409c-923f-522e04a8099f-db-sync-config-data\") pod \"barbican-db-sync-sm8gf\" (UID: \"e61fa54a-f23a-409c-923f-522e04a8099f\") " pod="openstack/barbican-db-sync-sm8gf" Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.343801 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-229dq"] Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.352656 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/e61fa54a-f23a-409c-923f-522e04a8099f-db-sync-config-data\") pod \"barbican-db-sync-sm8gf\" (UID: \"e61fa54a-f23a-409c-923f-522e04a8099f\") " pod="openstack/barbican-db-sync-sm8gf" Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.357696 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e61fa54a-f23a-409c-923f-522e04a8099f-combined-ca-bundle\") pod \"barbican-db-sync-sm8gf\" (UID: \"e61fa54a-f23a-409c-923f-522e04a8099f\") " pod="openstack/barbican-db-sync-sm8gf" Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.359342 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-p7g2s"] Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.389536 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b8pms\" (UniqueName: \"kubernetes.io/projected/e61fa54a-f23a-409c-923f-522e04a8099f-kube-api-access-b8pms\") pod \"barbican-db-sync-sm8gf\" (UID: \"e61fa54a-f23a-409c-923f-522e04a8099f\") " pod="openstack/barbican-db-sync-sm8gf" Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.439741 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.442009 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/fd30e6c5-1c0f-4ee4-9df1-c23b745ec136-config\") pod \"neutron-db-sync-p7g2s\" (UID: \"fd30e6c5-1c0f-4ee4-9df1-c23b745ec136\") " pod="openstack/neutron-db-sync-p7g2s" Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.442053 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1ad03482-854b-4e3a-8ee4-8e4f22b68d3c-ovsdbserver-nb\") pod \"dnsmasq-dns-5bfb76cfc-8k8x2\" (UID: \"1ad03482-854b-4e3a-8ee4-8e4f22b68d3c\") " pod="openstack/dnsmasq-dns-5bfb76cfc-8k8x2" Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.442084 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1ad03482-854b-4e3a-8ee4-8e4f22b68d3c-dns-svc\") pod \"dnsmasq-dns-5bfb76cfc-8k8x2\" (UID: \"1ad03482-854b-4e3a-8ee4-8e4f22b68d3c\") " pod="openstack/dnsmasq-dns-5bfb76cfc-8k8x2" Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.442106 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9d4jp\" (UniqueName: \"kubernetes.io/projected/2f7f0eb3-55c5-4dbc-8636-eba9a31b0a27-kube-api-access-9d4jp\") pod \"placement-db-sync-229dq\" (UID: \"2f7f0eb3-55c5-4dbc-8636-eba9a31b0a27\") " pod="openstack/placement-db-sync-229dq" Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.442140 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xr2tx\" (UniqueName: \"kubernetes.io/projected/fd30e6c5-1c0f-4ee4-9df1-c23b745ec136-kube-api-access-xr2tx\") pod \"neutron-db-sync-p7g2s\" (UID: \"fd30e6c5-1c0f-4ee4-9df1-c23b745ec136\") " pod="openstack/neutron-db-sync-p7g2s" Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.442161 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2f7f0eb3-55c5-4dbc-8636-eba9a31b0a27-scripts\") pod \"placement-db-sync-229dq\" (UID: \"2f7f0eb3-55c5-4dbc-8636-eba9a31b0a27\") " pod="openstack/placement-db-sync-229dq" Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.442211 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f7f0eb3-55c5-4dbc-8636-eba9a31b0a27-config-data\") pod \"placement-db-sync-229dq\" (UID: \"2f7f0eb3-55c5-4dbc-8636-eba9a31b0a27\") " pod="openstack/placement-db-sync-229dq" Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.442245 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f7f0eb3-55c5-4dbc-8636-eba9a31b0a27-combined-ca-bundle\") pod \"placement-db-sync-229dq\" (UID: \"2f7f0eb3-55c5-4dbc-8636-eba9a31b0a27\") " pod="openstack/placement-db-sync-229dq" Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.442280 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2f7f0eb3-55c5-4dbc-8636-eba9a31b0a27-logs\") pod \"placement-db-sync-229dq\" (UID: \"2f7f0eb3-55c5-4dbc-8636-eba9a31b0a27\") " pod="openstack/placement-db-sync-229dq" Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.442347 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd30e6c5-1c0f-4ee4-9df1-c23b745ec136-combined-ca-bundle\") pod \"neutron-db-sync-p7g2s\" (UID: \"fd30e6c5-1c0f-4ee4-9df1-c23b745ec136\") " pod="openstack/neutron-db-sync-p7g2s" Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.442496 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1ad03482-854b-4e3a-8ee4-8e4f22b68d3c-ovsdbserver-sb\") pod \"dnsmasq-dns-5bfb76cfc-8k8x2\" (UID: \"1ad03482-854b-4e3a-8ee4-8e4f22b68d3c\") " pod="openstack/dnsmasq-dns-5bfb76cfc-8k8x2" Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.442528 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pwn69\" (UniqueName: \"kubernetes.io/projected/1ad03482-854b-4e3a-8ee4-8e4f22b68d3c-kube-api-access-pwn69\") pod \"dnsmasq-dns-5bfb76cfc-8k8x2\" (UID: \"1ad03482-854b-4e3a-8ee4-8e4f22b68d3c\") " pod="openstack/dnsmasq-dns-5bfb76cfc-8k8x2" Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.442727 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1ad03482-854b-4e3a-8ee4-8e4f22b68d3c-config\") pod \"dnsmasq-dns-5bfb76cfc-8k8x2\" (UID: \"1ad03482-854b-4e3a-8ee4-8e4f22b68d3c\") " pod="openstack/dnsmasq-dns-5bfb76cfc-8k8x2" Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.449517 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.453573 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.454788 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.482789 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-dkrnx" Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.497353 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5bfb76cfc-8k8x2"] Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.507343 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-sm8gf" Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.510900 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.545186 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a518caeb-6b91-4b71-a47f-0d71d965af2e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a518caeb-6b91-4b71-a47f-0d71d965af2e\") " pod="openstack/ceilometer-0" Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.545245 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1ad03482-854b-4e3a-8ee4-8e4f22b68d3c-ovsdbserver-sb\") pod \"dnsmasq-dns-5bfb76cfc-8k8x2\" (UID: \"1ad03482-854b-4e3a-8ee4-8e4f22b68d3c\") " pod="openstack/dnsmasq-dns-5bfb76cfc-8k8x2" Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.545267 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a518caeb-6b91-4b71-a47f-0d71d965af2e-run-httpd\") pod \"ceilometer-0\" (UID: \"a518caeb-6b91-4b71-a47f-0d71d965af2e\") " pod="openstack/ceilometer-0" Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.545286 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pwn69\" (UniqueName: \"kubernetes.io/projected/1ad03482-854b-4e3a-8ee4-8e4f22b68d3c-kube-api-access-pwn69\") pod \"dnsmasq-dns-5bfb76cfc-8k8x2\" (UID: \"1ad03482-854b-4e3a-8ee4-8e4f22b68d3c\") " pod="openstack/dnsmasq-dns-5bfb76cfc-8k8x2" Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.545305 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gnldx\" (UniqueName: \"kubernetes.io/projected/a518caeb-6b91-4b71-a47f-0d71d965af2e-kube-api-access-gnldx\") pod \"ceilometer-0\" (UID: \"a518caeb-6b91-4b71-a47f-0d71d965af2e\") " pod="openstack/ceilometer-0" Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.545336 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1ad03482-854b-4e3a-8ee4-8e4f22b68d3c-config\") pod \"dnsmasq-dns-5bfb76cfc-8k8x2\" (UID: \"1ad03482-854b-4e3a-8ee4-8e4f22b68d3c\") " pod="openstack/dnsmasq-dns-5bfb76cfc-8k8x2" Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.547174 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1ad03482-854b-4e3a-8ee4-8e4f22b68d3c-ovsdbserver-sb\") pod \"dnsmasq-dns-5bfb76cfc-8k8x2\" (UID: \"1ad03482-854b-4e3a-8ee4-8e4f22b68d3c\") " pod="openstack/dnsmasq-dns-5bfb76cfc-8k8x2" Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.547554 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a518caeb-6b91-4b71-a47f-0d71d965af2e-scripts\") pod \"ceilometer-0\" (UID: \"a518caeb-6b91-4b71-a47f-0d71d965af2e\") " pod="openstack/ceilometer-0" Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.547608 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/fd30e6c5-1c0f-4ee4-9df1-c23b745ec136-config\") pod \"neutron-db-sync-p7g2s\" (UID: \"fd30e6c5-1c0f-4ee4-9df1-c23b745ec136\") " pod="openstack/neutron-db-sync-p7g2s" Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.547636 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1ad03482-854b-4e3a-8ee4-8e4f22b68d3c-ovsdbserver-nb\") pod \"dnsmasq-dns-5bfb76cfc-8k8x2\" (UID: \"1ad03482-854b-4e3a-8ee4-8e4f22b68d3c\") " pod="openstack/dnsmasq-dns-5bfb76cfc-8k8x2" Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.547903 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a518caeb-6b91-4b71-a47f-0d71d965af2e-log-httpd\") pod \"ceilometer-0\" (UID: \"a518caeb-6b91-4b71-a47f-0d71d965af2e\") " pod="openstack/ceilometer-0" Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.547969 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1ad03482-854b-4e3a-8ee4-8e4f22b68d3c-dns-svc\") pod \"dnsmasq-dns-5bfb76cfc-8k8x2\" (UID: \"1ad03482-854b-4e3a-8ee4-8e4f22b68d3c\") " pod="openstack/dnsmasq-dns-5bfb76cfc-8k8x2" Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.548000 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9d4jp\" (UniqueName: \"kubernetes.io/projected/2f7f0eb3-55c5-4dbc-8636-eba9a31b0a27-kube-api-access-9d4jp\") pod \"placement-db-sync-229dq\" (UID: \"2f7f0eb3-55c5-4dbc-8636-eba9a31b0a27\") " pod="openstack/placement-db-sync-229dq" Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.548175 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xr2tx\" (UniqueName: \"kubernetes.io/projected/fd30e6c5-1c0f-4ee4-9df1-c23b745ec136-kube-api-access-xr2tx\") pod \"neutron-db-sync-p7g2s\" (UID: \"fd30e6c5-1c0f-4ee4-9df1-c23b745ec136\") " pod="openstack/neutron-db-sync-p7g2s" Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.548203 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2f7f0eb3-55c5-4dbc-8636-eba9a31b0a27-scripts\") pod \"placement-db-sync-229dq\" (UID: \"2f7f0eb3-55c5-4dbc-8636-eba9a31b0a27\") " pod="openstack/placement-db-sync-229dq" Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.548220 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f7f0eb3-55c5-4dbc-8636-eba9a31b0a27-config-data\") pod \"placement-db-sync-229dq\" (UID: \"2f7f0eb3-55c5-4dbc-8636-eba9a31b0a27\") " pod="openstack/placement-db-sync-229dq" Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.548240 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f7f0eb3-55c5-4dbc-8636-eba9a31b0a27-combined-ca-bundle\") pod \"placement-db-sync-229dq\" (UID: \"2f7f0eb3-55c5-4dbc-8636-eba9a31b0a27\") " pod="openstack/placement-db-sync-229dq" Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.548259 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2f7f0eb3-55c5-4dbc-8636-eba9a31b0a27-logs\") pod \"placement-db-sync-229dq\" (UID: \"2f7f0eb3-55c5-4dbc-8636-eba9a31b0a27\") " pod="openstack/placement-db-sync-229dq" Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.548771 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd30e6c5-1c0f-4ee4-9df1-c23b745ec136-combined-ca-bundle\") pod \"neutron-db-sync-p7g2s\" (UID: \"fd30e6c5-1c0f-4ee4-9df1-c23b745ec136\") " pod="openstack/neutron-db-sync-p7g2s" Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.548848 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a518caeb-6b91-4b71-a47f-0d71d965af2e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a518caeb-6b91-4b71-a47f-0d71d965af2e\") " pod="openstack/ceilometer-0" Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.548910 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a518caeb-6b91-4b71-a47f-0d71d965af2e-config-data\") pod \"ceilometer-0\" (UID: \"a518caeb-6b91-4b71-a47f-0d71d965af2e\") " pod="openstack/ceilometer-0" Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.549724 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1ad03482-854b-4e3a-8ee4-8e4f22b68d3c-config\") pod \"dnsmasq-dns-5bfb76cfc-8k8x2\" (UID: \"1ad03482-854b-4e3a-8ee4-8e4f22b68d3c\") " pod="openstack/dnsmasq-dns-5bfb76cfc-8k8x2" Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.563976 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2f7f0eb3-55c5-4dbc-8636-eba9a31b0a27-logs\") pod \"placement-db-sync-229dq\" (UID: \"2f7f0eb3-55c5-4dbc-8636-eba9a31b0a27\") " pod="openstack/placement-db-sync-229dq" Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.567969 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1ad03482-854b-4e3a-8ee4-8e4f22b68d3c-dns-svc\") pod \"dnsmasq-dns-5bfb76cfc-8k8x2\" (UID: \"1ad03482-854b-4e3a-8ee4-8e4f22b68d3c\") " pod="openstack/dnsmasq-dns-5bfb76cfc-8k8x2" Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.568467 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2f7f0eb3-55c5-4dbc-8636-eba9a31b0a27-scripts\") pod \"placement-db-sync-229dq\" (UID: \"2f7f0eb3-55c5-4dbc-8636-eba9a31b0a27\") " pod="openstack/placement-db-sync-229dq" Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.573480 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1ad03482-854b-4e3a-8ee4-8e4f22b68d3c-ovsdbserver-nb\") pod \"dnsmasq-dns-5bfb76cfc-8k8x2\" (UID: \"1ad03482-854b-4e3a-8ee4-8e4f22b68d3c\") " pod="openstack/dnsmasq-dns-5bfb76cfc-8k8x2" Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.584681 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f7f0eb3-55c5-4dbc-8636-eba9a31b0a27-config-data\") pod \"placement-db-sync-229dq\" (UID: \"2f7f0eb3-55c5-4dbc-8636-eba9a31b0a27\") " pod="openstack/placement-db-sync-229dq" Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.588688 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd30e6c5-1c0f-4ee4-9df1-c23b745ec136-combined-ca-bundle\") pod \"neutron-db-sync-p7g2s\" (UID: \"fd30e6c5-1c0f-4ee4-9df1-c23b745ec136\") " pod="openstack/neutron-db-sync-p7g2s" Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.596160 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/fd30e6c5-1c0f-4ee4-9df1-c23b745ec136-config\") pod \"neutron-db-sync-p7g2s\" (UID: \"fd30e6c5-1c0f-4ee4-9df1-c23b745ec136\") " pod="openstack/neutron-db-sync-p7g2s" Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.598735 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9d4jp\" (UniqueName: \"kubernetes.io/projected/2f7f0eb3-55c5-4dbc-8636-eba9a31b0a27-kube-api-access-9d4jp\") pod \"placement-db-sync-229dq\" (UID: \"2f7f0eb3-55c5-4dbc-8636-eba9a31b0a27\") " pod="openstack/placement-db-sync-229dq" Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.600184 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xr2tx\" (UniqueName: \"kubernetes.io/projected/fd30e6c5-1c0f-4ee4-9df1-c23b745ec136-kube-api-access-xr2tx\") pod \"neutron-db-sync-p7g2s\" (UID: \"fd30e6c5-1c0f-4ee4-9df1-c23b745ec136\") " pod="openstack/neutron-db-sync-p7g2s" Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.604552 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f7f0eb3-55c5-4dbc-8636-eba9a31b0a27-combined-ca-bundle\") pod \"placement-db-sync-229dq\" (UID: \"2f7f0eb3-55c5-4dbc-8636-eba9a31b0a27\") " pod="openstack/placement-db-sync-229dq" Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.605138 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pwn69\" (UniqueName: \"kubernetes.io/projected/1ad03482-854b-4e3a-8ee4-8e4f22b68d3c-kube-api-access-pwn69\") pod \"dnsmasq-dns-5bfb76cfc-8k8x2\" (UID: \"1ad03482-854b-4e3a-8ee4-8e4f22b68d3c\") " pod="openstack/dnsmasq-dns-5bfb76cfc-8k8x2" Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.618794 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-229dq" Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.643250 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bfb76cfc-8k8x2" Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.659758 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a518caeb-6b91-4b71-a47f-0d71d965af2e-scripts\") pod \"ceilometer-0\" (UID: \"a518caeb-6b91-4b71-a47f-0d71d965af2e\") " pod="openstack/ceilometer-0" Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.659810 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a518caeb-6b91-4b71-a47f-0d71d965af2e-log-httpd\") pod \"ceilometer-0\" (UID: \"a518caeb-6b91-4b71-a47f-0d71d965af2e\") " pod="openstack/ceilometer-0" Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.659901 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a518caeb-6b91-4b71-a47f-0d71d965af2e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a518caeb-6b91-4b71-a47f-0d71d965af2e\") " pod="openstack/ceilometer-0" Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.659936 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a518caeb-6b91-4b71-a47f-0d71d965af2e-config-data\") pod \"ceilometer-0\" (UID: \"a518caeb-6b91-4b71-a47f-0d71d965af2e\") " pod="openstack/ceilometer-0" Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.659972 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a518caeb-6b91-4b71-a47f-0d71d965af2e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a518caeb-6b91-4b71-a47f-0d71d965af2e\") " pod="openstack/ceilometer-0" Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.660000 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a518caeb-6b91-4b71-a47f-0d71d965af2e-run-httpd\") pod \"ceilometer-0\" (UID: \"a518caeb-6b91-4b71-a47f-0d71d965af2e\") " pod="openstack/ceilometer-0" Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.660016 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gnldx\" (UniqueName: \"kubernetes.io/projected/a518caeb-6b91-4b71-a47f-0d71d965af2e-kube-api-access-gnldx\") pod \"ceilometer-0\" (UID: \"a518caeb-6b91-4b71-a47f-0d71d965af2e\") " pod="openstack/ceilometer-0" Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.663269 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a518caeb-6b91-4b71-a47f-0d71d965af2e-run-httpd\") pod \"ceilometer-0\" (UID: \"a518caeb-6b91-4b71-a47f-0d71d965af2e\") " pod="openstack/ceilometer-0" Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.664101 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a518caeb-6b91-4b71-a47f-0d71d965af2e-scripts\") pod \"ceilometer-0\" (UID: \"a518caeb-6b91-4b71-a47f-0d71d965af2e\") " pod="openstack/ceilometer-0" Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.665717 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a518caeb-6b91-4b71-a47f-0d71d965af2e-config-data\") pod \"ceilometer-0\" (UID: \"a518caeb-6b91-4b71-a47f-0d71d965af2e\") " pod="openstack/ceilometer-0" Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.665791 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a518caeb-6b91-4b71-a47f-0d71d965af2e-log-httpd\") pod \"ceilometer-0\" (UID: \"a518caeb-6b91-4b71-a47f-0d71d965af2e\") " pod="openstack/ceilometer-0" Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.672202 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a518caeb-6b91-4b71-a47f-0d71d965af2e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a518caeb-6b91-4b71-a47f-0d71d965af2e\") " pod="openstack/ceilometer-0" Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.674664 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a518caeb-6b91-4b71-a47f-0d71d965af2e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a518caeb-6b91-4b71-a47f-0d71d965af2e\") " pod="openstack/ceilometer-0" Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.683245 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gnldx\" (UniqueName: \"kubernetes.io/projected/a518caeb-6b91-4b71-a47f-0d71d965af2e-kube-api-access-gnldx\") pod \"ceilometer-0\" (UID: \"a518caeb-6b91-4b71-a47f-0d71d965af2e\") " pod="openstack/ceilometer-0" Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.809266 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.869573 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-nk659"] Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.881470 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-p7g2s" Feb 02 09:24:58 crc kubenswrapper[4764]: I0202 09:24:58.936519 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7bc867d685-gw7hp"] Feb 02 09:24:59 crc kubenswrapper[4764]: I0202 09:24:59.199565 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-dkrnx"] Feb 02 09:24:59 crc kubenswrapper[4764]: W0202 09:24:59.209639 4764 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6b3d3314_2f16_4dc3_aa52_5f7f8576813a.slice/crio-0260dc4698947ea66a729c434ded2d4f27469d3098b4e9a0bd7baf2466867ada WatchSource:0}: Error finding container 0260dc4698947ea66a729c434ded2d4f27469d3098b4e9a0bd7baf2466867ada: Status 404 returned error can't find the container with id 0260dc4698947ea66a729c434ded2d4f27469d3098b4e9a0bd7baf2466867ada Feb 02 09:24:59 crc kubenswrapper[4764]: W0202 09:24:59.214705 4764 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode61fa54a_f23a_409c_923f_522e04a8099f.slice/crio-f511998e84a5998777e047e56bfbef4f594c30f51da0d4596f7c24a44eacd196 WatchSource:0}: Error finding container f511998e84a5998777e047e56bfbef4f594c30f51da0d4596f7c24a44eacd196: Status 404 returned error can't find the container with id f511998e84a5998777e047e56bfbef4f594c30f51da0d4596f7c24a44eacd196 Feb 02 09:24:59 crc kubenswrapper[4764]: I0202 09:24:59.218896 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-sm8gf"] Feb 02 09:24:59 crc kubenswrapper[4764]: I0202 09:24:59.388057 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-229dq"] Feb 02 09:24:59 crc kubenswrapper[4764]: I0202 09:24:59.389440 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5bfb76cfc-8k8x2"] Feb 02 09:24:59 crc kubenswrapper[4764]: I0202 09:24:59.453538 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-p7g2s"] Feb 02 09:24:59 crc kubenswrapper[4764]: I0202 09:24:59.467474 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-dkrnx" event={"ID":"6b3d3314-2f16-4dc3-aa52-5f7f8576813a","Type":"ContainerStarted","Data":"0260dc4698947ea66a729c434ded2d4f27469d3098b4e9a0bd7baf2466867ada"} Feb 02 09:24:59 crc kubenswrapper[4764]: I0202 09:24:59.471050 4764 generic.go:334] "Generic (PLEG): container finished" podID="8b367c30-2cb1-40c4-b259-d9abbfe4cd16" containerID="92ee32a1b5b83aebf98a0185b0293031ddd07a1b68bd12ec7c4a4bdaecd08004" exitCode=0 Feb 02 09:24:59 crc kubenswrapper[4764]: I0202 09:24:59.471222 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bc867d685-gw7hp" event={"ID":"8b367c30-2cb1-40c4-b259-d9abbfe4cd16","Type":"ContainerDied","Data":"92ee32a1b5b83aebf98a0185b0293031ddd07a1b68bd12ec7c4a4bdaecd08004"} Feb 02 09:24:59 crc kubenswrapper[4764]: I0202 09:24:59.471252 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bc867d685-gw7hp" event={"ID":"8b367c30-2cb1-40c4-b259-d9abbfe4cd16","Type":"ContainerStarted","Data":"deecf5243fdb0f5a611a3726ebc76b96b07db6f4735b2b10372efd38f814b845"} Feb 02 09:24:59 crc kubenswrapper[4764]: I0202 09:24:59.497587 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-sm8gf" event={"ID":"e61fa54a-f23a-409c-923f-522e04a8099f","Type":"ContainerStarted","Data":"f511998e84a5998777e047e56bfbef4f594c30f51da0d4596f7c24a44eacd196"} Feb 02 09:24:59 crc kubenswrapper[4764]: I0202 09:24:59.502017 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-229dq" event={"ID":"2f7f0eb3-55c5-4dbc-8636-eba9a31b0a27","Type":"ContainerStarted","Data":"78ddb63cbddf7138fea265f3af07ddc4eab968da6cfc8c842b00d9dfd20e05c7"} Feb 02 09:24:59 crc kubenswrapper[4764]: I0202 09:24:59.524623 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-nk659" event={"ID":"9eb8eedc-a81b-4988-acc5-7a4ef9744aa2","Type":"ContainerStarted","Data":"074b62c56ecca7f047050603bbafc675a69dad1f862acb1313c25578ceb1bc02"} Feb 02 09:24:59 crc kubenswrapper[4764]: I0202 09:24:59.524677 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-nk659" event={"ID":"9eb8eedc-a81b-4988-acc5-7a4ef9744aa2","Type":"ContainerStarted","Data":"2f71edaf28e73208a13719dfb64c72f6975c6c14845e9938351a9f2855b2c576"} Feb 02 09:24:59 crc kubenswrapper[4764]: I0202 09:24:59.552364 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-nk659" podStartSLOduration=2.552345276 podStartE2EDuration="2.552345276s" podCreationTimestamp="2026-02-02 09:24:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:24:59.552225873 +0000 UTC m=+1062.485949961" watchObservedRunningTime="2026-02-02 09:24:59.552345276 +0000 UTC m=+1062.486069384" Feb 02 09:24:59 crc kubenswrapper[4764]: I0202 09:24:59.562295 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bfb76cfc-8k8x2" event={"ID":"1ad03482-854b-4e3a-8ee4-8e4f22b68d3c","Type":"ContainerStarted","Data":"a3197f7eb91ec309af68cce1627830dbbaa2e63867fd7c07d919736fbfe9788c"} Feb 02 09:24:59 crc kubenswrapper[4764]: I0202 09:24:59.626801 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 02 09:24:59 crc kubenswrapper[4764]: I0202 09:24:59.976902 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bc867d685-gw7hp" Feb 02 09:24:59 crc kubenswrapper[4764]: I0202 09:24:59.983084 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 02 09:25:00 crc kubenswrapper[4764]: I0202 09:25:00.001153 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8b367c30-2cb1-40c4-b259-d9abbfe4cd16-ovsdbserver-nb\") pod \"8b367c30-2cb1-40c4-b259-d9abbfe4cd16\" (UID: \"8b367c30-2cb1-40c4-b259-d9abbfe4cd16\") " Feb 02 09:25:00 crc kubenswrapper[4764]: I0202 09:25:00.001233 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pmghh\" (UniqueName: \"kubernetes.io/projected/8b367c30-2cb1-40c4-b259-d9abbfe4cd16-kube-api-access-pmghh\") pod \"8b367c30-2cb1-40c4-b259-d9abbfe4cd16\" (UID: \"8b367c30-2cb1-40c4-b259-d9abbfe4cd16\") " Feb 02 09:25:00 crc kubenswrapper[4764]: I0202 09:25:00.001289 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8b367c30-2cb1-40c4-b259-d9abbfe4cd16-ovsdbserver-sb\") pod \"8b367c30-2cb1-40c4-b259-d9abbfe4cd16\" (UID: \"8b367c30-2cb1-40c4-b259-d9abbfe4cd16\") " Feb 02 09:25:00 crc kubenswrapper[4764]: I0202 09:25:00.001414 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8b367c30-2cb1-40c4-b259-d9abbfe4cd16-config\") pod \"8b367c30-2cb1-40c4-b259-d9abbfe4cd16\" (UID: \"8b367c30-2cb1-40c4-b259-d9abbfe4cd16\") " Feb 02 09:25:00 crc kubenswrapper[4764]: I0202 09:25:00.001446 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8b367c30-2cb1-40c4-b259-d9abbfe4cd16-dns-svc\") pod \"8b367c30-2cb1-40c4-b259-d9abbfe4cd16\" (UID: \"8b367c30-2cb1-40c4-b259-d9abbfe4cd16\") " Feb 02 09:25:00 crc kubenswrapper[4764]: I0202 09:25:00.081665 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8b367c30-2cb1-40c4-b259-d9abbfe4cd16-kube-api-access-pmghh" (OuterVolumeSpecName: "kube-api-access-pmghh") pod "8b367c30-2cb1-40c4-b259-d9abbfe4cd16" (UID: "8b367c30-2cb1-40c4-b259-d9abbfe4cd16"). InnerVolumeSpecName "kube-api-access-pmghh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:25:00 crc kubenswrapper[4764]: I0202 09:25:00.081899 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8b367c30-2cb1-40c4-b259-d9abbfe4cd16-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "8b367c30-2cb1-40c4-b259-d9abbfe4cd16" (UID: "8b367c30-2cb1-40c4-b259-d9abbfe4cd16"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:25:00 crc kubenswrapper[4764]: I0202 09:25:00.082216 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8b367c30-2cb1-40c4-b259-d9abbfe4cd16-config" (OuterVolumeSpecName: "config") pod "8b367c30-2cb1-40c4-b259-d9abbfe4cd16" (UID: "8b367c30-2cb1-40c4-b259-d9abbfe4cd16"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:25:00 crc kubenswrapper[4764]: I0202 09:25:00.104927 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8b367c30-2cb1-40c4-b259-d9abbfe4cd16-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "8b367c30-2cb1-40c4-b259-d9abbfe4cd16" (UID: "8b367c30-2cb1-40c4-b259-d9abbfe4cd16"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:25:00 crc kubenswrapper[4764]: I0202 09:25:00.105027 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8b367c30-2cb1-40c4-b259-d9abbfe4cd16-ovsdbserver-nb\") pod \"8b367c30-2cb1-40c4-b259-d9abbfe4cd16\" (UID: \"8b367c30-2cb1-40c4-b259-d9abbfe4cd16\") " Feb 02 09:25:00 crc kubenswrapper[4764]: I0202 09:25:00.105485 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pmghh\" (UniqueName: \"kubernetes.io/projected/8b367c30-2cb1-40c4-b259-d9abbfe4cd16-kube-api-access-pmghh\") on node \"crc\" DevicePath \"\"" Feb 02 09:25:00 crc kubenswrapper[4764]: I0202 09:25:00.105497 4764 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8b367c30-2cb1-40c4-b259-d9abbfe4cd16-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 02 09:25:00 crc kubenswrapper[4764]: I0202 09:25:00.105508 4764 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8b367c30-2cb1-40c4-b259-d9abbfe4cd16-config\") on node \"crc\" DevicePath \"\"" Feb 02 09:25:00 crc kubenswrapper[4764]: W0202 09:25:00.105587 4764 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/8b367c30-2cb1-40c4-b259-d9abbfe4cd16/volumes/kubernetes.io~configmap/ovsdbserver-nb Feb 02 09:25:00 crc kubenswrapper[4764]: I0202 09:25:00.105598 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8b367c30-2cb1-40c4-b259-d9abbfe4cd16-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "8b367c30-2cb1-40c4-b259-d9abbfe4cd16" (UID: "8b367c30-2cb1-40c4-b259-d9abbfe4cd16"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:25:00 crc kubenswrapper[4764]: I0202 09:25:00.121693 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8b367c30-2cb1-40c4-b259-d9abbfe4cd16-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "8b367c30-2cb1-40c4-b259-d9abbfe4cd16" (UID: "8b367c30-2cb1-40c4-b259-d9abbfe4cd16"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:25:00 crc kubenswrapper[4764]: I0202 09:25:00.208995 4764 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8b367c30-2cb1-40c4-b259-d9abbfe4cd16-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 02 09:25:00 crc kubenswrapper[4764]: I0202 09:25:00.209035 4764 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8b367c30-2cb1-40c4-b259-d9abbfe4cd16-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 02 09:25:00 crc kubenswrapper[4764]: I0202 09:25:00.580859 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-p7g2s" event={"ID":"fd30e6c5-1c0f-4ee4-9df1-c23b745ec136","Type":"ContainerStarted","Data":"adb69964a6b9f244142facf775fd53dfc1a686481ec2d2b42286a7edf9112752"} Feb 02 09:25:00 crc kubenswrapper[4764]: I0202 09:25:00.580908 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-p7g2s" event={"ID":"fd30e6c5-1c0f-4ee4-9df1-c23b745ec136","Type":"ContainerStarted","Data":"46d600cc6b65e2ac379d155f0bb36dc26c310a8d46febaf499d20f0c05426f49"} Feb 02 09:25:00 crc kubenswrapper[4764]: I0202 09:25:00.585242 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a518caeb-6b91-4b71-a47f-0d71d965af2e","Type":"ContainerStarted","Data":"49077403dc69b61caa0b13b5af115287cd97803e61b32b6ae77094c97d68f0cf"} Feb 02 09:25:00 crc kubenswrapper[4764]: I0202 09:25:00.590399 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bc867d685-gw7hp" event={"ID":"8b367c30-2cb1-40c4-b259-d9abbfe4cd16","Type":"ContainerDied","Data":"deecf5243fdb0f5a611a3726ebc76b96b07db6f4735b2b10372efd38f814b845"} Feb 02 09:25:00 crc kubenswrapper[4764]: I0202 09:25:00.590438 4764 scope.go:117] "RemoveContainer" containerID="92ee32a1b5b83aebf98a0185b0293031ddd07a1b68bd12ec7c4a4bdaecd08004" Feb 02 09:25:00 crc kubenswrapper[4764]: I0202 09:25:00.590543 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bc867d685-gw7hp" Feb 02 09:25:00 crc kubenswrapper[4764]: I0202 09:25:00.610203 4764 generic.go:334] "Generic (PLEG): container finished" podID="1ad03482-854b-4e3a-8ee4-8e4f22b68d3c" containerID="8a7877158d5931d89dac739abbe12ca8fb2986dcae9b4a46319b73f7ba6d04d9" exitCode=0 Feb 02 09:25:00 crc kubenswrapper[4764]: I0202 09:25:00.610301 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bfb76cfc-8k8x2" event={"ID":"1ad03482-854b-4e3a-8ee4-8e4f22b68d3c","Type":"ContainerDied","Data":"8a7877158d5931d89dac739abbe12ca8fb2986dcae9b4a46319b73f7ba6d04d9"} Feb 02 09:25:00 crc kubenswrapper[4764]: I0202 09:25:00.611402 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-p7g2s" podStartSLOduration=2.6113791170000002 podStartE2EDuration="2.611379117s" podCreationTimestamp="2026-02-02 09:24:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:25:00.608861731 +0000 UTC m=+1063.542585819" watchObservedRunningTime="2026-02-02 09:25:00.611379117 +0000 UTC m=+1063.545103205" Feb 02 09:25:00 crc kubenswrapper[4764]: I0202 09:25:00.707696 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7bc867d685-gw7hp"] Feb 02 09:25:00 crc kubenswrapper[4764]: I0202 09:25:00.716705 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7bc867d685-gw7hp"] Feb 02 09:25:01 crc kubenswrapper[4764]: I0202 09:25:01.635610 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bfb76cfc-8k8x2" event={"ID":"1ad03482-854b-4e3a-8ee4-8e4f22b68d3c","Type":"ContainerStarted","Data":"8114f849c9b5b2c77035ec1db7e306cb25897aa6ff4fabe90927e8dd131746b6"} Feb 02 09:25:01 crc kubenswrapper[4764]: I0202 09:25:01.637976 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5bfb76cfc-8k8x2" Feb 02 09:25:01 crc kubenswrapper[4764]: I0202 09:25:01.672518 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5bfb76cfc-8k8x2" podStartSLOduration=3.672477214 podStartE2EDuration="3.672477214s" podCreationTimestamp="2026-02-02 09:24:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:25:01.66288561 +0000 UTC m=+1064.596609698" watchObservedRunningTime="2026-02-02 09:25:01.672477214 +0000 UTC m=+1064.606201312" Feb 02 09:25:01 crc kubenswrapper[4764]: I0202 09:25:01.835750 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8b367c30-2cb1-40c4-b259-d9abbfe4cd16" path="/var/lib/kubelet/pods/8b367c30-2cb1-40c4-b259-d9abbfe4cd16/volumes" Feb 02 09:25:05 crc kubenswrapper[4764]: I0202 09:25:05.685029 4764 generic.go:334] "Generic (PLEG): container finished" podID="9eb8eedc-a81b-4988-acc5-7a4ef9744aa2" containerID="074b62c56ecca7f047050603bbafc675a69dad1f862acb1313c25578ceb1bc02" exitCode=0 Feb 02 09:25:05 crc kubenswrapper[4764]: I0202 09:25:05.685115 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-nk659" event={"ID":"9eb8eedc-a81b-4988-acc5-7a4ef9744aa2","Type":"ContainerDied","Data":"074b62c56ecca7f047050603bbafc675a69dad1f862acb1313c25578ceb1bc02"} Feb 02 09:25:08 crc kubenswrapper[4764]: I0202 09:25:08.646337 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5bfb76cfc-8k8x2" Feb 02 09:25:08 crc kubenswrapper[4764]: I0202 09:25:08.756192 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-58976fdf49-bb6rk"] Feb 02 09:25:08 crc kubenswrapper[4764]: I0202 09:25:08.756840 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-58976fdf49-bb6rk" podUID="30d22725-a706-485b-8cc2-964b86137ac1" containerName="dnsmasq-dns" containerID="cri-o://3160f95e976a6f6e0632f2d7adc40045a33aab5f58543831305f2e7c08360f26" gracePeriod=10 Feb 02 09:25:09 crc kubenswrapper[4764]: I0202 09:25:09.764771 4764 generic.go:334] "Generic (PLEG): container finished" podID="30d22725-a706-485b-8cc2-964b86137ac1" containerID="3160f95e976a6f6e0632f2d7adc40045a33aab5f58543831305f2e7c08360f26" exitCode=0 Feb 02 09:25:09 crc kubenswrapper[4764]: I0202 09:25:09.764825 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58976fdf49-bb6rk" event={"ID":"30d22725-a706-485b-8cc2-964b86137ac1","Type":"ContainerDied","Data":"3160f95e976a6f6e0632f2d7adc40045a33aab5f58543831305f2e7c08360f26"} Feb 02 09:25:11 crc kubenswrapper[4764]: I0202 09:25:11.037512 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-nk659" Feb 02 09:25:11 crc kubenswrapper[4764]: I0202 09:25:11.176857 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9eb8eedc-a81b-4988-acc5-7a4ef9744aa2-fernet-keys\") pod \"9eb8eedc-a81b-4988-acc5-7a4ef9744aa2\" (UID: \"9eb8eedc-a81b-4988-acc5-7a4ef9744aa2\") " Feb 02 09:25:11 crc kubenswrapper[4764]: I0202 09:25:11.176965 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9eb8eedc-a81b-4988-acc5-7a4ef9744aa2-config-data\") pod \"9eb8eedc-a81b-4988-acc5-7a4ef9744aa2\" (UID: \"9eb8eedc-a81b-4988-acc5-7a4ef9744aa2\") " Feb 02 09:25:11 crc kubenswrapper[4764]: I0202 09:25:11.177010 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9eb8eedc-a81b-4988-acc5-7a4ef9744aa2-scripts\") pod \"9eb8eedc-a81b-4988-acc5-7a4ef9744aa2\" (UID: \"9eb8eedc-a81b-4988-acc5-7a4ef9744aa2\") " Feb 02 09:25:11 crc kubenswrapper[4764]: I0202 09:25:11.177057 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9eb8eedc-a81b-4988-acc5-7a4ef9744aa2-combined-ca-bundle\") pod \"9eb8eedc-a81b-4988-acc5-7a4ef9744aa2\" (UID: \"9eb8eedc-a81b-4988-acc5-7a4ef9744aa2\") " Feb 02 09:25:11 crc kubenswrapper[4764]: I0202 09:25:11.177135 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wprzq\" (UniqueName: \"kubernetes.io/projected/9eb8eedc-a81b-4988-acc5-7a4ef9744aa2-kube-api-access-wprzq\") pod \"9eb8eedc-a81b-4988-acc5-7a4ef9744aa2\" (UID: \"9eb8eedc-a81b-4988-acc5-7a4ef9744aa2\") " Feb 02 09:25:11 crc kubenswrapper[4764]: I0202 09:25:11.177193 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/9eb8eedc-a81b-4988-acc5-7a4ef9744aa2-credential-keys\") pod \"9eb8eedc-a81b-4988-acc5-7a4ef9744aa2\" (UID: \"9eb8eedc-a81b-4988-acc5-7a4ef9744aa2\") " Feb 02 09:25:11 crc kubenswrapper[4764]: I0202 09:25:11.185907 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9eb8eedc-a81b-4988-acc5-7a4ef9744aa2-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "9eb8eedc-a81b-4988-acc5-7a4ef9744aa2" (UID: "9eb8eedc-a81b-4988-acc5-7a4ef9744aa2"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:25:11 crc kubenswrapper[4764]: I0202 09:25:11.185945 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9eb8eedc-a81b-4988-acc5-7a4ef9744aa2-kube-api-access-wprzq" (OuterVolumeSpecName: "kube-api-access-wprzq") pod "9eb8eedc-a81b-4988-acc5-7a4ef9744aa2" (UID: "9eb8eedc-a81b-4988-acc5-7a4ef9744aa2"). InnerVolumeSpecName "kube-api-access-wprzq". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:25:11 crc kubenswrapper[4764]: I0202 09:25:11.201172 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9eb8eedc-a81b-4988-acc5-7a4ef9744aa2-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "9eb8eedc-a81b-4988-acc5-7a4ef9744aa2" (UID: "9eb8eedc-a81b-4988-acc5-7a4ef9744aa2"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:25:11 crc kubenswrapper[4764]: I0202 09:25:11.202079 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9eb8eedc-a81b-4988-acc5-7a4ef9744aa2-scripts" (OuterVolumeSpecName: "scripts") pod "9eb8eedc-a81b-4988-acc5-7a4ef9744aa2" (UID: "9eb8eedc-a81b-4988-acc5-7a4ef9744aa2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:25:11 crc kubenswrapper[4764]: I0202 09:25:11.209594 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9eb8eedc-a81b-4988-acc5-7a4ef9744aa2-config-data" (OuterVolumeSpecName: "config-data") pod "9eb8eedc-a81b-4988-acc5-7a4ef9744aa2" (UID: "9eb8eedc-a81b-4988-acc5-7a4ef9744aa2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:25:11 crc kubenswrapper[4764]: I0202 09:25:11.217331 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9eb8eedc-a81b-4988-acc5-7a4ef9744aa2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9eb8eedc-a81b-4988-acc5-7a4ef9744aa2" (UID: "9eb8eedc-a81b-4988-acc5-7a4ef9744aa2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:25:11 crc kubenswrapper[4764]: I0202 09:25:11.279254 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wprzq\" (UniqueName: \"kubernetes.io/projected/9eb8eedc-a81b-4988-acc5-7a4ef9744aa2-kube-api-access-wprzq\") on node \"crc\" DevicePath \"\"" Feb 02 09:25:11 crc kubenswrapper[4764]: I0202 09:25:11.279308 4764 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/9eb8eedc-a81b-4988-acc5-7a4ef9744aa2-credential-keys\") on node \"crc\" DevicePath \"\"" Feb 02 09:25:11 crc kubenswrapper[4764]: I0202 09:25:11.279321 4764 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9eb8eedc-a81b-4988-acc5-7a4ef9744aa2-fernet-keys\") on node \"crc\" DevicePath \"\"" Feb 02 09:25:11 crc kubenswrapper[4764]: I0202 09:25:11.279333 4764 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9eb8eedc-a81b-4988-acc5-7a4ef9744aa2-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 09:25:11 crc kubenswrapper[4764]: I0202 09:25:11.279344 4764 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9eb8eedc-a81b-4988-acc5-7a4ef9744aa2-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 09:25:11 crc kubenswrapper[4764]: I0202 09:25:11.279356 4764 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9eb8eedc-a81b-4988-acc5-7a4ef9744aa2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 09:25:11 crc kubenswrapper[4764]: I0202 09:25:11.781854 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-nk659" event={"ID":"9eb8eedc-a81b-4988-acc5-7a4ef9744aa2","Type":"ContainerDied","Data":"2f71edaf28e73208a13719dfb64c72f6975c6c14845e9938351a9f2855b2c576"} Feb 02 09:25:11 crc kubenswrapper[4764]: I0202 09:25:11.782249 4764 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2f71edaf28e73208a13719dfb64c72f6975c6c14845e9938351a9f2855b2c576" Feb 02 09:25:11 crc kubenswrapper[4764]: I0202 09:25:11.782188 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-nk659" Feb 02 09:25:11 crc kubenswrapper[4764]: I0202 09:25:11.888580 4764 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-58976fdf49-bb6rk" podUID="30d22725-a706-485b-8cc2-964b86137ac1" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.121:5353: connect: connection refused" Feb 02 09:25:12 crc kubenswrapper[4764]: I0202 09:25:12.142427 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-nk659"] Feb 02 09:25:12 crc kubenswrapper[4764]: I0202 09:25:12.147556 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-nk659"] Feb 02 09:25:12 crc kubenswrapper[4764]: I0202 09:25:12.246596 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-phkvk"] Feb 02 09:25:12 crc kubenswrapper[4764]: E0202 09:25:12.246887 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9eb8eedc-a81b-4988-acc5-7a4ef9744aa2" containerName="keystone-bootstrap" Feb 02 09:25:12 crc kubenswrapper[4764]: I0202 09:25:12.246904 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="9eb8eedc-a81b-4988-acc5-7a4ef9744aa2" containerName="keystone-bootstrap" Feb 02 09:25:12 crc kubenswrapper[4764]: E0202 09:25:12.246915 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b367c30-2cb1-40c4-b259-d9abbfe4cd16" containerName="init" Feb 02 09:25:12 crc kubenswrapper[4764]: I0202 09:25:12.246922 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b367c30-2cb1-40c4-b259-d9abbfe4cd16" containerName="init" Feb 02 09:25:12 crc kubenswrapper[4764]: I0202 09:25:12.247207 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="9eb8eedc-a81b-4988-acc5-7a4ef9744aa2" containerName="keystone-bootstrap" Feb 02 09:25:12 crc kubenswrapper[4764]: I0202 09:25:12.247225 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="8b367c30-2cb1-40c4-b259-d9abbfe4cd16" containerName="init" Feb 02 09:25:12 crc kubenswrapper[4764]: I0202 09:25:12.247710 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-phkvk" Feb 02 09:25:12 crc kubenswrapper[4764]: I0202 09:25:12.250001 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Feb 02 09:25:12 crc kubenswrapper[4764]: I0202 09:25:12.250281 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-dds8z" Feb 02 09:25:12 crc kubenswrapper[4764]: I0202 09:25:12.251224 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Feb 02 09:25:12 crc kubenswrapper[4764]: I0202 09:25:12.251422 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Feb 02 09:25:12 crc kubenswrapper[4764]: I0202 09:25:12.252188 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Feb 02 09:25:12 crc kubenswrapper[4764]: I0202 09:25:12.256852 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-phkvk"] Feb 02 09:25:12 crc kubenswrapper[4764]: I0202 09:25:12.399360 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4537c8c1-f5dc-4974-9885-527d6c85e100-config-data\") pod \"keystone-bootstrap-phkvk\" (UID: \"4537c8c1-f5dc-4974-9885-527d6c85e100\") " pod="openstack/keystone-bootstrap-phkvk" Feb 02 09:25:12 crc kubenswrapper[4764]: I0202 09:25:12.399422 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4537c8c1-f5dc-4974-9885-527d6c85e100-combined-ca-bundle\") pod \"keystone-bootstrap-phkvk\" (UID: \"4537c8c1-f5dc-4974-9885-527d6c85e100\") " pod="openstack/keystone-bootstrap-phkvk" Feb 02 09:25:12 crc kubenswrapper[4764]: I0202 09:25:12.399445 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4537c8c1-f5dc-4974-9885-527d6c85e100-fernet-keys\") pod \"keystone-bootstrap-phkvk\" (UID: \"4537c8c1-f5dc-4974-9885-527d6c85e100\") " pod="openstack/keystone-bootstrap-phkvk" Feb 02 09:25:12 crc kubenswrapper[4764]: I0202 09:25:12.399467 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4537c8c1-f5dc-4974-9885-527d6c85e100-scripts\") pod \"keystone-bootstrap-phkvk\" (UID: \"4537c8c1-f5dc-4974-9885-527d6c85e100\") " pod="openstack/keystone-bootstrap-phkvk" Feb 02 09:25:12 crc kubenswrapper[4764]: I0202 09:25:12.399525 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k2g4l\" (UniqueName: \"kubernetes.io/projected/4537c8c1-f5dc-4974-9885-527d6c85e100-kube-api-access-k2g4l\") pod \"keystone-bootstrap-phkvk\" (UID: \"4537c8c1-f5dc-4974-9885-527d6c85e100\") " pod="openstack/keystone-bootstrap-phkvk" Feb 02 09:25:12 crc kubenswrapper[4764]: I0202 09:25:12.399575 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/4537c8c1-f5dc-4974-9885-527d6c85e100-credential-keys\") pod \"keystone-bootstrap-phkvk\" (UID: \"4537c8c1-f5dc-4974-9885-527d6c85e100\") " pod="openstack/keystone-bootstrap-phkvk" Feb 02 09:25:12 crc kubenswrapper[4764]: I0202 09:25:12.501410 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/4537c8c1-f5dc-4974-9885-527d6c85e100-credential-keys\") pod \"keystone-bootstrap-phkvk\" (UID: \"4537c8c1-f5dc-4974-9885-527d6c85e100\") " pod="openstack/keystone-bootstrap-phkvk" Feb 02 09:25:12 crc kubenswrapper[4764]: I0202 09:25:12.501488 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4537c8c1-f5dc-4974-9885-527d6c85e100-config-data\") pod \"keystone-bootstrap-phkvk\" (UID: \"4537c8c1-f5dc-4974-9885-527d6c85e100\") " pod="openstack/keystone-bootstrap-phkvk" Feb 02 09:25:12 crc kubenswrapper[4764]: I0202 09:25:12.501549 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4537c8c1-f5dc-4974-9885-527d6c85e100-combined-ca-bundle\") pod \"keystone-bootstrap-phkvk\" (UID: \"4537c8c1-f5dc-4974-9885-527d6c85e100\") " pod="openstack/keystone-bootstrap-phkvk" Feb 02 09:25:12 crc kubenswrapper[4764]: I0202 09:25:12.501584 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4537c8c1-f5dc-4974-9885-527d6c85e100-fernet-keys\") pod \"keystone-bootstrap-phkvk\" (UID: \"4537c8c1-f5dc-4974-9885-527d6c85e100\") " pod="openstack/keystone-bootstrap-phkvk" Feb 02 09:25:12 crc kubenswrapper[4764]: I0202 09:25:12.501619 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4537c8c1-f5dc-4974-9885-527d6c85e100-scripts\") pod \"keystone-bootstrap-phkvk\" (UID: \"4537c8c1-f5dc-4974-9885-527d6c85e100\") " pod="openstack/keystone-bootstrap-phkvk" Feb 02 09:25:12 crc kubenswrapper[4764]: I0202 09:25:12.501706 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k2g4l\" (UniqueName: \"kubernetes.io/projected/4537c8c1-f5dc-4974-9885-527d6c85e100-kube-api-access-k2g4l\") pod \"keystone-bootstrap-phkvk\" (UID: \"4537c8c1-f5dc-4974-9885-527d6c85e100\") " pod="openstack/keystone-bootstrap-phkvk" Feb 02 09:25:12 crc kubenswrapper[4764]: I0202 09:25:12.508800 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/4537c8c1-f5dc-4974-9885-527d6c85e100-credential-keys\") pod \"keystone-bootstrap-phkvk\" (UID: \"4537c8c1-f5dc-4974-9885-527d6c85e100\") " pod="openstack/keystone-bootstrap-phkvk" Feb 02 09:25:12 crc kubenswrapper[4764]: I0202 09:25:12.508863 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4537c8c1-f5dc-4974-9885-527d6c85e100-combined-ca-bundle\") pod \"keystone-bootstrap-phkvk\" (UID: \"4537c8c1-f5dc-4974-9885-527d6c85e100\") " pod="openstack/keystone-bootstrap-phkvk" Feb 02 09:25:12 crc kubenswrapper[4764]: I0202 09:25:12.508995 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4537c8c1-f5dc-4974-9885-527d6c85e100-fernet-keys\") pod \"keystone-bootstrap-phkvk\" (UID: \"4537c8c1-f5dc-4974-9885-527d6c85e100\") " pod="openstack/keystone-bootstrap-phkvk" Feb 02 09:25:12 crc kubenswrapper[4764]: I0202 09:25:12.509561 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4537c8c1-f5dc-4974-9885-527d6c85e100-scripts\") pod \"keystone-bootstrap-phkvk\" (UID: \"4537c8c1-f5dc-4974-9885-527d6c85e100\") " pod="openstack/keystone-bootstrap-phkvk" Feb 02 09:25:12 crc kubenswrapper[4764]: I0202 09:25:12.511664 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4537c8c1-f5dc-4974-9885-527d6c85e100-config-data\") pod \"keystone-bootstrap-phkvk\" (UID: \"4537c8c1-f5dc-4974-9885-527d6c85e100\") " pod="openstack/keystone-bootstrap-phkvk" Feb 02 09:25:12 crc kubenswrapper[4764]: I0202 09:25:12.518639 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k2g4l\" (UniqueName: \"kubernetes.io/projected/4537c8c1-f5dc-4974-9885-527d6c85e100-kube-api-access-k2g4l\") pod \"keystone-bootstrap-phkvk\" (UID: \"4537c8c1-f5dc-4974-9885-527d6c85e100\") " pod="openstack/keystone-bootstrap-phkvk" Feb 02 09:25:12 crc kubenswrapper[4764]: I0202 09:25:12.571102 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-phkvk" Feb 02 09:25:13 crc kubenswrapper[4764]: I0202 09:25:13.843775 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9eb8eedc-a81b-4988-acc5-7a4ef9744aa2" path="/var/lib/kubelet/pods/9eb8eedc-a81b-4988-acc5-7a4ef9744aa2/volumes" Feb 02 09:25:16 crc kubenswrapper[4764]: I0202 09:25:16.889271 4764 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-58976fdf49-bb6rk" podUID="30d22725-a706-485b-8cc2-964b86137ac1" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.121:5353: connect: connection refused" Feb 02 09:25:20 crc kubenswrapper[4764]: E0202 09:25:20.624327 4764 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ceilometer-central@sha256:5ac8ede62671a3b3695cf29bd3a6f124f27c93d1730f9030cc3daa05034d4af4" Feb 02 09:25:20 crc kubenswrapper[4764]: E0202 09:25:20.624905 4764 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:quay.io/podified-antelope-centos9/openstack-ceilometer-central@sha256:5ac8ede62671a3b3695cf29bd3a6f124f27c93d1730f9030cc3daa05034d4af4,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n4h68bhd7h67fh5d5h5dh97h5bh689h56ch579hd4hb7h9bh67bh549h599h5dbh5bfhc6h5f7h7ch5c5hc8h77hbch56dhd4h4h55fh5c9h69q,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-gnldx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(a518caeb-6b91-4b71-a47f-0d71d965af2e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 02 09:25:21 crc kubenswrapper[4764]: I0202 09:25:21.893224 4764 generic.go:334] "Generic (PLEG): container finished" podID="fd30e6c5-1c0f-4ee4-9df1-c23b745ec136" containerID="adb69964a6b9f244142facf775fd53dfc1a686481ec2d2b42286a7edf9112752" exitCode=0 Feb 02 09:25:21 crc kubenswrapper[4764]: I0202 09:25:21.893352 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-p7g2s" event={"ID":"fd30e6c5-1c0f-4ee4-9df1-c23b745ec136","Type":"ContainerDied","Data":"adb69964a6b9f244142facf775fd53dfc1a686481ec2d2b42286a7edf9112752"} Feb 02 09:25:21 crc kubenswrapper[4764]: E0202 09:25:21.955253 4764 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:3fa6e687aa002b92fedbfe2c1ccaa2906b399c58d17bf9ecece2c4cd69a0210b" Feb 02 09:25:21 crc kubenswrapper[4764]: E0202 09:25:21.955611 4764 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:3fa6e687aa002b92fedbfe2c1ccaa2906b399c58d17bf9ecece2c4cd69a0210b,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-hgkdq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-dkrnx_openstack(6b3d3314-2f16-4dc3-aa52-5f7f8576813a): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 02 09:25:21 crc kubenswrapper[4764]: E0202 09:25:21.957792 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-dkrnx" podUID="6b3d3314-2f16-4dc3-aa52-5f7f8576813a" Feb 02 09:25:22 crc kubenswrapper[4764]: I0202 09:25:22.154522 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58976fdf49-bb6rk" Feb 02 09:25:22 crc kubenswrapper[4764]: I0202 09:25:22.332091 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/30d22725-a706-485b-8cc2-964b86137ac1-dns-svc\") pod \"30d22725-a706-485b-8cc2-964b86137ac1\" (UID: \"30d22725-a706-485b-8cc2-964b86137ac1\") " Feb 02 09:25:22 crc kubenswrapper[4764]: I0202 09:25:22.332162 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/30d22725-a706-485b-8cc2-964b86137ac1-ovsdbserver-nb\") pod \"30d22725-a706-485b-8cc2-964b86137ac1\" (UID: \"30d22725-a706-485b-8cc2-964b86137ac1\") " Feb 02 09:25:22 crc kubenswrapper[4764]: I0202 09:25:22.332286 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/30d22725-a706-485b-8cc2-964b86137ac1-ovsdbserver-sb\") pod \"30d22725-a706-485b-8cc2-964b86137ac1\" (UID: \"30d22725-a706-485b-8cc2-964b86137ac1\") " Feb 02 09:25:22 crc kubenswrapper[4764]: I0202 09:25:22.332323 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cbcg4\" (UniqueName: \"kubernetes.io/projected/30d22725-a706-485b-8cc2-964b86137ac1-kube-api-access-cbcg4\") pod \"30d22725-a706-485b-8cc2-964b86137ac1\" (UID: \"30d22725-a706-485b-8cc2-964b86137ac1\") " Feb 02 09:25:22 crc kubenswrapper[4764]: I0202 09:25:22.332398 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/30d22725-a706-485b-8cc2-964b86137ac1-config\") pod \"30d22725-a706-485b-8cc2-964b86137ac1\" (UID: \"30d22725-a706-485b-8cc2-964b86137ac1\") " Feb 02 09:25:22 crc kubenswrapper[4764]: I0202 09:25:22.376665 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/30d22725-a706-485b-8cc2-964b86137ac1-kube-api-access-cbcg4" (OuterVolumeSpecName: "kube-api-access-cbcg4") pod "30d22725-a706-485b-8cc2-964b86137ac1" (UID: "30d22725-a706-485b-8cc2-964b86137ac1"). InnerVolumeSpecName "kube-api-access-cbcg4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:25:22 crc kubenswrapper[4764]: I0202 09:25:22.408218 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/30d22725-a706-485b-8cc2-964b86137ac1-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "30d22725-a706-485b-8cc2-964b86137ac1" (UID: "30d22725-a706-485b-8cc2-964b86137ac1"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:25:22 crc kubenswrapper[4764]: I0202 09:25:22.410530 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-phkvk"] Feb 02 09:25:22 crc kubenswrapper[4764]: I0202 09:25:22.411534 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/30d22725-a706-485b-8cc2-964b86137ac1-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "30d22725-a706-485b-8cc2-964b86137ac1" (UID: "30d22725-a706-485b-8cc2-964b86137ac1"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:25:22 crc kubenswrapper[4764]: I0202 09:25:22.412912 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/30d22725-a706-485b-8cc2-964b86137ac1-config" (OuterVolumeSpecName: "config") pod "30d22725-a706-485b-8cc2-964b86137ac1" (UID: "30d22725-a706-485b-8cc2-964b86137ac1"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:25:22 crc kubenswrapper[4764]: I0202 09:25:22.423695 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/30d22725-a706-485b-8cc2-964b86137ac1-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "30d22725-a706-485b-8cc2-964b86137ac1" (UID: "30d22725-a706-485b-8cc2-964b86137ac1"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:25:22 crc kubenswrapper[4764]: I0202 09:25:22.434686 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cbcg4\" (UniqueName: \"kubernetes.io/projected/30d22725-a706-485b-8cc2-964b86137ac1-kube-api-access-cbcg4\") on node \"crc\" DevicePath \"\"" Feb 02 09:25:22 crc kubenswrapper[4764]: I0202 09:25:22.434715 4764 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/30d22725-a706-485b-8cc2-964b86137ac1-config\") on node \"crc\" DevicePath \"\"" Feb 02 09:25:22 crc kubenswrapper[4764]: I0202 09:25:22.434724 4764 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/30d22725-a706-485b-8cc2-964b86137ac1-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 02 09:25:22 crc kubenswrapper[4764]: I0202 09:25:22.434734 4764 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/30d22725-a706-485b-8cc2-964b86137ac1-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 02 09:25:22 crc kubenswrapper[4764]: I0202 09:25:22.434743 4764 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/30d22725-a706-485b-8cc2-964b86137ac1-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 02 09:25:22 crc kubenswrapper[4764]: W0202 09:25:22.668211 4764 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4537c8c1_f5dc_4974_9885_527d6c85e100.slice/crio-174e9cfc20869b1928c9becfde27c30d04dfa06aea1b1367b4cc81493ad1be72 WatchSource:0}: Error finding container 174e9cfc20869b1928c9becfde27c30d04dfa06aea1b1367b4cc81493ad1be72: Status 404 returned error can't find the container with id 174e9cfc20869b1928c9becfde27c30d04dfa06aea1b1367b4cc81493ad1be72 Feb 02 09:25:22 crc kubenswrapper[4764]: I0202 09:25:22.674707 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Feb 02 09:25:22 crc kubenswrapper[4764]: I0202 09:25:22.902031 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-phkvk" event={"ID":"4537c8c1-f5dc-4974-9885-527d6c85e100","Type":"ContainerStarted","Data":"7c9ba2df7010e52fb30b7d6b65df5f1f4f2d3f54edd5411f7806169a53553d20"} Feb 02 09:25:22 crc kubenswrapper[4764]: I0202 09:25:22.902097 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-phkvk" event={"ID":"4537c8c1-f5dc-4974-9885-527d6c85e100","Type":"ContainerStarted","Data":"174e9cfc20869b1928c9becfde27c30d04dfa06aea1b1367b4cc81493ad1be72"} Feb 02 09:25:22 crc kubenswrapper[4764]: I0202 09:25:22.903833 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-sm8gf" event={"ID":"e61fa54a-f23a-409c-923f-522e04a8099f","Type":"ContainerStarted","Data":"d9716bc9e16c96c00ea61d05b88a5f3eb7f109dd1e6457854009a15ff57725d1"} Feb 02 09:25:22 crc kubenswrapper[4764]: I0202 09:25:22.906838 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-229dq" event={"ID":"2f7f0eb3-55c5-4dbc-8636-eba9a31b0a27","Type":"ContainerStarted","Data":"380944daad0b77acae8f51a0b218175b09a49e178b6975b54acdf9ed0e99a75b"} Feb 02 09:25:22 crc kubenswrapper[4764]: I0202 09:25:22.909304 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58976fdf49-bb6rk" event={"ID":"30d22725-a706-485b-8cc2-964b86137ac1","Type":"ContainerDied","Data":"7a566970df23dee365ff4021a5e463f656969b25cc50ccec9520791c93fc4ba6"} Feb 02 09:25:22 crc kubenswrapper[4764]: I0202 09:25:22.909359 4764 scope.go:117] "RemoveContainer" containerID="3160f95e976a6f6e0632f2d7adc40045a33aab5f58543831305f2e7c08360f26" Feb 02 09:25:22 crc kubenswrapper[4764]: I0202 09:25:22.909367 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58976fdf49-bb6rk" Feb 02 09:25:22 crc kubenswrapper[4764]: I0202 09:25:22.914283 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a518caeb-6b91-4b71-a47f-0d71d965af2e","Type":"ContainerStarted","Data":"2eb65d6eba11ecb361786b367a01950ec0b46c28f3ac4e5936e72f214d4f9acd"} Feb 02 09:25:22 crc kubenswrapper[4764]: E0202 09:25:22.917860 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:3fa6e687aa002b92fedbfe2c1ccaa2906b399c58d17bf9ecece2c4cd69a0210b\\\"\"" pod="openstack/cinder-db-sync-dkrnx" podUID="6b3d3314-2f16-4dc3-aa52-5f7f8576813a" Feb 02 09:25:22 crc kubenswrapper[4764]: I0202 09:25:22.932688 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-phkvk" podStartSLOduration=10.932663022 podStartE2EDuration="10.932663022s" podCreationTimestamp="2026-02-02 09:25:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:25:22.929852208 +0000 UTC m=+1085.863576296" watchObservedRunningTime="2026-02-02 09:25:22.932663022 +0000 UTC m=+1085.866387110" Feb 02 09:25:22 crc kubenswrapper[4764]: I0202 09:25:22.969619 4764 scope.go:117] "RemoveContainer" containerID="12bd136d4e0414940957c9b807bf9d845d345d00ab87fc4e0292591dce89decb" Feb 02 09:25:22 crc kubenswrapper[4764]: I0202 09:25:22.978841 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-sm8gf" podStartSLOduration=2.258225117 podStartE2EDuration="24.978820621s" podCreationTimestamp="2026-02-02 09:24:58 +0000 UTC" firstStartedPulling="2026-02-02 09:24:59.216872153 +0000 UTC m=+1062.150596241" lastFinishedPulling="2026-02-02 09:25:21.937467647 +0000 UTC m=+1084.871191745" observedRunningTime="2026-02-02 09:25:22.975327629 +0000 UTC m=+1085.909051717" watchObservedRunningTime="2026-02-02 09:25:22.978820621 +0000 UTC m=+1085.912544709" Feb 02 09:25:22 crc kubenswrapper[4764]: I0202 09:25:22.998834 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-229dq" podStartSLOduration=2.527666087 podStartE2EDuration="24.99881616s" podCreationTimestamp="2026-02-02 09:24:58 +0000 UTC" firstStartedPulling="2026-02-02 09:24:59.432982153 +0000 UTC m=+1062.366706241" lastFinishedPulling="2026-02-02 09:25:21.904132226 +0000 UTC m=+1084.837856314" observedRunningTime="2026-02-02 09:25:22.992798211 +0000 UTC m=+1085.926522299" watchObservedRunningTime="2026-02-02 09:25:22.99881616 +0000 UTC m=+1085.932540248" Feb 02 09:25:23 crc kubenswrapper[4764]: I0202 09:25:23.019012 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-58976fdf49-bb6rk"] Feb 02 09:25:23 crc kubenswrapper[4764]: I0202 09:25:23.031540 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-58976fdf49-bb6rk"] Feb 02 09:25:23 crc kubenswrapper[4764]: I0202 09:25:23.404071 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-p7g2s" Feb 02 09:25:23 crc kubenswrapper[4764]: I0202 09:25:23.566734 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/fd30e6c5-1c0f-4ee4-9df1-c23b745ec136-config\") pod \"fd30e6c5-1c0f-4ee4-9df1-c23b745ec136\" (UID: \"fd30e6c5-1c0f-4ee4-9df1-c23b745ec136\") " Feb 02 09:25:23 crc kubenswrapper[4764]: I0202 09:25:23.567493 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd30e6c5-1c0f-4ee4-9df1-c23b745ec136-combined-ca-bundle\") pod \"fd30e6c5-1c0f-4ee4-9df1-c23b745ec136\" (UID: \"fd30e6c5-1c0f-4ee4-9df1-c23b745ec136\") " Feb 02 09:25:23 crc kubenswrapper[4764]: I0202 09:25:23.567650 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xr2tx\" (UniqueName: \"kubernetes.io/projected/fd30e6c5-1c0f-4ee4-9df1-c23b745ec136-kube-api-access-xr2tx\") pod \"fd30e6c5-1c0f-4ee4-9df1-c23b745ec136\" (UID: \"fd30e6c5-1c0f-4ee4-9df1-c23b745ec136\") " Feb 02 09:25:23 crc kubenswrapper[4764]: I0202 09:25:23.574140 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fd30e6c5-1c0f-4ee4-9df1-c23b745ec136-kube-api-access-xr2tx" (OuterVolumeSpecName: "kube-api-access-xr2tx") pod "fd30e6c5-1c0f-4ee4-9df1-c23b745ec136" (UID: "fd30e6c5-1c0f-4ee4-9df1-c23b745ec136"). InnerVolumeSpecName "kube-api-access-xr2tx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:25:23 crc kubenswrapper[4764]: I0202 09:25:23.589341 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd30e6c5-1c0f-4ee4-9df1-c23b745ec136-config" (OuterVolumeSpecName: "config") pod "fd30e6c5-1c0f-4ee4-9df1-c23b745ec136" (UID: "fd30e6c5-1c0f-4ee4-9df1-c23b745ec136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:25:23 crc kubenswrapper[4764]: I0202 09:25:23.631125 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd30e6c5-1c0f-4ee4-9df1-c23b745ec136-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fd30e6c5-1c0f-4ee4-9df1-c23b745ec136" (UID: "fd30e6c5-1c0f-4ee4-9df1-c23b745ec136"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:25:23 crc kubenswrapper[4764]: I0202 09:25:23.670144 4764 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd30e6c5-1c0f-4ee4-9df1-c23b745ec136-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 09:25:23 crc kubenswrapper[4764]: I0202 09:25:23.670189 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xr2tx\" (UniqueName: \"kubernetes.io/projected/fd30e6c5-1c0f-4ee4-9df1-c23b745ec136-kube-api-access-xr2tx\") on node \"crc\" DevicePath \"\"" Feb 02 09:25:23 crc kubenswrapper[4764]: I0202 09:25:23.670221 4764 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/fd30e6c5-1c0f-4ee4-9df1-c23b745ec136-config\") on node \"crc\" DevicePath \"\"" Feb 02 09:25:23 crc kubenswrapper[4764]: I0202 09:25:23.844562 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="30d22725-a706-485b-8cc2-964b86137ac1" path="/var/lib/kubelet/pods/30d22725-a706-485b-8cc2-964b86137ac1/volumes" Feb 02 09:25:23 crc kubenswrapper[4764]: I0202 09:25:23.949520 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-p7g2s" Feb 02 09:25:23 crc kubenswrapper[4764]: I0202 09:25:23.951062 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-p7g2s" event={"ID":"fd30e6c5-1c0f-4ee4-9df1-c23b745ec136","Type":"ContainerDied","Data":"46d600cc6b65e2ac379d155f0bb36dc26c310a8d46febaf499d20f0c05426f49"} Feb 02 09:25:23 crc kubenswrapper[4764]: I0202 09:25:23.951135 4764 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="46d600cc6b65e2ac379d155f0bb36dc26c310a8d46febaf499d20f0c05426f49" Feb 02 09:25:24 crc kubenswrapper[4764]: I0202 09:25:24.100846 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-64d8599669-wbmlc"] Feb 02 09:25:24 crc kubenswrapper[4764]: E0202 09:25:24.125989 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30d22725-a706-485b-8cc2-964b86137ac1" containerName="init" Feb 02 09:25:24 crc kubenswrapper[4764]: I0202 09:25:24.126521 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="30d22725-a706-485b-8cc2-964b86137ac1" containerName="init" Feb 02 09:25:24 crc kubenswrapper[4764]: E0202 09:25:24.126603 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd30e6c5-1c0f-4ee4-9df1-c23b745ec136" containerName="neutron-db-sync" Feb 02 09:25:24 crc kubenswrapper[4764]: I0202 09:25:24.126657 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd30e6c5-1c0f-4ee4-9df1-c23b745ec136" containerName="neutron-db-sync" Feb 02 09:25:24 crc kubenswrapper[4764]: E0202 09:25:24.126721 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30d22725-a706-485b-8cc2-964b86137ac1" containerName="dnsmasq-dns" Feb 02 09:25:24 crc kubenswrapper[4764]: I0202 09:25:24.126785 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="30d22725-a706-485b-8cc2-964b86137ac1" containerName="dnsmasq-dns" Feb 02 09:25:24 crc kubenswrapper[4764]: I0202 09:25:24.127124 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="30d22725-a706-485b-8cc2-964b86137ac1" containerName="dnsmasq-dns" Feb 02 09:25:24 crc kubenswrapper[4764]: I0202 09:25:24.127208 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd30e6c5-1c0f-4ee4-9df1-c23b745ec136" containerName="neutron-db-sync" Feb 02 09:25:24 crc kubenswrapper[4764]: I0202 09:25:24.128081 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64d8599669-wbmlc" Feb 02 09:25:24 crc kubenswrapper[4764]: I0202 09:25:24.130894 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-64d8599669-wbmlc"] Feb 02 09:25:24 crc kubenswrapper[4764]: I0202 09:25:24.259855 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-7bd869cccb-wnw6g"] Feb 02 09:25:24 crc kubenswrapper[4764]: I0202 09:25:24.269880 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7bd869cccb-wnw6g" Feb 02 09:25:24 crc kubenswrapper[4764]: I0202 09:25:24.276583 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Feb 02 09:25:24 crc kubenswrapper[4764]: I0202 09:25:24.277074 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Feb 02 09:25:24 crc kubenswrapper[4764]: I0202 09:25:24.277230 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Feb 02 09:25:24 crc kubenswrapper[4764]: I0202 09:25:24.277518 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-hpwwg" Feb 02 09:25:24 crc kubenswrapper[4764]: I0202 09:25:24.285238 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ldw77\" (UniqueName: \"kubernetes.io/projected/0482366b-96dd-493a-98fb-c9cc90254d1e-kube-api-access-ldw77\") pod \"dnsmasq-dns-64d8599669-wbmlc\" (UID: \"0482366b-96dd-493a-98fb-c9cc90254d1e\") " pod="openstack/dnsmasq-dns-64d8599669-wbmlc" Feb 02 09:25:24 crc kubenswrapper[4764]: I0202 09:25:24.285281 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0482366b-96dd-493a-98fb-c9cc90254d1e-ovsdbserver-sb\") pod \"dnsmasq-dns-64d8599669-wbmlc\" (UID: \"0482366b-96dd-493a-98fb-c9cc90254d1e\") " pod="openstack/dnsmasq-dns-64d8599669-wbmlc" Feb 02 09:25:24 crc kubenswrapper[4764]: I0202 09:25:24.285322 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0482366b-96dd-493a-98fb-c9cc90254d1e-config\") pod \"dnsmasq-dns-64d8599669-wbmlc\" (UID: \"0482366b-96dd-493a-98fb-c9cc90254d1e\") " pod="openstack/dnsmasq-dns-64d8599669-wbmlc" Feb 02 09:25:24 crc kubenswrapper[4764]: I0202 09:25:24.285343 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0482366b-96dd-493a-98fb-c9cc90254d1e-dns-svc\") pod \"dnsmasq-dns-64d8599669-wbmlc\" (UID: \"0482366b-96dd-493a-98fb-c9cc90254d1e\") " pod="openstack/dnsmasq-dns-64d8599669-wbmlc" Feb 02 09:25:24 crc kubenswrapper[4764]: I0202 09:25:24.285393 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0482366b-96dd-493a-98fb-c9cc90254d1e-ovsdbserver-nb\") pod \"dnsmasq-dns-64d8599669-wbmlc\" (UID: \"0482366b-96dd-493a-98fb-c9cc90254d1e\") " pod="openstack/dnsmasq-dns-64d8599669-wbmlc" Feb 02 09:25:24 crc kubenswrapper[4764]: I0202 09:25:24.293171 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7bd869cccb-wnw6g"] Feb 02 09:25:24 crc kubenswrapper[4764]: I0202 09:25:24.386684 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/d63e0dd6-3a9b-4ac4-8a81-a0a50fc4c2bb-httpd-config\") pod \"neutron-7bd869cccb-wnw6g\" (UID: \"d63e0dd6-3a9b-4ac4-8a81-a0a50fc4c2bb\") " pod="openstack/neutron-7bd869cccb-wnw6g" Feb 02 09:25:24 crc kubenswrapper[4764]: I0202 09:25:24.386972 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ldw77\" (UniqueName: \"kubernetes.io/projected/0482366b-96dd-493a-98fb-c9cc90254d1e-kube-api-access-ldw77\") pod \"dnsmasq-dns-64d8599669-wbmlc\" (UID: \"0482366b-96dd-493a-98fb-c9cc90254d1e\") " pod="openstack/dnsmasq-dns-64d8599669-wbmlc" Feb 02 09:25:24 crc kubenswrapper[4764]: I0202 09:25:24.387066 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0482366b-96dd-493a-98fb-c9cc90254d1e-ovsdbserver-sb\") pod \"dnsmasq-dns-64d8599669-wbmlc\" (UID: \"0482366b-96dd-493a-98fb-c9cc90254d1e\") " pod="openstack/dnsmasq-dns-64d8599669-wbmlc" Feb 02 09:25:24 crc kubenswrapper[4764]: I0202 09:25:24.387137 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/d63e0dd6-3a9b-4ac4-8a81-a0a50fc4c2bb-config\") pod \"neutron-7bd869cccb-wnw6g\" (UID: \"d63e0dd6-3a9b-4ac4-8a81-a0a50fc4c2bb\") " pod="openstack/neutron-7bd869cccb-wnw6g" Feb 02 09:25:24 crc kubenswrapper[4764]: I0202 09:25:24.387212 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kk9dj\" (UniqueName: \"kubernetes.io/projected/d63e0dd6-3a9b-4ac4-8a81-a0a50fc4c2bb-kube-api-access-kk9dj\") pod \"neutron-7bd869cccb-wnw6g\" (UID: \"d63e0dd6-3a9b-4ac4-8a81-a0a50fc4c2bb\") " pod="openstack/neutron-7bd869cccb-wnw6g" Feb 02 09:25:24 crc kubenswrapper[4764]: I0202 09:25:24.387327 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0482366b-96dd-493a-98fb-c9cc90254d1e-config\") pod \"dnsmasq-dns-64d8599669-wbmlc\" (UID: \"0482366b-96dd-493a-98fb-c9cc90254d1e\") " pod="openstack/dnsmasq-dns-64d8599669-wbmlc" Feb 02 09:25:24 crc kubenswrapper[4764]: I0202 09:25:24.387420 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0482366b-96dd-493a-98fb-c9cc90254d1e-dns-svc\") pod \"dnsmasq-dns-64d8599669-wbmlc\" (UID: \"0482366b-96dd-493a-98fb-c9cc90254d1e\") " pod="openstack/dnsmasq-dns-64d8599669-wbmlc" Feb 02 09:25:24 crc kubenswrapper[4764]: I0202 09:25:24.387526 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0482366b-96dd-493a-98fb-c9cc90254d1e-ovsdbserver-nb\") pod \"dnsmasq-dns-64d8599669-wbmlc\" (UID: \"0482366b-96dd-493a-98fb-c9cc90254d1e\") " pod="openstack/dnsmasq-dns-64d8599669-wbmlc" Feb 02 09:25:24 crc kubenswrapper[4764]: I0202 09:25:24.387614 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d63e0dd6-3a9b-4ac4-8a81-a0a50fc4c2bb-ovndb-tls-certs\") pod \"neutron-7bd869cccb-wnw6g\" (UID: \"d63e0dd6-3a9b-4ac4-8a81-a0a50fc4c2bb\") " pod="openstack/neutron-7bd869cccb-wnw6g" Feb 02 09:25:24 crc kubenswrapper[4764]: I0202 09:25:24.387729 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d63e0dd6-3a9b-4ac4-8a81-a0a50fc4c2bb-combined-ca-bundle\") pod \"neutron-7bd869cccb-wnw6g\" (UID: \"d63e0dd6-3a9b-4ac4-8a81-a0a50fc4c2bb\") " pod="openstack/neutron-7bd869cccb-wnw6g" Feb 02 09:25:24 crc kubenswrapper[4764]: I0202 09:25:24.388956 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0482366b-96dd-493a-98fb-c9cc90254d1e-ovsdbserver-sb\") pod \"dnsmasq-dns-64d8599669-wbmlc\" (UID: \"0482366b-96dd-493a-98fb-c9cc90254d1e\") " pod="openstack/dnsmasq-dns-64d8599669-wbmlc" Feb 02 09:25:24 crc kubenswrapper[4764]: I0202 09:25:24.389755 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0482366b-96dd-493a-98fb-c9cc90254d1e-config\") pod \"dnsmasq-dns-64d8599669-wbmlc\" (UID: \"0482366b-96dd-493a-98fb-c9cc90254d1e\") " pod="openstack/dnsmasq-dns-64d8599669-wbmlc" Feb 02 09:25:24 crc kubenswrapper[4764]: I0202 09:25:24.390508 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0482366b-96dd-493a-98fb-c9cc90254d1e-dns-svc\") pod \"dnsmasq-dns-64d8599669-wbmlc\" (UID: \"0482366b-96dd-493a-98fb-c9cc90254d1e\") " pod="openstack/dnsmasq-dns-64d8599669-wbmlc" Feb 02 09:25:24 crc kubenswrapper[4764]: I0202 09:25:24.390514 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0482366b-96dd-493a-98fb-c9cc90254d1e-ovsdbserver-nb\") pod \"dnsmasq-dns-64d8599669-wbmlc\" (UID: \"0482366b-96dd-493a-98fb-c9cc90254d1e\") " pod="openstack/dnsmasq-dns-64d8599669-wbmlc" Feb 02 09:25:24 crc kubenswrapper[4764]: I0202 09:25:24.416708 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ldw77\" (UniqueName: \"kubernetes.io/projected/0482366b-96dd-493a-98fb-c9cc90254d1e-kube-api-access-ldw77\") pod \"dnsmasq-dns-64d8599669-wbmlc\" (UID: \"0482366b-96dd-493a-98fb-c9cc90254d1e\") " pod="openstack/dnsmasq-dns-64d8599669-wbmlc" Feb 02 09:25:24 crc kubenswrapper[4764]: I0202 09:25:24.479544 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64d8599669-wbmlc" Feb 02 09:25:24 crc kubenswrapper[4764]: I0202 09:25:24.489212 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/d63e0dd6-3a9b-4ac4-8a81-a0a50fc4c2bb-config\") pod \"neutron-7bd869cccb-wnw6g\" (UID: \"d63e0dd6-3a9b-4ac4-8a81-a0a50fc4c2bb\") " pod="openstack/neutron-7bd869cccb-wnw6g" Feb 02 09:25:24 crc kubenswrapper[4764]: I0202 09:25:24.489252 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kk9dj\" (UniqueName: \"kubernetes.io/projected/d63e0dd6-3a9b-4ac4-8a81-a0a50fc4c2bb-kube-api-access-kk9dj\") pod \"neutron-7bd869cccb-wnw6g\" (UID: \"d63e0dd6-3a9b-4ac4-8a81-a0a50fc4c2bb\") " pod="openstack/neutron-7bd869cccb-wnw6g" Feb 02 09:25:24 crc kubenswrapper[4764]: I0202 09:25:24.489312 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d63e0dd6-3a9b-4ac4-8a81-a0a50fc4c2bb-ovndb-tls-certs\") pod \"neutron-7bd869cccb-wnw6g\" (UID: \"d63e0dd6-3a9b-4ac4-8a81-a0a50fc4c2bb\") " pod="openstack/neutron-7bd869cccb-wnw6g" Feb 02 09:25:24 crc kubenswrapper[4764]: I0202 09:25:24.489334 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d63e0dd6-3a9b-4ac4-8a81-a0a50fc4c2bb-combined-ca-bundle\") pod \"neutron-7bd869cccb-wnw6g\" (UID: \"d63e0dd6-3a9b-4ac4-8a81-a0a50fc4c2bb\") " pod="openstack/neutron-7bd869cccb-wnw6g" Feb 02 09:25:24 crc kubenswrapper[4764]: I0202 09:25:24.489409 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/d63e0dd6-3a9b-4ac4-8a81-a0a50fc4c2bb-httpd-config\") pod \"neutron-7bd869cccb-wnw6g\" (UID: \"d63e0dd6-3a9b-4ac4-8a81-a0a50fc4c2bb\") " pod="openstack/neutron-7bd869cccb-wnw6g" Feb 02 09:25:24 crc kubenswrapper[4764]: I0202 09:25:24.494855 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d63e0dd6-3a9b-4ac4-8a81-a0a50fc4c2bb-ovndb-tls-certs\") pod \"neutron-7bd869cccb-wnw6g\" (UID: \"d63e0dd6-3a9b-4ac4-8a81-a0a50fc4c2bb\") " pod="openstack/neutron-7bd869cccb-wnw6g" Feb 02 09:25:24 crc kubenswrapper[4764]: I0202 09:25:24.495788 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/d63e0dd6-3a9b-4ac4-8a81-a0a50fc4c2bb-httpd-config\") pod \"neutron-7bd869cccb-wnw6g\" (UID: \"d63e0dd6-3a9b-4ac4-8a81-a0a50fc4c2bb\") " pod="openstack/neutron-7bd869cccb-wnw6g" Feb 02 09:25:24 crc kubenswrapper[4764]: I0202 09:25:24.496193 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d63e0dd6-3a9b-4ac4-8a81-a0a50fc4c2bb-combined-ca-bundle\") pod \"neutron-7bd869cccb-wnw6g\" (UID: \"d63e0dd6-3a9b-4ac4-8a81-a0a50fc4c2bb\") " pod="openstack/neutron-7bd869cccb-wnw6g" Feb 02 09:25:24 crc kubenswrapper[4764]: I0202 09:25:24.512706 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/d63e0dd6-3a9b-4ac4-8a81-a0a50fc4c2bb-config\") pod \"neutron-7bd869cccb-wnw6g\" (UID: \"d63e0dd6-3a9b-4ac4-8a81-a0a50fc4c2bb\") " pod="openstack/neutron-7bd869cccb-wnw6g" Feb 02 09:25:24 crc kubenswrapper[4764]: I0202 09:25:24.525744 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kk9dj\" (UniqueName: \"kubernetes.io/projected/d63e0dd6-3a9b-4ac4-8a81-a0a50fc4c2bb-kube-api-access-kk9dj\") pod \"neutron-7bd869cccb-wnw6g\" (UID: \"d63e0dd6-3a9b-4ac4-8a81-a0a50fc4c2bb\") " pod="openstack/neutron-7bd869cccb-wnw6g" Feb 02 09:25:24 crc kubenswrapper[4764]: I0202 09:25:24.592368 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7bd869cccb-wnw6g" Feb 02 09:25:25 crc kubenswrapper[4764]: I0202 09:25:25.020308 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-64d8599669-wbmlc"] Feb 02 09:25:25 crc kubenswrapper[4764]: I0202 09:25:25.326053 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7bd869cccb-wnw6g"] Feb 02 09:25:25 crc kubenswrapper[4764]: I0202 09:25:25.990708 4764 generic.go:334] "Generic (PLEG): container finished" podID="0482366b-96dd-493a-98fb-c9cc90254d1e" containerID="93ee852f1e7b5f4f874e25c5e628d6544c821ed4fdbbbee14374c3937bdd8907" exitCode=0 Feb 02 09:25:25 crc kubenswrapper[4764]: I0202 09:25:25.991008 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64d8599669-wbmlc" event={"ID":"0482366b-96dd-493a-98fb-c9cc90254d1e","Type":"ContainerDied","Data":"93ee852f1e7b5f4f874e25c5e628d6544c821ed4fdbbbee14374c3937bdd8907"} Feb 02 09:25:25 crc kubenswrapper[4764]: I0202 09:25:25.991034 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64d8599669-wbmlc" event={"ID":"0482366b-96dd-493a-98fb-c9cc90254d1e","Type":"ContainerStarted","Data":"fe22a164d3f6ac15c7aacd92d740d287efb9340b2c933a000e014503af88ed0e"} Feb 02 09:25:26 crc kubenswrapper[4764]: I0202 09:25:26.014317 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7bd869cccb-wnw6g" event={"ID":"d63e0dd6-3a9b-4ac4-8a81-a0a50fc4c2bb","Type":"ContainerStarted","Data":"7db2f812a41d2d8d5ff08564eb909dcecd25bc8f721bb2a9e01ad5ae0a5c26a5"} Feb 02 09:25:26 crc kubenswrapper[4764]: I0202 09:25:26.014362 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7bd869cccb-wnw6g" event={"ID":"d63e0dd6-3a9b-4ac4-8a81-a0a50fc4c2bb","Type":"ContainerStarted","Data":"b85d54c7bce23ff6e47f6ec753906bcc21fe69a94f2b437e86386fa7b0a474fc"} Feb 02 09:25:26 crc kubenswrapper[4764]: I0202 09:25:26.014372 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7bd869cccb-wnw6g" event={"ID":"d63e0dd6-3a9b-4ac4-8a81-a0a50fc4c2bb","Type":"ContainerStarted","Data":"e4a0856af1974d6db7b83e8050802367153fe4f357d12e03c204b4e1cdb41a7b"} Feb 02 09:25:26 crc kubenswrapper[4764]: I0202 09:25:26.015171 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-7bd869cccb-wnw6g" Feb 02 09:25:26 crc kubenswrapper[4764]: I0202 09:25:26.079220 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-7bd869cccb-wnw6g" podStartSLOduration=2.079202289 podStartE2EDuration="2.079202289s" podCreationTimestamp="2026-02-02 09:25:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:25:26.068404163 +0000 UTC m=+1089.002128241" watchObservedRunningTime="2026-02-02 09:25:26.079202289 +0000 UTC m=+1089.012926377" Feb 02 09:25:26 crc kubenswrapper[4764]: I0202 09:25:26.565203 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-69b69cc889-m4fck"] Feb 02 09:25:26 crc kubenswrapper[4764]: I0202 09:25:26.567128 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-69b69cc889-m4fck" Feb 02 09:25:26 crc kubenswrapper[4764]: I0202 09:25:26.573423 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Feb 02 09:25:26 crc kubenswrapper[4764]: I0202 09:25:26.573623 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Feb 02 09:25:26 crc kubenswrapper[4764]: I0202 09:25:26.585059 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-69b69cc889-m4fck"] Feb 02 09:25:26 crc kubenswrapper[4764]: I0202 09:25:26.738444 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hmggt\" (UniqueName: \"kubernetes.io/projected/364d30e9-d4ef-4c84-b82f-673b41ffe9ff-kube-api-access-hmggt\") pod \"neutron-69b69cc889-m4fck\" (UID: \"364d30e9-d4ef-4c84-b82f-673b41ffe9ff\") " pod="openstack/neutron-69b69cc889-m4fck" Feb 02 09:25:26 crc kubenswrapper[4764]: I0202 09:25:26.738496 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/364d30e9-d4ef-4c84-b82f-673b41ffe9ff-ovndb-tls-certs\") pod \"neutron-69b69cc889-m4fck\" (UID: \"364d30e9-d4ef-4c84-b82f-673b41ffe9ff\") " pod="openstack/neutron-69b69cc889-m4fck" Feb 02 09:25:26 crc kubenswrapper[4764]: I0202 09:25:26.738520 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/364d30e9-d4ef-4c84-b82f-673b41ffe9ff-httpd-config\") pod \"neutron-69b69cc889-m4fck\" (UID: \"364d30e9-d4ef-4c84-b82f-673b41ffe9ff\") " pod="openstack/neutron-69b69cc889-m4fck" Feb 02 09:25:26 crc kubenswrapper[4764]: I0202 09:25:26.738535 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/364d30e9-d4ef-4c84-b82f-673b41ffe9ff-combined-ca-bundle\") pod \"neutron-69b69cc889-m4fck\" (UID: \"364d30e9-d4ef-4c84-b82f-673b41ffe9ff\") " pod="openstack/neutron-69b69cc889-m4fck" Feb 02 09:25:26 crc kubenswrapper[4764]: I0202 09:25:26.738552 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/364d30e9-d4ef-4c84-b82f-673b41ffe9ff-internal-tls-certs\") pod \"neutron-69b69cc889-m4fck\" (UID: \"364d30e9-d4ef-4c84-b82f-673b41ffe9ff\") " pod="openstack/neutron-69b69cc889-m4fck" Feb 02 09:25:26 crc kubenswrapper[4764]: I0202 09:25:26.738646 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/364d30e9-d4ef-4c84-b82f-673b41ffe9ff-config\") pod \"neutron-69b69cc889-m4fck\" (UID: \"364d30e9-d4ef-4c84-b82f-673b41ffe9ff\") " pod="openstack/neutron-69b69cc889-m4fck" Feb 02 09:25:26 crc kubenswrapper[4764]: I0202 09:25:26.739225 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/364d30e9-d4ef-4c84-b82f-673b41ffe9ff-public-tls-certs\") pod \"neutron-69b69cc889-m4fck\" (UID: \"364d30e9-d4ef-4c84-b82f-673b41ffe9ff\") " pod="openstack/neutron-69b69cc889-m4fck" Feb 02 09:25:26 crc kubenswrapper[4764]: I0202 09:25:26.843392 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/364d30e9-d4ef-4c84-b82f-673b41ffe9ff-public-tls-certs\") pod \"neutron-69b69cc889-m4fck\" (UID: \"364d30e9-d4ef-4c84-b82f-673b41ffe9ff\") " pod="openstack/neutron-69b69cc889-m4fck" Feb 02 09:25:26 crc kubenswrapper[4764]: I0202 09:25:26.843457 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hmggt\" (UniqueName: \"kubernetes.io/projected/364d30e9-d4ef-4c84-b82f-673b41ffe9ff-kube-api-access-hmggt\") pod \"neutron-69b69cc889-m4fck\" (UID: \"364d30e9-d4ef-4c84-b82f-673b41ffe9ff\") " pod="openstack/neutron-69b69cc889-m4fck" Feb 02 09:25:26 crc kubenswrapper[4764]: I0202 09:25:26.843475 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/364d30e9-d4ef-4c84-b82f-673b41ffe9ff-ovndb-tls-certs\") pod \"neutron-69b69cc889-m4fck\" (UID: \"364d30e9-d4ef-4c84-b82f-673b41ffe9ff\") " pod="openstack/neutron-69b69cc889-m4fck" Feb 02 09:25:26 crc kubenswrapper[4764]: I0202 09:25:26.843497 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/364d30e9-d4ef-4c84-b82f-673b41ffe9ff-httpd-config\") pod \"neutron-69b69cc889-m4fck\" (UID: \"364d30e9-d4ef-4c84-b82f-673b41ffe9ff\") " pod="openstack/neutron-69b69cc889-m4fck" Feb 02 09:25:26 crc kubenswrapper[4764]: I0202 09:25:26.843512 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/364d30e9-d4ef-4c84-b82f-673b41ffe9ff-combined-ca-bundle\") pod \"neutron-69b69cc889-m4fck\" (UID: \"364d30e9-d4ef-4c84-b82f-673b41ffe9ff\") " pod="openstack/neutron-69b69cc889-m4fck" Feb 02 09:25:26 crc kubenswrapper[4764]: I0202 09:25:26.843528 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/364d30e9-d4ef-4c84-b82f-673b41ffe9ff-internal-tls-certs\") pod \"neutron-69b69cc889-m4fck\" (UID: \"364d30e9-d4ef-4c84-b82f-673b41ffe9ff\") " pod="openstack/neutron-69b69cc889-m4fck" Feb 02 09:25:26 crc kubenswrapper[4764]: I0202 09:25:26.843574 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/364d30e9-d4ef-4c84-b82f-673b41ffe9ff-config\") pod \"neutron-69b69cc889-m4fck\" (UID: \"364d30e9-d4ef-4c84-b82f-673b41ffe9ff\") " pod="openstack/neutron-69b69cc889-m4fck" Feb 02 09:25:26 crc kubenswrapper[4764]: I0202 09:25:26.851125 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/364d30e9-d4ef-4c84-b82f-673b41ffe9ff-config\") pod \"neutron-69b69cc889-m4fck\" (UID: \"364d30e9-d4ef-4c84-b82f-673b41ffe9ff\") " pod="openstack/neutron-69b69cc889-m4fck" Feb 02 09:25:26 crc kubenswrapper[4764]: I0202 09:25:26.852018 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/364d30e9-d4ef-4c84-b82f-673b41ffe9ff-combined-ca-bundle\") pod \"neutron-69b69cc889-m4fck\" (UID: \"364d30e9-d4ef-4c84-b82f-673b41ffe9ff\") " pod="openstack/neutron-69b69cc889-m4fck" Feb 02 09:25:26 crc kubenswrapper[4764]: I0202 09:25:26.852645 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/364d30e9-d4ef-4c84-b82f-673b41ffe9ff-ovndb-tls-certs\") pod \"neutron-69b69cc889-m4fck\" (UID: \"364d30e9-d4ef-4c84-b82f-673b41ffe9ff\") " pod="openstack/neutron-69b69cc889-m4fck" Feb 02 09:25:26 crc kubenswrapper[4764]: I0202 09:25:26.853110 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/364d30e9-d4ef-4c84-b82f-673b41ffe9ff-internal-tls-certs\") pod \"neutron-69b69cc889-m4fck\" (UID: \"364d30e9-d4ef-4c84-b82f-673b41ffe9ff\") " pod="openstack/neutron-69b69cc889-m4fck" Feb 02 09:25:26 crc kubenswrapper[4764]: I0202 09:25:26.853563 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/364d30e9-d4ef-4c84-b82f-673b41ffe9ff-public-tls-certs\") pod \"neutron-69b69cc889-m4fck\" (UID: \"364d30e9-d4ef-4c84-b82f-673b41ffe9ff\") " pod="openstack/neutron-69b69cc889-m4fck" Feb 02 09:25:26 crc kubenswrapper[4764]: I0202 09:25:26.853961 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/364d30e9-d4ef-4c84-b82f-673b41ffe9ff-httpd-config\") pod \"neutron-69b69cc889-m4fck\" (UID: \"364d30e9-d4ef-4c84-b82f-673b41ffe9ff\") " pod="openstack/neutron-69b69cc889-m4fck" Feb 02 09:25:26 crc kubenswrapper[4764]: I0202 09:25:26.869363 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hmggt\" (UniqueName: \"kubernetes.io/projected/364d30e9-d4ef-4c84-b82f-673b41ffe9ff-kube-api-access-hmggt\") pod \"neutron-69b69cc889-m4fck\" (UID: \"364d30e9-d4ef-4c84-b82f-673b41ffe9ff\") " pod="openstack/neutron-69b69cc889-m4fck" Feb 02 09:25:26 crc kubenswrapper[4764]: I0202 09:25:26.889339 4764 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-58976fdf49-bb6rk" podUID="30d22725-a706-485b-8cc2-964b86137ac1" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.121:5353: i/o timeout" Feb 02 09:25:26 crc kubenswrapper[4764]: I0202 09:25:26.911399 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-69b69cc889-m4fck" Feb 02 09:25:27 crc kubenswrapper[4764]: I0202 09:25:27.023439 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64d8599669-wbmlc" event={"ID":"0482366b-96dd-493a-98fb-c9cc90254d1e","Type":"ContainerStarted","Data":"47268c7c0203a63ca5fa59eed2fee46a95043af36c15272dcc23c51f4a290b68"} Feb 02 09:25:27 crc kubenswrapper[4764]: I0202 09:25:27.061364 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-64d8599669-wbmlc" podStartSLOduration=3.061333578 podStartE2EDuration="3.061333578s" podCreationTimestamp="2026-02-02 09:25:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:25:27.050230825 +0000 UTC m=+1089.983954923" watchObservedRunningTime="2026-02-02 09:25:27.061333578 +0000 UTC m=+1089.995057656" Feb 02 09:25:28 crc kubenswrapper[4764]: I0202 09:25:28.034417 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-64d8599669-wbmlc" Feb 02 09:25:29 crc kubenswrapper[4764]: I0202 09:25:29.041271 4764 generic.go:334] "Generic (PLEG): container finished" podID="e61fa54a-f23a-409c-923f-522e04a8099f" containerID="d9716bc9e16c96c00ea61d05b88a5f3eb7f109dd1e6457854009a15ff57725d1" exitCode=0 Feb 02 09:25:29 crc kubenswrapper[4764]: I0202 09:25:29.041331 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-sm8gf" event={"ID":"e61fa54a-f23a-409c-923f-522e04a8099f","Type":"ContainerDied","Data":"d9716bc9e16c96c00ea61d05b88a5f3eb7f109dd1e6457854009a15ff57725d1"} Feb 02 09:25:29 crc kubenswrapper[4764]: I0202 09:25:29.044053 4764 generic.go:334] "Generic (PLEG): container finished" podID="2f7f0eb3-55c5-4dbc-8636-eba9a31b0a27" containerID="380944daad0b77acae8f51a0b218175b09a49e178b6975b54acdf9ed0e99a75b" exitCode=0 Feb 02 09:25:29 crc kubenswrapper[4764]: I0202 09:25:29.044133 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-229dq" event={"ID":"2f7f0eb3-55c5-4dbc-8636-eba9a31b0a27","Type":"ContainerDied","Data":"380944daad0b77acae8f51a0b218175b09a49e178b6975b54acdf9ed0e99a75b"} Feb 02 09:25:31 crc kubenswrapper[4764]: I0202 09:25:31.066373 4764 generic.go:334] "Generic (PLEG): container finished" podID="4537c8c1-f5dc-4974-9885-527d6c85e100" containerID="7c9ba2df7010e52fb30b7d6b65df5f1f4f2d3f54edd5411f7806169a53553d20" exitCode=0 Feb 02 09:25:31 crc kubenswrapper[4764]: I0202 09:25:31.075097 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-phkvk" event={"ID":"4537c8c1-f5dc-4974-9885-527d6c85e100","Type":"ContainerDied","Data":"7c9ba2df7010e52fb30b7d6b65df5f1f4f2d3f54edd5411f7806169a53553d20"} Feb 02 09:25:31 crc kubenswrapper[4764]: I0202 09:25:31.189670 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-229dq" Feb 02 09:25:31 crc kubenswrapper[4764]: I0202 09:25:31.193621 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-sm8gf" Feb 02 09:25:31 crc kubenswrapper[4764]: I0202 09:25:31.259036 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2f7f0eb3-55c5-4dbc-8636-eba9a31b0a27-logs\") pod \"2f7f0eb3-55c5-4dbc-8636-eba9a31b0a27\" (UID: \"2f7f0eb3-55c5-4dbc-8636-eba9a31b0a27\") " Feb 02 09:25:31 crc kubenswrapper[4764]: I0202 09:25:31.259090 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f7f0eb3-55c5-4dbc-8636-eba9a31b0a27-combined-ca-bundle\") pod \"2f7f0eb3-55c5-4dbc-8636-eba9a31b0a27\" (UID: \"2f7f0eb3-55c5-4dbc-8636-eba9a31b0a27\") " Feb 02 09:25:31 crc kubenswrapper[4764]: I0202 09:25:31.259113 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e61fa54a-f23a-409c-923f-522e04a8099f-combined-ca-bundle\") pod \"e61fa54a-f23a-409c-923f-522e04a8099f\" (UID: \"e61fa54a-f23a-409c-923f-522e04a8099f\") " Feb 02 09:25:31 crc kubenswrapper[4764]: I0202 09:25:31.261150 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2f7f0eb3-55c5-4dbc-8636-eba9a31b0a27-logs" (OuterVolumeSpecName: "logs") pod "2f7f0eb3-55c5-4dbc-8636-eba9a31b0a27" (UID: "2f7f0eb3-55c5-4dbc-8636-eba9a31b0a27"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 09:25:31 crc kubenswrapper[4764]: I0202 09:25:31.317683 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f7f0eb3-55c5-4dbc-8636-eba9a31b0a27-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2f7f0eb3-55c5-4dbc-8636-eba9a31b0a27" (UID: "2f7f0eb3-55c5-4dbc-8636-eba9a31b0a27"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:25:31 crc kubenswrapper[4764]: I0202 09:25:31.318083 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e61fa54a-f23a-409c-923f-522e04a8099f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e61fa54a-f23a-409c-923f-522e04a8099f" (UID: "e61fa54a-f23a-409c-923f-522e04a8099f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:25:31 crc kubenswrapper[4764]: I0202 09:25:31.360779 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b8pms\" (UniqueName: \"kubernetes.io/projected/e61fa54a-f23a-409c-923f-522e04a8099f-kube-api-access-b8pms\") pod \"e61fa54a-f23a-409c-923f-522e04a8099f\" (UID: \"e61fa54a-f23a-409c-923f-522e04a8099f\") " Feb 02 09:25:31 crc kubenswrapper[4764]: I0202 09:25:31.360900 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9d4jp\" (UniqueName: \"kubernetes.io/projected/2f7f0eb3-55c5-4dbc-8636-eba9a31b0a27-kube-api-access-9d4jp\") pod \"2f7f0eb3-55c5-4dbc-8636-eba9a31b0a27\" (UID: \"2f7f0eb3-55c5-4dbc-8636-eba9a31b0a27\") " Feb 02 09:25:31 crc kubenswrapper[4764]: I0202 09:25:31.361001 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/e61fa54a-f23a-409c-923f-522e04a8099f-db-sync-config-data\") pod \"e61fa54a-f23a-409c-923f-522e04a8099f\" (UID: \"e61fa54a-f23a-409c-923f-522e04a8099f\") " Feb 02 09:25:31 crc kubenswrapper[4764]: I0202 09:25:31.361064 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f7f0eb3-55c5-4dbc-8636-eba9a31b0a27-config-data\") pod \"2f7f0eb3-55c5-4dbc-8636-eba9a31b0a27\" (UID: \"2f7f0eb3-55c5-4dbc-8636-eba9a31b0a27\") " Feb 02 09:25:31 crc kubenswrapper[4764]: I0202 09:25:31.361209 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2f7f0eb3-55c5-4dbc-8636-eba9a31b0a27-scripts\") pod \"2f7f0eb3-55c5-4dbc-8636-eba9a31b0a27\" (UID: \"2f7f0eb3-55c5-4dbc-8636-eba9a31b0a27\") " Feb 02 09:25:31 crc kubenswrapper[4764]: I0202 09:25:31.361554 4764 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2f7f0eb3-55c5-4dbc-8636-eba9a31b0a27-logs\") on node \"crc\" DevicePath \"\"" Feb 02 09:25:31 crc kubenswrapper[4764]: I0202 09:25:31.361572 4764 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f7f0eb3-55c5-4dbc-8636-eba9a31b0a27-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 09:25:31 crc kubenswrapper[4764]: I0202 09:25:31.361583 4764 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e61fa54a-f23a-409c-923f-522e04a8099f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 09:25:31 crc kubenswrapper[4764]: I0202 09:25:31.364528 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e61fa54a-f23a-409c-923f-522e04a8099f-kube-api-access-b8pms" (OuterVolumeSpecName: "kube-api-access-b8pms") pod "e61fa54a-f23a-409c-923f-522e04a8099f" (UID: "e61fa54a-f23a-409c-923f-522e04a8099f"). InnerVolumeSpecName "kube-api-access-b8pms". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:25:31 crc kubenswrapper[4764]: I0202 09:25:31.365221 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e61fa54a-f23a-409c-923f-522e04a8099f-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "e61fa54a-f23a-409c-923f-522e04a8099f" (UID: "e61fa54a-f23a-409c-923f-522e04a8099f"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:25:31 crc kubenswrapper[4764]: I0202 09:25:31.365654 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f7f0eb3-55c5-4dbc-8636-eba9a31b0a27-kube-api-access-9d4jp" (OuterVolumeSpecName: "kube-api-access-9d4jp") pod "2f7f0eb3-55c5-4dbc-8636-eba9a31b0a27" (UID: "2f7f0eb3-55c5-4dbc-8636-eba9a31b0a27"). InnerVolumeSpecName "kube-api-access-9d4jp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:25:31 crc kubenswrapper[4764]: I0202 09:25:31.366821 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f7f0eb3-55c5-4dbc-8636-eba9a31b0a27-scripts" (OuterVolumeSpecName: "scripts") pod "2f7f0eb3-55c5-4dbc-8636-eba9a31b0a27" (UID: "2f7f0eb3-55c5-4dbc-8636-eba9a31b0a27"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:25:31 crc kubenswrapper[4764]: I0202 09:25:31.393900 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f7f0eb3-55c5-4dbc-8636-eba9a31b0a27-config-data" (OuterVolumeSpecName: "config-data") pod "2f7f0eb3-55c5-4dbc-8636-eba9a31b0a27" (UID: "2f7f0eb3-55c5-4dbc-8636-eba9a31b0a27"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:25:31 crc kubenswrapper[4764]: I0202 09:25:31.462341 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b8pms\" (UniqueName: \"kubernetes.io/projected/e61fa54a-f23a-409c-923f-522e04a8099f-kube-api-access-b8pms\") on node \"crc\" DevicePath \"\"" Feb 02 09:25:31 crc kubenswrapper[4764]: I0202 09:25:31.462707 4764 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/e61fa54a-f23a-409c-923f-522e04a8099f-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 09:25:31 crc kubenswrapper[4764]: I0202 09:25:31.462717 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9d4jp\" (UniqueName: \"kubernetes.io/projected/2f7f0eb3-55c5-4dbc-8636-eba9a31b0a27-kube-api-access-9d4jp\") on node \"crc\" DevicePath \"\"" Feb 02 09:25:31 crc kubenswrapper[4764]: I0202 09:25:31.462727 4764 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f7f0eb3-55c5-4dbc-8636-eba9a31b0a27-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 09:25:31 crc kubenswrapper[4764]: I0202 09:25:31.462736 4764 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2f7f0eb3-55c5-4dbc-8636-eba9a31b0a27-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 09:25:32 crc kubenswrapper[4764]: I0202 09:25:32.076721 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-229dq" event={"ID":"2f7f0eb3-55c5-4dbc-8636-eba9a31b0a27","Type":"ContainerDied","Data":"78ddb63cbddf7138fea265f3af07ddc4eab968da6cfc8c842b00d9dfd20e05c7"} Feb 02 09:25:32 crc kubenswrapper[4764]: I0202 09:25:32.076790 4764 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="78ddb63cbddf7138fea265f3af07ddc4eab968da6cfc8c842b00d9dfd20e05c7" Feb 02 09:25:32 crc kubenswrapper[4764]: I0202 09:25:32.076883 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-229dq" Feb 02 09:25:32 crc kubenswrapper[4764]: I0202 09:25:32.079105 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-sm8gf" Feb 02 09:25:32 crc kubenswrapper[4764]: I0202 09:25:32.079440 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-sm8gf" event={"ID":"e61fa54a-f23a-409c-923f-522e04a8099f","Type":"ContainerDied","Data":"f511998e84a5998777e047e56bfbef4f594c30f51da0d4596f7c24a44eacd196"} Feb 02 09:25:32 crc kubenswrapper[4764]: I0202 09:25:32.079463 4764 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f511998e84a5998777e047e56bfbef4f594c30f51da0d4596f7c24a44eacd196" Feb 02 09:25:32 crc kubenswrapper[4764]: I0202 09:25:32.329076 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-65fcd88786-dggfp"] Feb 02 09:25:32 crc kubenswrapper[4764]: E0202 09:25:32.329667 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f7f0eb3-55c5-4dbc-8636-eba9a31b0a27" containerName="placement-db-sync" Feb 02 09:25:32 crc kubenswrapper[4764]: I0202 09:25:32.329687 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f7f0eb3-55c5-4dbc-8636-eba9a31b0a27" containerName="placement-db-sync" Feb 02 09:25:32 crc kubenswrapper[4764]: E0202 09:25:32.329716 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e61fa54a-f23a-409c-923f-522e04a8099f" containerName="barbican-db-sync" Feb 02 09:25:32 crc kubenswrapper[4764]: I0202 09:25:32.329740 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="e61fa54a-f23a-409c-923f-522e04a8099f" containerName="barbican-db-sync" Feb 02 09:25:32 crc kubenswrapper[4764]: I0202 09:25:32.329983 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="e61fa54a-f23a-409c-923f-522e04a8099f" containerName="barbican-db-sync" Feb 02 09:25:32 crc kubenswrapper[4764]: I0202 09:25:32.330004 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f7f0eb3-55c5-4dbc-8636-eba9a31b0a27" containerName="placement-db-sync" Feb 02 09:25:32 crc kubenswrapper[4764]: I0202 09:25:32.331058 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-65fcd88786-dggfp" Feb 02 09:25:32 crc kubenswrapper[4764]: I0202 09:25:32.338586 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Feb 02 09:25:32 crc kubenswrapper[4764]: I0202 09:25:32.338879 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Feb 02 09:25:32 crc kubenswrapper[4764]: I0202 09:25:32.340102 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-w8vss" Feb 02 09:25:32 crc kubenswrapper[4764]: I0202 09:25:32.340142 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Feb 02 09:25:32 crc kubenswrapper[4764]: I0202 09:25:32.340546 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Feb 02 09:25:32 crc kubenswrapper[4764]: I0202 09:25:32.356997 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-65fcd88786-dggfp"] Feb 02 09:25:32 crc kubenswrapper[4764]: I0202 09:25:32.472371 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-6db5f6d5bf-pxtw5"] Feb 02 09:25:32 crc kubenswrapper[4764]: I0202 09:25:32.475109 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-6db5f6d5bf-pxtw5" Feb 02 09:25:32 crc kubenswrapper[4764]: I0202 09:25:32.478683 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dedd7b1f-7ec4-4a7a-a703-7e228f7767b1-scripts\") pod \"placement-65fcd88786-dggfp\" (UID: \"dedd7b1f-7ec4-4a7a-a703-7e228f7767b1\") " pod="openstack/placement-65fcd88786-dggfp" Feb 02 09:25:32 crc kubenswrapper[4764]: I0202 09:25:32.478748 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dedd7b1f-7ec4-4a7a-a703-7e228f7767b1-combined-ca-bundle\") pod \"placement-65fcd88786-dggfp\" (UID: \"dedd7b1f-7ec4-4a7a-a703-7e228f7767b1\") " pod="openstack/placement-65fcd88786-dggfp" Feb 02 09:25:32 crc kubenswrapper[4764]: I0202 09:25:32.478776 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jvkzr\" (UniqueName: \"kubernetes.io/projected/dedd7b1f-7ec4-4a7a-a703-7e228f7767b1-kube-api-access-jvkzr\") pod \"placement-65fcd88786-dggfp\" (UID: \"dedd7b1f-7ec4-4a7a-a703-7e228f7767b1\") " pod="openstack/placement-65fcd88786-dggfp" Feb 02 09:25:32 crc kubenswrapper[4764]: I0202 09:25:32.478793 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dedd7b1f-7ec4-4a7a-a703-7e228f7767b1-config-data\") pod \"placement-65fcd88786-dggfp\" (UID: \"dedd7b1f-7ec4-4a7a-a703-7e228f7767b1\") " pod="openstack/placement-65fcd88786-dggfp" Feb 02 09:25:32 crc kubenswrapper[4764]: I0202 09:25:32.478816 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/dedd7b1f-7ec4-4a7a-a703-7e228f7767b1-public-tls-certs\") pod \"placement-65fcd88786-dggfp\" (UID: \"dedd7b1f-7ec4-4a7a-a703-7e228f7767b1\") " pod="openstack/placement-65fcd88786-dggfp" Feb 02 09:25:32 crc kubenswrapper[4764]: I0202 09:25:32.478863 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/dedd7b1f-7ec4-4a7a-a703-7e228f7767b1-internal-tls-certs\") pod \"placement-65fcd88786-dggfp\" (UID: \"dedd7b1f-7ec4-4a7a-a703-7e228f7767b1\") " pod="openstack/placement-65fcd88786-dggfp" Feb 02 09:25:32 crc kubenswrapper[4764]: I0202 09:25:32.478898 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dedd7b1f-7ec4-4a7a-a703-7e228f7767b1-logs\") pod \"placement-65fcd88786-dggfp\" (UID: \"dedd7b1f-7ec4-4a7a-a703-7e228f7767b1\") " pod="openstack/placement-65fcd88786-dggfp" Feb 02 09:25:32 crc kubenswrapper[4764]: I0202 09:25:32.480945 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Feb 02 09:25:32 crc kubenswrapper[4764]: I0202 09:25:32.483663 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Feb 02 09:25:32 crc kubenswrapper[4764]: I0202 09:25:32.486188 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-g8sjd" Feb 02 09:25:32 crc kubenswrapper[4764]: I0202 09:25:32.535351 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-6db5f6d5bf-pxtw5"] Feb 02 09:25:32 crc kubenswrapper[4764]: I0202 09:25:32.548381 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-6d6c6fd44b-2hfzf"] Feb 02 09:25:32 crc kubenswrapper[4764]: I0202 09:25:32.549729 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-6d6c6fd44b-2hfzf" Feb 02 09:25:32 crc kubenswrapper[4764]: I0202 09:25:32.559556 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Feb 02 09:25:32 crc kubenswrapper[4764]: I0202 09:25:32.580875 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jvkzr\" (UniqueName: \"kubernetes.io/projected/dedd7b1f-7ec4-4a7a-a703-7e228f7767b1-kube-api-access-jvkzr\") pod \"placement-65fcd88786-dggfp\" (UID: \"dedd7b1f-7ec4-4a7a-a703-7e228f7767b1\") " pod="openstack/placement-65fcd88786-dggfp" Feb 02 09:25:32 crc kubenswrapper[4764]: I0202 09:25:32.581815 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dedd7b1f-7ec4-4a7a-a703-7e228f7767b1-config-data\") pod \"placement-65fcd88786-dggfp\" (UID: \"dedd7b1f-7ec4-4a7a-a703-7e228f7767b1\") " pod="openstack/placement-65fcd88786-dggfp" Feb 02 09:25:32 crc kubenswrapper[4764]: I0202 09:25:32.581845 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ebb36bb-8032-42ea-8954-7716e5a81301-config-data\") pod \"barbican-worker-6db5f6d5bf-pxtw5\" (UID: \"4ebb36bb-8032-42ea-8954-7716e5a81301\") " pod="openstack/barbican-worker-6db5f6d5bf-pxtw5" Feb 02 09:25:32 crc kubenswrapper[4764]: I0202 09:25:32.581866 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/dedd7b1f-7ec4-4a7a-a703-7e228f7767b1-public-tls-certs\") pod \"placement-65fcd88786-dggfp\" (UID: \"dedd7b1f-7ec4-4a7a-a703-7e228f7767b1\") " pod="openstack/placement-65fcd88786-dggfp" Feb 02 09:25:32 crc kubenswrapper[4764]: I0202 09:25:32.581909 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/dedd7b1f-7ec4-4a7a-a703-7e228f7767b1-internal-tls-certs\") pod \"placement-65fcd88786-dggfp\" (UID: \"dedd7b1f-7ec4-4a7a-a703-7e228f7767b1\") " pod="openstack/placement-65fcd88786-dggfp" Feb 02 09:25:32 crc kubenswrapper[4764]: I0202 09:25:32.581925 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cb7vb\" (UniqueName: \"kubernetes.io/projected/4ebb36bb-8032-42ea-8954-7716e5a81301-kube-api-access-cb7vb\") pod \"barbican-worker-6db5f6d5bf-pxtw5\" (UID: \"4ebb36bb-8032-42ea-8954-7716e5a81301\") " pod="openstack/barbican-worker-6db5f6d5bf-pxtw5" Feb 02 09:25:32 crc kubenswrapper[4764]: I0202 09:25:32.581966 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dedd7b1f-7ec4-4a7a-a703-7e228f7767b1-logs\") pod \"placement-65fcd88786-dggfp\" (UID: \"dedd7b1f-7ec4-4a7a-a703-7e228f7767b1\") " pod="openstack/placement-65fcd88786-dggfp" Feb 02 09:25:32 crc kubenswrapper[4764]: I0202 09:25:32.582005 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dedd7b1f-7ec4-4a7a-a703-7e228f7767b1-scripts\") pod \"placement-65fcd88786-dggfp\" (UID: \"dedd7b1f-7ec4-4a7a-a703-7e228f7767b1\") " pod="openstack/placement-65fcd88786-dggfp" Feb 02 09:25:32 crc kubenswrapper[4764]: I0202 09:25:32.582021 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4ebb36bb-8032-42ea-8954-7716e5a81301-config-data-custom\") pod \"barbican-worker-6db5f6d5bf-pxtw5\" (UID: \"4ebb36bb-8032-42ea-8954-7716e5a81301\") " pod="openstack/barbican-worker-6db5f6d5bf-pxtw5" Feb 02 09:25:32 crc kubenswrapper[4764]: I0202 09:25:32.582044 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ebb36bb-8032-42ea-8954-7716e5a81301-combined-ca-bundle\") pod \"barbican-worker-6db5f6d5bf-pxtw5\" (UID: \"4ebb36bb-8032-42ea-8954-7716e5a81301\") " pod="openstack/barbican-worker-6db5f6d5bf-pxtw5" Feb 02 09:25:32 crc kubenswrapper[4764]: I0202 09:25:32.582075 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4ebb36bb-8032-42ea-8954-7716e5a81301-logs\") pod \"barbican-worker-6db5f6d5bf-pxtw5\" (UID: \"4ebb36bb-8032-42ea-8954-7716e5a81301\") " pod="openstack/barbican-worker-6db5f6d5bf-pxtw5" Feb 02 09:25:32 crc kubenswrapper[4764]: I0202 09:25:32.582092 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dedd7b1f-7ec4-4a7a-a703-7e228f7767b1-combined-ca-bundle\") pod \"placement-65fcd88786-dggfp\" (UID: \"dedd7b1f-7ec4-4a7a-a703-7e228f7767b1\") " pod="openstack/placement-65fcd88786-dggfp" Feb 02 09:25:32 crc kubenswrapper[4764]: I0202 09:25:32.583075 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dedd7b1f-7ec4-4a7a-a703-7e228f7767b1-logs\") pod \"placement-65fcd88786-dggfp\" (UID: \"dedd7b1f-7ec4-4a7a-a703-7e228f7767b1\") " pod="openstack/placement-65fcd88786-dggfp" Feb 02 09:25:32 crc kubenswrapper[4764]: I0202 09:25:32.612912 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-6d6c6fd44b-2hfzf"] Feb 02 09:25:32 crc kubenswrapper[4764]: I0202 09:25:32.616057 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/dedd7b1f-7ec4-4a7a-a703-7e228f7767b1-public-tls-certs\") pod \"placement-65fcd88786-dggfp\" (UID: \"dedd7b1f-7ec4-4a7a-a703-7e228f7767b1\") " pod="openstack/placement-65fcd88786-dggfp" Feb 02 09:25:32 crc kubenswrapper[4764]: I0202 09:25:32.617058 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dedd7b1f-7ec4-4a7a-a703-7e228f7767b1-scripts\") pod \"placement-65fcd88786-dggfp\" (UID: \"dedd7b1f-7ec4-4a7a-a703-7e228f7767b1\") " pod="openstack/placement-65fcd88786-dggfp" Feb 02 09:25:32 crc kubenswrapper[4764]: I0202 09:25:32.617554 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dedd7b1f-7ec4-4a7a-a703-7e228f7767b1-combined-ca-bundle\") pod \"placement-65fcd88786-dggfp\" (UID: \"dedd7b1f-7ec4-4a7a-a703-7e228f7767b1\") " pod="openstack/placement-65fcd88786-dggfp" Feb 02 09:25:32 crc kubenswrapper[4764]: I0202 09:25:32.625487 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/dedd7b1f-7ec4-4a7a-a703-7e228f7767b1-internal-tls-certs\") pod \"placement-65fcd88786-dggfp\" (UID: \"dedd7b1f-7ec4-4a7a-a703-7e228f7767b1\") " pod="openstack/placement-65fcd88786-dggfp" Feb 02 09:25:32 crc kubenswrapper[4764]: I0202 09:25:32.647530 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jvkzr\" (UniqueName: \"kubernetes.io/projected/dedd7b1f-7ec4-4a7a-a703-7e228f7767b1-kube-api-access-jvkzr\") pod \"placement-65fcd88786-dggfp\" (UID: \"dedd7b1f-7ec4-4a7a-a703-7e228f7767b1\") " pod="openstack/placement-65fcd88786-dggfp" Feb 02 09:25:32 crc kubenswrapper[4764]: I0202 09:25:32.649923 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dedd7b1f-7ec4-4a7a-a703-7e228f7767b1-config-data\") pod \"placement-65fcd88786-dggfp\" (UID: \"dedd7b1f-7ec4-4a7a-a703-7e228f7767b1\") " pod="openstack/placement-65fcd88786-dggfp" Feb 02 09:25:32 crc kubenswrapper[4764]: I0202 09:25:32.673435 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-65fcd88786-dggfp" Feb 02 09:25:32 crc kubenswrapper[4764]: I0202 09:25:32.684664 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rsq9n\" (UniqueName: \"kubernetes.io/projected/4f12af32-00c9-437a-8e0e-805caa8df878-kube-api-access-rsq9n\") pod \"barbican-keystone-listener-6d6c6fd44b-2hfzf\" (UID: \"4f12af32-00c9-437a-8e0e-805caa8df878\") " pod="openstack/barbican-keystone-listener-6d6c6fd44b-2hfzf" Feb 02 09:25:32 crc kubenswrapper[4764]: I0202 09:25:32.684715 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4ebb36bb-8032-42ea-8954-7716e5a81301-config-data-custom\") pod \"barbican-worker-6db5f6d5bf-pxtw5\" (UID: \"4ebb36bb-8032-42ea-8954-7716e5a81301\") " pod="openstack/barbican-worker-6db5f6d5bf-pxtw5" Feb 02 09:25:32 crc kubenswrapper[4764]: I0202 09:25:32.684744 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ebb36bb-8032-42ea-8954-7716e5a81301-combined-ca-bundle\") pod \"barbican-worker-6db5f6d5bf-pxtw5\" (UID: \"4ebb36bb-8032-42ea-8954-7716e5a81301\") " pod="openstack/barbican-worker-6db5f6d5bf-pxtw5" Feb 02 09:25:32 crc kubenswrapper[4764]: I0202 09:25:32.684790 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4ebb36bb-8032-42ea-8954-7716e5a81301-logs\") pod \"barbican-worker-6db5f6d5bf-pxtw5\" (UID: \"4ebb36bb-8032-42ea-8954-7716e5a81301\") " pod="openstack/barbican-worker-6db5f6d5bf-pxtw5" Feb 02 09:25:32 crc kubenswrapper[4764]: I0202 09:25:32.684819 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ebb36bb-8032-42ea-8954-7716e5a81301-config-data\") pod \"barbican-worker-6db5f6d5bf-pxtw5\" (UID: \"4ebb36bb-8032-42ea-8954-7716e5a81301\") " pod="openstack/barbican-worker-6db5f6d5bf-pxtw5" Feb 02 09:25:32 crc kubenswrapper[4764]: I0202 09:25:32.684863 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4f12af32-00c9-437a-8e0e-805caa8df878-config-data\") pod \"barbican-keystone-listener-6d6c6fd44b-2hfzf\" (UID: \"4f12af32-00c9-437a-8e0e-805caa8df878\") " pod="openstack/barbican-keystone-listener-6d6c6fd44b-2hfzf" Feb 02 09:25:32 crc kubenswrapper[4764]: I0202 09:25:32.684890 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cb7vb\" (UniqueName: \"kubernetes.io/projected/4ebb36bb-8032-42ea-8954-7716e5a81301-kube-api-access-cb7vb\") pod \"barbican-worker-6db5f6d5bf-pxtw5\" (UID: \"4ebb36bb-8032-42ea-8954-7716e5a81301\") " pod="openstack/barbican-worker-6db5f6d5bf-pxtw5" Feb 02 09:25:32 crc kubenswrapper[4764]: I0202 09:25:32.684918 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f12af32-00c9-437a-8e0e-805caa8df878-combined-ca-bundle\") pod \"barbican-keystone-listener-6d6c6fd44b-2hfzf\" (UID: \"4f12af32-00c9-437a-8e0e-805caa8df878\") " pod="openstack/barbican-keystone-listener-6d6c6fd44b-2hfzf" Feb 02 09:25:32 crc kubenswrapper[4764]: I0202 09:25:32.684979 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4f12af32-00c9-437a-8e0e-805caa8df878-logs\") pod \"barbican-keystone-listener-6d6c6fd44b-2hfzf\" (UID: \"4f12af32-00c9-437a-8e0e-805caa8df878\") " pod="openstack/barbican-keystone-listener-6d6c6fd44b-2hfzf" Feb 02 09:25:32 crc kubenswrapper[4764]: I0202 09:25:32.685004 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4f12af32-00c9-437a-8e0e-805caa8df878-config-data-custom\") pod \"barbican-keystone-listener-6d6c6fd44b-2hfzf\" (UID: \"4f12af32-00c9-437a-8e0e-805caa8df878\") " pod="openstack/barbican-keystone-listener-6d6c6fd44b-2hfzf" Feb 02 09:25:32 crc kubenswrapper[4764]: I0202 09:25:32.687108 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4ebb36bb-8032-42ea-8954-7716e5a81301-logs\") pod \"barbican-worker-6db5f6d5bf-pxtw5\" (UID: \"4ebb36bb-8032-42ea-8954-7716e5a81301\") " pod="openstack/barbican-worker-6db5f6d5bf-pxtw5" Feb 02 09:25:32 crc kubenswrapper[4764]: I0202 09:25:32.710615 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ebb36bb-8032-42ea-8954-7716e5a81301-combined-ca-bundle\") pod \"barbican-worker-6db5f6d5bf-pxtw5\" (UID: \"4ebb36bb-8032-42ea-8954-7716e5a81301\") " pod="openstack/barbican-worker-6db5f6d5bf-pxtw5" Feb 02 09:25:32 crc kubenswrapper[4764]: I0202 09:25:32.722996 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cb7vb\" (UniqueName: \"kubernetes.io/projected/4ebb36bb-8032-42ea-8954-7716e5a81301-kube-api-access-cb7vb\") pod \"barbican-worker-6db5f6d5bf-pxtw5\" (UID: \"4ebb36bb-8032-42ea-8954-7716e5a81301\") " pod="openstack/barbican-worker-6db5f6d5bf-pxtw5" Feb 02 09:25:32 crc kubenswrapper[4764]: I0202 09:25:32.727797 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4ebb36bb-8032-42ea-8954-7716e5a81301-config-data-custom\") pod \"barbican-worker-6db5f6d5bf-pxtw5\" (UID: \"4ebb36bb-8032-42ea-8954-7716e5a81301\") " pod="openstack/barbican-worker-6db5f6d5bf-pxtw5" Feb 02 09:25:32 crc kubenswrapper[4764]: I0202 09:25:32.744699 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ebb36bb-8032-42ea-8954-7716e5a81301-config-data\") pod \"barbican-worker-6db5f6d5bf-pxtw5\" (UID: \"4ebb36bb-8032-42ea-8954-7716e5a81301\") " pod="openstack/barbican-worker-6db5f6d5bf-pxtw5" Feb 02 09:25:32 crc kubenswrapper[4764]: I0202 09:25:32.789825 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4f12af32-00c9-437a-8e0e-805caa8df878-config-data\") pod \"barbican-keystone-listener-6d6c6fd44b-2hfzf\" (UID: \"4f12af32-00c9-437a-8e0e-805caa8df878\") " pod="openstack/barbican-keystone-listener-6d6c6fd44b-2hfzf" Feb 02 09:25:32 crc kubenswrapper[4764]: I0202 09:25:32.789905 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f12af32-00c9-437a-8e0e-805caa8df878-combined-ca-bundle\") pod \"barbican-keystone-listener-6d6c6fd44b-2hfzf\" (UID: \"4f12af32-00c9-437a-8e0e-805caa8df878\") " pod="openstack/barbican-keystone-listener-6d6c6fd44b-2hfzf" Feb 02 09:25:32 crc kubenswrapper[4764]: I0202 09:25:32.790004 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4f12af32-00c9-437a-8e0e-805caa8df878-logs\") pod \"barbican-keystone-listener-6d6c6fd44b-2hfzf\" (UID: \"4f12af32-00c9-437a-8e0e-805caa8df878\") " pod="openstack/barbican-keystone-listener-6d6c6fd44b-2hfzf" Feb 02 09:25:32 crc kubenswrapper[4764]: I0202 09:25:32.790033 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4f12af32-00c9-437a-8e0e-805caa8df878-config-data-custom\") pod \"barbican-keystone-listener-6d6c6fd44b-2hfzf\" (UID: \"4f12af32-00c9-437a-8e0e-805caa8df878\") " pod="openstack/barbican-keystone-listener-6d6c6fd44b-2hfzf" Feb 02 09:25:32 crc kubenswrapper[4764]: I0202 09:25:32.790060 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rsq9n\" (UniqueName: \"kubernetes.io/projected/4f12af32-00c9-437a-8e0e-805caa8df878-kube-api-access-rsq9n\") pod \"barbican-keystone-listener-6d6c6fd44b-2hfzf\" (UID: \"4f12af32-00c9-437a-8e0e-805caa8df878\") " pod="openstack/barbican-keystone-listener-6d6c6fd44b-2hfzf" Feb 02 09:25:32 crc kubenswrapper[4764]: I0202 09:25:32.791649 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4f12af32-00c9-437a-8e0e-805caa8df878-logs\") pod \"barbican-keystone-listener-6d6c6fd44b-2hfzf\" (UID: \"4f12af32-00c9-437a-8e0e-805caa8df878\") " pod="openstack/barbican-keystone-listener-6d6c6fd44b-2hfzf" Feb 02 09:25:32 crc kubenswrapper[4764]: I0202 09:25:32.799135 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f12af32-00c9-437a-8e0e-805caa8df878-combined-ca-bundle\") pod \"barbican-keystone-listener-6d6c6fd44b-2hfzf\" (UID: \"4f12af32-00c9-437a-8e0e-805caa8df878\") " pod="openstack/barbican-keystone-listener-6d6c6fd44b-2hfzf" Feb 02 09:25:32 crc kubenswrapper[4764]: I0202 09:25:32.799369 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-6db5f6d5bf-pxtw5" Feb 02 09:25:32 crc kubenswrapper[4764]: I0202 09:25:32.800623 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4f12af32-00c9-437a-8e0e-805caa8df878-config-data\") pod \"barbican-keystone-listener-6d6c6fd44b-2hfzf\" (UID: \"4f12af32-00c9-437a-8e0e-805caa8df878\") " pod="openstack/barbican-keystone-listener-6d6c6fd44b-2hfzf" Feb 02 09:25:32 crc kubenswrapper[4764]: I0202 09:25:32.799250 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4f12af32-00c9-437a-8e0e-805caa8df878-config-data-custom\") pod \"barbican-keystone-listener-6d6c6fd44b-2hfzf\" (UID: \"4f12af32-00c9-437a-8e0e-805caa8df878\") " pod="openstack/barbican-keystone-listener-6d6c6fd44b-2hfzf" Feb 02 09:25:32 crc kubenswrapper[4764]: I0202 09:25:32.832158 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rsq9n\" (UniqueName: \"kubernetes.io/projected/4f12af32-00c9-437a-8e0e-805caa8df878-kube-api-access-rsq9n\") pod \"barbican-keystone-listener-6d6c6fd44b-2hfzf\" (UID: \"4f12af32-00c9-437a-8e0e-805caa8df878\") " pod="openstack/barbican-keystone-listener-6d6c6fd44b-2hfzf" Feb 02 09:25:32 crc kubenswrapper[4764]: I0202 09:25:32.878518 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-6d6c6fd44b-2hfzf" Feb 02 09:25:32 crc kubenswrapper[4764]: I0202 09:25:32.963701 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-64d8599669-wbmlc"] Feb 02 09:25:32 crc kubenswrapper[4764]: I0202 09:25:32.964091 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-64d8599669-wbmlc" podUID="0482366b-96dd-493a-98fb-c9cc90254d1e" containerName="dnsmasq-dns" containerID="cri-o://47268c7c0203a63ca5fa59eed2fee46a95043af36c15272dcc23c51f4a290b68" gracePeriod=10 Feb 02 09:25:32 crc kubenswrapper[4764]: I0202 09:25:32.965726 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-64d8599669-wbmlc" Feb 02 09:25:33 crc kubenswrapper[4764]: I0202 09:25:33.077753 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-7c55c99774-xk9j7"] Feb 02 09:25:33 crc kubenswrapper[4764]: I0202 09:25:33.079463 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7c55c99774-xk9j7" Feb 02 09:25:33 crc kubenswrapper[4764]: I0202 09:25:33.129542 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Feb 02 09:25:33 crc kubenswrapper[4764]: I0202 09:25:33.191082 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-7c55c99774-xk9j7"] Feb 02 09:25:33 crc kubenswrapper[4764]: I0202 09:25:33.221179 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-74cf444645-8gljr"] Feb 02 09:25:33 crc kubenswrapper[4764]: I0202 09:25:33.233537 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/075a47f1-9307-494f-8268-c544497cf25b-config-data\") pod \"barbican-api-7c55c99774-xk9j7\" (UID: \"075a47f1-9307-494f-8268-c544497cf25b\") " pod="openstack/barbican-api-7c55c99774-xk9j7" Feb 02 09:25:33 crc kubenswrapper[4764]: I0202 09:25:33.234012 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2nzjv\" (UniqueName: \"kubernetes.io/projected/075a47f1-9307-494f-8268-c544497cf25b-kube-api-access-2nzjv\") pod \"barbican-api-7c55c99774-xk9j7\" (UID: \"075a47f1-9307-494f-8268-c544497cf25b\") " pod="openstack/barbican-api-7c55c99774-xk9j7" Feb 02 09:25:33 crc kubenswrapper[4764]: I0202 09:25:33.234175 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/075a47f1-9307-494f-8268-c544497cf25b-config-data-custom\") pod \"barbican-api-7c55c99774-xk9j7\" (UID: \"075a47f1-9307-494f-8268-c544497cf25b\") " pod="openstack/barbican-api-7c55c99774-xk9j7" Feb 02 09:25:33 crc kubenswrapper[4764]: I0202 09:25:33.235996 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/075a47f1-9307-494f-8268-c544497cf25b-logs\") pod \"barbican-api-7c55c99774-xk9j7\" (UID: \"075a47f1-9307-494f-8268-c544497cf25b\") " pod="openstack/barbican-api-7c55c99774-xk9j7" Feb 02 09:25:33 crc kubenswrapper[4764]: I0202 09:25:33.236148 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/075a47f1-9307-494f-8268-c544497cf25b-combined-ca-bundle\") pod \"barbican-api-7c55c99774-xk9j7\" (UID: \"075a47f1-9307-494f-8268-c544497cf25b\") " pod="openstack/barbican-api-7c55c99774-xk9j7" Feb 02 09:25:33 crc kubenswrapper[4764]: I0202 09:25:33.235691 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74cf444645-8gljr" Feb 02 09:25:33 crc kubenswrapper[4764]: I0202 09:25:33.327486 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-74cf444645-8gljr"] Feb 02 09:25:33 crc kubenswrapper[4764]: I0202 09:25:33.338590 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2nzjv\" (UniqueName: \"kubernetes.io/projected/075a47f1-9307-494f-8268-c544497cf25b-kube-api-access-2nzjv\") pod \"barbican-api-7c55c99774-xk9j7\" (UID: \"075a47f1-9307-494f-8268-c544497cf25b\") " pod="openstack/barbican-api-7c55c99774-xk9j7" Feb 02 09:25:33 crc kubenswrapper[4764]: I0202 09:25:33.338658 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/075a47f1-9307-494f-8268-c544497cf25b-config-data-custom\") pod \"barbican-api-7c55c99774-xk9j7\" (UID: \"075a47f1-9307-494f-8268-c544497cf25b\") " pod="openstack/barbican-api-7c55c99774-xk9j7" Feb 02 09:25:33 crc kubenswrapper[4764]: I0202 09:25:33.338738 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/64a8ff87-c26b-49c9-9a34-0b1833b6b7a5-dns-svc\") pod \"dnsmasq-dns-74cf444645-8gljr\" (UID: \"64a8ff87-c26b-49c9-9a34-0b1833b6b7a5\") " pod="openstack/dnsmasq-dns-74cf444645-8gljr" Feb 02 09:25:33 crc kubenswrapper[4764]: I0202 09:25:33.338758 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/64a8ff87-c26b-49c9-9a34-0b1833b6b7a5-ovsdbserver-nb\") pod \"dnsmasq-dns-74cf444645-8gljr\" (UID: \"64a8ff87-c26b-49c9-9a34-0b1833b6b7a5\") " pod="openstack/dnsmasq-dns-74cf444645-8gljr" Feb 02 09:25:33 crc kubenswrapper[4764]: I0202 09:25:33.338778 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/075a47f1-9307-494f-8268-c544497cf25b-logs\") pod \"barbican-api-7c55c99774-xk9j7\" (UID: \"075a47f1-9307-494f-8268-c544497cf25b\") " pod="openstack/barbican-api-7c55c99774-xk9j7" Feb 02 09:25:33 crc kubenswrapper[4764]: I0202 09:25:33.338884 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/075a47f1-9307-494f-8268-c544497cf25b-combined-ca-bundle\") pod \"barbican-api-7c55c99774-xk9j7\" (UID: \"075a47f1-9307-494f-8268-c544497cf25b\") " pod="openstack/barbican-api-7c55c99774-xk9j7" Feb 02 09:25:33 crc kubenswrapper[4764]: I0202 09:25:33.338960 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/075a47f1-9307-494f-8268-c544497cf25b-config-data\") pod \"barbican-api-7c55c99774-xk9j7\" (UID: \"075a47f1-9307-494f-8268-c544497cf25b\") " pod="openstack/barbican-api-7c55c99774-xk9j7" Feb 02 09:25:33 crc kubenswrapper[4764]: I0202 09:25:33.338985 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/64a8ff87-c26b-49c9-9a34-0b1833b6b7a5-ovsdbserver-sb\") pod \"dnsmasq-dns-74cf444645-8gljr\" (UID: \"64a8ff87-c26b-49c9-9a34-0b1833b6b7a5\") " pod="openstack/dnsmasq-dns-74cf444645-8gljr" Feb 02 09:25:33 crc kubenswrapper[4764]: I0202 09:25:33.339001 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7t727\" (UniqueName: \"kubernetes.io/projected/64a8ff87-c26b-49c9-9a34-0b1833b6b7a5-kube-api-access-7t727\") pod \"dnsmasq-dns-74cf444645-8gljr\" (UID: \"64a8ff87-c26b-49c9-9a34-0b1833b6b7a5\") " pod="openstack/dnsmasq-dns-74cf444645-8gljr" Feb 02 09:25:33 crc kubenswrapper[4764]: I0202 09:25:33.339019 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/64a8ff87-c26b-49c9-9a34-0b1833b6b7a5-config\") pod \"dnsmasq-dns-74cf444645-8gljr\" (UID: \"64a8ff87-c26b-49c9-9a34-0b1833b6b7a5\") " pod="openstack/dnsmasq-dns-74cf444645-8gljr" Feb 02 09:25:33 crc kubenswrapper[4764]: I0202 09:25:33.342796 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/075a47f1-9307-494f-8268-c544497cf25b-logs\") pod \"barbican-api-7c55c99774-xk9j7\" (UID: \"075a47f1-9307-494f-8268-c544497cf25b\") " pod="openstack/barbican-api-7c55c99774-xk9j7" Feb 02 09:25:33 crc kubenswrapper[4764]: I0202 09:25:33.357133 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/075a47f1-9307-494f-8268-c544497cf25b-combined-ca-bundle\") pod \"barbican-api-7c55c99774-xk9j7\" (UID: \"075a47f1-9307-494f-8268-c544497cf25b\") " pod="openstack/barbican-api-7c55c99774-xk9j7" Feb 02 09:25:33 crc kubenswrapper[4764]: I0202 09:25:33.357575 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/075a47f1-9307-494f-8268-c544497cf25b-config-data\") pod \"barbican-api-7c55c99774-xk9j7\" (UID: \"075a47f1-9307-494f-8268-c544497cf25b\") " pod="openstack/barbican-api-7c55c99774-xk9j7" Feb 02 09:25:33 crc kubenswrapper[4764]: I0202 09:25:33.363986 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/075a47f1-9307-494f-8268-c544497cf25b-config-data-custom\") pod \"barbican-api-7c55c99774-xk9j7\" (UID: \"075a47f1-9307-494f-8268-c544497cf25b\") " pod="openstack/barbican-api-7c55c99774-xk9j7" Feb 02 09:25:33 crc kubenswrapper[4764]: I0202 09:25:33.385542 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2nzjv\" (UniqueName: \"kubernetes.io/projected/075a47f1-9307-494f-8268-c544497cf25b-kube-api-access-2nzjv\") pod \"barbican-api-7c55c99774-xk9j7\" (UID: \"075a47f1-9307-494f-8268-c544497cf25b\") " pod="openstack/barbican-api-7c55c99774-xk9j7" Feb 02 09:25:33 crc kubenswrapper[4764]: I0202 09:25:33.440347 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/64a8ff87-c26b-49c9-9a34-0b1833b6b7a5-ovsdbserver-sb\") pod \"dnsmasq-dns-74cf444645-8gljr\" (UID: \"64a8ff87-c26b-49c9-9a34-0b1833b6b7a5\") " pod="openstack/dnsmasq-dns-74cf444645-8gljr" Feb 02 09:25:33 crc kubenswrapper[4764]: I0202 09:25:33.441574 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7t727\" (UniqueName: \"kubernetes.io/projected/64a8ff87-c26b-49c9-9a34-0b1833b6b7a5-kube-api-access-7t727\") pod \"dnsmasq-dns-74cf444645-8gljr\" (UID: \"64a8ff87-c26b-49c9-9a34-0b1833b6b7a5\") " pod="openstack/dnsmasq-dns-74cf444645-8gljr" Feb 02 09:25:33 crc kubenswrapper[4764]: I0202 09:25:33.441665 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/64a8ff87-c26b-49c9-9a34-0b1833b6b7a5-config\") pod \"dnsmasq-dns-74cf444645-8gljr\" (UID: \"64a8ff87-c26b-49c9-9a34-0b1833b6b7a5\") " pod="openstack/dnsmasq-dns-74cf444645-8gljr" Feb 02 09:25:33 crc kubenswrapper[4764]: I0202 09:25:33.441343 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/64a8ff87-c26b-49c9-9a34-0b1833b6b7a5-ovsdbserver-sb\") pod \"dnsmasq-dns-74cf444645-8gljr\" (UID: \"64a8ff87-c26b-49c9-9a34-0b1833b6b7a5\") " pod="openstack/dnsmasq-dns-74cf444645-8gljr" Feb 02 09:25:33 crc kubenswrapper[4764]: I0202 09:25:33.441924 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/64a8ff87-c26b-49c9-9a34-0b1833b6b7a5-dns-svc\") pod \"dnsmasq-dns-74cf444645-8gljr\" (UID: \"64a8ff87-c26b-49c9-9a34-0b1833b6b7a5\") " pod="openstack/dnsmasq-dns-74cf444645-8gljr" Feb 02 09:25:33 crc kubenswrapper[4764]: I0202 09:25:33.442012 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/64a8ff87-c26b-49c9-9a34-0b1833b6b7a5-ovsdbserver-nb\") pod \"dnsmasq-dns-74cf444645-8gljr\" (UID: \"64a8ff87-c26b-49c9-9a34-0b1833b6b7a5\") " pod="openstack/dnsmasq-dns-74cf444645-8gljr" Feb 02 09:25:33 crc kubenswrapper[4764]: I0202 09:25:33.442422 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/64a8ff87-c26b-49c9-9a34-0b1833b6b7a5-config\") pod \"dnsmasq-dns-74cf444645-8gljr\" (UID: \"64a8ff87-c26b-49c9-9a34-0b1833b6b7a5\") " pod="openstack/dnsmasq-dns-74cf444645-8gljr" Feb 02 09:25:33 crc kubenswrapper[4764]: I0202 09:25:33.442923 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/64a8ff87-c26b-49c9-9a34-0b1833b6b7a5-ovsdbserver-nb\") pod \"dnsmasq-dns-74cf444645-8gljr\" (UID: \"64a8ff87-c26b-49c9-9a34-0b1833b6b7a5\") " pod="openstack/dnsmasq-dns-74cf444645-8gljr" Feb 02 09:25:33 crc kubenswrapper[4764]: I0202 09:25:33.443193 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/64a8ff87-c26b-49c9-9a34-0b1833b6b7a5-dns-svc\") pod \"dnsmasq-dns-74cf444645-8gljr\" (UID: \"64a8ff87-c26b-49c9-9a34-0b1833b6b7a5\") " pod="openstack/dnsmasq-dns-74cf444645-8gljr" Feb 02 09:25:33 crc kubenswrapper[4764]: I0202 09:25:33.460478 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7c55c99774-xk9j7" Feb 02 09:25:33 crc kubenswrapper[4764]: I0202 09:25:33.485924 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7t727\" (UniqueName: \"kubernetes.io/projected/64a8ff87-c26b-49c9-9a34-0b1833b6b7a5-kube-api-access-7t727\") pod \"dnsmasq-dns-74cf444645-8gljr\" (UID: \"64a8ff87-c26b-49c9-9a34-0b1833b6b7a5\") " pod="openstack/dnsmasq-dns-74cf444645-8gljr" Feb 02 09:25:33 crc kubenswrapper[4764]: I0202 09:25:33.589310 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74cf444645-8gljr" Feb 02 09:25:34 crc kubenswrapper[4764]: I0202 09:25:34.134291 4764 generic.go:334] "Generic (PLEG): container finished" podID="0482366b-96dd-493a-98fb-c9cc90254d1e" containerID="47268c7c0203a63ca5fa59eed2fee46a95043af36c15272dcc23c51f4a290b68" exitCode=0 Feb 02 09:25:34 crc kubenswrapper[4764]: I0202 09:25:34.134336 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64d8599669-wbmlc" event={"ID":"0482366b-96dd-493a-98fb-c9cc90254d1e","Type":"ContainerDied","Data":"47268c7c0203a63ca5fa59eed2fee46a95043af36c15272dcc23c51f4a290b68"} Feb 02 09:25:34 crc kubenswrapper[4764]: I0202 09:25:34.482102 4764 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-64d8599669-wbmlc" podUID="0482366b-96dd-493a-98fb-c9cc90254d1e" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.138:5353: connect: connection refused" Feb 02 09:25:34 crc kubenswrapper[4764]: I0202 09:25:34.945827 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-phkvk" Feb 02 09:25:35 crc kubenswrapper[4764]: I0202 09:25:35.076336 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4537c8c1-f5dc-4974-9885-527d6c85e100-combined-ca-bundle\") pod \"4537c8c1-f5dc-4974-9885-527d6c85e100\" (UID: \"4537c8c1-f5dc-4974-9885-527d6c85e100\") " Feb 02 09:25:35 crc kubenswrapper[4764]: I0202 09:25:35.077264 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4537c8c1-f5dc-4974-9885-527d6c85e100-config-data\") pod \"4537c8c1-f5dc-4974-9885-527d6c85e100\" (UID: \"4537c8c1-f5dc-4974-9885-527d6c85e100\") " Feb 02 09:25:35 crc kubenswrapper[4764]: I0202 09:25:35.077292 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4537c8c1-f5dc-4974-9885-527d6c85e100-scripts\") pod \"4537c8c1-f5dc-4974-9885-527d6c85e100\" (UID: \"4537c8c1-f5dc-4974-9885-527d6c85e100\") " Feb 02 09:25:35 crc kubenswrapper[4764]: I0202 09:25:35.077318 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k2g4l\" (UniqueName: \"kubernetes.io/projected/4537c8c1-f5dc-4974-9885-527d6c85e100-kube-api-access-k2g4l\") pod \"4537c8c1-f5dc-4974-9885-527d6c85e100\" (UID: \"4537c8c1-f5dc-4974-9885-527d6c85e100\") " Feb 02 09:25:35 crc kubenswrapper[4764]: I0202 09:25:35.077405 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/4537c8c1-f5dc-4974-9885-527d6c85e100-credential-keys\") pod \"4537c8c1-f5dc-4974-9885-527d6c85e100\" (UID: \"4537c8c1-f5dc-4974-9885-527d6c85e100\") " Feb 02 09:25:35 crc kubenswrapper[4764]: I0202 09:25:35.077501 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4537c8c1-f5dc-4974-9885-527d6c85e100-fernet-keys\") pod \"4537c8c1-f5dc-4974-9885-527d6c85e100\" (UID: \"4537c8c1-f5dc-4974-9885-527d6c85e100\") " Feb 02 09:25:35 crc kubenswrapper[4764]: I0202 09:25:35.087171 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4537c8c1-f5dc-4974-9885-527d6c85e100-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "4537c8c1-f5dc-4974-9885-527d6c85e100" (UID: "4537c8c1-f5dc-4974-9885-527d6c85e100"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:25:35 crc kubenswrapper[4764]: I0202 09:25:35.093335 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4537c8c1-f5dc-4974-9885-527d6c85e100-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "4537c8c1-f5dc-4974-9885-527d6c85e100" (UID: "4537c8c1-f5dc-4974-9885-527d6c85e100"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:25:35 crc kubenswrapper[4764]: I0202 09:25:35.094555 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4537c8c1-f5dc-4974-9885-527d6c85e100-kube-api-access-k2g4l" (OuterVolumeSpecName: "kube-api-access-k2g4l") pod "4537c8c1-f5dc-4974-9885-527d6c85e100" (UID: "4537c8c1-f5dc-4974-9885-527d6c85e100"). InnerVolumeSpecName "kube-api-access-k2g4l". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:25:35 crc kubenswrapper[4764]: I0202 09:25:35.095191 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4537c8c1-f5dc-4974-9885-527d6c85e100-scripts" (OuterVolumeSpecName: "scripts") pod "4537c8c1-f5dc-4974-9885-527d6c85e100" (UID: "4537c8c1-f5dc-4974-9885-527d6c85e100"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:25:35 crc kubenswrapper[4764]: I0202 09:25:35.183888 4764 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4537c8c1-f5dc-4974-9885-527d6c85e100-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 09:25:35 crc kubenswrapper[4764]: I0202 09:25:35.183977 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k2g4l\" (UniqueName: \"kubernetes.io/projected/4537c8c1-f5dc-4974-9885-527d6c85e100-kube-api-access-k2g4l\") on node \"crc\" DevicePath \"\"" Feb 02 09:25:35 crc kubenswrapper[4764]: I0202 09:25:35.183986 4764 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/4537c8c1-f5dc-4974-9885-527d6c85e100-credential-keys\") on node \"crc\" DevicePath \"\"" Feb 02 09:25:35 crc kubenswrapper[4764]: I0202 09:25:35.183994 4764 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4537c8c1-f5dc-4974-9885-527d6c85e100-fernet-keys\") on node \"crc\" DevicePath \"\"" Feb 02 09:25:35 crc kubenswrapper[4764]: I0202 09:25:35.231158 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-phkvk" Feb 02 09:25:35 crc kubenswrapper[4764]: I0202 09:25:35.233592 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-phkvk" event={"ID":"4537c8c1-f5dc-4974-9885-527d6c85e100","Type":"ContainerDied","Data":"174e9cfc20869b1928c9becfde27c30d04dfa06aea1b1367b4cc81493ad1be72"} Feb 02 09:25:35 crc kubenswrapper[4764]: I0202 09:25:35.233624 4764 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="174e9cfc20869b1928c9becfde27c30d04dfa06aea1b1367b4cc81493ad1be72" Feb 02 09:25:35 crc kubenswrapper[4764]: I0202 09:25:35.235769 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4537c8c1-f5dc-4974-9885-527d6c85e100-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4537c8c1-f5dc-4974-9885-527d6c85e100" (UID: "4537c8c1-f5dc-4974-9885-527d6c85e100"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:25:35 crc kubenswrapper[4764]: I0202 09:25:35.255463 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4537c8c1-f5dc-4974-9885-527d6c85e100-config-data" (OuterVolumeSpecName: "config-data") pod "4537c8c1-f5dc-4974-9885-527d6c85e100" (UID: "4537c8c1-f5dc-4974-9885-527d6c85e100"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:25:35 crc kubenswrapper[4764]: I0202 09:25:35.285817 4764 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4537c8c1-f5dc-4974-9885-527d6c85e100-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 09:25:35 crc kubenswrapper[4764]: I0202 09:25:35.286241 4764 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4537c8c1-f5dc-4974-9885-527d6c85e100-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 09:25:35 crc kubenswrapper[4764]: I0202 09:25:35.456608 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64d8599669-wbmlc" Feb 02 09:25:35 crc kubenswrapper[4764]: I0202 09:25:35.601569 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0482366b-96dd-493a-98fb-c9cc90254d1e-dns-svc\") pod \"0482366b-96dd-493a-98fb-c9cc90254d1e\" (UID: \"0482366b-96dd-493a-98fb-c9cc90254d1e\") " Feb 02 09:25:35 crc kubenswrapper[4764]: I0202 09:25:35.601628 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ldw77\" (UniqueName: \"kubernetes.io/projected/0482366b-96dd-493a-98fb-c9cc90254d1e-kube-api-access-ldw77\") pod \"0482366b-96dd-493a-98fb-c9cc90254d1e\" (UID: \"0482366b-96dd-493a-98fb-c9cc90254d1e\") " Feb 02 09:25:35 crc kubenswrapper[4764]: I0202 09:25:35.601692 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0482366b-96dd-493a-98fb-c9cc90254d1e-config\") pod \"0482366b-96dd-493a-98fb-c9cc90254d1e\" (UID: \"0482366b-96dd-493a-98fb-c9cc90254d1e\") " Feb 02 09:25:35 crc kubenswrapper[4764]: I0202 09:25:35.601728 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0482366b-96dd-493a-98fb-c9cc90254d1e-ovsdbserver-sb\") pod \"0482366b-96dd-493a-98fb-c9cc90254d1e\" (UID: \"0482366b-96dd-493a-98fb-c9cc90254d1e\") " Feb 02 09:25:35 crc kubenswrapper[4764]: I0202 09:25:35.601762 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0482366b-96dd-493a-98fb-c9cc90254d1e-ovsdbserver-nb\") pod \"0482366b-96dd-493a-98fb-c9cc90254d1e\" (UID: \"0482366b-96dd-493a-98fb-c9cc90254d1e\") " Feb 02 09:25:35 crc kubenswrapper[4764]: I0202 09:25:35.642219 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0482366b-96dd-493a-98fb-c9cc90254d1e-kube-api-access-ldw77" (OuterVolumeSpecName: "kube-api-access-ldw77") pod "0482366b-96dd-493a-98fb-c9cc90254d1e" (UID: "0482366b-96dd-493a-98fb-c9cc90254d1e"). InnerVolumeSpecName "kube-api-access-ldw77". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:25:35 crc kubenswrapper[4764]: I0202 09:25:35.652570 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-7b6fbdbb96-fgr8m"] Feb 02 09:25:35 crc kubenswrapper[4764]: E0202 09:25:35.653084 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4537c8c1-f5dc-4974-9885-527d6c85e100" containerName="keystone-bootstrap" Feb 02 09:25:35 crc kubenswrapper[4764]: I0202 09:25:35.653737 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="4537c8c1-f5dc-4974-9885-527d6c85e100" containerName="keystone-bootstrap" Feb 02 09:25:35 crc kubenswrapper[4764]: E0202 09:25:35.653763 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0482366b-96dd-493a-98fb-c9cc90254d1e" containerName="init" Feb 02 09:25:35 crc kubenswrapper[4764]: I0202 09:25:35.653769 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="0482366b-96dd-493a-98fb-c9cc90254d1e" containerName="init" Feb 02 09:25:35 crc kubenswrapper[4764]: E0202 09:25:35.653779 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0482366b-96dd-493a-98fb-c9cc90254d1e" containerName="dnsmasq-dns" Feb 02 09:25:35 crc kubenswrapper[4764]: I0202 09:25:35.653785 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="0482366b-96dd-493a-98fb-c9cc90254d1e" containerName="dnsmasq-dns" Feb 02 09:25:35 crc kubenswrapper[4764]: I0202 09:25:35.653960 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="4537c8c1-f5dc-4974-9885-527d6c85e100" containerName="keystone-bootstrap" Feb 02 09:25:35 crc kubenswrapper[4764]: I0202 09:25:35.653974 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="0482366b-96dd-493a-98fb-c9cc90254d1e" containerName="dnsmasq-dns" Feb 02 09:25:35 crc kubenswrapper[4764]: I0202 09:25:35.654809 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7b6fbdbb96-fgr8m" Feb 02 09:25:35 crc kubenswrapper[4764]: I0202 09:25:35.663132 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Feb 02 09:25:35 crc kubenswrapper[4764]: I0202 09:25:35.663337 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Feb 02 09:25:35 crc kubenswrapper[4764]: I0202 09:25:35.703896 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9sxl9\" (UniqueName: \"kubernetes.io/projected/40e640ad-6fb1-4c56-9b33-14fe28103b07-kube-api-access-9sxl9\") pod \"barbican-api-7b6fbdbb96-fgr8m\" (UID: \"40e640ad-6fb1-4c56-9b33-14fe28103b07\") " pod="openstack/barbican-api-7b6fbdbb96-fgr8m" Feb 02 09:25:35 crc kubenswrapper[4764]: I0202 09:25:35.704404 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-7b6fbdbb96-fgr8m"] Feb 02 09:25:35 crc kubenswrapper[4764]: I0202 09:25:35.703970 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/40e640ad-6fb1-4c56-9b33-14fe28103b07-public-tls-certs\") pod \"barbican-api-7b6fbdbb96-fgr8m\" (UID: \"40e640ad-6fb1-4c56-9b33-14fe28103b07\") " pod="openstack/barbican-api-7b6fbdbb96-fgr8m" Feb 02 09:25:35 crc kubenswrapper[4764]: I0202 09:25:35.705912 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/40e640ad-6fb1-4c56-9b33-14fe28103b07-internal-tls-certs\") pod \"barbican-api-7b6fbdbb96-fgr8m\" (UID: \"40e640ad-6fb1-4c56-9b33-14fe28103b07\") " pod="openstack/barbican-api-7b6fbdbb96-fgr8m" Feb 02 09:25:35 crc kubenswrapper[4764]: I0202 09:25:35.706088 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/40e640ad-6fb1-4c56-9b33-14fe28103b07-config-data\") pod \"barbican-api-7b6fbdbb96-fgr8m\" (UID: \"40e640ad-6fb1-4c56-9b33-14fe28103b07\") " pod="openstack/barbican-api-7b6fbdbb96-fgr8m" Feb 02 09:25:35 crc kubenswrapper[4764]: I0202 09:25:35.706297 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/40e640ad-6fb1-4c56-9b33-14fe28103b07-logs\") pod \"barbican-api-7b6fbdbb96-fgr8m\" (UID: \"40e640ad-6fb1-4c56-9b33-14fe28103b07\") " pod="openstack/barbican-api-7b6fbdbb96-fgr8m" Feb 02 09:25:35 crc kubenswrapper[4764]: I0202 09:25:35.706466 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40e640ad-6fb1-4c56-9b33-14fe28103b07-combined-ca-bundle\") pod \"barbican-api-7b6fbdbb96-fgr8m\" (UID: \"40e640ad-6fb1-4c56-9b33-14fe28103b07\") " pod="openstack/barbican-api-7b6fbdbb96-fgr8m" Feb 02 09:25:35 crc kubenswrapper[4764]: I0202 09:25:35.706554 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/40e640ad-6fb1-4c56-9b33-14fe28103b07-config-data-custom\") pod \"barbican-api-7b6fbdbb96-fgr8m\" (UID: \"40e640ad-6fb1-4c56-9b33-14fe28103b07\") " pod="openstack/barbican-api-7b6fbdbb96-fgr8m" Feb 02 09:25:35 crc kubenswrapper[4764]: I0202 09:25:35.706665 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ldw77\" (UniqueName: \"kubernetes.io/projected/0482366b-96dd-493a-98fb-c9cc90254d1e-kube-api-access-ldw77\") on node \"crc\" DevicePath \"\"" Feb 02 09:25:35 crc kubenswrapper[4764]: I0202 09:25:35.715364 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-74cf444645-8gljr"] Feb 02 09:25:35 crc kubenswrapper[4764]: I0202 09:25:35.724923 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-65fcd88786-dggfp"] Feb 02 09:25:35 crc kubenswrapper[4764]: W0202 09:25:35.755261 4764 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod64a8ff87_c26b_49c9_9a34_0b1833b6b7a5.slice/crio-2faf3e8aa7d0587946beb4a728098bf586b0031811a48af2f1746583d014ddef WatchSource:0}: Error finding container 2faf3e8aa7d0587946beb4a728098bf586b0031811a48af2f1746583d014ddef: Status 404 returned error can't find the container with id 2faf3e8aa7d0587946beb4a728098bf586b0031811a48af2f1746583d014ddef Feb 02 09:25:35 crc kubenswrapper[4764]: I0202 09:25:35.761416 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0482366b-96dd-493a-98fb-c9cc90254d1e-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "0482366b-96dd-493a-98fb-c9cc90254d1e" (UID: "0482366b-96dd-493a-98fb-c9cc90254d1e"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:25:35 crc kubenswrapper[4764]: I0202 09:25:35.770313 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0482366b-96dd-493a-98fb-c9cc90254d1e-config" (OuterVolumeSpecName: "config") pod "0482366b-96dd-493a-98fb-c9cc90254d1e" (UID: "0482366b-96dd-493a-98fb-c9cc90254d1e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:25:35 crc kubenswrapper[4764]: I0202 09:25:35.772340 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0482366b-96dd-493a-98fb-c9cc90254d1e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "0482366b-96dd-493a-98fb-c9cc90254d1e" (UID: "0482366b-96dd-493a-98fb-c9cc90254d1e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:25:35 crc kubenswrapper[4764]: I0202 09:25:35.785093 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0482366b-96dd-493a-98fb-c9cc90254d1e-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "0482366b-96dd-493a-98fb-c9cc90254d1e" (UID: "0482366b-96dd-493a-98fb-c9cc90254d1e"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:25:35 crc kubenswrapper[4764]: I0202 09:25:35.810879 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9sxl9\" (UniqueName: \"kubernetes.io/projected/40e640ad-6fb1-4c56-9b33-14fe28103b07-kube-api-access-9sxl9\") pod \"barbican-api-7b6fbdbb96-fgr8m\" (UID: \"40e640ad-6fb1-4c56-9b33-14fe28103b07\") " pod="openstack/barbican-api-7b6fbdbb96-fgr8m" Feb 02 09:25:35 crc kubenswrapper[4764]: I0202 09:25:35.810955 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/40e640ad-6fb1-4c56-9b33-14fe28103b07-public-tls-certs\") pod \"barbican-api-7b6fbdbb96-fgr8m\" (UID: \"40e640ad-6fb1-4c56-9b33-14fe28103b07\") " pod="openstack/barbican-api-7b6fbdbb96-fgr8m" Feb 02 09:25:35 crc kubenswrapper[4764]: I0202 09:25:35.810977 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/40e640ad-6fb1-4c56-9b33-14fe28103b07-internal-tls-certs\") pod \"barbican-api-7b6fbdbb96-fgr8m\" (UID: \"40e640ad-6fb1-4c56-9b33-14fe28103b07\") " pod="openstack/barbican-api-7b6fbdbb96-fgr8m" Feb 02 09:25:35 crc kubenswrapper[4764]: I0202 09:25:35.811003 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/40e640ad-6fb1-4c56-9b33-14fe28103b07-config-data\") pod \"barbican-api-7b6fbdbb96-fgr8m\" (UID: \"40e640ad-6fb1-4c56-9b33-14fe28103b07\") " pod="openstack/barbican-api-7b6fbdbb96-fgr8m" Feb 02 09:25:35 crc kubenswrapper[4764]: I0202 09:25:35.811095 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/40e640ad-6fb1-4c56-9b33-14fe28103b07-logs\") pod \"barbican-api-7b6fbdbb96-fgr8m\" (UID: \"40e640ad-6fb1-4c56-9b33-14fe28103b07\") " pod="openstack/barbican-api-7b6fbdbb96-fgr8m" Feb 02 09:25:35 crc kubenswrapper[4764]: I0202 09:25:35.811155 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40e640ad-6fb1-4c56-9b33-14fe28103b07-combined-ca-bundle\") pod \"barbican-api-7b6fbdbb96-fgr8m\" (UID: \"40e640ad-6fb1-4c56-9b33-14fe28103b07\") " pod="openstack/barbican-api-7b6fbdbb96-fgr8m" Feb 02 09:25:35 crc kubenswrapper[4764]: I0202 09:25:35.811195 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/40e640ad-6fb1-4c56-9b33-14fe28103b07-config-data-custom\") pod \"barbican-api-7b6fbdbb96-fgr8m\" (UID: \"40e640ad-6fb1-4c56-9b33-14fe28103b07\") " pod="openstack/barbican-api-7b6fbdbb96-fgr8m" Feb 02 09:25:35 crc kubenswrapper[4764]: I0202 09:25:35.811253 4764 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0482366b-96dd-493a-98fb-c9cc90254d1e-config\") on node \"crc\" DevicePath \"\"" Feb 02 09:25:35 crc kubenswrapper[4764]: I0202 09:25:35.811263 4764 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0482366b-96dd-493a-98fb-c9cc90254d1e-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 02 09:25:35 crc kubenswrapper[4764]: I0202 09:25:35.811273 4764 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0482366b-96dd-493a-98fb-c9cc90254d1e-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 02 09:25:35 crc kubenswrapper[4764]: I0202 09:25:35.811282 4764 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0482366b-96dd-493a-98fb-c9cc90254d1e-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 02 09:25:35 crc kubenswrapper[4764]: I0202 09:25:35.813384 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/40e640ad-6fb1-4c56-9b33-14fe28103b07-logs\") pod \"barbican-api-7b6fbdbb96-fgr8m\" (UID: \"40e640ad-6fb1-4c56-9b33-14fe28103b07\") " pod="openstack/barbican-api-7b6fbdbb96-fgr8m" Feb 02 09:25:35 crc kubenswrapper[4764]: I0202 09:25:35.816916 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/40e640ad-6fb1-4c56-9b33-14fe28103b07-internal-tls-certs\") pod \"barbican-api-7b6fbdbb96-fgr8m\" (UID: \"40e640ad-6fb1-4c56-9b33-14fe28103b07\") " pod="openstack/barbican-api-7b6fbdbb96-fgr8m" Feb 02 09:25:35 crc kubenswrapper[4764]: I0202 09:25:35.821709 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40e640ad-6fb1-4c56-9b33-14fe28103b07-combined-ca-bundle\") pod \"barbican-api-7b6fbdbb96-fgr8m\" (UID: \"40e640ad-6fb1-4c56-9b33-14fe28103b07\") " pod="openstack/barbican-api-7b6fbdbb96-fgr8m" Feb 02 09:25:35 crc kubenswrapper[4764]: I0202 09:25:35.821966 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/40e640ad-6fb1-4c56-9b33-14fe28103b07-public-tls-certs\") pod \"barbican-api-7b6fbdbb96-fgr8m\" (UID: \"40e640ad-6fb1-4c56-9b33-14fe28103b07\") " pod="openstack/barbican-api-7b6fbdbb96-fgr8m" Feb 02 09:25:35 crc kubenswrapper[4764]: I0202 09:25:35.822077 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/40e640ad-6fb1-4c56-9b33-14fe28103b07-config-data-custom\") pod \"barbican-api-7b6fbdbb96-fgr8m\" (UID: \"40e640ad-6fb1-4c56-9b33-14fe28103b07\") " pod="openstack/barbican-api-7b6fbdbb96-fgr8m" Feb 02 09:25:35 crc kubenswrapper[4764]: I0202 09:25:35.822360 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/40e640ad-6fb1-4c56-9b33-14fe28103b07-config-data\") pod \"barbican-api-7b6fbdbb96-fgr8m\" (UID: \"40e640ad-6fb1-4c56-9b33-14fe28103b07\") " pod="openstack/barbican-api-7b6fbdbb96-fgr8m" Feb 02 09:25:35 crc kubenswrapper[4764]: I0202 09:25:35.831090 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9sxl9\" (UniqueName: \"kubernetes.io/projected/40e640ad-6fb1-4c56-9b33-14fe28103b07-kube-api-access-9sxl9\") pod \"barbican-api-7b6fbdbb96-fgr8m\" (UID: \"40e640ad-6fb1-4c56-9b33-14fe28103b07\") " pod="openstack/barbican-api-7b6fbdbb96-fgr8m" Feb 02 09:25:36 crc kubenswrapper[4764]: I0202 09:25:36.000556 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7b6fbdbb96-fgr8m" Feb 02 09:25:36 crc kubenswrapper[4764]: I0202 09:25:36.015001 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-6db5f6d5bf-pxtw5"] Feb 02 09:25:36 crc kubenswrapper[4764]: I0202 09:25:36.097001 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-6d6c6fd44b-2hfzf"] Feb 02 09:25:36 crc kubenswrapper[4764]: I0202 09:25:36.135440 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-7c55c99774-xk9j7"] Feb 02 09:25:36 crc kubenswrapper[4764]: W0202 09:25:36.180480 4764 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod075a47f1_9307_494f_8268_c544497cf25b.slice/crio-d5e01390ca284556900867d6beb52fd76ba6f0d7eb0a2abc10071f834486b312 WatchSource:0}: Error finding container d5e01390ca284556900867d6beb52fd76ba6f0d7eb0a2abc10071f834486b312: Status 404 returned error can't find the container with id d5e01390ca284556900867d6beb52fd76ba6f0d7eb0a2abc10071f834486b312 Feb 02 09:25:36 crc kubenswrapper[4764]: I0202 09:25:36.254126 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-6d6c6fd44b-2hfzf" event={"ID":"4f12af32-00c9-437a-8e0e-805caa8df878","Type":"ContainerStarted","Data":"d088edd9082f11f04848b13b2a6a0159d98d19ea524eaee71d140dc187193e16"} Feb 02 09:25:36 crc kubenswrapper[4764]: I0202 09:25:36.271206 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64d8599669-wbmlc" event={"ID":"0482366b-96dd-493a-98fb-c9cc90254d1e","Type":"ContainerDied","Data":"fe22a164d3f6ac15c7aacd92d740d287efb9340b2c933a000e014503af88ed0e"} Feb 02 09:25:36 crc kubenswrapper[4764]: I0202 09:25:36.271269 4764 scope.go:117] "RemoveContainer" containerID="47268c7c0203a63ca5fa59eed2fee46a95043af36c15272dcc23c51f4a290b68" Feb 02 09:25:36 crc kubenswrapper[4764]: I0202 09:25:36.271437 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64d8599669-wbmlc" Feb 02 09:25:36 crc kubenswrapper[4764]: I0202 09:25:36.289397 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-b47bfbff6-m265b"] Feb 02 09:25:36 crc kubenswrapper[4764]: I0202 09:25:36.290872 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-b47bfbff6-m265b" Feb 02 09:25:36 crc kubenswrapper[4764]: I0202 09:25:36.292523 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-6db5f6d5bf-pxtw5" event={"ID":"4ebb36bb-8032-42ea-8954-7716e5a81301","Type":"ContainerStarted","Data":"1834b62df8a95754f2ba8e5735004a0526bf197a1915bd2b88293ce268ee1dc3"} Feb 02 09:25:36 crc kubenswrapper[4764]: I0202 09:25:36.302797 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a518caeb-6b91-4b71-a47f-0d71d965af2e","Type":"ContainerStarted","Data":"babe286652a8ae251bf5bcfe389a56fa30eaefd0cd0ff5f5bcb6999e85408847"} Feb 02 09:25:36 crc kubenswrapper[4764]: I0202 09:25:36.304119 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-65fcd88786-dggfp" event={"ID":"dedd7b1f-7ec4-4a7a-a703-7e228f7767b1","Type":"ContainerStarted","Data":"d7df5c1af9c56a5971f2d2cda7bbcd4f2d4a1b32db5eeeb2b19911d615045fd5"} Feb 02 09:25:36 crc kubenswrapper[4764]: I0202 09:25:36.305779 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7c55c99774-xk9j7" event={"ID":"075a47f1-9307-494f-8268-c544497cf25b","Type":"ContainerStarted","Data":"d5e01390ca284556900867d6beb52fd76ba6f0d7eb0a2abc10071f834486b312"} Feb 02 09:25:36 crc kubenswrapper[4764]: I0202 09:25:36.306823 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74cf444645-8gljr" event={"ID":"64a8ff87-c26b-49c9-9a34-0b1833b6b7a5","Type":"ContainerStarted","Data":"2faf3e8aa7d0587946beb4a728098bf586b0031811a48af2f1746583d014ddef"} Feb 02 09:25:36 crc kubenswrapper[4764]: I0202 09:25:36.307100 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Feb 02 09:25:36 crc kubenswrapper[4764]: I0202 09:25:36.307996 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Feb 02 09:25:36 crc kubenswrapper[4764]: I0202 09:25:36.308240 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Feb 02 09:25:36 crc kubenswrapper[4764]: I0202 09:25:36.308427 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-dds8z" Feb 02 09:25:36 crc kubenswrapper[4764]: I0202 09:25:36.308622 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Feb 02 09:25:36 crc kubenswrapper[4764]: I0202 09:25:36.308829 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Feb 02 09:25:36 crc kubenswrapper[4764]: I0202 09:25:36.355593 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-64d8599669-wbmlc"] Feb 02 09:25:36 crc kubenswrapper[4764]: I0202 09:25:36.369107 4764 scope.go:117] "RemoveContainer" containerID="93ee852f1e7b5f4f874e25c5e628d6544c821ed4fdbbbee14374c3937bdd8907" Feb 02 09:25:36 crc kubenswrapper[4764]: I0202 09:25:36.382000 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-b47bfbff6-m265b"] Feb 02 09:25:36 crc kubenswrapper[4764]: I0202 09:25:36.393011 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-64d8599669-wbmlc"] Feb 02 09:25:36 crc kubenswrapper[4764]: I0202 09:25:36.408979 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-69b69cc889-m4fck"] Feb 02 09:25:36 crc kubenswrapper[4764]: I0202 09:25:36.448321 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/45891ac1-d70d-4c3d-af07-6a65855a9281-credential-keys\") pod \"keystone-b47bfbff6-m265b\" (UID: \"45891ac1-d70d-4c3d-af07-6a65855a9281\") " pod="openstack/keystone-b47bfbff6-m265b" Feb 02 09:25:36 crc kubenswrapper[4764]: I0202 09:25:36.448387 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45891ac1-d70d-4c3d-af07-6a65855a9281-combined-ca-bundle\") pod \"keystone-b47bfbff6-m265b\" (UID: \"45891ac1-d70d-4c3d-af07-6a65855a9281\") " pod="openstack/keystone-b47bfbff6-m265b" Feb 02 09:25:36 crc kubenswrapper[4764]: I0202 09:25:36.448410 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/45891ac1-d70d-4c3d-af07-6a65855a9281-internal-tls-certs\") pod \"keystone-b47bfbff6-m265b\" (UID: \"45891ac1-d70d-4c3d-af07-6a65855a9281\") " pod="openstack/keystone-b47bfbff6-m265b" Feb 02 09:25:36 crc kubenswrapper[4764]: I0202 09:25:36.448440 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/45891ac1-d70d-4c3d-af07-6a65855a9281-fernet-keys\") pod \"keystone-b47bfbff6-m265b\" (UID: \"45891ac1-d70d-4c3d-af07-6a65855a9281\") " pod="openstack/keystone-b47bfbff6-m265b" Feb 02 09:25:36 crc kubenswrapper[4764]: I0202 09:25:36.448462 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5rfsd\" (UniqueName: \"kubernetes.io/projected/45891ac1-d70d-4c3d-af07-6a65855a9281-kube-api-access-5rfsd\") pod \"keystone-b47bfbff6-m265b\" (UID: \"45891ac1-d70d-4c3d-af07-6a65855a9281\") " pod="openstack/keystone-b47bfbff6-m265b" Feb 02 09:25:36 crc kubenswrapper[4764]: I0202 09:25:36.448495 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/45891ac1-d70d-4c3d-af07-6a65855a9281-public-tls-certs\") pod \"keystone-b47bfbff6-m265b\" (UID: \"45891ac1-d70d-4c3d-af07-6a65855a9281\") " pod="openstack/keystone-b47bfbff6-m265b" Feb 02 09:25:36 crc kubenswrapper[4764]: I0202 09:25:36.448522 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/45891ac1-d70d-4c3d-af07-6a65855a9281-scripts\") pod \"keystone-b47bfbff6-m265b\" (UID: \"45891ac1-d70d-4c3d-af07-6a65855a9281\") " pod="openstack/keystone-b47bfbff6-m265b" Feb 02 09:25:36 crc kubenswrapper[4764]: I0202 09:25:36.448567 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/45891ac1-d70d-4c3d-af07-6a65855a9281-config-data\") pod \"keystone-b47bfbff6-m265b\" (UID: \"45891ac1-d70d-4c3d-af07-6a65855a9281\") " pod="openstack/keystone-b47bfbff6-m265b" Feb 02 09:25:36 crc kubenswrapper[4764]: I0202 09:25:36.551011 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/45891ac1-d70d-4c3d-af07-6a65855a9281-credential-keys\") pod \"keystone-b47bfbff6-m265b\" (UID: \"45891ac1-d70d-4c3d-af07-6a65855a9281\") " pod="openstack/keystone-b47bfbff6-m265b" Feb 02 09:25:36 crc kubenswrapper[4764]: I0202 09:25:36.551073 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45891ac1-d70d-4c3d-af07-6a65855a9281-combined-ca-bundle\") pod \"keystone-b47bfbff6-m265b\" (UID: \"45891ac1-d70d-4c3d-af07-6a65855a9281\") " pod="openstack/keystone-b47bfbff6-m265b" Feb 02 09:25:36 crc kubenswrapper[4764]: I0202 09:25:36.551093 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/45891ac1-d70d-4c3d-af07-6a65855a9281-internal-tls-certs\") pod \"keystone-b47bfbff6-m265b\" (UID: \"45891ac1-d70d-4c3d-af07-6a65855a9281\") " pod="openstack/keystone-b47bfbff6-m265b" Feb 02 09:25:36 crc kubenswrapper[4764]: I0202 09:25:36.551121 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/45891ac1-d70d-4c3d-af07-6a65855a9281-fernet-keys\") pod \"keystone-b47bfbff6-m265b\" (UID: \"45891ac1-d70d-4c3d-af07-6a65855a9281\") " pod="openstack/keystone-b47bfbff6-m265b" Feb 02 09:25:36 crc kubenswrapper[4764]: I0202 09:25:36.551155 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5rfsd\" (UniqueName: \"kubernetes.io/projected/45891ac1-d70d-4c3d-af07-6a65855a9281-kube-api-access-5rfsd\") pod \"keystone-b47bfbff6-m265b\" (UID: \"45891ac1-d70d-4c3d-af07-6a65855a9281\") " pod="openstack/keystone-b47bfbff6-m265b" Feb 02 09:25:36 crc kubenswrapper[4764]: I0202 09:25:36.551190 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/45891ac1-d70d-4c3d-af07-6a65855a9281-public-tls-certs\") pod \"keystone-b47bfbff6-m265b\" (UID: \"45891ac1-d70d-4c3d-af07-6a65855a9281\") " pod="openstack/keystone-b47bfbff6-m265b" Feb 02 09:25:36 crc kubenswrapper[4764]: I0202 09:25:36.551217 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/45891ac1-d70d-4c3d-af07-6a65855a9281-scripts\") pod \"keystone-b47bfbff6-m265b\" (UID: \"45891ac1-d70d-4c3d-af07-6a65855a9281\") " pod="openstack/keystone-b47bfbff6-m265b" Feb 02 09:25:36 crc kubenswrapper[4764]: I0202 09:25:36.551255 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/45891ac1-d70d-4c3d-af07-6a65855a9281-config-data\") pod \"keystone-b47bfbff6-m265b\" (UID: \"45891ac1-d70d-4c3d-af07-6a65855a9281\") " pod="openstack/keystone-b47bfbff6-m265b" Feb 02 09:25:36 crc kubenswrapper[4764]: I0202 09:25:36.569715 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45891ac1-d70d-4c3d-af07-6a65855a9281-combined-ca-bundle\") pod \"keystone-b47bfbff6-m265b\" (UID: \"45891ac1-d70d-4c3d-af07-6a65855a9281\") " pod="openstack/keystone-b47bfbff6-m265b" Feb 02 09:25:36 crc kubenswrapper[4764]: I0202 09:25:36.575370 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/45891ac1-d70d-4c3d-af07-6a65855a9281-scripts\") pod \"keystone-b47bfbff6-m265b\" (UID: \"45891ac1-d70d-4c3d-af07-6a65855a9281\") " pod="openstack/keystone-b47bfbff6-m265b" Feb 02 09:25:36 crc kubenswrapper[4764]: I0202 09:25:36.577609 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/45891ac1-d70d-4c3d-af07-6a65855a9281-fernet-keys\") pod \"keystone-b47bfbff6-m265b\" (UID: \"45891ac1-d70d-4c3d-af07-6a65855a9281\") " pod="openstack/keystone-b47bfbff6-m265b" Feb 02 09:25:36 crc kubenswrapper[4764]: I0202 09:25:36.581533 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/45891ac1-d70d-4c3d-af07-6a65855a9281-config-data\") pod \"keystone-b47bfbff6-m265b\" (UID: \"45891ac1-d70d-4c3d-af07-6a65855a9281\") " pod="openstack/keystone-b47bfbff6-m265b" Feb 02 09:25:36 crc kubenswrapper[4764]: I0202 09:25:36.581854 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/45891ac1-d70d-4c3d-af07-6a65855a9281-internal-tls-certs\") pod \"keystone-b47bfbff6-m265b\" (UID: \"45891ac1-d70d-4c3d-af07-6a65855a9281\") " pod="openstack/keystone-b47bfbff6-m265b" Feb 02 09:25:36 crc kubenswrapper[4764]: I0202 09:25:36.616421 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/45891ac1-d70d-4c3d-af07-6a65855a9281-credential-keys\") pod \"keystone-b47bfbff6-m265b\" (UID: \"45891ac1-d70d-4c3d-af07-6a65855a9281\") " pod="openstack/keystone-b47bfbff6-m265b" Feb 02 09:25:36 crc kubenswrapper[4764]: I0202 09:25:36.646515 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/45891ac1-d70d-4c3d-af07-6a65855a9281-public-tls-certs\") pod \"keystone-b47bfbff6-m265b\" (UID: \"45891ac1-d70d-4c3d-af07-6a65855a9281\") " pod="openstack/keystone-b47bfbff6-m265b" Feb 02 09:25:36 crc kubenswrapper[4764]: I0202 09:25:36.648670 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5rfsd\" (UniqueName: \"kubernetes.io/projected/45891ac1-d70d-4c3d-af07-6a65855a9281-kube-api-access-5rfsd\") pod \"keystone-b47bfbff6-m265b\" (UID: \"45891ac1-d70d-4c3d-af07-6a65855a9281\") " pod="openstack/keystone-b47bfbff6-m265b" Feb 02 09:25:36 crc kubenswrapper[4764]: I0202 09:25:36.906962 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-b47bfbff6-m265b" Feb 02 09:25:36 crc kubenswrapper[4764]: I0202 09:25:36.997499 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-7b6fbdbb96-fgr8m"] Feb 02 09:25:37 crc kubenswrapper[4764]: I0202 09:25:37.330744 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-dkrnx" event={"ID":"6b3d3314-2f16-4dc3-aa52-5f7f8576813a","Type":"ContainerStarted","Data":"3a7e2521eea6d5fe7739f7c62d9ec21dc8b57406acc029a0ee0d71a25fd330b2"} Feb 02 09:25:37 crc kubenswrapper[4764]: I0202 09:25:37.344576 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-65fcd88786-dggfp" event={"ID":"dedd7b1f-7ec4-4a7a-a703-7e228f7767b1","Type":"ContainerStarted","Data":"990d29535b531c195dc668ed0f9c16b9592d389ae625fafcb51d55c1040391d2"} Feb 02 09:25:37 crc kubenswrapper[4764]: I0202 09:25:37.355806 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7c55c99774-xk9j7" event={"ID":"075a47f1-9307-494f-8268-c544497cf25b","Type":"ContainerStarted","Data":"5d0b70e768483117b47f431cc740738a434e093686f250c01880a75ce4848ada"} Feb 02 09:25:37 crc kubenswrapper[4764]: I0202 09:25:37.355951 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-dkrnx" podStartSLOduration=3.374577863 podStartE2EDuration="39.355919799s" podCreationTimestamp="2026-02-02 09:24:58 +0000 UTC" firstStartedPulling="2026-02-02 09:24:59.211796759 +0000 UTC m=+1062.145520837" lastFinishedPulling="2026-02-02 09:25:35.193138695 +0000 UTC m=+1098.126862773" observedRunningTime="2026-02-02 09:25:37.353556606 +0000 UTC m=+1100.287280694" watchObservedRunningTime="2026-02-02 09:25:37.355919799 +0000 UTC m=+1100.289643887" Feb 02 09:25:37 crc kubenswrapper[4764]: I0202 09:25:37.376200 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7b6fbdbb96-fgr8m" event={"ID":"40e640ad-6fb1-4c56-9b33-14fe28103b07","Type":"ContainerStarted","Data":"d67e68a11ee8449b0e97f4f27db9fd2f842d895b18942e639dfc4dc74232b8c0"} Feb 02 09:25:37 crc kubenswrapper[4764]: I0202 09:25:37.380886 4764 generic.go:334] "Generic (PLEG): container finished" podID="64a8ff87-c26b-49c9-9a34-0b1833b6b7a5" containerID="b516111137de647525a54a53b73b341f4425b981f9a2c917699bda63892f062c" exitCode=0 Feb 02 09:25:37 crc kubenswrapper[4764]: I0202 09:25:37.380989 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74cf444645-8gljr" event={"ID":"64a8ff87-c26b-49c9-9a34-0b1833b6b7a5","Type":"ContainerDied","Data":"b516111137de647525a54a53b73b341f4425b981f9a2c917699bda63892f062c"} Feb 02 09:25:37 crc kubenswrapper[4764]: I0202 09:25:37.397372 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-69b69cc889-m4fck" event={"ID":"364d30e9-d4ef-4c84-b82f-673b41ffe9ff","Type":"ContainerStarted","Data":"46ac0362dd33241a49276203ed445721439244f4f312c3c5389ac4126d87b43f"} Feb 02 09:25:37 crc kubenswrapper[4764]: I0202 09:25:37.397409 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-69b69cc889-m4fck" event={"ID":"364d30e9-d4ef-4c84-b82f-673b41ffe9ff","Type":"ContainerStarted","Data":"edb74c3a670b8b4d80206645680542a348eec13833c041852e92d949d0bffb54"} Feb 02 09:25:37 crc kubenswrapper[4764]: I0202 09:25:37.636274 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-b47bfbff6-m265b"] Feb 02 09:25:37 crc kubenswrapper[4764]: I0202 09:25:37.843143 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0482366b-96dd-493a-98fb-c9cc90254d1e" path="/var/lib/kubelet/pods/0482366b-96dd-493a-98fb-c9cc90254d1e/volumes" Feb 02 09:25:38 crc kubenswrapper[4764]: I0202 09:25:38.457234 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7c55c99774-xk9j7" event={"ID":"075a47f1-9307-494f-8268-c544497cf25b","Type":"ContainerStarted","Data":"46fbf04fb098db11dba046c564e34e636b9baab622642a8e64fd55a7fa860a7e"} Feb 02 09:25:38 crc kubenswrapper[4764]: I0202 09:25:38.457510 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-7c55c99774-xk9j7" Feb 02 09:25:38 crc kubenswrapper[4764]: I0202 09:25:38.457531 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-7c55c99774-xk9j7" Feb 02 09:25:38 crc kubenswrapper[4764]: I0202 09:25:38.461444 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7b6fbdbb96-fgr8m" event={"ID":"40e640ad-6fb1-4c56-9b33-14fe28103b07","Type":"ContainerStarted","Data":"23ab8753e3ef9d74bf5c41dc1f2f62af8d159606e3323372c877898d35fe5001"} Feb 02 09:25:38 crc kubenswrapper[4764]: I0202 09:25:38.461478 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7b6fbdbb96-fgr8m" event={"ID":"40e640ad-6fb1-4c56-9b33-14fe28103b07","Type":"ContainerStarted","Data":"dc70e7817c8136c324242956b2cb32f6723b34276a53f37132eb9a7cf1641096"} Feb 02 09:25:38 crc kubenswrapper[4764]: I0202 09:25:38.461530 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-7b6fbdbb96-fgr8m" Feb 02 09:25:38 crc kubenswrapper[4764]: I0202 09:25:38.461541 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-7b6fbdbb96-fgr8m" Feb 02 09:25:38 crc kubenswrapper[4764]: I0202 09:25:38.469592 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74cf444645-8gljr" event={"ID":"64a8ff87-c26b-49c9-9a34-0b1833b6b7a5","Type":"ContainerStarted","Data":"28b2d3a2705c93f075ce7d5742d6e8515c70e94f15bbe822bd302c82836394ae"} Feb 02 09:25:38 crc kubenswrapper[4764]: I0202 09:25:38.470341 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-74cf444645-8gljr" Feb 02 09:25:38 crc kubenswrapper[4764]: I0202 09:25:38.491605 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-69b69cc889-m4fck" event={"ID":"364d30e9-d4ef-4c84-b82f-673b41ffe9ff","Type":"ContainerStarted","Data":"502205bc8cf64e06eabe340873459163683f16c8530ea72161ca0adfc975501a"} Feb 02 09:25:38 crc kubenswrapper[4764]: I0202 09:25:38.492123 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-69b69cc889-m4fck" Feb 02 09:25:38 crc kubenswrapper[4764]: I0202 09:25:38.512040 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-b47bfbff6-m265b" event={"ID":"45891ac1-d70d-4c3d-af07-6a65855a9281","Type":"ContainerStarted","Data":"37b6007aed5a5396507f84d1c9130b41d9a3942f030553c3417851a17bee3876"} Feb 02 09:25:38 crc kubenswrapper[4764]: I0202 09:25:38.512080 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-b47bfbff6-m265b" event={"ID":"45891ac1-d70d-4c3d-af07-6a65855a9281","Type":"ContainerStarted","Data":"686a68f6534218e40bdf23d4f2db8cbfbbd3d4a06fe201d08b355fc3a06e29c8"} Feb 02 09:25:38 crc kubenswrapper[4764]: I0202 09:25:38.512342 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-b47bfbff6-m265b" Feb 02 09:25:38 crc kubenswrapper[4764]: I0202 09:25:38.520685 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-65fcd88786-dggfp" event={"ID":"dedd7b1f-7ec4-4a7a-a703-7e228f7767b1","Type":"ContainerStarted","Data":"a3f865f4635480487ff622f30424f2b9a0ebc2e202ceb9b00303e7ebcdad3716"} Feb 02 09:25:38 crc kubenswrapper[4764]: I0202 09:25:38.521585 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-65fcd88786-dggfp" Feb 02 09:25:38 crc kubenswrapper[4764]: I0202 09:25:38.521613 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-65fcd88786-dggfp" Feb 02 09:25:38 crc kubenswrapper[4764]: I0202 09:25:38.524280 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-7c55c99774-xk9j7" podStartSLOduration=5.524257871 podStartE2EDuration="5.524257871s" podCreationTimestamp="2026-02-02 09:25:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:25:38.488604889 +0000 UTC m=+1101.422328977" watchObservedRunningTime="2026-02-02 09:25:38.524257871 +0000 UTC m=+1101.457981959" Feb 02 09:25:38 crc kubenswrapper[4764]: I0202 09:25:38.529769 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-7b6fbdbb96-fgr8m" podStartSLOduration=3.52974723 podStartE2EDuration="3.52974723s" podCreationTimestamp="2026-02-02 09:25:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:25:38.517183218 +0000 UTC m=+1101.450907306" watchObservedRunningTime="2026-02-02 09:25:38.52974723 +0000 UTC m=+1101.463471318" Feb 02 09:25:38 crc kubenswrapper[4764]: I0202 09:25:38.541146 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-74cf444645-8gljr" podStartSLOduration=5.54113211 podStartE2EDuration="5.54113211s" podCreationTimestamp="2026-02-02 09:25:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:25:38.540409121 +0000 UTC m=+1101.474133209" watchObservedRunningTime="2026-02-02 09:25:38.54113211 +0000 UTC m=+1101.474856198" Feb 02 09:25:38 crc kubenswrapper[4764]: I0202 09:25:38.566763 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-65fcd88786-dggfp" podStartSLOduration=6.5667492979999995 podStartE2EDuration="6.566749298s" podCreationTimestamp="2026-02-02 09:25:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:25:38.564083056 +0000 UTC m=+1101.497807144" watchObservedRunningTime="2026-02-02 09:25:38.566749298 +0000 UTC m=+1101.500473386" Feb 02 09:25:38 crc kubenswrapper[4764]: I0202 09:25:38.605822 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-b47bfbff6-m265b" podStartSLOduration=2.605803193 podStartE2EDuration="2.605803193s" podCreationTimestamp="2026-02-02 09:25:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:25:38.582151918 +0000 UTC m=+1101.515876006" watchObservedRunningTime="2026-02-02 09:25:38.605803193 +0000 UTC m=+1101.539527281" Feb 02 09:25:38 crc kubenswrapper[4764]: I0202 09:25:38.621025 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-69b69cc889-m4fck" podStartSLOduration=12.621008207 podStartE2EDuration="12.621008207s" podCreationTimestamp="2026-02-02 09:25:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:25:38.603649604 +0000 UTC m=+1101.537373692" watchObservedRunningTime="2026-02-02 09:25:38.621008207 +0000 UTC m=+1101.554732295" Feb 02 09:25:40 crc kubenswrapper[4764]: I0202 09:25:40.541042 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-6d6c6fd44b-2hfzf" event={"ID":"4f12af32-00c9-437a-8e0e-805caa8df878","Type":"ContainerStarted","Data":"c9e3885c81fca81f2a481c177ff87282be988d2fe765f222b6301ad7278e1331"} Feb 02 09:25:40 crc kubenswrapper[4764]: I0202 09:25:40.546871 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-6db5f6d5bf-pxtw5" event={"ID":"4ebb36bb-8032-42ea-8954-7716e5a81301","Type":"ContainerStarted","Data":"2e2b3632dba3173bf5217987f806dceddb390cb12830850b19e2e1784e431682"} Feb 02 09:25:41 crc kubenswrapper[4764]: I0202 09:25:41.564361 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-6d6c6fd44b-2hfzf" event={"ID":"4f12af32-00c9-437a-8e0e-805caa8df878","Type":"ContainerStarted","Data":"90aa0957e88cd020632dd7468ddffe7b8338bced2a1080b5399b0538a9434ff6"} Feb 02 09:25:41 crc kubenswrapper[4764]: I0202 09:25:41.567395 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-6db5f6d5bf-pxtw5" event={"ID":"4ebb36bb-8032-42ea-8954-7716e5a81301","Type":"ContainerStarted","Data":"530aca16f5dc00ab072aeb59ad0d8b2d8bea184774c6c58f4786ffc31305fe78"} Feb 02 09:25:41 crc kubenswrapper[4764]: I0202 09:25:41.596909 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-6d6c6fd44b-2hfzf" podStartSLOduration=5.613381832 podStartE2EDuration="9.596887495s" podCreationTimestamp="2026-02-02 09:25:32 +0000 UTC" firstStartedPulling="2026-02-02 09:25:36.179328501 +0000 UTC m=+1099.113052589" lastFinishedPulling="2026-02-02 09:25:40.162834174 +0000 UTC m=+1103.096558252" observedRunningTime="2026-02-02 09:25:41.587746655 +0000 UTC m=+1104.521470753" watchObservedRunningTime="2026-02-02 09:25:41.596887495 +0000 UTC m=+1104.530611583" Feb 02 09:25:41 crc kubenswrapper[4764]: I0202 09:25:41.605299 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-6db5f6d5bf-pxtw5" podStartSLOduration=5.509663111 podStartE2EDuration="9.605282053s" podCreationTimestamp="2026-02-02 09:25:32 +0000 UTC" firstStartedPulling="2026-02-02 09:25:36.063811459 +0000 UTC m=+1098.997535547" lastFinishedPulling="2026-02-02 09:25:40.159430401 +0000 UTC m=+1103.093154489" observedRunningTime="2026-02-02 09:25:41.601958993 +0000 UTC m=+1104.535683091" watchObservedRunningTime="2026-02-02 09:25:41.605282053 +0000 UTC m=+1104.539006141" Feb 02 09:25:41 crc kubenswrapper[4764]: I0202 09:25:41.752256 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-5686dc79cc-sb69r"] Feb 02 09:25:41 crc kubenswrapper[4764]: I0202 09:25:41.753786 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-5686dc79cc-sb69r" Feb 02 09:25:41 crc kubenswrapper[4764]: I0202 09:25:41.766890 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-855c89b68b-ptbcl"] Feb 02 09:25:41 crc kubenswrapper[4764]: I0202 09:25:41.768230 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-855c89b68b-ptbcl" Feb 02 09:25:41 crc kubenswrapper[4764]: I0202 09:25:41.797224 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-855c89b68b-ptbcl"] Feb 02 09:25:41 crc kubenswrapper[4764]: I0202 09:25:41.842760 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-5686dc79cc-sb69r"] Feb 02 09:25:41 crc kubenswrapper[4764]: I0202 09:25:41.864809 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-7c55c99774-xk9j7"] Feb 02 09:25:41 crc kubenswrapper[4764]: I0202 09:25:41.865070 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-7c55c99774-xk9j7" podUID="075a47f1-9307-494f-8268-c544497cf25b" containerName="barbican-api-log" containerID="cri-o://5d0b70e768483117b47f431cc740738a434e093686f250c01880a75ce4848ada" gracePeriod=30 Feb 02 09:25:41 crc kubenswrapper[4764]: I0202 09:25:41.865741 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-7c55c99774-xk9j7" podUID="075a47f1-9307-494f-8268-c544497cf25b" containerName="barbican-api" containerID="cri-o://46fbf04fb098db11dba046c564e34e636b9baab622642a8e64fd55a7fa860a7e" gracePeriod=30 Feb 02 09:25:41 crc kubenswrapper[4764]: I0202 09:25:41.885870 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4e3b3433-3d64-44ce-9a3a-14fcd0bdaa6d-config-data-custom\") pod \"barbican-worker-5686dc79cc-sb69r\" (UID: \"4e3b3433-3d64-44ce-9a3a-14fcd0bdaa6d\") " pod="openstack/barbican-worker-5686dc79cc-sb69r" Feb 02 09:25:41 crc kubenswrapper[4764]: I0202 09:25:41.886004 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e3b3433-3d64-44ce-9a3a-14fcd0bdaa6d-config-data\") pod \"barbican-worker-5686dc79cc-sb69r\" (UID: \"4e3b3433-3d64-44ce-9a3a-14fcd0bdaa6d\") " pod="openstack/barbican-worker-5686dc79cc-sb69r" Feb 02 09:25:41 crc kubenswrapper[4764]: I0202 09:25:41.886052 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7vfjl\" (UniqueName: \"kubernetes.io/projected/4e3b3433-3d64-44ce-9a3a-14fcd0bdaa6d-kube-api-access-7vfjl\") pod \"barbican-worker-5686dc79cc-sb69r\" (UID: \"4e3b3433-3d64-44ce-9a3a-14fcd0bdaa6d\") " pod="openstack/barbican-worker-5686dc79cc-sb69r" Feb 02 09:25:41 crc kubenswrapper[4764]: I0202 09:25:41.886085 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ln6h5\" (UniqueName: \"kubernetes.io/projected/b004a188-196b-4812-a9e6-e398062cf4d3-kube-api-access-ln6h5\") pod \"barbican-keystone-listener-855c89b68b-ptbcl\" (UID: \"b004a188-196b-4812-a9e6-e398062cf4d3\") " pod="openstack/barbican-keystone-listener-855c89b68b-ptbcl" Feb 02 09:25:41 crc kubenswrapper[4764]: I0202 09:25:41.886100 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b004a188-196b-4812-a9e6-e398062cf4d3-combined-ca-bundle\") pod \"barbican-keystone-listener-855c89b68b-ptbcl\" (UID: \"b004a188-196b-4812-a9e6-e398062cf4d3\") " pod="openstack/barbican-keystone-listener-855c89b68b-ptbcl" Feb 02 09:25:41 crc kubenswrapper[4764]: I0202 09:25:41.886124 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e3b3433-3d64-44ce-9a3a-14fcd0bdaa6d-combined-ca-bundle\") pod \"barbican-worker-5686dc79cc-sb69r\" (UID: \"4e3b3433-3d64-44ce-9a3a-14fcd0bdaa6d\") " pod="openstack/barbican-worker-5686dc79cc-sb69r" Feb 02 09:25:41 crc kubenswrapper[4764]: I0202 09:25:41.886153 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b004a188-196b-4812-a9e6-e398062cf4d3-logs\") pod \"barbican-keystone-listener-855c89b68b-ptbcl\" (UID: \"b004a188-196b-4812-a9e6-e398062cf4d3\") " pod="openstack/barbican-keystone-listener-855c89b68b-ptbcl" Feb 02 09:25:41 crc kubenswrapper[4764]: I0202 09:25:41.886172 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4e3b3433-3d64-44ce-9a3a-14fcd0bdaa6d-logs\") pod \"barbican-worker-5686dc79cc-sb69r\" (UID: \"4e3b3433-3d64-44ce-9a3a-14fcd0bdaa6d\") " pod="openstack/barbican-worker-5686dc79cc-sb69r" Feb 02 09:25:41 crc kubenswrapper[4764]: I0202 09:25:41.886199 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b004a188-196b-4812-a9e6-e398062cf4d3-config-data-custom\") pod \"barbican-keystone-listener-855c89b68b-ptbcl\" (UID: \"b004a188-196b-4812-a9e6-e398062cf4d3\") " pod="openstack/barbican-keystone-listener-855c89b68b-ptbcl" Feb 02 09:25:41 crc kubenswrapper[4764]: I0202 09:25:41.886219 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b004a188-196b-4812-a9e6-e398062cf4d3-config-data\") pod \"barbican-keystone-listener-855c89b68b-ptbcl\" (UID: \"b004a188-196b-4812-a9e6-e398062cf4d3\") " pod="openstack/barbican-keystone-listener-855c89b68b-ptbcl" Feb 02 09:25:41 crc kubenswrapper[4764]: I0202 09:25:41.905812 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-797597c45b-9mkpp"] Feb 02 09:25:41 crc kubenswrapper[4764]: I0202 09:25:41.907149 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-797597c45b-9mkpp" Feb 02 09:25:41 crc kubenswrapper[4764]: I0202 09:25:41.918282 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-797597c45b-9mkpp"] Feb 02 09:25:41 crc kubenswrapper[4764]: I0202 09:25:41.987812 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e3b3433-3d64-44ce-9a3a-14fcd0bdaa6d-config-data\") pod \"barbican-worker-5686dc79cc-sb69r\" (UID: \"4e3b3433-3d64-44ce-9a3a-14fcd0bdaa6d\") " pod="openstack/barbican-worker-5686dc79cc-sb69r" Feb 02 09:25:41 crc kubenswrapper[4764]: I0202 09:25:41.987888 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f06f976d-5e29-4e78-81f4-81de01ae8391-config-data-custom\") pod \"barbican-api-797597c45b-9mkpp\" (UID: \"f06f976d-5e29-4e78-81f4-81de01ae8391\") " pod="openstack/barbican-api-797597c45b-9mkpp" Feb 02 09:25:41 crc kubenswrapper[4764]: I0202 09:25:41.987920 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7vfjl\" (UniqueName: \"kubernetes.io/projected/4e3b3433-3d64-44ce-9a3a-14fcd0bdaa6d-kube-api-access-7vfjl\") pod \"barbican-worker-5686dc79cc-sb69r\" (UID: \"4e3b3433-3d64-44ce-9a3a-14fcd0bdaa6d\") " pod="openstack/barbican-worker-5686dc79cc-sb69r" Feb 02 09:25:41 crc kubenswrapper[4764]: I0202 09:25:41.987986 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ln6h5\" (UniqueName: \"kubernetes.io/projected/b004a188-196b-4812-a9e6-e398062cf4d3-kube-api-access-ln6h5\") pod \"barbican-keystone-listener-855c89b68b-ptbcl\" (UID: \"b004a188-196b-4812-a9e6-e398062cf4d3\") " pod="openstack/barbican-keystone-listener-855c89b68b-ptbcl" Feb 02 09:25:41 crc kubenswrapper[4764]: I0202 09:25:41.988007 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b004a188-196b-4812-a9e6-e398062cf4d3-combined-ca-bundle\") pod \"barbican-keystone-listener-855c89b68b-ptbcl\" (UID: \"b004a188-196b-4812-a9e6-e398062cf4d3\") " pod="openstack/barbican-keystone-listener-855c89b68b-ptbcl" Feb 02 09:25:41 crc kubenswrapper[4764]: I0202 09:25:41.988051 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f06f976d-5e29-4e78-81f4-81de01ae8391-public-tls-certs\") pod \"barbican-api-797597c45b-9mkpp\" (UID: \"f06f976d-5e29-4e78-81f4-81de01ae8391\") " pod="openstack/barbican-api-797597c45b-9mkpp" Feb 02 09:25:41 crc kubenswrapper[4764]: I0202 09:25:41.988077 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f06f976d-5e29-4e78-81f4-81de01ae8391-logs\") pod \"barbican-api-797597c45b-9mkpp\" (UID: \"f06f976d-5e29-4e78-81f4-81de01ae8391\") " pod="openstack/barbican-api-797597c45b-9mkpp" Feb 02 09:25:41 crc kubenswrapper[4764]: I0202 09:25:41.988113 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e3b3433-3d64-44ce-9a3a-14fcd0bdaa6d-combined-ca-bundle\") pod \"barbican-worker-5686dc79cc-sb69r\" (UID: \"4e3b3433-3d64-44ce-9a3a-14fcd0bdaa6d\") " pod="openstack/barbican-worker-5686dc79cc-sb69r" Feb 02 09:25:41 crc kubenswrapper[4764]: I0202 09:25:41.988132 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f06f976d-5e29-4e78-81f4-81de01ae8391-config-data\") pod \"barbican-api-797597c45b-9mkpp\" (UID: \"f06f976d-5e29-4e78-81f4-81de01ae8391\") " pod="openstack/barbican-api-797597c45b-9mkpp" Feb 02 09:25:41 crc kubenswrapper[4764]: I0202 09:25:41.988149 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2t7pz\" (UniqueName: \"kubernetes.io/projected/f06f976d-5e29-4e78-81f4-81de01ae8391-kube-api-access-2t7pz\") pod \"barbican-api-797597c45b-9mkpp\" (UID: \"f06f976d-5e29-4e78-81f4-81de01ae8391\") " pod="openstack/barbican-api-797597c45b-9mkpp" Feb 02 09:25:41 crc kubenswrapper[4764]: I0202 09:25:41.988191 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b004a188-196b-4812-a9e6-e398062cf4d3-logs\") pod \"barbican-keystone-listener-855c89b68b-ptbcl\" (UID: \"b004a188-196b-4812-a9e6-e398062cf4d3\") " pod="openstack/barbican-keystone-listener-855c89b68b-ptbcl" Feb 02 09:25:41 crc kubenswrapper[4764]: I0202 09:25:41.988222 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4e3b3433-3d64-44ce-9a3a-14fcd0bdaa6d-logs\") pod \"barbican-worker-5686dc79cc-sb69r\" (UID: \"4e3b3433-3d64-44ce-9a3a-14fcd0bdaa6d\") " pod="openstack/barbican-worker-5686dc79cc-sb69r" Feb 02 09:25:41 crc kubenswrapper[4764]: I0202 09:25:41.988289 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b004a188-196b-4812-a9e6-e398062cf4d3-config-data-custom\") pod \"barbican-keystone-listener-855c89b68b-ptbcl\" (UID: \"b004a188-196b-4812-a9e6-e398062cf4d3\") " pod="openstack/barbican-keystone-listener-855c89b68b-ptbcl" Feb 02 09:25:41 crc kubenswrapper[4764]: I0202 09:25:41.988311 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b004a188-196b-4812-a9e6-e398062cf4d3-config-data\") pod \"barbican-keystone-listener-855c89b68b-ptbcl\" (UID: \"b004a188-196b-4812-a9e6-e398062cf4d3\") " pod="openstack/barbican-keystone-listener-855c89b68b-ptbcl" Feb 02 09:25:41 crc kubenswrapper[4764]: I0202 09:25:41.988361 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f06f976d-5e29-4e78-81f4-81de01ae8391-combined-ca-bundle\") pod \"barbican-api-797597c45b-9mkpp\" (UID: \"f06f976d-5e29-4e78-81f4-81de01ae8391\") " pod="openstack/barbican-api-797597c45b-9mkpp" Feb 02 09:25:41 crc kubenswrapper[4764]: I0202 09:25:41.988380 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4e3b3433-3d64-44ce-9a3a-14fcd0bdaa6d-config-data-custom\") pod \"barbican-worker-5686dc79cc-sb69r\" (UID: \"4e3b3433-3d64-44ce-9a3a-14fcd0bdaa6d\") " pod="openstack/barbican-worker-5686dc79cc-sb69r" Feb 02 09:25:41 crc kubenswrapper[4764]: I0202 09:25:41.988405 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f06f976d-5e29-4e78-81f4-81de01ae8391-internal-tls-certs\") pod \"barbican-api-797597c45b-9mkpp\" (UID: \"f06f976d-5e29-4e78-81f4-81de01ae8391\") " pod="openstack/barbican-api-797597c45b-9mkpp" Feb 02 09:25:41 crc kubenswrapper[4764]: I0202 09:25:41.992392 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b004a188-196b-4812-a9e6-e398062cf4d3-logs\") pod \"barbican-keystone-listener-855c89b68b-ptbcl\" (UID: \"b004a188-196b-4812-a9e6-e398062cf4d3\") " pod="openstack/barbican-keystone-listener-855c89b68b-ptbcl" Feb 02 09:25:41 crc kubenswrapper[4764]: I0202 09:25:41.992842 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4e3b3433-3d64-44ce-9a3a-14fcd0bdaa6d-logs\") pod \"barbican-worker-5686dc79cc-sb69r\" (UID: \"4e3b3433-3d64-44ce-9a3a-14fcd0bdaa6d\") " pod="openstack/barbican-worker-5686dc79cc-sb69r" Feb 02 09:25:42 crc kubenswrapper[4764]: I0202 09:25:42.007192 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e3b3433-3d64-44ce-9a3a-14fcd0bdaa6d-config-data\") pod \"barbican-worker-5686dc79cc-sb69r\" (UID: \"4e3b3433-3d64-44ce-9a3a-14fcd0bdaa6d\") " pod="openstack/barbican-worker-5686dc79cc-sb69r" Feb 02 09:25:42 crc kubenswrapper[4764]: I0202 09:25:42.007506 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e3b3433-3d64-44ce-9a3a-14fcd0bdaa6d-combined-ca-bundle\") pod \"barbican-worker-5686dc79cc-sb69r\" (UID: \"4e3b3433-3d64-44ce-9a3a-14fcd0bdaa6d\") " pod="openstack/barbican-worker-5686dc79cc-sb69r" Feb 02 09:25:42 crc kubenswrapper[4764]: I0202 09:25:42.008871 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b004a188-196b-4812-a9e6-e398062cf4d3-combined-ca-bundle\") pod \"barbican-keystone-listener-855c89b68b-ptbcl\" (UID: \"b004a188-196b-4812-a9e6-e398062cf4d3\") " pod="openstack/barbican-keystone-listener-855c89b68b-ptbcl" Feb 02 09:25:42 crc kubenswrapper[4764]: I0202 09:25:42.009357 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4e3b3433-3d64-44ce-9a3a-14fcd0bdaa6d-config-data-custom\") pod \"barbican-worker-5686dc79cc-sb69r\" (UID: \"4e3b3433-3d64-44ce-9a3a-14fcd0bdaa6d\") " pod="openstack/barbican-worker-5686dc79cc-sb69r" Feb 02 09:25:42 crc kubenswrapper[4764]: I0202 09:25:42.015214 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b004a188-196b-4812-a9e6-e398062cf4d3-config-data-custom\") pod \"barbican-keystone-listener-855c89b68b-ptbcl\" (UID: \"b004a188-196b-4812-a9e6-e398062cf4d3\") " pod="openstack/barbican-keystone-listener-855c89b68b-ptbcl" Feb 02 09:25:42 crc kubenswrapper[4764]: I0202 09:25:42.015920 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b004a188-196b-4812-a9e6-e398062cf4d3-config-data\") pod \"barbican-keystone-listener-855c89b68b-ptbcl\" (UID: \"b004a188-196b-4812-a9e6-e398062cf4d3\") " pod="openstack/barbican-keystone-listener-855c89b68b-ptbcl" Feb 02 09:25:42 crc kubenswrapper[4764]: I0202 09:25:42.023985 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ln6h5\" (UniqueName: \"kubernetes.io/projected/b004a188-196b-4812-a9e6-e398062cf4d3-kube-api-access-ln6h5\") pod \"barbican-keystone-listener-855c89b68b-ptbcl\" (UID: \"b004a188-196b-4812-a9e6-e398062cf4d3\") " pod="openstack/barbican-keystone-listener-855c89b68b-ptbcl" Feb 02 09:25:42 crc kubenswrapper[4764]: I0202 09:25:42.024095 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7vfjl\" (UniqueName: \"kubernetes.io/projected/4e3b3433-3d64-44ce-9a3a-14fcd0bdaa6d-kube-api-access-7vfjl\") pod \"barbican-worker-5686dc79cc-sb69r\" (UID: \"4e3b3433-3d64-44ce-9a3a-14fcd0bdaa6d\") " pod="openstack/barbican-worker-5686dc79cc-sb69r" Feb 02 09:25:42 crc kubenswrapper[4764]: I0202 09:25:42.070453 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-5686dc79cc-sb69r" Feb 02 09:25:42 crc kubenswrapper[4764]: I0202 09:25:42.085298 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-855c89b68b-ptbcl" Feb 02 09:25:42 crc kubenswrapper[4764]: I0202 09:25:42.093884 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f06f976d-5e29-4e78-81f4-81de01ae8391-logs\") pod \"barbican-api-797597c45b-9mkpp\" (UID: \"f06f976d-5e29-4e78-81f4-81de01ae8391\") " pod="openstack/barbican-api-797597c45b-9mkpp" Feb 02 09:25:42 crc kubenswrapper[4764]: I0202 09:25:42.093972 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f06f976d-5e29-4e78-81f4-81de01ae8391-config-data\") pod \"barbican-api-797597c45b-9mkpp\" (UID: \"f06f976d-5e29-4e78-81f4-81de01ae8391\") " pod="openstack/barbican-api-797597c45b-9mkpp" Feb 02 09:25:42 crc kubenswrapper[4764]: I0202 09:25:42.093997 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2t7pz\" (UniqueName: \"kubernetes.io/projected/f06f976d-5e29-4e78-81f4-81de01ae8391-kube-api-access-2t7pz\") pod \"barbican-api-797597c45b-9mkpp\" (UID: \"f06f976d-5e29-4e78-81f4-81de01ae8391\") " pod="openstack/barbican-api-797597c45b-9mkpp" Feb 02 09:25:42 crc kubenswrapper[4764]: I0202 09:25:42.094062 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f06f976d-5e29-4e78-81f4-81de01ae8391-combined-ca-bundle\") pod \"barbican-api-797597c45b-9mkpp\" (UID: \"f06f976d-5e29-4e78-81f4-81de01ae8391\") " pod="openstack/barbican-api-797597c45b-9mkpp" Feb 02 09:25:42 crc kubenswrapper[4764]: I0202 09:25:42.094090 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f06f976d-5e29-4e78-81f4-81de01ae8391-internal-tls-certs\") pod \"barbican-api-797597c45b-9mkpp\" (UID: \"f06f976d-5e29-4e78-81f4-81de01ae8391\") " pod="openstack/barbican-api-797597c45b-9mkpp" Feb 02 09:25:42 crc kubenswrapper[4764]: I0202 09:25:42.094138 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f06f976d-5e29-4e78-81f4-81de01ae8391-config-data-custom\") pod \"barbican-api-797597c45b-9mkpp\" (UID: \"f06f976d-5e29-4e78-81f4-81de01ae8391\") " pod="openstack/barbican-api-797597c45b-9mkpp" Feb 02 09:25:42 crc kubenswrapper[4764]: I0202 09:25:42.094181 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f06f976d-5e29-4e78-81f4-81de01ae8391-public-tls-certs\") pod \"barbican-api-797597c45b-9mkpp\" (UID: \"f06f976d-5e29-4e78-81f4-81de01ae8391\") " pod="openstack/barbican-api-797597c45b-9mkpp" Feb 02 09:25:42 crc kubenswrapper[4764]: I0202 09:25:42.094851 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f06f976d-5e29-4e78-81f4-81de01ae8391-logs\") pod \"barbican-api-797597c45b-9mkpp\" (UID: \"f06f976d-5e29-4e78-81f4-81de01ae8391\") " pod="openstack/barbican-api-797597c45b-9mkpp" Feb 02 09:25:42 crc kubenswrapper[4764]: I0202 09:25:42.100887 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f06f976d-5e29-4e78-81f4-81de01ae8391-combined-ca-bundle\") pod \"barbican-api-797597c45b-9mkpp\" (UID: \"f06f976d-5e29-4e78-81f4-81de01ae8391\") " pod="openstack/barbican-api-797597c45b-9mkpp" Feb 02 09:25:42 crc kubenswrapper[4764]: I0202 09:25:42.105883 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f06f976d-5e29-4e78-81f4-81de01ae8391-config-data-custom\") pod \"barbican-api-797597c45b-9mkpp\" (UID: \"f06f976d-5e29-4e78-81f4-81de01ae8391\") " pod="openstack/barbican-api-797597c45b-9mkpp" Feb 02 09:25:42 crc kubenswrapper[4764]: I0202 09:25:42.106255 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f06f976d-5e29-4e78-81f4-81de01ae8391-internal-tls-certs\") pod \"barbican-api-797597c45b-9mkpp\" (UID: \"f06f976d-5e29-4e78-81f4-81de01ae8391\") " pod="openstack/barbican-api-797597c45b-9mkpp" Feb 02 09:25:42 crc kubenswrapper[4764]: I0202 09:25:42.111988 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f06f976d-5e29-4e78-81f4-81de01ae8391-config-data\") pod \"barbican-api-797597c45b-9mkpp\" (UID: \"f06f976d-5e29-4e78-81f4-81de01ae8391\") " pod="openstack/barbican-api-797597c45b-9mkpp" Feb 02 09:25:42 crc kubenswrapper[4764]: I0202 09:25:42.112403 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f06f976d-5e29-4e78-81f4-81de01ae8391-public-tls-certs\") pod \"barbican-api-797597c45b-9mkpp\" (UID: \"f06f976d-5e29-4e78-81f4-81de01ae8391\") " pod="openstack/barbican-api-797597c45b-9mkpp" Feb 02 09:25:42 crc kubenswrapper[4764]: I0202 09:25:42.113373 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2t7pz\" (UniqueName: \"kubernetes.io/projected/f06f976d-5e29-4e78-81f4-81de01ae8391-kube-api-access-2t7pz\") pod \"barbican-api-797597c45b-9mkpp\" (UID: \"f06f976d-5e29-4e78-81f4-81de01ae8391\") " pod="openstack/barbican-api-797597c45b-9mkpp" Feb 02 09:25:42 crc kubenswrapper[4764]: I0202 09:25:42.174364 4764 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-7c55c99774-xk9j7" podUID="075a47f1-9307-494f-8268-c544497cf25b" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.144:9311/healthcheck\": read tcp 10.217.0.2:47112->10.217.0.144:9311: read: connection reset by peer" Feb 02 09:25:42 crc kubenswrapper[4764]: I0202 09:25:42.261344 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-797597c45b-9mkpp" Feb 02 09:25:42 crc kubenswrapper[4764]: I0202 09:25:42.604293 4764 generic.go:334] "Generic (PLEG): container finished" podID="075a47f1-9307-494f-8268-c544497cf25b" containerID="46fbf04fb098db11dba046c564e34e636b9baab622642a8e64fd55a7fa860a7e" exitCode=0 Feb 02 09:25:42 crc kubenswrapper[4764]: I0202 09:25:42.604591 4764 generic.go:334] "Generic (PLEG): container finished" podID="075a47f1-9307-494f-8268-c544497cf25b" containerID="5d0b70e768483117b47f431cc740738a434e093686f250c01880a75ce4848ada" exitCode=143 Feb 02 09:25:42 crc kubenswrapper[4764]: I0202 09:25:42.604989 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7c55c99774-xk9j7" event={"ID":"075a47f1-9307-494f-8268-c544497cf25b","Type":"ContainerDied","Data":"46fbf04fb098db11dba046c564e34e636b9baab622642a8e64fd55a7fa860a7e"} Feb 02 09:25:42 crc kubenswrapper[4764]: I0202 09:25:42.605081 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7c55c99774-xk9j7" event={"ID":"075a47f1-9307-494f-8268-c544497cf25b","Type":"ContainerDied","Data":"5d0b70e768483117b47f431cc740738a434e093686f250c01880a75ce4848ada"} Feb 02 09:25:42 crc kubenswrapper[4764]: I0202 09:25:42.764824 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-855c89b68b-ptbcl"] Feb 02 09:25:42 crc kubenswrapper[4764]: I0202 09:25:42.948100 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-5686dc79cc-sb69r"] Feb 02 09:25:43 crc kubenswrapper[4764]: I0202 09:25:43.088237 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7c55c99774-xk9j7" Feb 02 09:25:43 crc kubenswrapper[4764]: I0202 09:25:43.131322 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-797597c45b-9mkpp"] Feb 02 09:25:43 crc kubenswrapper[4764]: I0202 09:25:43.239521 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/075a47f1-9307-494f-8268-c544497cf25b-logs\") pod \"075a47f1-9307-494f-8268-c544497cf25b\" (UID: \"075a47f1-9307-494f-8268-c544497cf25b\") " Feb 02 09:25:43 crc kubenswrapper[4764]: I0202 09:25:43.239610 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2nzjv\" (UniqueName: \"kubernetes.io/projected/075a47f1-9307-494f-8268-c544497cf25b-kube-api-access-2nzjv\") pod \"075a47f1-9307-494f-8268-c544497cf25b\" (UID: \"075a47f1-9307-494f-8268-c544497cf25b\") " Feb 02 09:25:43 crc kubenswrapper[4764]: I0202 09:25:43.239646 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/075a47f1-9307-494f-8268-c544497cf25b-config-data-custom\") pod \"075a47f1-9307-494f-8268-c544497cf25b\" (UID: \"075a47f1-9307-494f-8268-c544497cf25b\") " Feb 02 09:25:43 crc kubenswrapper[4764]: I0202 09:25:43.239674 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/075a47f1-9307-494f-8268-c544497cf25b-config-data\") pod \"075a47f1-9307-494f-8268-c544497cf25b\" (UID: \"075a47f1-9307-494f-8268-c544497cf25b\") " Feb 02 09:25:43 crc kubenswrapper[4764]: I0202 09:25:43.239727 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/075a47f1-9307-494f-8268-c544497cf25b-combined-ca-bundle\") pod \"075a47f1-9307-494f-8268-c544497cf25b\" (UID: \"075a47f1-9307-494f-8268-c544497cf25b\") " Feb 02 09:25:43 crc kubenswrapper[4764]: I0202 09:25:43.249582 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/075a47f1-9307-494f-8268-c544497cf25b-logs" (OuterVolumeSpecName: "logs") pod "075a47f1-9307-494f-8268-c544497cf25b" (UID: "075a47f1-9307-494f-8268-c544497cf25b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 09:25:43 crc kubenswrapper[4764]: I0202 09:25:43.257187 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/075a47f1-9307-494f-8268-c544497cf25b-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "075a47f1-9307-494f-8268-c544497cf25b" (UID: "075a47f1-9307-494f-8268-c544497cf25b"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:25:43 crc kubenswrapper[4764]: I0202 09:25:43.259671 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/075a47f1-9307-494f-8268-c544497cf25b-kube-api-access-2nzjv" (OuterVolumeSpecName: "kube-api-access-2nzjv") pod "075a47f1-9307-494f-8268-c544497cf25b" (UID: "075a47f1-9307-494f-8268-c544497cf25b"). InnerVolumeSpecName "kube-api-access-2nzjv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:25:43 crc kubenswrapper[4764]: I0202 09:25:43.296108 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/075a47f1-9307-494f-8268-c544497cf25b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "075a47f1-9307-494f-8268-c544497cf25b" (UID: "075a47f1-9307-494f-8268-c544497cf25b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:25:43 crc kubenswrapper[4764]: I0202 09:25:43.348043 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2nzjv\" (UniqueName: \"kubernetes.io/projected/075a47f1-9307-494f-8268-c544497cf25b-kube-api-access-2nzjv\") on node \"crc\" DevicePath \"\"" Feb 02 09:25:43 crc kubenswrapper[4764]: I0202 09:25:43.348073 4764 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/075a47f1-9307-494f-8268-c544497cf25b-config-data-custom\") on node \"crc\" DevicePath \"\"" Feb 02 09:25:43 crc kubenswrapper[4764]: I0202 09:25:43.348082 4764 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/075a47f1-9307-494f-8268-c544497cf25b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 09:25:43 crc kubenswrapper[4764]: I0202 09:25:43.348091 4764 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/075a47f1-9307-494f-8268-c544497cf25b-logs\") on node \"crc\" DevicePath \"\"" Feb 02 09:25:43 crc kubenswrapper[4764]: I0202 09:25:43.373168 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/075a47f1-9307-494f-8268-c544497cf25b-config-data" (OuterVolumeSpecName: "config-data") pod "075a47f1-9307-494f-8268-c544497cf25b" (UID: "075a47f1-9307-494f-8268-c544497cf25b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:25:43 crc kubenswrapper[4764]: I0202 09:25:43.449993 4764 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/075a47f1-9307-494f-8268-c544497cf25b-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 09:25:43 crc kubenswrapper[4764]: I0202 09:25:43.526626 4764 patch_prober.go:28] interesting pod/machine-config-daemon-4ndm4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 09:25:43 crc kubenswrapper[4764]: I0202 09:25:43.526676 4764 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 09:25:43 crc kubenswrapper[4764]: I0202 09:25:43.591107 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-74cf444645-8gljr" Feb 02 09:25:43 crc kubenswrapper[4764]: I0202 09:25:43.646491 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5686dc79cc-sb69r" event={"ID":"4e3b3433-3d64-44ce-9a3a-14fcd0bdaa6d","Type":"ContainerStarted","Data":"a8c1cdaee4241ef8edd1967f8527d30057cd3cf2dd0a1400d272d28cf898fbe1"} Feb 02 09:25:43 crc kubenswrapper[4764]: I0202 09:25:43.646601 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5686dc79cc-sb69r" event={"ID":"4e3b3433-3d64-44ce-9a3a-14fcd0bdaa6d","Type":"ContainerStarted","Data":"a5582ebf100c2753a353124a2716d2334f4faf07b0571b9de3c9a9e41e74636b"} Feb 02 09:25:43 crc kubenswrapper[4764]: I0202 09:25:43.681219 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5bfb76cfc-8k8x2"] Feb 02 09:25:43 crc kubenswrapper[4764]: I0202 09:25:43.681442 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5bfb76cfc-8k8x2" podUID="1ad03482-854b-4e3a-8ee4-8e4f22b68d3c" containerName="dnsmasq-dns" containerID="cri-o://8114f849c9b5b2c77035ec1db7e306cb25897aa6ff4fabe90927e8dd131746b6" gracePeriod=10 Feb 02 09:25:43 crc kubenswrapper[4764]: I0202 09:25:43.689375 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7c55c99774-xk9j7" event={"ID":"075a47f1-9307-494f-8268-c544497cf25b","Type":"ContainerDied","Data":"d5e01390ca284556900867d6beb52fd76ba6f0d7eb0a2abc10071f834486b312"} Feb 02 09:25:43 crc kubenswrapper[4764]: I0202 09:25:43.689421 4764 scope.go:117] "RemoveContainer" containerID="46fbf04fb098db11dba046c564e34e636b9baab622642a8e64fd55a7fa860a7e" Feb 02 09:25:43 crc kubenswrapper[4764]: I0202 09:25:43.689543 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7c55c99774-xk9j7" Feb 02 09:25:43 crc kubenswrapper[4764]: I0202 09:25:43.735917 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-855c89b68b-ptbcl" event={"ID":"b004a188-196b-4812-a9e6-e398062cf4d3","Type":"ContainerStarted","Data":"34939dc8441170817341f34a0d22176b25e0f6f5c228ffcc60dd2c24edb6ed68"} Feb 02 09:25:43 crc kubenswrapper[4764]: I0202 09:25:43.735990 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-855c89b68b-ptbcl" event={"ID":"b004a188-196b-4812-a9e6-e398062cf4d3","Type":"ContainerStarted","Data":"8afbf10cbbb420fc656552791a55e5ba8cf0f3bfdbf23d0e5d06a28008efb0a7"} Feb 02 09:25:43 crc kubenswrapper[4764]: I0202 09:25:43.756682 4764 scope.go:117] "RemoveContainer" containerID="5d0b70e768483117b47f431cc740738a434e093686f250c01880a75ce4848ada" Feb 02 09:25:43 crc kubenswrapper[4764]: I0202 09:25:43.756838 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-797597c45b-9mkpp" event={"ID":"f06f976d-5e29-4e78-81f4-81de01ae8391","Type":"ContainerStarted","Data":"0297dea885c523a915f8d39094de23c564f47da06ff7a74c778a2fc6ac88cffe"} Feb 02 09:25:43 crc kubenswrapper[4764]: I0202 09:25:43.756862 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-797597c45b-9mkpp" event={"ID":"f06f976d-5e29-4e78-81f4-81de01ae8391","Type":"ContainerStarted","Data":"7b37a3588b2ca39b7293823e1266e83bf60230cc38c0b515eac95f0bb8c90346"} Feb 02 09:25:43 crc kubenswrapper[4764]: I0202 09:25:43.771094 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-7c55c99774-xk9j7"] Feb 02 09:25:43 crc kubenswrapper[4764]: I0202 09:25:43.789345 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-7c55c99774-xk9j7"] Feb 02 09:25:43 crc kubenswrapper[4764]: I0202 09:25:43.859920 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="075a47f1-9307-494f-8268-c544497cf25b" path="/var/lib/kubelet/pods/075a47f1-9307-494f-8268-c544497cf25b/volumes" Feb 02 09:25:44 crc kubenswrapper[4764]: I0202 09:25:44.308451 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bfb76cfc-8k8x2" Feb 02 09:25:44 crc kubenswrapper[4764]: I0202 09:25:44.374548 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pwn69\" (UniqueName: \"kubernetes.io/projected/1ad03482-854b-4e3a-8ee4-8e4f22b68d3c-kube-api-access-pwn69\") pod \"1ad03482-854b-4e3a-8ee4-8e4f22b68d3c\" (UID: \"1ad03482-854b-4e3a-8ee4-8e4f22b68d3c\") " Feb 02 09:25:44 crc kubenswrapper[4764]: I0202 09:25:44.374601 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1ad03482-854b-4e3a-8ee4-8e4f22b68d3c-ovsdbserver-sb\") pod \"1ad03482-854b-4e3a-8ee4-8e4f22b68d3c\" (UID: \"1ad03482-854b-4e3a-8ee4-8e4f22b68d3c\") " Feb 02 09:25:44 crc kubenswrapper[4764]: I0202 09:25:44.374766 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1ad03482-854b-4e3a-8ee4-8e4f22b68d3c-ovsdbserver-nb\") pod \"1ad03482-854b-4e3a-8ee4-8e4f22b68d3c\" (UID: \"1ad03482-854b-4e3a-8ee4-8e4f22b68d3c\") " Feb 02 09:25:44 crc kubenswrapper[4764]: I0202 09:25:44.374807 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1ad03482-854b-4e3a-8ee4-8e4f22b68d3c-dns-svc\") pod \"1ad03482-854b-4e3a-8ee4-8e4f22b68d3c\" (UID: \"1ad03482-854b-4e3a-8ee4-8e4f22b68d3c\") " Feb 02 09:25:44 crc kubenswrapper[4764]: I0202 09:25:44.374829 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1ad03482-854b-4e3a-8ee4-8e4f22b68d3c-config\") pod \"1ad03482-854b-4e3a-8ee4-8e4f22b68d3c\" (UID: \"1ad03482-854b-4e3a-8ee4-8e4f22b68d3c\") " Feb 02 09:25:44 crc kubenswrapper[4764]: I0202 09:25:44.385149 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1ad03482-854b-4e3a-8ee4-8e4f22b68d3c-kube-api-access-pwn69" (OuterVolumeSpecName: "kube-api-access-pwn69") pod "1ad03482-854b-4e3a-8ee4-8e4f22b68d3c" (UID: "1ad03482-854b-4e3a-8ee4-8e4f22b68d3c"). InnerVolumeSpecName "kube-api-access-pwn69". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:25:44 crc kubenswrapper[4764]: I0202 09:25:44.442618 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1ad03482-854b-4e3a-8ee4-8e4f22b68d3c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "1ad03482-854b-4e3a-8ee4-8e4f22b68d3c" (UID: "1ad03482-854b-4e3a-8ee4-8e4f22b68d3c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:25:44 crc kubenswrapper[4764]: I0202 09:25:44.480039 4764 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1ad03482-854b-4e3a-8ee4-8e4f22b68d3c-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 02 09:25:44 crc kubenswrapper[4764]: I0202 09:25:44.480071 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pwn69\" (UniqueName: \"kubernetes.io/projected/1ad03482-854b-4e3a-8ee4-8e4f22b68d3c-kube-api-access-pwn69\") on node \"crc\" DevicePath \"\"" Feb 02 09:25:44 crc kubenswrapper[4764]: I0202 09:25:44.504891 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1ad03482-854b-4e3a-8ee4-8e4f22b68d3c-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "1ad03482-854b-4e3a-8ee4-8e4f22b68d3c" (UID: "1ad03482-854b-4e3a-8ee4-8e4f22b68d3c"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:25:44 crc kubenswrapper[4764]: I0202 09:25:44.529185 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1ad03482-854b-4e3a-8ee4-8e4f22b68d3c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "1ad03482-854b-4e3a-8ee4-8e4f22b68d3c" (UID: "1ad03482-854b-4e3a-8ee4-8e4f22b68d3c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:25:44 crc kubenswrapper[4764]: I0202 09:25:44.551633 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1ad03482-854b-4e3a-8ee4-8e4f22b68d3c-config" (OuterVolumeSpecName: "config") pod "1ad03482-854b-4e3a-8ee4-8e4f22b68d3c" (UID: "1ad03482-854b-4e3a-8ee4-8e4f22b68d3c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:25:44 crc kubenswrapper[4764]: I0202 09:25:44.583191 4764 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1ad03482-854b-4e3a-8ee4-8e4f22b68d3c-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 02 09:25:44 crc kubenswrapper[4764]: I0202 09:25:44.583221 4764 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1ad03482-854b-4e3a-8ee4-8e4f22b68d3c-config\") on node \"crc\" DevicePath \"\"" Feb 02 09:25:44 crc kubenswrapper[4764]: I0202 09:25:44.583231 4764 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1ad03482-854b-4e3a-8ee4-8e4f22b68d3c-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 02 09:25:44 crc kubenswrapper[4764]: I0202 09:25:44.770630 4764 generic.go:334] "Generic (PLEG): container finished" podID="1ad03482-854b-4e3a-8ee4-8e4f22b68d3c" containerID="8114f849c9b5b2c77035ec1db7e306cb25897aa6ff4fabe90927e8dd131746b6" exitCode=0 Feb 02 09:25:44 crc kubenswrapper[4764]: I0202 09:25:44.770692 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bfb76cfc-8k8x2" event={"ID":"1ad03482-854b-4e3a-8ee4-8e4f22b68d3c","Type":"ContainerDied","Data":"8114f849c9b5b2c77035ec1db7e306cb25897aa6ff4fabe90927e8dd131746b6"} Feb 02 09:25:44 crc kubenswrapper[4764]: I0202 09:25:44.770714 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bfb76cfc-8k8x2" event={"ID":"1ad03482-854b-4e3a-8ee4-8e4f22b68d3c","Type":"ContainerDied","Data":"a3197f7eb91ec309af68cce1627830dbbaa2e63867fd7c07d919736fbfe9788c"} Feb 02 09:25:44 crc kubenswrapper[4764]: I0202 09:25:44.770732 4764 scope.go:117] "RemoveContainer" containerID="8114f849c9b5b2c77035ec1db7e306cb25897aa6ff4fabe90927e8dd131746b6" Feb 02 09:25:44 crc kubenswrapper[4764]: I0202 09:25:44.770811 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bfb76cfc-8k8x2" Feb 02 09:25:44 crc kubenswrapper[4764]: I0202 09:25:44.790093 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-855c89b68b-ptbcl" event={"ID":"b004a188-196b-4812-a9e6-e398062cf4d3","Type":"ContainerStarted","Data":"f6d3f5c7772df19fe961e9995af19916d1a6ae961a5ab7519b662b70853de195"} Feb 02 09:25:44 crc kubenswrapper[4764]: I0202 09:25:44.811550 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-797597c45b-9mkpp" event={"ID":"f06f976d-5e29-4e78-81f4-81de01ae8391","Type":"ContainerStarted","Data":"359dba813ca71d6278d53a5b1e6bc5259d22cc4fc46f4993180a35412d073c50"} Feb 02 09:25:44 crc kubenswrapper[4764]: I0202 09:25:44.812308 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-797597c45b-9mkpp" Feb 02 09:25:44 crc kubenswrapper[4764]: I0202 09:25:44.812345 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-797597c45b-9mkpp" Feb 02 09:25:44 crc kubenswrapper[4764]: I0202 09:25:44.818098 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5bfb76cfc-8k8x2"] Feb 02 09:25:44 crc kubenswrapper[4764]: I0202 09:25:44.828347 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5686dc79cc-sb69r" event={"ID":"4e3b3433-3d64-44ce-9a3a-14fcd0bdaa6d","Type":"ContainerStarted","Data":"65339fd235fe926cc79898c7928f8b77d678244856d4c4d81f5103a228cfdb34"} Feb 02 09:25:44 crc kubenswrapper[4764]: I0202 09:25:44.832845 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-855c89b68b-ptbcl" podStartSLOduration=3.832829129 podStartE2EDuration="3.832829129s" podCreationTimestamp="2026-02-02 09:25:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:25:44.82405657 +0000 UTC m=+1107.757780658" watchObservedRunningTime="2026-02-02 09:25:44.832829129 +0000 UTC m=+1107.766553217" Feb 02 09:25:44 crc kubenswrapper[4764]: I0202 09:25:44.834173 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5bfb76cfc-8k8x2"] Feb 02 09:25:44 crc kubenswrapper[4764]: I0202 09:25:44.872018 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-keystone-listener-6d6c6fd44b-2hfzf"] Feb 02 09:25:44 crc kubenswrapper[4764]: I0202 09:25:44.872233 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-keystone-listener-6d6c6fd44b-2hfzf" podUID="4f12af32-00c9-437a-8e0e-805caa8df878" containerName="barbican-keystone-listener-log" containerID="cri-o://c9e3885c81fca81f2a481c177ff87282be988d2fe765f222b6301ad7278e1331" gracePeriod=30 Feb 02 09:25:44 crc kubenswrapper[4764]: I0202 09:25:44.872348 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-keystone-listener-6d6c6fd44b-2hfzf" podUID="4f12af32-00c9-437a-8e0e-805caa8df878" containerName="barbican-keystone-listener" containerID="cri-o://90aa0957e88cd020632dd7468ddffe7b8338bced2a1080b5399b0538a9434ff6" gracePeriod=30 Feb 02 09:25:44 crc kubenswrapper[4764]: I0202 09:25:44.889415 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-797597c45b-9mkpp" podStartSLOduration=3.889395941 podStartE2EDuration="3.889395941s" podCreationTimestamp="2026-02-02 09:25:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:25:44.860506713 +0000 UTC m=+1107.794230801" watchObservedRunningTime="2026-02-02 09:25:44.889395941 +0000 UTC m=+1107.823120029" Feb 02 09:25:44 crc kubenswrapper[4764]: I0202 09:25:44.897557 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-5686dc79cc-sb69r" podStartSLOduration=3.897540813 podStartE2EDuration="3.897540813s" podCreationTimestamp="2026-02-02 09:25:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:25:44.884353333 +0000 UTC m=+1107.818077421" watchObservedRunningTime="2026-02-02 09:25:44.897540813 +0000 UTC m=+1107.831264901" Feb 02 09:25:44 crc kubenswrapper[4764]: I0202 09:25:44.928367 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-worker-6db5f6d5bf-pxtw5"] Feb 02 09:25:44 crc kubenswrapper[4764]: I0202 09:25:44.928561 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-worker-6db5f6d5bf-pxtw5" podUID="4ebb36bb-8032-42ea-8954-7716e5a81301" containerName="barbican-worker-log" containerID="cri-o://2e2b3632dba3173bf5217987f806dceddb390cb12830850b19e2e1784e431682" gracePeriod=30 Feb 02 09:25:44 crc kubenswrapper[4764]: I0202 09:25:44.928703 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-worker-6db5f6d5bf-pxtw5" podUID="4ebb36bb-8032-42ea-8954-7716e5a81301" containerName="barbican-worker" containerID="cri-o://530aca16f5dc00ab072aeb59ad0d8b2d8bea184774c6c58f4786ffc31305fe78" gracePeriod=30 Feb 02 09:25:45 crc kubenswrapper[4764]: I0202 09:25:45.836412 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1ad03482-854b-4e3a-8ee4-8e4f22b68d3c" path="/var/lib/kubelet/pods/1ad03482-854b-4e3a-8ee4-8e4f22b68d3c/volumes" Feb 02 09:25:45 crc kubenswrapper[4764]: I0202 09:25:45.843692 4764 generic.go:334] "Generic (PLEG): container finished" podID="4ebb36bb-8032-42ea-8954-7716e5a81301" containerID="530aca16f5dc00ab072aeb59ad0d8b2d8bea184774c6c58f4786ffc31305fe78" exitCode=0 Feb 02 09:25:45 crc kubenswrapper[4764]: I0202 09:25:45.843722 4764 generic.go:334] "Generic (PLEG): container finished" podID="4ebb36bb-8032-42ea-8954-7716e5a81301" containerID="2e2b3632dba3173bf5217987f806dceddb390cb12830850b19e2e1784e431682" exitCode=143 Feb 02 09:25:45 crc kubenswrapper[4764]: I0202 09:25:45.843773 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-6db5f6d5bf-pxtw5" event={"ID":"4ebb36bb-8032-42ea-8954-7716e5a81301","Type":"ContainerDied","Data":"530aca16f5dc00ab072aeb59ad0d8b2d8bea184774c6c58f4786ffc31305fe78"} Feb 02 09:25:45 crc kubenswrapper[4764]: I0202 09:25:45.843817 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-6db5f6d5bf-pxtw5" event={"ID":"4ebb36bb-8032-42ea-8954-7716e5a81301","Type":"ContainerDied","Data":"2e2b3632dba3173bf5217987f806dceddb390cb12830850b19e2e1784e431682"} Feb 02 09:25:45 crc kubenswrapper[4764]: I0202 09:25:45.846331 4764 generic.go:334] "Generic (PLEG): container finished" podID="4f12af32-00c9-437a-8e0e-805caa8df878" containerID="90aa0957e88cd020632dd7468ddffe7b8338bced2a1080b5399b0538a9434ff6" exitCode=0 Feb 02 09:25:45 crc kubenswrapper[4764]: I0202 09:25:45.846354 4764 generic.go:334] "Generic (PLEG): container finished" podID="4f12af32-00c9-437a-8e0e-805caa8df878" containerID="c9e3885c81fca81f2a481c177ff87282be988d2fe765f222b6301ad7278e1331" exitCode=143 Feb 02 09:25:45 crc kubenswrapper[4764]: I0202 09:25:45.846431 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-6d6c6fd44b-2hfzf" event={"ID":"4f12af32-00c9-437a-8e0e-805caa8df878","Type":"ContainerDied","Data":"90aa0957e88cd020632dd7468ddffe7b8338bced2a1080b5399b0538a9434ff6"} Feb 02 09:25:45 crc kubenswrapper[4764]: I0202 09:25:45.846471 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-6d6c6fd44b-2hfzf" event={"ID":"4f12af32-00c9-437a-8e0e-805caa8df878","Type":"ContainerDied","Data":"c9e3885c81fca81f2a481c177ff87282be988d2fe765f222b6301ad7278e1331"} Feb 02 09:25:46 crc kubenswrapper[4764]: I0202 09:25:46.864326 4764 generic.go:334] "Generic (PLEG): container finished" podID="6b3d3314-2f16-4dc3-aa52-5f7f8576813a" containerID="3a7e2521eea6d5fe7739f7c62d9ec21dc8b57406acc029a0ee0d71a25fd330b2" exitCode=0 Feb 02 09:25:46 crc kubenswrapper[4764]: I0202 09:25:46.864441 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-dkrnx" event={"ID":"6b3d3314-2f16-4dc3-aa52-5f7f8576813a","Type":"ContainerDied","Data":"3a7e2521eea6d5fe7739f7c62d9ec21dc8b57406acc029a0ee0d71a25fd330b2"} Feb 02 09:25:48 crc kubenswrapper[4764]: I0202 09:25:48.005568 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-7b6fbdbb96-fgr8m" Feb 02 09:25:48 crc kubenswrapper[4764]: I0202 09:25:48.028467 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-7b6fbdbb96-fgr8m" Feb 02 09:25:49 crc kubenswrapper[4764]: I0202 09:25:49.904552 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-dkrnx" event={"ID":"6b3d3314-2f16-4dc3-aa52-5f7f8576813a","Type":"ContainerDied","Data":"0260dc4698947ea66a729c434ded2d4f27469d3098b4e9a0bd7baf2466867ada"} Feb 02 09:25:49 crc kubenswrapper[4764]: I0202 09:25:49.904968 4764 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0260dc4698947ea66a729c434ded2d4f27469d3098b4e9a0bd7baf2466867ada" Feb 02 09:25:49 crc kubenswrapper[4764]: I0202 09:25:49.909761 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-dkrnx" Feb 02 09:25:50 crc kubenswrapper[4764]: I0202 09:25:49.999245 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hgkdq\" (UniqueName: \"kubernetes.io/projected/6b3d3314-2f16-4dc3-aa52-5f7f8576813a-kube-api-access-hgkdq\") pod \"6b3d3314-2f16-4dc3-aa52-5f7f8576813a\" (UID: \"6b3d3314-2f16-4dc3-aa52-5f7f8576813a\") " Feb 02 09:25:50 crc kubenswrapper[4764]: I0202 09:25:49.999371 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b3d3314-2f16-4dc3-aa52-5f7f8576813a-scripts\") pod \"6b3d3314-2f16-4dc3-aa52-5f7f8576813a\" (UID: \"6b3d3314-2f16-4dc3-aa52-5f7f8576813a\") " Feb 02 09:25:50 crc kubenswrapper[4764]: I0202 09:25:49.999389 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6b3d3314-2f16-4dc3-aa52-5f7f8576813a-etc-machine-id\") pod \"6b3d3314-2f16-4dc3-aa52-5f7f8576813a\" (UID: \"6b3d3314-2f16-4dc3-aa52-5f7f8576813a\") " Feb 02 09:25:50 crc kubenswrapper[4764]: I0202 09:25:49.999405 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b3d3314-2f16-4dc3-aa52-5f7f8576813a-combined-ca-bundle\") pod \"6b3d3314-2f16-4dc3-aa52-5f7f8576813a\" (UID: \"6b3d3314-2f16-4dc3-aa52-5f7f8576813a\") " Feb 02 09:25:50 crc kubenswrapper[4764]: I0202 09:25:49.999468 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b3d3314-2f16-4dc3-aa52-5f7f8576813a-config-data\") pod \"6b3d3314-2f16-4dc3-aa52-5f7f8576813a\" (UID: \"6b3d3314-2f16-4dc3-aa52-5f7f8576813a\") " Feb 02 09:25:50 crc kubenswrapper[4764]: I0202 09:25:49.999600 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/6b3d3314-2f16-4dc3-aa52-5f7f8576813a-db-sync-config-data\") pod \"6b3d3314-2f16-4dc3-aa52-5f7f8576813a\" (UID: \"6b3d3314-2f16-4dc3-aa52-5f7f8576813a\") " Feb 02 09:25:50 crc kubenswrapper[4764]: I0202 09:25:50.000051 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6b3d3314-2f16-4dc3-aa52-5f7f8576813a-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "6b3d3314-2f16-4dc3-aa52-5f7f8576813a" (UID: "6b3d3314-2f16-4dc3-aa52-5f7f8576813a"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 09:25:50 crc kubenswrapper[4764]: I0202 09:25:50.006736 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b3d3314-2f16-4dc3-aa52-5f7f8576813a-scripts" (OuterVolumeSpecName: "scripts") pod "6b3d3314-2f16-4dc3-aa52-5f7f8576813a" (UID: "6b3d3314-2f16-4dc3-aa52-5f7f8576813a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:25:50 crc kubenswrapper[4764]: I0202 09:25:50.025214 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6b3d3314-2f16-4dc3-aa52-5f7f8576813a-kube-api-access-hgkdq" (OuterVolumeSpecName: "kube-api-access-hgkdq") pod "6b3d3314-2f16-4dc3-aa52-5f7f8576813a" (UID: "6b3d3314-2f16-4dc3-aa52-5f7f8576813a"). InnerVolumeSpecName "kube-api-access-hgkdq". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:25:50 crc kubenswrapper[4764]: I0202 09:25:50.026543 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b3d3314-2f16-4dc3-aa52-5f7f8576813a-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "6b3d3314-2f16-4dc3-aa52-5f7f8576813a" (UID: "6b3d3314-2f16-4dc3-aa52-5f7f8576813a"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:25:50 crc kubenswrapper[4764]: I0202 09:25:50.096279 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b3d3314-2f16-4dc3-aa52-5f7f8576813a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6b3d3314-2f16-4dc3-aa52-5f7f8576813a" (UID: "6b3d3314-2f16-4dc3-aa52-5f7f8576813a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:25:50 crc kubenswrapper[4764]: I0202 09:25:50.105639 4764 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b3d3314-2f16-4dc3-aa52-5f7f8576813a-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 09:25:50 crc kubenswrapper[4764]: I0202 09:25:50.105666 4764 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6b3d3314-2f16-4dc3-aa52-5f7f8576813a-etc-machine-id\") on node \"crc\" DevicePath \"\"" Feb 02 09:25:50 crc kubenswrapper[4764]: I0202 09:25:50.105677 4764 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b3d3314-2f16-4dc3-aa52-5f7f8576813a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 09:25:50 crc kubenswrapper[4764]: I0202 09:25:50.105686 4764 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/6b3d3314-2f16-4dc3-aa52-5f7f8576813a-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 09:25:50 crc kubenswrapper[4764]: I0202 09:25:50.105694 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hgkdq\" (UniqueName: \"kubernetes.io/projected/6b3d3314-2f16-4dc3-aa52-5f7f8576813a-kube-api-access-hgkdq\") on node \"crc\" DevicePath \"\"" Feb 02 09:25:50 crc kubenswrapper[4764]: I0202 09:25:50.141237 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b3d3314-2f16-4dc3-aa52-5f7f8576813a-config-data" (OuterVolumeSpecName: "config-data") pod "6b3d3314-2f16-4dc3-aa52-5f7f8576813a" (UID: "6b3d3314-2f16-4dc3-aa52-5f7f8576813a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:25:50 crc kubenswrapper[4764]: I0202 09:25:50.215838 4764 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b3d3314-2f16-4dc3-aa52-5f7f8576813a-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 09:25:50 crc kubenswrapper[4764]: I0202 09:25:50.487047 4764 scope.go:117] "RemoveContainer" containerID="8a7877158d5931d89dac739abbe12ca8fb2986dcae9b4a46319b73f7ba6d04d9" Feb 02 09:25:50 crc kubenswrapper[4764]: I0202 09:25:50.502124 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-6db5f6d5bf-pxtw5" Feb 02 09:25:50 crc kubenswrapper[4764]: I0202 09:25:50.509973 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-6d6c6fd44b-2hfzf" Feb 02 09:25:50 crc kubenswrapper[4764]: I0202 09:25:50.573348 4764 scope.go:117] "RemoveContainer" containerID="8114f849c9b5b2c77035ec1db7e306cb25897aa6ff4fabe90927e8dd131746b6" Feb 02 09:25:50 crc kubenswrapper[4764]: E0202 09:25:50.574119 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8114f849c9b5b2c77035ec1db7e306cb25897aa6ff4fabe90927e8dd131746b6\": container with ID starting with 8114f849c9b5b2c77035ec1db7e306cb25897aa6ff4fabe90927e8dd131746b6 not found: ID does not exist" containerID="8114f849c9b5b2c77035ec1db7e306cb25897aa6ff4fabe90927e8dd131746b6" Feb 02 09:25:50 crc kubenswrapper[4764]: I0202 09:25:50.574179 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8114f849c9b5b2c77035ec1db7e306cb25897aa6ff4fabe90927e8dd131746b6"} err="failed to get container status \"8114f849c9b5b2c77035ec1db7e306cb25897aa6ff4fabe90927e8dd131746b6\": rpc error: code = NotFound desc = could not find container \"8114f849c9b5b2c77035ec1db7e306cb25897aa6ff4fabe90927e8dd131746b6\": container with ID starting with 8114f849c9b5b2c77035ec1db7e306cb25897aa6ff4fabe90927e8dd131746b6 not found: ID does not exist" Feb 02 09:25:50 crc kubenswrapper[4764]: I0202 09:25:50.574227 4764 scope.go:117] "RemoveContainer" containerID="8a7877158d5931d89dac739abbe12ca8fb2986dcae9b4a46319b73f7ba6d04d9" Feb 02 09:25:50 crc kubenswrapper[4764]: E0202 09:25:50.574518 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8a7877158d5931d89dac739abbe12ca8fb2986dcae9b4a46319b73f7ba6d04d9\": container with ID starting with 8a7877158d5931d89dac739abbe12ca8fb2986dcae9b4a46319b73f7ba6d04d9 not found: ID does not exist" containerID="8a7877158d5931d89dac739abbe12ca8fb2986dcae9b4a46319b73f7ba6d04d9" Feb 02 09:25:50 crc kubenswrapper[4764]: I0202 09:25:50.574563 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8a7877158d5931d89dac739abbe12ca8fb2986dcae9b4a46319b73f7ba6d04d9"} err="failed to get container status \"8a7877158d5931d89dac739abbe12ca8fb2986dcae9b4a46319b73f7ba6d04d9\": rpc error: code = NotFound desc = could not find container \"8a7877158d5931d89dac739abbe12ca8fb2986dcae9b4a46319b73f7ba6d04d9\": container with ID starting with 8a7877158d5931d89dac739abbe12ca8fb2986dcae9b4a46319b73f7ba6d04d9 not found: ID does not exist" Feb 02 09:25:50 crc kubenswrapper[4764]: I0202 09:25:50.624843 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f12af32-00c9-437a-8e0e-805caa8df878-combined-ca-bundle\") pod \"4f12af32-00c9-437a-8e0e-805caa8df878\" (UID: \"4f12af32-00c9-437a-8e0e-805caa8df878\") " Feb 02 09:25:50 crc kubenswrapper[4764]: I0202 09:25:50.625374 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4f12af32-00c9-437a-8e0e-805caa8df878-config-data-custom\") pod \"4f12af32-00c9-437a-8e0e-805caa8df878\" (UID: \"4f12af32-00c9-437a-8e0e-805caa8df878\") " Feb 02 09:25:50 crc kubenswrapper[4764]: I0202 09:25:50.625795 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4ebb36bb-8032-42ea-8954-7716e5a81301-config-data-custom\") pod \"4ebb36bb-8032-42ea-8954-7716e5a81301\" (UID: \"4ebb36bb-8032-42ea-8954-7716e5a81301\") " Feb 02 09:25:50 crc kubenswrapper[4764]: I0202 09:25:50.625825 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cb7vb\" (UniqueName: \"kubernetes.io/projected/4ebb36bb-8032-42ea-8954-7716e5a81301-kube-api-access-cb7vb\") pod \"4ebb36bb-8032-42ea-8954-7716e5a81301\" (UID: \"4ebb36bb-8032-42ea-8954-7716e5a81301\") " Feb 02 09:25:50 crc kubenswrapper[4764]: I0202 09:25:50.626331 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ebb36bb-8032-42ea-8954-7716e5a81301-combined-ca-bundle\") pod \"4ebb36bb-8032-42ea-8954-7716e5a81301\" (UID: \"4ebb36bb-8032-42ea-8954-7716e5a81301\") " Feb 02 09:25:50 crc kubenswrapper[4764]: I0202 09:25:50.626363 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rsq9n\" (UniqueName: \"kubernetes.io/projected/4f12af32-00c9-437a-8e0e-805caa8df878-kube-api-access-rsq9n\") pod \"4f12af32-00c9-437a-8e0e-805caa8df878\" (UID: \"4f12af32-00c9-437a-8e0e-805caa8df878\") " Feb 02 09:25:50 crc kubenswrapper[4764]: I0202 09:25:50.626409 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4ebb36bb-8032-42ea-8954-7716e5a81301-logs\") pod \"4ebb36bb-8032-42ea-8954-7716e5a81301\" (UID: \"4ebb36bb-8032-42ea-8954-7716e5a81301\") " Feb 02 09:25:50 crc kubenswrapper[4764]: I0202 09:25:50.626440 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4f12af32-00c9-437a-8e0e-805caa8df878-logs\") pod \"4f12af32-00c9-437a-8e0e-805caa8df878\" (UID: \"4f12af32-00c9-437a-8e0e-805caa8df878\") " Feb 02 09:25:50 crc kubenswrapper[4764]: I0202 09:25:50.626468 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ebb36bb-8032-42ea-8954-7716e5a81301-config-data\") pod \"4ebb36bb-8032-42ea-8954-7716e5a81301\" (UID: \"4ebb36bb-8032-42ea-8954-7716e5a81301\") " Feb 02 09:25:50 crc kubenswrapper[4764]: I0202 09:25:50.626512 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4f12af32-00c9-437a-8e0e-805caa8df878-config-data\") pod \"4f12af32-00c9-437a-8e0e-805caa8df878\" (UID: \"4f12af32-00c9-437a-8e0e-805caa8df878\") " Feb 02 09:25:50 crc kubenswrapper[4764]: I0202 09:25:50.627037 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4ebb36bb-8032-42ea-8954-7716e5a81301-logs" (OuterVolumeSpecName: "logs") pod "4ebb36bb-8032-42ea-8954-7716e5a81301" (UID: "4ebb36bb-8032-42ea-8954-7716e5a81301"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 09:25:50 crc kubenswrapper[4764]: I0202 09:25:50.627154 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4f12af32-00c9-437a-8e0e-805caa8df878-logs" (OuterVolumeSpecName: "logs") pod "4f12af32-00c9-437a-8e0e-805caa8df878" (UID: "4f12af32-00c9-437a-8e0e-805caa8df878"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 09:25:50 crc kubenswrapper[4764]: I0202 09:25:50.627688 4764 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4ebb36bb-8032-42ea-8954-7716e5a81301-logs\") on node \"crc\" DevicePath \"\"" Feb 02 09:25:50 crc kubenswrapper[4764]: I0202 09:25:50.627713 4764 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4f12af32-00c9-437a-8e0e-805caa8df878-logs\") on node \"crc\" DevicePath \"\"" Feb 02 09:25:50 crc kubenswrapper[4764]: I0202 09:25:50.628601 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f12af32-00c9-437a-8e0e-805caa8df878-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "4f12af32-00c9-437a-8e0e-805caa8df878" (UID: "4f12af32-00c9-437a-8e0e-805caa8df878"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:25:50 crc kubenswrapper[4764]: I0202 09:25:50.630158 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ebb36bb-8032-42ea-8954-7716e5a81301-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "4ebb36bb-8032-42ea-8954-7716e5a81301" (UID: "4ebb36bb-8032-42ea-8954-7716e5a81301"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:25:50 crc kubenswrapper[4764]: I0202 09:25:50.630279 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4f12af32-00c9-437a-8e0e-805caa8df878-kube-api-access-rsq9n" (OuterVolumeSpecName: "kube-api-access-rsq9n") pod "4f12af32-00c9-437a-8e0e-805caa8df878" (UID: "4f12af32-00c9-437a-8e0e-805caa8df878"). InnerVolumeSpecName "kube-api-access-rsq9n". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:25:50 crc kubenswrapper[4764]: I0202 09:25:50.631227 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ebb36bb-8032-42ea-8954-7716e5a81301-kube-api-access-cb7vb" (OuterVolumeSpecName: "kube-api-access-cb7vb") pod "4ebb36bb-8032-42ea-8954-7716e5a81301" (UID: "4ebb36bb-8032-42ea-8954-7716e5a81301"). InnerVolumeSpecName "kube-api-access-cb7vb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:25:50 crc kubenswrapper[4764]: I0202 09:25:50.645743 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ebb36bb-8032-42ea-8954-7716e5a81301-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4ebb36bb-8032-42ea-8954-7716e5a81301" (UID: "4ebb36bb-8032-42ea-8954-7716e5a81301"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:25:50 crc kubenswrapper[4764]: I0202 09:25:50.649665 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f12af32-00c9-437a-8e0e-805caa8df878-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4f12af32-00c9-437a-8e0e-805caa8df878" (UID: "4f12af32-00c9-437a-8e0e-805caa8df878"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:25:50 crc kubenswrapper[4764]: I0202 09:25:50.675005 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ebb36bb-8032-42ea-8954-7716e5a81301-config-data" (OuterVolumeSpecName: "config-data") pod "4ebb36bb-8032-42ea-8954-7716e5a81301" (UID: "4ebb36bb-8032-42ea-8954-7716e5a81301"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:25:50 crc kubenswrapper[4764]: I0202 09:25:50.676078 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f12af32-00c9-437a-8e0e-805caa8df878-config-data" (OuterVolumeSpecName: "config-data") pod "4f12af32-00c9-437a-8e0e-805caa8df878" (UID: "4f12af32-00c9-437a-8e0e-805caa8df878"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:25:50 crc kubenswrapper[4764]: I0202 09:25:50.729883 4764 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4f12af32-00c9-437a-8e0e-805caa8df878-config-data-custom\") on node \"crc\" DevicePath \"\"" Feb 02 09:25:50 crc kubenswrapper[4764]: I0202 09:25:50.729913 4764 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4ebb36bb-8032-42ea-8954-7716e5a81301-config-data-custom\") on node \"crc\" DevicePath \"\"" Feb 02 09:25:50 crc kubenswrapper[4764]: I0202 09:25:50.729923 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cb7vb\" (UniqueName: \"kubernetes.io/projected/4ebb36bb-8032-42ea-8954-7716e5a81301-kube-api-access-cb7vb\") on node \"crc\" DevicePath \"\"" Feb 02 09:25:50 crc kubenswrapper[4764]: I0202 09:25:50.729949 4764 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ebb36bb-8032-42ea-8954-7716e5a81301-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 09:25:50 crc kubenswrapper[4764]: I0202 09:25:50.729961 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rsq9n\" (UniqueName: \"kubernetes.io/projected/4f12af32-00c9-437a-8e0e-805caa8df878-kube-api-access-rsq9n\") on node \"crc\" DevicePath \"\"" Feb 02 09:25:50 crc kubenswrapper[4764]: I0202 09:25:50.729972 4764 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ebb36bb-8032-42ea-8954-7716e5a81301-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 09:25:50 crc kubenswrapper[4764]: I0202 09:25:50.729983 4764 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4f12af32-00c9-437a-8e0e-805caa8df878-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 09:25:50 crc kubenswrapper[4764]: I0202 09:25:50.729993 4764 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f12af32-00c9-437a-8e0e-805caa8df878-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 09:25:50 crc kubenswrapper[4764]: E0202 09:25:50.750729 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ceilometer-0" podUID="a518caeb-6b91-4b71-a47f-0d71d965af2e" Feb 02 09:25:50 crc kubenswrapper[4764]: I0202 09:25:50.925577 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a518caeb-6b91-4b71-a47f-0d71d965af2e","Type":"ContainerStarted","Data":"61b9bc35586b452625d8a3057d23d4edbbe7c69316a8ec1ebf5860267f1aadd3"} Feb 02 09:25:50 crc kubenswrapper[4764]: I0202 09:25:50.925704 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a518caeb-6b91-4b71-a47f-0d71d965af2e" containerName="ceilometer-notification-agent" containerID="cri-o://2eb65d6eba11ecb361786b367a01950ec0b46c28f3ac4e5936e72f214d4f9acd" gracePeriod=30 Feb 02 09:25:50 crc kubenswrapper[4764]: I0202 09:25:50.925755 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 02 09:25:50 crc kubenswrapper[4764]: I0202 09:25:50.925756 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a518caeb-6b91-4b71-a47f-0d71d965af2e" containerName="proxy-httpd" containerID="cri-o://61b9bc35586b452625d8a3057d23d4edbbe7c69316a8ec1ebf5860267f1aadd3" gracePeriod=30 Feb 02 09:25:50 crc kubenswrapper[4764]: I0202 09:25:50.926013 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a518caeb-6b91-4b71-a47f-0d71d965af2e" containerName="sg-core" containerID="cri-o://babe286652a8ae251bf5bcfe389a56fa30eaefd0cd0ff5f5bcb6999e85408847" gracePeriod=30 Feb 02 09:25:50 crc kubenswrapper[4764]: I0202 09:25:50.940826 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-6d6c6fd44b-2hfzf" event={"ID":"4f12af32-00c9-437a-8e0e-805caa8df878","Type":"ContainerDied","Data":"d088edd9082f11f04848b13b2a6a0159d98d19ea524eaee71d140dc187193e16"} Feb 02 09:25:50 crc kubenswrapper[4764]: I0202 09:25:50.940886 4764 scope.go:117] "RemoveContainer" containerID="90aa0957e88cd020632dd7468ddffe7b8338bced2a1080b5399b0538a9434ff6" Feb 02 09:25:50 crc kubenswrapper[4764]: I0202 09:25:50.941149 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-6d6c6fd44b-2hfzf" Feb 02 09:25:50 crc kubenswrapper[4764]: I0202 09:25:50.958458 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-dkrnx" Feb 02 09:25:50 crc kubenswrapper[4764]: I0202 09:25:50.958511 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-6db5f6d5bf-pxtw5" event={"ID":"4ebb36bb-8032-42ea-8954-7716e5a81301","Type":"ContainerDied","Data":"1834b62df8a95754f2ba8e5735004a0526bf197a1915bd2b88293ce268ee1dc3"} Feb 02 09:25:50 crc kubenswrapper[4764]: I0202 09:25:50.958784 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-6db5f6d5bf-pxtw5" Feb 02 09:25:50 crc kubenswrapper[4764]: I0202 09:25:50.987552 4764 scope.go:117] "RemoveContainer" containerID="c9e3885c81fca81f2a481c177ff87282be988d2fe765f222b6301ad7278e1331" Feb 02 09:25:51 crc kubenswrapper[4764]: I0202 09:25:51.014547 4764 scope.go:117] "RemoveContainer" containerID="530aca16f5dc00ab072aeb59ad0d8b2d8bea184774c6c58f4786ffc31305fe78" Feb 02 09:25:51 crc kubenswrapper[4764]: I0202 09:25:51.039193 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-keystone-listener-6d6c6fd44b-2hfzf"] Feb 02 09:25:51 crc kubenswrapper[4764]: I0202 09:25:51.040549 4764 scope.go:117] "RemoveContainer" containerID="2e2b3632dba3173bf5217987f806dceddb390cb12830850b19e2e1784e431682" Feb 02 09:25:51 crc kubenswrapper[4764]: I0202 09:25:51.068006 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-keystone-listener-6d6c6fd44b-2hfzf"] Feb 02 09:25:51 crc kubenswrapper[4764]: I0202 09:25:51.078618 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-worker-6db5f6d5bf-pxtw5"] Feb 02 09:25:51 crc kubenswrapper[4764]: I0202 09:25:51.099559 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-worker-6db5f6d5bf-pxtw5"] Feb 02 09:25:51 crc kubenswrapper[4764]: I0202 09:25:51.269437 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Feb 02 09:25:51 crc kubenswrapper[4764]: E0202 09:25:51.269859 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b3d3314-2f16-4dc3-aa52-5f7f8576813a" containerName="cinder-db-sync" Feb 02 09:25:51 crc kubenswrapper[4764]: I0202 09:25:51.269882 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b3d3314-2f16-4dc3-aa52-5f7f8576813a" containerName="cinder-db-sync" Feb 02 09:25:51 crc kubenswrapper[4764]: E0202 09:25:51.269904 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ad03482-854b-4e3a-8ee4-8e4f22b68d3c" containerName="init" Feb 02 09:25:51 crc kubenswrapper[4764]: I0202 09:25:51.269912 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ad03482-854b-4e3a-8ee4-8e4f22b68d3c" containerName="init" Feb 02 09:25:51 crc kubenswrapper[4764]: E0202 09:25:51.269928 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f12af32-00c9-437a-8e0e-805caa8df878" containerName="barbican-keystone-listener-log" Feb 02 09:25:51 crc kubenswrapper[4764]: I0202 09:25:51.269957 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f12af32-00c9-437a-8e0e-805caa8df878" containerName="barbican-keystone-listener-log" Feb 02 09:25:51 crc kubenswrapper[4764]: E0202 09:25:51.269970 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f12af32-00c9-437a-8e0e-805caa8df878" containerName="barbican-keystone-listener" Feb 02 09:25:51 crc kubenswrapper[4764]: I0202 09:25:51.269977 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f12af32-00c9-437a-8e0e-805caa8df878" containerName="barbican-keystone-listener" Feb 02 09:25:51 crc kubenswrapper[4764]: E0202 09:25:51.269990 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ad03482-854b-4e3a-8ee4-8e4f22b68d3c" containerName="dnsmasq-dns" Feb 02 09:25:51 crc kubenswrapper[4764]: I0202 09:25:51.269997 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ad03482-854b-4e3a-8ee4-8e4f22b68d3c" containerName="dnsmasq-dns" Feb 02 09:25:51 crc kubenswrapper[4764]: E0202 09:25:51.270009 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="075a47f1-9307-494f-8268-c544497cf25b" containerName="barbican-api-log" Feb 02 09:25:51 crc kubenswrapper[4764]: I0202 09:25:51.270016 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="075a47f1-9307-494f-8268-c544497cf25b" containerName="barbican-api-log" Feb 02 09:25:51 crc kubenswrapper[4764]: E0202 09:25:51.270029 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ebb36bb-8032-42ea-8954-7716e5a81301" containerName="barbican-worker-log" Feb 02 09:25:51 crc kubenswrapper[4764]: I0202 09:25:51.270036 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ebb36bb-8032-42ea-8954-7716e5a81301" containerName="barbican-worker-log" Feb 02 09:25:51 crc kubenswrapper[4764]: E0202 09:25:51.270049 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ebb36bb-8032-42ea-8954-7716e5a81301" containerName="barbican-worker" Feb 02 09:25:51 crc kubenswrapper[4764]: I0202 09:25:51.270055 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ebb36bb-8032-42ea-8954-7716e5a81301" containerName="barbican-worker" Feb 02 09:25:51 crc kubenswrapper[4764]: E0202 09:25:51.270071 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="075a47f1-9307-494f-8268-c544497cf25b" containerName="barbican-api" Feb 02 09:25:51 crc kubenswrapper[4764]: I0202 09:25:51.270081 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="075a47f1-9307-494f-8268-c544497cf25b" containerName="barbican-api" Feb 02 09:25:51 crc kubenswrapper[4764]: I0202 09:25:51.270254 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="075a47f1-9307-494f-8268-c544497cf25b" containerName="barbican-api" Feb 02 09:25:51 crc kubenswrapper[4764]: I0202 09:25:51.270273 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ad03482-854b-4e3a-8ee4-8e4f22b68d3c" containerName="dnsmasq-dns" Feb 02 09:25:51 crc kubenswrapper[4764]: I0202 09:25:51.270286 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f12af32-00c9-437a-8e0e-805caa8df878" containerName="barbican-keystone-listener-log" Feb 02 09:25:51 crc kubenswrapper[4764]: I0202 09:25:51.270296 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f12af32-00c9-437a-8e0e-805caa8df878" containerName="barbican-keystone-listener" Feb 02 09:25:51 crc kubenswrapper[4764]: I0202 09:25:51.270306 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="075a47f1-9307-494f-8268-c544497cf25b" containerName="barbican-api-log" Feb 02 09:25:51 crc kubenswrapper[4764]: I0202 09:25:51.270319 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ebb36bb-8032-42ea-8954-7716e5a81301" containerName="barbican-worker-log" Feb 02 09:25:51 crc kubenswrapper[4764]: I0202 09:25:51.270329 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b3d3314-2f16-4dc3-aa52-5f7f8576813a" containerName="cinder-db-sync" Feb 02 09:25:51 crc kubenswrapper[4764]: I0202 09:25:51.270339 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ebb36bb-8032-42ea-8954-7716e5a81301" containerName="barbican-worker" Feb 02 09:25:51 crc kubenswrapper[4764]: I0202 09:25:51.271193 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Feb 02 09:25:51 crc kubenswrapper[4764]: I0202 09:25:51.273875 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Feb 02 09:25:51 crc kubenswrapper[4764]: I0202 09:25:51.274047 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-4ftqb" Feb 02 09:25:51 crc kubenswrapper[4764]: I0202 09:25:51.274216 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Feb 02 09:25:51 crc kubenswrapper[4764]: I0202 09:25:51.274394 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Feb 02 09:25:51 crc kubenswrapper[4764]: I0202 09:25:51.294294 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 02 09:25:51 crc kubenswrapper[4764]: I0202 09:25:51.347140 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-555cfdffc9-rw6ht"] Feb 02 09:25:51 crc kubenswrapper[4764]: I0202 09:25:51.364997 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-555cfdffc9-rw6ht" Feb 02 09:25:51 crc kubenswrapper[4764]: I0202 09:25:51.396468 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-555cfdffc9-rw6ht"] Feb 02 09:25:51 crc kubenswrapper[4764]: I0202 09:25:51.458267 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9wzfx\" (UniqueName: \"kubernetes.io/projected/2c66c255-9454-496c-baf3-5c331d5a2f16-kube-api-access-9wzfx\") pod \"dnsmasq-dns-555cfdffc9-rw6ht\" (UID: \"2c66c255-9454-496c-baf3-5c331d5a2f16\") " pod="openstack/dnsmasq-dns-555cfdffc9-rw6ht" Feb 02 09:25:51 crc kubenswrapper[4764]: I0202 09:25:51.458313 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d509e876-498c-4e04-95af-bdb84c334a99-scripts\") pod \"cinder-scheduler-0\" (UID: \"d509e876-498c-4e04-95af-bdb84c334a99\") " pod="openstack/cinder-scheduler-0" Feb 02 09:25:51 crc kubenswrapper[4764]: I0202 09:25:51.458334 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2c66c255-9454-496c-baf3-5c331d5a2f16-ovsdbserver-sb\") pod \"dnsmasq-dns-555cfdffc9-rw6ht\" (UID: \"2c66c255-9454-496c-baf3-5c331d5a2f16\") " pod="openstack/dnsmasq-dns-555cfdffc9-rw6ht" Feb 02 09:25:51 crc kubenswrapper[4764]: I0202 09:25:51.458357 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2c66c255-9454-496c-baf3-5c331d5a2f16-dns-svc\") pod \"dnsmasq-dns-555cfdffc9-rw6ht\" (UID: \"2c66c255-9454-496c-baf3-5c331d5a2f16\") " pod="openstack/dnsmasq-dns-555cfdffc9-rw6ht" Feb 02 09:25:51 crc kubenswrapper[4764]: I0202 09:25:51.458388 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d509e876-498c-4e04-95af-bdb84c334a99-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"d509e876-498c-4e04-95af-bdb84c334a99\") " pod="openstack/cinder-scheduler-0" Feb 02 09:25:51 crc kubenswrapper[4764]: I0202 09:25:51.458407 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2c66c255-9454-496c-baf3-5c331d5a2f16-config\") pod \"dnsmasq-dns-555cfdffc9-rw6ht\" (UID: \"2c66c255-9454-496c-baf3-5c331d5a2f16\") " pod="openstack/dnsmasq-dns-555cfdffc9-rw6ht" Feb 02 09:25:51 crc kubenswrapper[4764]: I0202 09:25:51.458434 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d509e876-498c-4e04-95af-bdb84c334a99-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"d509e876-498c-4e04-95af-bdb84c334a99\") " pod="openstack/cinder-scheduler-0" Feb 02 09:25:51 crc kubenswrapper[4764]: I0202 09:25:51.458459 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2c66c255-9454-496c-baf3-5c331d5a2f16-ovsdbserver-nb\") pod \"dnsmasq-dns-555cfdffc9-rw6ht\" (UID: \"2c66c255-9454-496c-baf3-5c331d5a2f16\") " pod="openstack/dnsmasq-dns-555cfdffc9-rw6ht" Feb 02 09:25:51 crc kubenswrapper[4764]: I0202 09:25:51.458483 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d509e876-498c-4e04-95af-bdb84c334a99-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"d509e876-498c-4e04-95af-bdb84c334a99\") " pod="openstack/cinder-scheduler-0" Feb 02 09:25:51 crc kubenswrapper[4764]: I0202 09:25:51.458518 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d509e876-498c-4e04-95af-bdb84c334a99-config-data\") pod \"cinder-scheduler-0\" (UID: \"d509e876-498c-4e04-95af-bdb84c334a99\") " pod="openstack/cinder-scheduler-0" Feb 02 09:25:51 crc kubenswrapper[4764]: I0202 09:25:51.458568 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tjcrn\" (UniqueName: \"kubernetes.io/projected/d509e876-498c-4e04-95af-bdb84c334a99-kube-api-access-tjcrn\") pod \"cinder-scheduler-0\" (UID: \"d509e876-498c-4e04-95af-bdb84c334a99\") " pod="openstack/cinder-scheduler-0" Feb 02 09:25:51 crc kubenswrapper[4764]: I0202 09:25:51.534151 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Feb 02 09:25:51 crc kubenswrapper[4764]: I0202 09:25:51.535545 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 02 09:25:51 crc kubenswrapper[4764]: I0202 09:25:51.543229 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Feb 02 09:25:51 crc kubenswrapper[4764]: I0202 09:25:51.556475 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Feb 02 09:25:51 crc kubenswrapper[4764]: I0202 09:25:51.560178 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d509e876-498c-4e04-95af-bdb84c334a99-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"d509e876-498c-4e04-95af-bdb84c334a99\") " pod="openstack/cinder-scheduler-0" Feb 02 09:25:51 crc kubenswrapper[4764]: I0202 09:25:51.560237 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2c66c255-9454-496c-baf3-5c331d5a2f16-config\") pod \"dnsmasq-dns-555cfdffc9-rw6ht\" (UID: \"2c66c255-9454-496c-baf3-5c331d5a2f16\") " pod="openstack/dnsmasq-dns-555cfdffc9-rw6ht" Feb 02 09:25:51 crc kubenswrapper[4764]: I0202 09:25:51.560269 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d509e876-498c-4e04-95af-bdb84c334a99-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"d509e876-498c-4e04-95af-bdb84c334a99\") " pod="openstack/cinder-scheduler-0" Feb 02 09:25:51 crc kubenswrapper[4764]: I0202 09:25:51.560291 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2c66c255-9454-496c-baf3-5c331d5a2f16-ovsdbserver-nb\") pod \"dnsmasq-dns-555cfdffc9-rw6ht\" (UID: \"2c66c255-9454-496c-baf3-5c331d5a2f16\") " pod="openstack/dnsmasq-dns-555cfdffc9-rw6ht" Feb 02 09:25:51 crc kubenswrapper[4764]: I0202 09:25:51.560316 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d509e876-498c-4e04-95af-bdb84c334a99-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"d509e876-498c-4e04-95af-bdb84c334a99\") " pod="openstack/cinder-scheduler-0" Feb 02 09:25:51 crc kubenswrapper[4764]: I0202 09:25:51.560343 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d509e876-498c-4e04-95af-bdb84c334a99-config-data\") pod \"cinder-scheduler-0\" (UID: \"d509e876-498c-4e04-95af-bdb84c334a99\") " pod="openstack/cinder-scheduler-0" Feb 02 09:25:51 crc kubenswrapper[4764]: I0202 09:25:51.560398 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tjcrn\" (UniqueName: \"kubernetes.io/projected/d509e876-498c-4e04-95af-bdb84c334a99-kube-api-access-tjcrn\") pod \"cinder-scheduler-0\" (UID: \"d509e876-498c-4e04-95af-bdb84c334a99\") " pod="openstack/cinder-scheduler-0" Feb 02 09:25:51 crc kubenswrapper[4764]: I0202 09:25:51.560451 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9wzfx\" (UniqueName: \"kubernetes.io/projected/2c66c255-9454-496c-baf3-5c331d5a2f16-kube-api-access-9wzfx\") pod \"dnsmasq-dns-555cfdffc9-rw6ht\" (UID: \"2c66c255-9454-496c-baf3-5c331d5a2f16\") " pod="openstack/dnsmasq-dns-555cfdffc9-rw6ht" Feb 02 09:25:51 crc kubenswrapper[4764]: I0202 09:25:51.560467 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d509e876-498c-4e04-95af-bdb84c334a99-scripts\") pod \"cinder-scheduler-0\" (UID: \"d509e876-498c-4e04-95af-bdb84c334a99\") " pod="openstack/cinder-scheduler-0" Feb 02 09:25:51 crc kubenswrapper[4764]: I0202 09:25:51.560484 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2c66c255-9454-496c-baf3-5c331d5a2f16-ovsdbserver-sb\") pod \"dnsmasq-dns-555cfdffc9-rw6ht\" (UID: \"2c66c255-9454-496c-baf3-5c331d5a2f16\") " pod="openstack/dnsmasq-dns-555cfdffc9-rw6ht" Feb 02 09:25:51 crc kubenswrapper[4764]: I0202 09:25:51.560505 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2c66c255-9454-496c-baf3-5c331d5a2f16-dns-svc\") pod \"dnsmasq-dns-555cfdffc9-rw6ht\" (UID: \"2c66c255-9454-496c-baf3-5c331d5a2f16\") " pod="openstack/dnsmasq-dns-555cfdffc9-rw6ht" Feb 02 09:25:51 crc kubenswrapper[4764]: I0202 09:25:51.561446 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2c66c255-9454-496c-baf3-5c331d5a2f16-dns-svc\") pod \"dnsmasq-dns-555cfdffc9-rw6ht\" (UID: \"2c66c255-9454-496c-baf3-5c331d5a2f16\") " pod="openstack/dnsmasq-dns-555cfdffc9-rw6ht" Feb 02 09:25:51 crc kubenswrapper[4764]: I0202 09:25:51.562336 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2c66c255-9454-496c-baf3-5c331d5a2f16-config\") pod \"dnsmasq-dns-555cfdffc9-rw6ht\" (UID: \"2c66c255-9454-496c-baf3-5c331d5a2f16\") " pod="openstack/dnsmasq-dns-555cfdffc9-rw6ht" Feb 02 09:25:51 crc kubenswrapper[4764]: I0202 09:25:51.562829 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2c66c255-9454-496c-baf3-5c331d5a2f16-ovsdbserver-sb\") pod \"dnsmasq-dns-555cfdffc9-rw6ht\" (UID: \"2c66c255-9454-496c-baf3-5c331d5a2f16\") " pod="openstack/dnsmasq-dns-555cfdffc9-rw6ht" Feb 02 09:25:51 crc kubenswrapper[4764]: I0202 09:25:51.563084 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d509e876-498c-4e04-95af-bdb84c334a99-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"d509e876-498c-4e04-95af-bdb84c334a99\") " pod="openstack/cinder-scheduler-0" Feb 02 09:25:51 crc kubenswrapper[4764]: I0202 09:25:51.563509 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2c66c255-9454-496c-baf3-5c331d5a2f16-ovsdbserver-nb\") pod \"dnsmasq-dns-555cfdffc9-rw6ht\" (UID: \"2c66c255-9454-496c-baf3-5c331d5a2f16\") " pod="openstack/dnsmasq-dns-555cfdffc9-rw6ht" Feb 02 09:25:51 crc kubenswrapper[4764]: I0202 09:25:51.566460 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d509e876-498c-4e04-95af-bdb84c334a99-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"d509e876-498c-4e04-95af-bdb84c334a99\") " pod="openstack/cinder-scheduler-0" Feb 02 09:25:51 crc kubenswrapper[4764]: I0202 09:25:51.579051 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d509e876-498c-4e04-95af-bdb84c334a99-scripts\") pod \"cinder-scheduler-0\" (UID: \"d509e876-498c-4e04-95af-bdb84c334a99\") " pod="openstack/cinder-scheduler-0" Feb 02 09:25:51 crc kubenswrapper[4764]: I0202 09:25:51.580420 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d509e876-498c-4e04-95af-bdb84c334a99-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"d509e876-498c-4e04-95af-bdb84c334a99\") " pod="openstack/cinder-scheduler-0" Feb 02 09:25:51 crc kubenswrapper[4764]: I0202 09:25:51.582356 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9wzfx\" (UniqueName: \"kubernetes.io/projected/2c66c255-9454-496c-baf3-5c331d5a2f16-kube-api-access-9wzfx\") pod \"dnsmasq-dns-555cfdffc9-rw6ht\" (UID: \"2c66c255-9454-496c-baf3-5c331d5a2f16\") " pod="openstack/dnsmasq-dns-555cfdffc9-rw6ht" Feb 02 09:25:51 crc kubenswrapper[4764]: I0202 09:25:51.594036 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d509e876-498c-4e04-95af-bdb84c334a99-config-data\") pod \"cinder-scheduler-0\" (UID: \"d509e876-498c-4e04-95af-bdb84c334a99\") " pod="openstack/cinder-scheduler-0" Feb 02 09:25:51 crc kubenswrapper[4764]: I0202 09:25:51.599204 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tjcrn\" (UniqueName: \"kubernetes.io/projected/d509e876-498c-4e04-95af-bdb84c334a99-kube-api-access-tjcrn\") pod \"cinder-scheduler-0\" (UID: \"d509e876-498c-4e04-95af-bdb84c334a99\") " pod="openstack/cinder-scheduler-0" Feb 02 09:25:51 crc kubenswrapper[4764]: I0202 09:25:51.662603 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6181d3f4-677f-40dd-b165-c409d72bcb7e-config-data\") pod \"cinder-api-0\" (UID: \"6181d3f4-677f-40dd-b165-c409d72bcb7e\") " pod="openstack/cinder-api-0" Feb 02 09:25:51 crc kubenswrapper[4764]: I0202 09:25:51.662677 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6181d3f4-677f-40dd-b165-c409d72bcb7e-config-data-custom\") pod \"cinder-api-0\" (UID: \"6181d3f4-677f-40dd-b165-c409d72bcb7e\") " pod="openstack/cinder-api-0" Feb 02 09:25:51 crc kubenswrapper[4764]: I0202 09:25:51.662953 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6181d3f4-677f-40dd-b165-c409d72bcb7e-scripts\") pod \"cinder-api-0\" (UID: \"6181d3f4-677f-40dd-b165-c409d72bcb7e\") " pod="openstack/cinder-api-0" Feb 02 09:25:51 crc kubenswrapper[4764]: I0202 09:25:51.663009 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6181d3f4-677f-40dd-b165-c409d72bcb7e-etc-machine-id\") pod \"cinder-api-0\" (UID: \"6181d3f4-677f-40dd-b165-c409d72bcb7e\") " pod="openstack/cinder-api-0" Feb 02 09:25:51 crc kubenswrapper[4764]: I0202 09:25:51.663155 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pkpw5\" (UniqueName: \"kubernetes.io/projected/6181d3f4-677f-40dd-b165-c409d72bcb7e-kube-api-access-pkpw5\") pod \"cinder-api-0\" (UID: \"6181d3f4-677f-40dd-b165-c409d72bcb7e\") " pod="openstack/cinder-api-0" Feb 02 09:25:51 crc kubenswrapper[4764]: I0202 09:25:51.663206 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6181d3f4-677f-40dd-b165-c409d72bcb7e-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"6181d3f4-677f-40dd-b165-c409d72bcb7e\") " pod="openstack/cinder-api-0" Feb 02 09:25:51 crc kubenswrapper[4764]: I0202 09:25:51.663296 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6181d3f4-677f-40dd-b165-c409d72bcb7e-logs\") pod \"cinder-api-0\" (UID: \"6181d3f4-677f-40dd-b165-c409d72bcb7e\") " pod="openstack/cinder-api-0" Feb 02 09:25:51 crc kubenswrapper[4764]: I0202 09:25:51.694688 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-555cfdffc9-rw6ht" Feb 02 09:25:51 crc kubenswrapper[4764]: I0202 09:25:51.764890 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6181d3f4-677f-40dd-b165-c409d72bcb7e-scripts\") pod \"cinder-api-0\" (UID: \"6181d3f4-677f-40dd-b165-c409d72bcb7e\") " pod="openstack/cinder-api-0" Feb 02 09:25:51 crc kubenswrapper[4764]: I0202 09:25:51.765276 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6181d3f4-677f-40dd-b165-c409d72bcb7e-etc-machine-id\") pod \"cinder-api-0\" (UID: \"6181d3f4-677f-40dd-b165-c409d72bcb7e\") " pod="openstack/cinder-api-0" Feb 02 09:25:51 crc kubenswrapper[4764]: I0202 09:25:51.765320 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pkpw5\" (UniqueName: \"kubernetes.io/projected/6181d3f4-677f-40dd-b165-c409d72bcb7e-kube-api-access-pkpw5\") pod \"cinder-api-0\" (UID: \"6181d3f4-677f-40dd-b165-c409d72bcb7e\") " pod="openstack/cinder-api-0" Feb 02 09:25:51 crc kubenswrapper[4764]: I0202 09:25:51.765345 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6181d3f4-677f-40dd-b165-c409d72bcb7e-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"6181d3f4-677f-40dd-b165-c409d72bcb7e\") " pod="openstack/cinder-api-0" Feb 02 09:25:51 crc kubenswrapper[4764]: I0202 09:25:51.765382 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6181d3f4-677f-40dd-b165-c409d72bcb7e-logs\") pod \"cinder-api-0\" (UID: \"6181d3f4-677f-40dd-b165-c409d72bcb7e\") " pod="openstack/cinder-api-0" Feb 02 09:25:51 crc kubenswrapper[4764]: I0202 09:25:51.765412 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6181d3f4-677f-40dd-b165-c409d72bcb7e-config-data\") pod \"cinder-api-0\" (UID: \"6181d3f4-677f-40dd-b165-c409d72bcb7e\") " pod="openstack/cinder-api-0" Feb 02 09:25:51 crc kubenswrapper[4764]: I0202 09:25:51.765440 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6181d3f4-677f-40dd-b165-c409d72bcb7e-config-data-custom\") pod \"cinder-api-0\" (UID: \"6181d3f4-677f-40dd-b165-c409d72bcb7e\") " pod="openstack/cinder-api-0" Feb 02 09:25:51 crc kubenswrapper[4764]: I0202 09:25:51.765439 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6181d3f4-677f-40dd-b165-c409d72bcb7e-etc-machine-id\") pod \"cinder-api-0\" (UID: \"6181d3f4-677f-40dd-b165-c409d72bcb7e\") " pod="openstack/cinder-api-0" Feb 02 09:25:51 crc kubenswrapper[4764]: I0202 09:25:51.765905 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6181d3f4-677f-40dd-b165-c409d72bcb7e-logs\") pod \"cinder-api-0\" (UID: \"6181d3f4-677f-40dd-b165-c409d72bcb7e\") " pod="openstack/cinder-api-0" Feb 02 09:25:51 crc kubenswrapper[4764]: I0202 09:25:51.768404 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6181d3f4-677f-40dd-b165-c409d72bcb7e-scripts\") pod \"cinder-api-0\" (UID: \"6181d3f4-677f-40dd-b165-c409d72bcb7e\") " pod="openstack/cinder-api-0" Feb 02 09:25:51 crc kubenswrapper[4764]: I0202 09:25:51.770191 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6181d3f4-677f-40dd-b165-c409d72bcb7e-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"6181d3f4-677f-40dd-b165-c409d72bcb7e\") " pod="openstack/cinder-api-0" Feb 02 09:25:51 crc kubenswrapper[4764]: I0202 09:25:51.770871 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6181d3f4-677f-40dd-b165-c409d72bcb7e-config-data-custom\") pod \"cinder-api-0\" (UID: \"6181d3f4-677f-40dd-b165-c409d72bcb7e\") " pod="openstack/cinder-api-0" Feb 02 09:25:51 crc kubenswrapper[4764]: I0202 09:25:51.771009 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6181d3f4-677f-40dd-b165-c409d72bcb7e-config-data\") pod \"cinder-api-0\" (UID: \"6181d3f4-677f-40dd-b165-c409d72bcb7e\") " pod="openstack/cinder-api-0" Feb 02 09:25:51 crc kubenswrapper[4764]: I0202 09:25:51.790247 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pkpw5\" (UniqueName: \"kubernetes.io/projected/6181d3f4-677f-40dd-b165-c409d72bcb7e-kube-api-access-pkpw5\") pod \"cinder-api-0\" (UID: \"6181d3f4-677f-40dd-b165-c409d72bcb7e\") " pod="openstack/cinder-api-0" Feb 02 09:25:51 crc kubenswrapper[4764]: I0202 09:25:51.837590 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4ebb36bb-8032-42ea-8954-7716e5a81301" path="/var/lib/kubelet/pods/4ebb36bb-8032-42ea-8954-7716e5a81301/volumes" Feb 02 09:25:51 crc kubenswrapper[4764]: I0202 09:25:51.838844 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4f12af32-00c9-437a-8e0e-805caa8df878" path="/var/lib/kubelet/pods/4f12af32-00c9-437a-8e0e-805caa8df878/volumes" Feb 02 09:25:51 crc kubenswrapper[4764]: I0202 09:25:51.857829 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 02 09:25:51 crc kubenswrapper[4764]: I0202 09:25:51.888075 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Feb 02 09:25:51 crc kubenswrapper[4764]: I0202 09:25:51.970720 4764 generic.go:334] "Generic (PLEG): container finished" podID="a518caeb-6b91-4b71-a47f-0d71d965af2e" containerID="61b9bc35586b452625d8a3057d23d4edbbe7c69316a8ec1ebf5860267f1aadd3" exitCode=0 Feb 02 09:25:51 crc kubenswrapper[4764]: I0202 09:25:51.970962 4764 generic.go:334] "Generic (PLEG): container finished" podID="a518caeb-6b91-4b71-a47f-0d71d965af2e" containerID="babe286652a8ae251bf5bcfe389a56fa30eaefd0cd0ff5f5bcb6999e85408847" exitCode=2 Feb 02 09:25:51 crc kubenswrapper[4764]: I0202 09:25:51.970980 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a518caeb-6b91-4b71-a47f-0d71d965af2e","Type":"ContainerDied","Data":"61b9bc35586b452625d8a3057d23d4edbbe7c69316a8ec1ebf5860267f1aadd3"} Feb 02 09:25:51 crc kubenswrapper[4764]: I0202 09:25:51.971000 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a518caeb-6b91-4b71-a47f-0d71d965af2e","Type":"ContainerDied","Data":"babe286652a8ae251bf5bcfe389a56fa30eaefd0cd0ff5f5bcb6999e85408847"} Feb 02 09:25:52 crc kubenswrapper[4764]: I0202 09:25:52.201757 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-555cfdffc9-rw6ht"] Feb 02 09:25:52 crc kubenswrapper[4764]: I0202 09:25:52.424163 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Feb 02 09:25:52 crc kubenswrapper[4764]: I0202 09:25:52.594586 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 02 09:25:53 crc kubenswrapper[4764]: I0202 09:25:53.015730 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"d509e876-498c-4e04-95af-bdb84c334a99","Type":"ContainerStarted","Data":"bc3465eeb4fca0708d8513f5a6cc31ab9f11387e9359a3b8c5d4c3cc5912bd7f"} Feb 02 09:25:53 crc kubenswrapper[4764]: I0202 09:25:53.032440 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"6181d3f4-677f-40dd-b165-c409d72bcb7e","Type":"ContainerStarted","Data":"c980acd8107f62ffd57dd275017aa041ffb3f29cd91353b20f5783ad5e3b12ba"} Feb 02 09:25:53 crc kubenswrapper[4764]: I0202 09:25:53.034825 4764 generic.go:334] "Generic (PLEG): container finished" podID="2c66c255-9454-496c-baf3-5c331d5a2f16" containerID="94f0f40f3100f7ed2dd2d67e2a6da28fe26a61d8d19d2128470d3558ac0876cf" exitCode=0 Feb 02 09:25:53 crc kubenswrapper[4764]: I0202 09:25:53.034854 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-555cfdffc9-rw6ht" event={"ID":"2c66c255-9454-496c-baf3-5c331d5a2f16","Type":"ContainerDied","Data":"94f0f40f3100f7ed2dd2d67e2a6da28fe26a61d8d19d2128470d3558ac0876cf"} Feb 02 09:25:53 crc kubenswrapper[4764]: I0202 09:25:53.034870 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-555cfdffc9-rw6ht" event={"ID":"2c66c255-9454-496c-baf3-5c331d5a2f16","Type":"ContainerStarted","Data":"e42ba78f4103f80a77aff2863838250ea8c45442eba5684c8498bdffd7fdaaa5"} Feb 02 09:25:53 crc kubenswrapper[4764]: I0202 09:25:53.611273 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Feb 02 09:25:54 crc kubenswrapper[4764]: I0202 09:25:54.049499 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"6181d3f4-677f-40dd-b165-c409d72bcb7e","Type":"ContainerStarted","Data":"294d44cc286817e23d8f3b1d9417fc50ada00fd33354cb6a949b1259efb86ba9"} Feb 02 09:25:54 crc kubenswrapper[4764]: I0202 09:25:54.053334 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-555cfdffc9-rw6ht" event={"ID":"2c66c255-9454-496c-baf3-5c331d5a2f16","Type":"ContainerStarted","Data":"5ce5a71993fab6383785294c33c51d3ada6114b29ce3fe0de247942ba4acefce"} Feb 02 09:25:54 crc kubenswrapper[4764]: I0202 09:25:54.054385 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-555cfdffc9-rw6ht" Feb 02 09:25:54 crc kubenswrapper[4764]: I0202 09:25:54.118588 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-797597c45b-9mkpp" Feb 02 09:25:54 crc kubenswrapper[4764]: I0202 09:25:54.142212 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-555cfdffc9-rw6ht" podStartSLOduration=3.142193444 podStartE2EDuration="3.142193444s" podCreationTimestamp="2026-02-02 09:25:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:25:54.089547639 +0000 UTC m=+1117.023271737" watchObservedRunningTime="2026-02-02 09:25:54.142193444 +0000 UTC m=+1117.075917532" Feb 02 09:25:54 crc kubenswrapper[4764]: I0202 09:25:54.469746 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-797597c45b-9mkpp" Feb 02 09:25:54 crc kubenswrapper[4764]: I0202 09:25:54.575761 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-7b6fbdbb96-fgr8m"] Feb 02 09:25:54 crc kubenswrapper[4764]: I0202 09:25:54.576147 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-7b6fbdbb96-fgr8m" podUID="40e640ad-6fb1-4c56-9b33-14fe28103b07" containerName="barbican-api-log" containerID="cri-o://dc70e7817c8136c324242956b2cb32f6723b34276a53f37132eb9a7cf1641096" gracePeriod=30 Feb 02 09:25:54 crc kubenswrapper[4764]: I0202 09:25:54.576595 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-7b6fbdbb96-fgr8m" podUID="40e640ad-6fb1-4c56-9b33-14fe28103b07" containerName="barbican-api" containerID="cri-o://23ab8753e3ef9d74bf5c41dc1f2f62af8d159606e3323372c877898d35fe5001" gracePeriod=30 Feb 02 09:25:54 crc kubenswrapper[4764]: I0202 09:25:54.622475 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-7bd869cccb-wnw6g" Feb 02 09:25:54 crc kubenswrapper[4764]: E0202 09:25:54.776808 4764 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod40e640ad_6fb1_4c56_9b33_14fe28103b07.slice/crio-dc70e7817c8136c324242956b2cb32f6723b34276a53f37132eb9a7cf1641096.scope\": RecentStats: unable to find data in memory cache]" Feb 02 09:25:54 crc kubenswrapper[4764]: I0202 09:25:54.973601 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-69b69cc889-m4fck"] Feb 02 09:25:54 crc kubenswrapper[4764]: I0202 09:25:54.974774 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-69b69cc889-m4fck" podUID="364d30e9-d4ef-4c84-b82f-673b41ffe9ff" containerName="neutron-api" containerID="cri-o://46ac0362dd33241a49276203ed445721439244f4f312c3c5389ac4126d87b43f" gracePeriod=30 Feb 02 09:25:54 crc kubenswrapper[4764]: I0202 09:25:54.978271 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-69b69cc889-m4fck" podUID="364d30e9-d4ef-4c84-b82f-673b41ffe9ff" containerName="neutron-httpd" containerID="cri-o://502205bc8cf64e06eabe340873459163683f16c8530ea72161ca0adfc975501a" gracePeriod=30 Feb 02 09:25:55 crc kubenswrapper[4764]: I0202 09:25:55.027985 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-6b9f774cdf-j8pdr"] Feb 02 09:25:55 crc kubenswrapper[4764]: I0202 09:25:55.029240 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6b9f774cdf-j8pdr" Feb 02 09:25:55 crc kubenswrapper[4764]: I0202 09:25:55.038539 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-69b69cc889-m4fck" Feb 02 09:25:55 crc kubenswrapper[4764]: I0202 09:25:55.066213 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6b9f774cdf-j8pdr"] Feb 02 09:25:55 crc kubenswrapper[4764]: I0202 09:25:55.105755 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"d509e876-498c-4e04-95af-bdb84c334a99","Type":"ContainerStarted","Data":"c301ea8b2f199dad2ba71f47cdbeb2bc337210f442107f205bf2728ba0595afd"} Feb 02 09:25:55 crc kubenswrapper[4764]: I0202 09:25:55.108617 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"6181d3f4-677f-40dd-b165-c409d72bcb7e","Type":"ContainerStarted","Data":"c6c4d99dacaf78f359aaf41aedad12c157da6d0de970ce57f1147991e1f002d5"} Feb 02 09:25:55 crc kubenswrapper[4764]: I0202 09:25:55.108743 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="6181d3f4-677f-40dd-b165-c409d72bcb7e" containerName="cinder-api-log" containerID="cri-o://294d44cc286817e23d8f3b1d9417fc50ada00fd33354cb6a949b1259efb86ba9" gracePeriod=30 Feb 02 09:25:55 crc kubenswrapper[4764]: I0202 09:25:55.109013 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Feb 02 09:25:55 crc kubenswrapper[4764]: I0202 09:25:55.109241 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="6181d3f4-677f-40dd-b165-c409d72bcb7e" containerName="cinder-api" containerID="cri-o://c6c4d99dacaf78f359aaf41aedad12c157da6d0de970ce57f1147991e1f002d5" gracePeriod=30 Feb 02 09:25:55 crc kubenswrapper[4764]: I0202 09:25:55.125297 4764 generic.go:334] "Generic (PLEG): container finished" podID="40e640ad-6fb1-4c56-9b33-14fe28103b07" containerID="dc70e7817c8136c324242956b2cb32f6723b34276a53f37132eb9a7cf1641096" exitCode=143 Feb 02 09:25:55 crc kubenswrapper[4764]: I0202 09:25:55.126122 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7b6fbdbb96-fgr8m" event={"ID":"40e640ad-6fb1-4c56-9b33-14fe28103b07","Type":"ContainerDied","Data":"dc70e7817c8136c324242956b2cb32f6723b34276a53f37132eb9a7cf1641096"} Feb 02 09:25:55 crc kubenswrapper[4764]: I0202 09:25:55.137567 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.137552519 podStartE2EDuration="4.137552519s" podCreationTimestamp="2026-02-02 09:25:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:25:55.13537965 +0000 UTC m=+1118.069103728" watchObservedRunningTime="2026-02-02 09:25:55.137552519 +0000 UTC m=+1118.071276607" Feb 02 09:25:55 crc kubenswrapper[4764]: I0202 09:25:55.140543 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c7553cbf-5c61-4003-bd4a-aad61da77950-public-tls-certs\") pod \"neutron-6b9f774cdf-j8pdr\" (UID: \"c7553cbf-5c61-4003-bd4a-aad61da77950\") " pod="openstack/neutron-6b9f774cdf-j8pdr" Feb 02 09:25:55 crc kubenswrapper[4764]: I0202 09:25:55.140601 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c7553cbf-5c61-4003-bd4a-aad61da77950-ovndb-tls-certs\") pod \"neutron-6b9f774cdf-j8pdr\" (UID: \"c7553cbf-5c61-4003-bd4a-aad61da77950\") " pod="openstack/neutron-6b9f774cdf-j8pdr" Feb 02 09:25:55 crc kubenswrapper[4764]: I0202 09:25:55.140638 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c7553cbf-5c61-4003-bd4a-aad61da77950-internal-tls-certs\") pod \"neutron-6b9f774cdf-j8pdr\" (UID: \"c7553cbf-5c61-4003-bd4a-aad61da77950\") " pod="openstack/neutron-6b9f774cdf-j8pdr" Feb 02 09:25:55 crc kubenswrapper[4764]: I0202 09:25:55.140676 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5dlww\" (UniqueName: \"kubernetes.io/projected/c7553cbf-5c61-4003-bd4a-aad61da77950-kube-api-access-5dlww\") pod \"neutron-6b9f774cdf-j8pdr\" (UID: \"c7553cbf-5c61-4003-bd4a-aad61da77950\") " pod="openstack/neutron-6b9f774cdf-j8pdr" Feb 02 09:25:55 crc kubenswrapper[4764]: I0202 09:25:55.140704 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/c7553cbf-5c61-4003-bd4a-aad61da77950-config\") pod \"neutron-6b9f774cdf-j8pdr\" (UID: \"c7553cbf-5c61-4003-bd4a-aad61da77950\") " pod="openstack/neutron-6b9f774cdf-j8pdr" Feb 02 09:25:55 crc kubenswrapper[4764]: I0202 09:25:55.140726 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c7553cbf-5c61-4003-bd4a-aad61da77950-combined-ca-bundle\") pod \"neutron-6b9f774cdf-j8pdr\" (UID: \"c7553cbf-5c61-4003-bd4a-aad61da77950\") " pod="openstack/neutron-6b9f774cdf-j8pdr" Feb 02 09:25:55 crc kubenswrapper[4764]: I0202 09:25:55.140742 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/c7553cbf-5c61-4003-bd4a-aad61da77950-httpd-config\") pod \"neutron-6b9f774cdf-j8pdr\" (UID: \"c7553cbf-5c61-4003-bd4a-aad61da77950\") " pod="openstack/neutron-6b9f774cdf-j8pdr" Feb 02 09:25:55 crc kubenswrapper[4764]: I0202 09:25:55.244267 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c7553cbf-5c61-4003-bd4a-aad61da77950-public-tls-certs\") pod \"neutron-6b9f774cdf-j8pdr\" (UID: \"c7553cbf-5c61-4003-bd4a-aad61da77950\") " pod="openstack/neutron-6b9f774cdf-j8pdr" Feb 02 09:25:55 crc kubenswrapper[4764]: I0202 09:25:55.244461 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c7553cbf-5c61-4003-bd4a-aad61da77950-ovndb-tls-certs\") pod \"neutron-6b9f774cdf-j8pdr\" (UID: \"c7553cbf-5c61-4003-bd4a-aad61da77950\") " pod="openstack/neutron-6b9f774cdf-j8pdr" Feb 02 09:25:55 crc kubenswrapper[4764]: I0202 09:25:55.244530 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c7553cbf-5c61-4003-bd4a-aad61da77950-internal-tls-certs\") pod \"neutron-6b9f774cdf-j8pdr\" (UID: \"c7553cbf-5c61-4003-bd4a-aad61da77950\") " pod="openstack/neutron-6b9f774cdf-j8pdr" Feb 02 09:25:55 crc kubenswrapper[4764]: I0202 09:25:55.244745 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5dlww\" (UniqueName: \"kubernetes.io/projected/c7553cbf-5c61-4003-bd4a-aad61da77950-kube-api-access-5dlww\") pod \"neutron-6b9f774cdf-j8pdr\" (UID: \"c7553cbf-5c61-4003-bd4a-aad61da77950\") " pod="openstack/neutron-6b9f774cdf-j8pdr" Feb 02 09:25:55 crc kubenswrapper[4764]: I0202 09:25:55.245196 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/c7553cbf-5c61-4003-bd4a-aad61da77950-config\") pod \"neutron-6b9f774cdf-j8pdr\" (UID: \"c7553cbf-5c61-4003-bd4a-aad61da77950\") " pod="openstack/neutron-6b9f774cdf-j8pdr" Feb 02 09:25:55 crc kubenswrapper[4764]: I0202 09:25:55.245250 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c7553cbf-5c61-4003-bd4a-aad61da77950-combined-ca-bundle\") pod \"neutron-6b9f774cdf-j8pdr\" (UID: \"c7553cbf-5c61-4003-bd4a-aad61da77950\") " pod="openstack/neutron-6b9f774cdf-j8pdr" Feb 02 09:25:55 crc kubenswrapper[4764]: I0202 09:25:55.245277 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/c7553cbf-5c61-4003-bd4a-aad61da77950-httpd-config\") pod \"neutron-6b9f774cdf-j8pdr\" (UID: \"c7553cbf-5c61-4003-bd4a-aad61da77950\") " pod="openstack/neutron-6b9f774cdf-j8pdr" Feb 02 09:25:55 crc kubenswrapper[4764]: I0202 09:25:55.256178 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/c7553cbf-5c61-4003-bd4a-aad61da77950-config\") pod \"neutron-6b9f774cdf-j8pdr\" (UID: \"c7553cbf-5c61-4003-bd4a-aad61da77950\") " pod="openstack/neutron-6b9f774cdf-j8pdr" Feb 02 09:25:55 crc kubenswrapper[4764]: I0202 09:25:55.258775 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c7553cbf-5c61-4003-bd4a-aad61da77950-public-tls-certs\") pod \"neutron-6b9f774cdf-j8pdr\" (UID: \"c7553cbf-5c61-4003-bd4a-aad61da77950\") " pod="openstack/neutron-6b9f774cdf-j8pdr" Feb 02 09:25:55 crc kubenswrapper[4764]: I0202 09:25:55.258800 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/c7553cbf-5c61-4003-bd4a-aad61da77950-httpd-config\") pod \"neutron-6b9f774cdf-j8pdr\" (UID: \"c7553cbf-5c61-4003-bd4a-aad61da77950\") " pod="openstack/neutron-6b9f774cdf-j8pdr" Feb 02 09:25:55 crc kubenswrapper[4764]: I0202 09:25:55.259615 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c7553cbf-5c61-4003-bd4a-aad61da77950-ovndb-tls-certs\") pod \"neutron-6b9f774cdf-j8pdr\" (UID: \"c7553cbf-5c61-4003-bd4a-aad61da77950\") " pod="openstack/neutron-6b9f774cdf-j8pdr" Feb 02 09:25:55 crc kubenswrapper[4764]: I0202 09:25:55.266660 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c7553cbf-5c61-4003-bd4a-aad61da77950-combined-ca-bundle\") pod \"neutron-6b9f774cdf-j8pdr\" (UID: \"c7553cbf-5c61-4003-bd4a-aad61da77950\") " pod="openstack/neutron-6b9f774cdf-j8pdr" Feb 02 09:25:55 crc kubenswrapper[4764]: I0202 09:25:55.267307 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c7553cbf-5c61-4003-bd4a-aad61da77950-internal-tls-certs\") pod \"neutron-6b9f774cdf-j8pdr\" (UID: \"c7553cbf-5c61-4003-bd4a-aad61da77950\") " pod="openstack/neutron-6b9f774cdf-j8pdr" Feb 02 09:25:55 crc kubenswrapper[4764]: I0202 09:25:55.280732 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5dlww\" (UniqueName: \"kubernetes.io/projected/c7553cbf-5c61-4003-bd4a-aad61da77950-kube-api-access-5dlww\") pod \"neutron-6b9f774cdf-j8pdr\" (UID: \"c7553cbf-5c61-4003-bd4a-aad61da77950\") " pod="openstack/neutron-6b9f774cdf-j8pdr" Feb 02 09:25:55 crc kubenswrapper[4764]: I0202 09:25:55.386306 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6b9f774cdf-j8pdr" Feb 02 09:25:56 crc kubenswrapper[4764]: I0202 09:25:56.022637 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6b9f774cdf-j8pdr"] Feb 02 09:25:56 crc kubenswrapper[4764]: I0202 09:25:56.139823 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"d509e876-498c-4e04-95af-bdb84c334a99","Type":"ContainerStarted","Data":"51619052573cea6bacf4e847a8df670a9036620cd4def95428f044400cefce9a"} Feb 02 09:25:56 crc kubenswrapper[4764]: I0202 09:25:56.141485 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6b9f774cdf-j8pdr" event={"ID":"c7553cbf-5c61-4003-bd4a-aad61da77950","Type":"ContainerStarted","Data":"ef0fba52970acca7a70508205d3d543d0e994f8d2d693b60723765383411ef88"} Feb 02 09:25:56 crc kubenswrapper[4764]: I0202 09:25:56.151143 4764 generic.go:334] "Generic (PLEG): container finished" podID="6181d3f4-677f-40dd-b165-c409d72bcb7e" containerID="c6c4d99dacaf78f359aaf41aedad12c157da6d0de970ce57f1147991e1f002d5" exitCode=0 Feb 02 09:25:56 crc kubenswrapper[4764]: I0202 09:25:56.151175 4764 generic.go:334] "Generic (PLEG): container finished" podID="6181d3f4-677f-40dd-b165-c409d72bcb7e" containerID="294d44cc286817e23d8f3b1d9417fc50ada00fd33354cb6a949b1259efb86ba9" exitCode=143 Feb 02 09:25:56 crc kubenswrapper[4764]: I0202 09:25:56.151233 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"6181d3f4-677f-40dd-b165-c409d72bcb7e","Type":"ContainerDied","Data":"c6c4d99dacaf78f359aaf41aedad12c157da6d0de970ce57f1147991e1f002d5"} Feb 02 09:25:56 crc kubenswrapper[4764]: I0202 09:25:56.151258 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"6181d3f4-677f-40dd-b165-c409d72bcb7e","Type":"ContainerDied","Data":"294d44cc286817e23d8f3b1d9417fc50ada00fd33354cb6a949b1259efb86ba9"} Feb 02 09:25:56 crc kubenswrapper[4764]: I0202 09:25:56.151271 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"6181d3f4-677f-40dd-b165-c409d72bcb7e","Type":"ContainerDied","Data":"c980acd8107f62ffd57dd275017aa041ffb3f29cd91353b20f5783ad5e3b12ba"} Feb 02 09:25:56 crc kubenswrapper[4764]: I0202 09:25:56.151283 4764 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c980acd8107f62ffd57dd275017aa041ffb3f29cd91353b20f5783ad5e3b12ba" Feb 02 09:25:56 crc kubenswrapper[4764]: I0202 09:25:56.160048 4764 generic.go:334] "Generic (PLEG): container finished" podID="364d30e9-d4ef-4c84-b82f-673b41ffe9ff" containerID="502205bc8cf64e06eabe340873459163683f16c8530ea72161ca0adfc975501a" exitCode=0 Feb 02 09:25:56 crc kubenswrapper[4764]: I0202 09:25:56.160951 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-69b69cc889-m4fck" event={"ID":"364d30e9-d4ef-4c84-b82f-673b41ffe9ff","Type":"ContainerDied","Data":"502205bc8cf64e06eabe340873459163683f16c8530ea72161ca0adfc975501a"} Feb 02 09:25:56 crc kubenswrapper[4764]: I0202 09:25:56.253596 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 02 09:25:56 crc kubenswrapper[4764]: I0202 09:25:56.286835 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=4.408159144 podStartE2EDuration="5.286818068s" podCreationTimestamp="2026-02-02 09:25:51 +0000 UTC" firstStartedPulling="2026-02-02 09:25:52.593898581 +0000 UTC m=+1115.527622669" lastFinishedPulling="2026-02-02 09:25:53.472557505 +0000 UTC m=+1116.406281593" observedRunningTime="2026-02-02 09:25:56.167356832 +0000 UTC m=+1119.101080920" watchObservedRunningTime="2026-02-02 09:25:56.286818068 +0000 UTC m=+1119.220542146" Feb 02 09:25:56 crc kubenswrapper[4764]: I0202 09:25:56.382407 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6181d3f4-677f-40dd-b165-c409d72bcb7e-config-data\") pod \"6181d3f4-677f-40dd-b165-c409d72bcb7e\" (UID: \"6181d3f4-677f-40dd-b165-c409d72bcb7e\") " Feb 02 09:25:56 crc kubenswrapper[4764]: I0202 09:25:56.382458 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6181d3f4-677f-40dd-b165-c409d72bcb7e-config-data-custom\") pod \"6181d3f4-677f-40dd-b165-c409d72bcb7e\" (UID: \"6181d3f4-677f-40dd-b165-c409d72bcb7e\") " Feb 02 09:25:56 crc kubenswrapper[4764]: I0202 09:25:56.382550 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6181d3f4-677f-40dd-b165-c409d72bcb7e-scripts\") pod \"6181d3f4-677f-40dd-b165-c409d72bcb7e\" (UID: \"6181d3f4-677f-40dd-b165-c409d72bcb7e\") " Feb 02 09:25:56 crc kubenswrapper[4764]: I0202 09:25:56.382586 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6181d3f4-677f-40dd-b165-c409d72bcb7e-logs\") pod \"6181d3f4-677f-40dd-b165-c409d72bcb7e\" (UID: \"6181d3f4-677f-40dd-b165-c409d72bcb7e\") " Feb 02 09:25:56 crc kubenswrapper[4764]: I0202 09:25:56.382610 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6181d3f4-677f-40dd-b165-c409d72bcb7e-etc-machine-id\") pod \"6181d3f4-677f-40dd-b165-c409d72bcb7e\" (UID: \"6181d3f4-677f-40dd-b165-c409d72bcb7e\") " Feb 02 09:25:56 crc kubenswrapper[4764]: I0202 09:25:56.382676 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pkpw5\" (UniqueName: \"kubernetes.io/projected/6181d3f4-677f-40dd-b165-c409d72bcb7e-kube-api-access-pkpw5\") pod \"6181d3f4-677f-40dd-b165-c409d72bcb7e\" (UID: \"6181d3f4-677f-40dd-b165-c409d72bcb7e\") " Feb 02 09:25:56 crc kubenswrapper[4764]: I0202 09:25:56.382697 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6181d3f4-677f-40dd-b165-c409d72bcb7e-combined-ca-bundle\") pod \"6181d3f4-677f-40dd-b165-c409d72bcb7e\" (UID: \"6181d3f4-677f-40dd-b165-c409d72bcb7e\") " Feb 02 09:25:56 crc kubenswrapper[4764]: I0202 09:25:56.384786 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6181d3f4-677f-40dd-b165-c409d72bcb7e-logs" (OuterVolumeSpecName: "logs") pod "6181d3f4-677f-40dd-b165-c409d72bcb7e" (UID: "6181d3f4-677f-40dd-b165-c409d72bcb7e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 09:25:56 crc kubenswrapper[4764]: I0202 09:25:56.385279 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6181d3f4-677f-40dd-b165-c409d72bcb7e-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "6181d3f4-677f-40dd-b165-c409d72bcb7e" (UID: "6181d3f4-677f-40dd-b165-c409d72bcb7e"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 09:25:56 crc kubenswrapper[4764]: I0202 09:25:56.390804 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6181d3f4-677f-40dd-b165-c409d72bcb7e-scripts" (OuterVolumeSpecName: "scripts") pod "6181d3f4-677f-40dd-b165-c409d72bcb7e" (UID: "6181d3f4-677f-40dd-b165-c409d72bcb7e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:25:56 crc kubenswrapper[4764]: I0202 09:25:56.407540 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6181d3f4-677f-40dd-b165-c409d72bcb7e-kube-api-access-pkpw5" (OuterVolumeSpecName: "kube-api-access-pkpw5") pod "6181d3f4-677f-40dd-b165-c409d72bcb7e" (UID: "6181d3f4-677f-40dd-b165-c409d72bcb7e"). InnerVolumeSpecName "kube-api-access-pkpw5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:25:56 crc kubenswrapper[4764]: I0202 09:25:56.407234 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6181d3f4-677f-40dd-b165-c409d72bcb7e-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "6181d3f4-677f-40dd-b165-c409d72bcb7e" (UID: "6181d3f4-677f-40dd-b165-c409d72bcb7e"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:25:56 crc kubenswrapper[4764]: I0202 09:25:56.434004 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6181d3f4-677f-40dd-b165-c409d72bcb7e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6181d3f4-677f-40dd-b165-c409d72bcb7e" (UID: "6181d3f4-677f-40dd-b165-c409d72bcb7e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:25:56 crc kubenswrapper[4764]: I0202 09:25:56.477563 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6181d3f4-677f-40dd-b165-c409d72bcb7e-config-data" (OuterVolumeSpecName: "config-data") pod "6181d3f4-677f-40dd-b165-c409d72bcb7e" (UID: "6181d3f4-677f-40dd-b165-c409d72bcb7e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:25:56 crc kubenswrapper[4764]: I0202 09:25:56.485091 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pkpw5\" (UniqueName: \"kubernetes.io/projected/6181d3f4-677f-40dd-b165-c409d72bcb7e-kube-api-access-pkpw5\") on node \"crc\" DevicePath \"\"" Feb 02 09:25:56 crc kubenswrapper[4764]: I0202 09:25:56.485115 4764 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6181d3f4-677f-40dd-b165-c409d72bcb7e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 09:25:56 crc kubenswrapper[4764]: I0202 09:25:56.485124 4764 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6181d3f4-677f-40dd-b165-c409d72bcb7e-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 09:25:56 crc kubenswrapper[4764]: I0202 09:25:56.485134 4764 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6181d3f4-677f-40dd-b165-c409d72bcb7e-config-data-custom\") on node \"crc\" DevicePath \"\"" Feb 02 09:25:56 crc kubenswrapper[4764]: I0202 09:25:56.485142 4764 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6181d3f4-677f-40dd-b165-c409d72bcb7e-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 09:25:56 crc kubenswrapper[4764]: I0202 09:25:56.485151 4764 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6181d3f4-677f-40dd-b165-c409d72bcb7e-logs\") on node \"crc\" DevicePath \"\"" Feb 02 09:25:56 crc kubenswrapper[4764]: I0202 09:25:56.485159 4764 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6181d3f4-677f-40dd-b165-c409d72bcb7e-etc-machine-id\") on node \"crc\" DevicePath \"\"" Feb 02 09:25:56 crc kubenswrapper[4764]: I0202 09:25:56.600018 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 02 09:25:56 crc kubenswrapper[4764]: I0202 09:25:56.687257 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a518caeb-6b91-4b71-a47f-0d71d965af2e-scripts\") pod \"a518caeb-6b91-4b71-a47f-0d71d965af2e\" (UID: \"a518caeb-6b91-4b71-a47f-0d71d965af2e\") " Feb 02 09:25:56 crc kubenswrapper[4764]: I0202 09:25:56.687318 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gnldx\" (UniqueName: \"kubernetes.io/projected/a518caeb-6b91-4b71-a47f-0d71d965af2e-kube-api-access-gnldx\") pod \"a518caeb-6b91-4b71-a47f-0d71d965af2e\" (UID: \"a518caeb-6b91-4b71-a47f-0d71d965af2e\") " Feb 02 09:25:56 crc kubenswrapper[4764]: I0202 09:25:56.687465 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a518caeb-6b91-4b71-a47f-0d71d965af2e-log-httpd\") pod \"a518caeb-6b91-4b71-a47f-0d71d965af2e\" (UID: \"a518caeb-6b91-4b71-a47f-0d71d965af2e\") " Feb 02 09:25:56 crc kubenswrapper[4764]: I0202 09:25:56.687491 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a518caeb-6b91-4b71-a47f-0d71d965af2e-config-data\") pod \"a518caeb-6b91-4b71-a47f-0d71d965af2e\" (UID: \"a518caeb-6b91-4b71-a47f-0d71d965af2e\") " Feb 02 09:25:56 crc kubenswrapper[4764]: I0202 09:25:56.687511 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a518caeb-6b91-4b71-a47f-0d71d965af2e-sg-core-conf-yaml\") pod \"a518caeb-6b91-4b71-a47f-0d71d965af2e\" (UID: \"a518caeb-6b91-4b71-a47f-0d71d965af2e\") " Feb 02 09:25:56 crc kubenswrapper[4764]: I0202 09:25:56.687535 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a518caeb-6b91-4b71-a47f-0d71d965af2e-run-httpd\") pod \"a518caeb-6b91-4b71-a47f-0d71d965af2e\" (UID: \"a518caeb-6b91-4b71-a47f-0d71d965af2e\") " Feb 02 09:25:56 crc kubenswrapper[4764]: I0202 09:25:56.687589 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a518caeb-6b91-4b71-a47f-0d71d965af2e-combined-ca-bundle\") pod \"a518caeb-6b91-4b71-a47f-0d71d965af2e\" (UID: \"a518caeb-6b91-4b71-a47f-0d71d965af2e\") " Feb 02 09:25:56 crc kubenswrapper[4764]: I0202 09:25:56.688881 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a518caeb-6b91-4b71-a47f-0d71d965af2e-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "a518caeb-6b91-4b71-a47f-0d71d965af2e" (UID: "a518caeb-6b91-4b71-a47f-0d71d965af2e"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 09:25:56 crc kubenswrapper[4764]: I0202 09:25:56.689078 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a518caeb-6b91-4b71-a47f-0d71d965af2e-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "a518caeb-6b91-4b71-a47f-0d71d965af2e" (UID: "a518caeb-6b91-4b71-a47f-0d71d965af2e"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 09:25:56 crc kubenswrapper[4764]: I0202 09:25:56.693083 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a518caeb-6b91-4b71-a47f-0d71d965af2e-scripts" (OuterVolumeSpecName: "scripts") pod "a518caeb-6b91-4b71-a47f-0d71d965af2e" (UID: "a518caeb-6b91-4b71-a47f-0d71d965af2e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:25:56 crc kubenswrapper[4764]: I0202 09:25:56.693227 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a518caeb-6b91-4b71-a47f-0d71d965af2e-kube-api-access-gnldx" (OuterVolumeSpecName: "kube-api-access-gnldx") pod "a518caeb-6b91-4b71-a47f-0d71d965af2e" (UID: "a518caeb-6b91-4b71-a47f-0d71d965af2e"). InnerVolumeSpecName "kube-api-access-gnldx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:25:56 crc kubenswrapper[4764]: I0202 09:25:56.743055 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a518caeb-6b91-4b71-a47f-0d71d965af2e-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "a518caeb-6b91-4b71-a47f-0d71d965af2e" (UID: "a518caeb-6b91-4b71-a47f-0d71d965af2e"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:25:56 crc kubenswrapper[4764]: I0202 09:25:56.774200 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a518caeb-6b91-4b71-a47f-0d71d965af2e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a518caeb-6b91-4b71-a47f-0d71d965af2e" (UID: "a518caeb-6b91-4b71-a47f-0d71d965af2e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:25:56 crc kubenswrapper[4764]: I0202 09:25:56.789253 4764 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a518caeb-6b91-4b71-a47f-0d71d965af2e-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 09:25:56 crc kubenswrapper[4764]: I0202 09:25:56.789397 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gnldx\" (UniqueName: \"kubernetes.io/projected/a518caeb-6b91-4b71-a47f-0d71d965af2e-kube-api-access-gnldx\") on node \"crc\" DevicePath \"\"" Feb 02 09:25:56 crc kubenswrapper[4764]: I0202 09:25:56.789475 4764 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a518caeb-6b91-4b71-a47f-0d71d965af2e-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 02 09:25:56 crc kubenswrapper[4764]: I0202 09:25:56.789546 4764 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a518caeb-6b91-4b71-a47f-0d71d965af2e-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 02 09:25:56 crc kubenswrapper[4764]: I0202 09:25:56.789602 4764 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a518caeb-6b91-4b71-a47f-0d71d965af2e-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 02 09:25:56 crc kubenswrapper[4764]: I0202 09:25:56.789672 4764 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a518caeb-6b91-4b71-a47f-0d71d965af2e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 09:25:56 crc kubenswrapper[4764]: I0202 09:25:56.790376 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a518caeb-6b91-4b71-a47f-0d71d965af2e-config-data" (OuterVolumeSpecName: "config-data") pod "a518caeb-6b91-4b71-a47f-0d71d965af2e" (UID: "a518caeb-6b91-4b71-a47f-0d71d965af2e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:25:56 crc kubenswrapper[4764]: I0202 09:25:56.889027 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Feb 02 09:25:56 crc kubenswrapper[4764]: I0202 09:25:56.891243 4764 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a518caeb-6b91-4b71-a47f-0d71d965af2e-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 09:25:56 crc kubenswrapper[4764]: I0202 09:25:56.919081 4764 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/neutron-69b69cc889-m4fck" podUID="364d30e9-d4ef-4c84-b82f-673b41ffe9ff" containerName="neutron-httpd" probeResult="failure" output="Get \"https://10.217.0.140:9696/\": dial tcp 10.217.0.140:9696: connect: connection refused" Feb 02 09:25:57 crc kubenswrapper[4764]: I0202 09:25:57.170829 4764 generic.go:334] "Generic (PLEG): container finished" podID="a518caeb-6b91-4b71-a47f-0d71d965af2e" containerID="2eb65d6eba11ecb361786b367a01950ec0b46c28f3ac4e5936e72f214d4f9acd" exitCode=0 Feb 02 09:25:57 crc kubenswrapper[4764]: I0202 09:25:57.170858 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 02 09:25:57 crc kubenswrapper[4764]: I0202 09:25:57.170875 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a518caeb-6b91-4b71-a47f-0d71d965af2e","Type":"ContainerDied","Data":"2eb65d6eba11ecb361786b367a01950ec0b46c28f3ac4e5936e72f214d4f9acd"} Feb 02 09:25:57 crc kubenswrapper[4764]: I0202 09:25:57.172399 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a518caeb-6b91-4b71-a47f-0d71d965af2e","Type":"ContainerDied","Data":"49077403dc69b61caa0b13b5af115287cd97803e61b32b6ae77094c97d68f0cf"} Feb 02 09:25:57 crc kubenswrapper[4764]: I0202 09:25:57.172439 4764 scope.go:117] "RemoveContainer" containerID="61b9bc35586b452625d8a3057d23d4edbbe7c69316a8ec1ebf5860267f1aadd3" Feb 02 09:25:57 crc kubenswrapper[4764]: I0202 09:25:57.174486 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6b9f774cdf-j8pdr" event={"ID":"c7553cbf-5c61-4003-bd4a-aad61da77950","Type":"ContainerStarted","Data":"6ac95cf2e9ee0c0ce62b29807db566647cba58fca74b9fdcb0ff8d2ea64b0e15"} Feb 02 09:25:57 crc kubenswrapper[4764]: I0202 09:25:57.174519 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6b9f774cdf-j8pdr" event={"ID":"c7553cbf-5c61-4003-bd4a-aad61da77950","Type":"ContainerStarted","Data":"17af8d7790b0d71c55ae084478c4cd06e3c85d5b862115d6169748fdf375d502"} Feb 02 09:25:57 crc kubenswrapper[4764]: I0202 09:25:57.174533 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 02 09:25:57 crc kubenswrapper[4764]: I0202 09:25:57.174776 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-6b9f774cdf-j8pdr" Feb 02 09:25:57 crc kubenswrapper[4764]: I0202 09:25:57.198540 4764 scope.go:117] "RemoveContainer" containerID="babe286652a8ae251bf5bcfe389a56fa30eaefd0cd0ff5f5bcb6999e85408847" Feb 02 09:25:57 crc kubenswrapper[4764]: I0202 09:25:57.221325 4764 scope.go:117] "RemoveContainer" containerID="2eb65d6eba11ecb361786b367a01950ec0b46c28f3ac4e5936e72f214d4f9acd" Feb 02 09:25:57 crc kubenswrapper[4764]: I0202 09:25:57.240378 4764 scope.go:117] "RemoveContainer" containerID="61b9bc35586b452625d8a3057d23d4edbbe7c69316a8ec1ebf5860267f1aadd3" Feb 02 09:25:57 crc kubenswrapper[4764]: E0202 09:25:57.244636 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"61b9bc35586b452625d8a3057d23d4edbbe7c69316a8ec1ebf5860267f1aadd3\": container with ID starting with 61b9bc35586b452625d8a3057d23d4edbbe7c69316a8ec1ebf5860267f1aadd3 not found: ID does not exist" containerID="61b9bc35586b452625d8a3057d23d4edbbe7c69316a8ec1ebf5860267f1aadd3" Feb 02 09:25:57 crc kubenswrapper[4764]: I0202 09:25:57.244758 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"61b9bc35586b452625d8a3057d23d4edbbe7c69316a8ec1ebf5860267f1aadd3"} err="failed to get container status \"61b9bc35586b452625d8a3057d23d4edbbe7c69316a8ec1ebf5860267f1aadd3\": rpc error: code = NotFound desc = could not find container \"61b9bc35586b452625d8a3057d23d4edbbe7c69316a8ec1ebf5860267f1aadd3\": container with ID starting with 61b9bc35586b452625d8a3057d23d4edbbe7c69316a8ec1ebf5860267f1aadd3 not found: ID does not exist" Feb 02 09:25:57 crc kubenswrapper[4764]: I0202 09:25:57.244889 4764 scope.go:117] "RemoveContainer" containerID="babe286652a8ae251bf5bcfe389a56fa30eaefd0cd0ff5f5bcb6999e85408847" Feb 02 09:25:57 crc kubenswrapper[4764]: E0202 09:25:57.245798 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"babe286652a8ae251bf5bcfe389a56fa30eaefd0cd0ff5f5bcb6999e85408847\": container with ID starting with babe286652a8ae251bf5bcfe389a56fa30eaefd0cd0ff5f5bcb6999e85408847 not found: ID does not exist" containerID="babe286652a8ae251bf5bcfe389a56fa30eaefd0cd0ff5f5bcb6999e85408847" Feb 02 09:25:57 crc kubenswrapper[4764]: I0202 09:25:57.245854 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"babe286652a8ae251bf5bcfe389a56fa30eaefd0cd0ff5f5bcb6999e85408847"} err="failed to get container status \"babe286652a8ae251bf5bcfe389a56fa30eaefd0cd0ff5f5bcb6999e85408847\": rpc error: code = NotFound desc = could not find container \"babe286652a8ae251bf5bcfe389a56fa30eaefd0cd0ff5f5bcb6999e85408847\": container with ID starting with babe286652a8ae251bf5bcfe389a56fa30eaefd0cd0ff5f5bcb6999e85408847 not found: ID does not exist" Feb 02 09:25:57 crc kubenswrapper[4764]: I0202 09:25:57.245887 4764 scope.go:117] "RemoveContainer" containerID="2eb65d6eba11ecb361786b367a01950ec0b46c28f3ac4e5936e72f214d4f9acd" Feb 02 09:25:57 crc kubenswrapper[4764]: E0202 09:25:57.246194 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2eb65d6eba11ecb361786b367a01950ec0b46c28f3ac4e5936e72f214d4f9acd\": container with ID starting with 2eb65d6eba11ecb361786b367a01950ec0b46c28f3ac4e5936e72f214d4f9acd not found: ID does not exist" containerID="2eb65d6eba11ecb361786b367a01950ec0b46c28f3ac4e5936e72f214d4f9acd" Feb 02 09:25:57 crc kubenswrapper[4764]: I0202 09:25:57.246271 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2eb65d6eba11ecb361786b367a01950ec0b46c28f3ac4e5936e72f214d4f9acd"} err="failed to get container status \"2eb65d6eba11ecb361786b367a01950ec0b46c28f3ac4e5936e72f214d4f9acd\": rpc error: code = NotFound desc = could not find container \"2eb65d6eba11ecb361786b367a01950ec0b46c28f3ac4e5936e72f214d4f9acd\": container with ID starting with 2eb65d6eba11ecb361786b367a01950ec0b46c28f3ac4e5936e72f214d4f9acd not found: ID does not exist" Feb 02 09:25:57 crc kubenswrapper[4764]: I0202 09:25:57.253664 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-6b9f774cdf-j8pdr" podStartSLOduration=2.253640444 podStartE2EDuration="2.253640444s" podCreationTimestamp="2026-02-02 09:25:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:25:57.208555176 +0000 UTC m=+1120.142279264" watchObservedRunningTime="2026-02-02 09:25:57.253640444 +0000 UTC m=+1120.187364532" Feb 02 09:25:57 crc kubenswrapper[4764]: I0202 09:25:57.284323 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 02 09:25:57 crc kubenswrapper[4764]: I0202 09:25:57.297277 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 02 09:25:57 crc kubenswrapper[4764]: I0202 09:25:57.305898 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Feb 02 09:25:57 crc kubenswrapper[4764]: I0202 09:25:57.323167 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Feb 02 09:25:57 crc kubenswrapper[4764]: I0202 09:25:57.328554 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 02 09:25:57 crc kubenswrapper[4764]: E0202 09:25:57.328974 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a518caeb-6b91-4b71-a47f-0d71d965af2e" containerName="sg-core" Feb 02 09:25:57 crc kubenswrapper[4764]: I0202 09:25:57.328996 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="a518caeb-6b91-4b71-a47f-0d71d965af2e" containerName="sg-core" Feb 02 09:25:57 crc kubenswrapper[4764]: E0202 09:25:57.329013 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a518caeb-6b91-4b71-a47f-0d71d965af2e" containerName="proxy-httpd" Feb 02 09:25:57 crc kubenswrapper[4764]: I0202 09:25:57.329019 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="a518caeb-6b91-4b71-a47f-0d71d965af2e" containerName="proxy-httpd" Feb 02 09:25:57 crc kubenswrapper[4764]: E0202 09:25:57.329040 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6181d3f4-677f-40dd-b165-c409d72bcb7e" containerName="cinder-api" Feb 02 09:25:57 crc kubenswrapper[4764]: I0202 09:25:57.329046 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="6181d3f4-677f-40dd-b165-c409d72bcb7e" containerName="cinder-api" Feb 02 09:25:57 crc kubenswrapper[4764]: E0202 09:25:57.329059 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a518caeb-6b91-4b71-a47f-0d71d965af2e" containerName="ceilometer-notification-agent" Feb 02 09:25:57 crc kubenswrapper[4764]: I0202 09:25:57.329065 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="a518caeb-6b91-4b71-a47f-0d71d965af2e" containerName="ceilometer-notification-agent" Feb 02 09:25:57 crc kubenswrapper[4764]: E0202 09:25:57.329085 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6181d3f4-677f-40dd-b165-c409d72bcb7e" containerName="cinder-api-log" Feb 02 09:25:57 crc kubenswrapper[4764]: I0202 09:25:57.329091 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="6181d3f4-677f-40dd-b165-c409d72bcb7e" containerName="cinder-api-log" Feb 02 09:25:57 crc kubenswrapper[4764]: I0202 09:25:57.329246 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="a518caeb-6b91-4b71-a47f-0d71d965af2e" containerName="sg-core" Feb 02 09:25:57 crc kubenswrapper[4764]: I0202 09:25:57.329258 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="6181d3f4-677f-40dd-b165-c409d72bcb7e" containerName="cinder-api-log" Feb 02 09:25:57 crc kubenswrapper[4764]: I0202 09:25:57.329271 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="6181d3f4-677f-40dd-b165-c409d72bcb7e" containerName="cinder-api" Feb 02 09:25:57 crc kubenswrapper[4764]: I0202 09:25:57.329283 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="a518caeb-6b91-4b71-a47f-0d71d965af2e" containerName="proxy-httpd" Feb 02 09:25:57 crc kubenswrapper[4764]: I0202 09:25:57.329294 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="a518caeb-6b91-4b71-a47f-0d71d965af2e" containerName="ceilometer-notification-agent" Feb 02 09:25:57 crc kubenswrapper[4764]: I0202 09:25:57.332101 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 02 09:25:57 crc kubenswrapper[4764]: I0202 09:25:57.334855 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 02 09:25:57 crc kubenswrapper[4764]: I0202 09:25:57.336175 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 02 09:25:57 crc kubenswrapper[4764]: I0202 09:25:57.336841 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Feb 02 09:25:57 crc kubenswrapper[4764]: I0202 09:25:57.340503 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 02 09:25:57 crc kubenswrapper[4764]: I0202 09:25:57.342639 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Feb 02 09:25:57 crc kubenswrapper[4764]: I0202 09:25:57.346087 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Feb 02 09:25:57 crc kubenswrapper[4764]: I0202 09:25:57.346240 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Feb 02 09:25:57 crc kubenswrapper[4764]: I0202 09:25:57.346384 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Feb 02 09:25:57 crc kubenswrapper[4764]: I0202 09:25:57.348602 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 02 09:25:57 crc kubenswrapper[4764]: I0202 09:25:57.399781 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a132428-772f-4d1a-bd11-bf9f69301bfa-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2a132428-772f-4d1a-bd11-bf9f69301bfa\") " pod="openstack/ceilometer-0" Feb 02 09:25:57 crc kubenswrapper[4764]: I0202 09:25:57.399820 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c74wc\" (UniqueName: \"kubernetes.io/projected/2a132428-772f-4d1a-bd11-bf9f69301bfa-kube-api-access-c74wc\") pod \"ceilometer-0\" (UID: \"2a132428-772f-4d1a-bd11-bf9f69301bfa\") " pod="openstack/ceilometer-0" Feb 02 09:25:57 crc kubenswrapper[4764]: I0202 09:25:57.399843 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b8e0fd6-1949-4b4f-bbc5-b663c51f08c4-config-data\") pod \"cinder-api-0\" (UID: \"4b8e0fd6-1949-4b4f-bbc5-b663c51f08c4\") " pod="openstack/cinder-api-0" Feb 02 09:25:57 crc kubenswrapper[4764]: I0202 09:25:57.399858 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2a132428-772f-4d1a-bd11-bf9f69301bfa-log-httpd\") pod \"ceilometer-0\" (UID: \"2a132428-772f-4d1a-bd11-bf9f69301bfa\") " pod="openstack/ceilometer-0" Feb 02 09:25:57 crc kubenswrapper[4764]: I0202 09:25:57.399900 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2a132428-772f-4d1a-bd11-bf9f69301bfa-run-httpd\") pod \"ceilometer-0\" (UID: \"2a132428-772f-4d1a-bd11-bf9f69301bfa\") " pod="openstack/ceilometer-0" Feb 02 09:25:57 crc kubenswrapper[4764]: I0202 09:25:57.399916 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b8e0fd6-1949-4b4f-bbc5-b663c51f08c4-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"4b8e0fd6-1949-4b4f-bbc5-b663c51f08c4\") " pod="openstack/cinder-api-0" Feb 02 09:25:57 crc kubenswrapper[4764]: I0202 09:25:57.399956 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s9chv\" (UniqueName: \"kubernetes.io/projected/4b8e0fd6-1949-4b4f-bbc5-b663c51f08c4-kube-api-access-s9chv\") pod \"cinder-api-0\" (UID: \"4b8e0fd6-1949-4b4f-bbc5-b663c51f08c4\") " pod="openstack/cinder-api-0" Feb 02 09:25:57 crc kubenswrapper[4764]: I0202 09:25:57.399982 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2a132428-772f-4d1a-bd11-bf9f69301bfa-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2a132428-772f-4d1a-bd11-bf9f69301bfa\") " pod="openstack/ceilometer-0" Feb 02 09:25:57 crc kubenswrapper[4764]: I0202 09:25:57.400012 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2a132428-772f-4d1a-bd11-bf9f69301bfa-scripts\") pod \"ceilometer-0\" (UID: \"2a132428-772f-4d1a-bd11-bf9f69301bfa\") " pod="openstack/ceilometer-0" Feb 02 09:25:57 crc kubenswrapper[4764]: I0202 09:25:57.400031 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a132428-772f-4d1a-bd11-bf9f69301bfa-config-data\") pod \"ceilometer-0\" (UID: \"2a132428-772f-4d1a-bd11-bf9f69301bfa\") " pod="openstack/ceilometer-0" Feb 02 09:25:57 crc kubenswrapper[4764]: I0202 09:25:57.400069 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4b8e0fd6-1949-4b4f-bbc5-b663c51f08c4-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"4b8e0fd6-1949-4b4f-bbc5-b663c51f08c4\") " pod="openstack/cinder-api-0" Feb 02 09:25:57 crc kubenswrapper[4764]: I0202 09:25:57.400088 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4b8e0fd6-1949-4b4f-bbc5-b663c51f08c4-etc-machine-id\") pod \"cinder-api-0\" (UID: \"4b8e0fd6-1949-4b4f-bbc5-b663c51f08c4\") " pod="openstack/cinder-api-0" Feb 02 09:25:57 crc kubenswrapper[4764]: I0202 09:25:57.400142 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4b8e0fd6-1949-4b4f-bbc5-b663c51f08c4-logs\") pod \"cinder-api-0\" (UID: \"4b8e0fd6-1949-4b4f-bbc5-b663c51f08c4\") " pod="openstack/cinder-api-0" Feb 02 09:25:57 crc kubenswrapper[4764]: I0202 09:25:57.400167 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4b8e0fd6-1949-4b4f-bbc5-b663c51f08c4-public-tls-certs\") pod \"cinder-api-0\" (UID: \"4b8e0fd6-1949-4b4f-bbc5-b663c51f08c4\") " pod="openstack/cinder-api-0" Feb 02 09:25:57 crc kubenswrapper[4764]: I0202 09:25:57.400187 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4b8e0fd6-1949-4b4f-bbc5-b663c51f08c4-scripts\") pod \"cinder-api-0\" (UID: \"4b8e0fd6-1949-4b4f-bbc5-b663c51f08c4\") " pod="openstack/cinder-api-0" Feb 02 09:25:57 crc kubenswrapper[4764]: I0202 09:25:57.400219 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4b8e0fd6-1949-4b4f-bbc5-b663c51f08c4-config-data-custom\") pod \"cinder-api-0\" (UID: \"4b8e0fd6-1949-4b4f-bbc5-b663c51f08c4\") " pod="openstack/cinder-api-0" Feb 02 09:25:57 crc kubenswrapper[4764]: I0202 09:25:57.502173 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b8e0fd6-1949-4b4f-bbc5-b663c51f08c4-config-data\") pod \"cinder-api-0\" (UID: \"4b8e0fd6-1949-4b4f-bbc5-b663c51f08c4\") " pod="openstack/cinder-api-0" Feb 02 09:25:57 crc kubenswrapper[4764]: I0202 09:25:57.502450 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2a132428-772f-4d1a-bd11-bf9f69301bfa-log-httpd\") pod \"ceilometer-0\" (UID: \"2a132428-772f-4d1a-bd11-bf9f69301bfa\") " pod="openstack/ceilometer-0" Feb 02 09:25:57 crc kubenswrapper[4764]: I0202 09:25:57.502526 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2a132428-772f-4d1a-bd11-bf9f69301bfa-run-httpd\") pod \"ceilometer-0\" (UID: \"2a132428-772f-4d1a-bd11-bf9f69301bfa\") " pod="openstack/ceilometer-0" Feb 02 09:25:57 crc kubenswrapper[4764]: I0202 09:25:57.502597 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b8e0fd6-1949-4b4f-bbc5-b663c51f08c4-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"4b8e0fd6-1949-4b4f-bbc5-b663c51f08c4\") " pod="openstack/cinder-api-0" Feb 02 09:25:57 crc kubenswrapper[4764]: I0202 09:25:57.502683 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s9chv\" (UniqueName: \"kubernetes.io/projected/4b8e0fd6-1949-4b4f-bbc5-b663c51f08c4-kube-api-access-s9chv\") pod \"cinder-api-0\" (UID: \"4b8e0fd6-1949-4b4f-bbc5-b663c51f08c4\") " pod="openstack/cinder-api-0" Feb 02 09:25:57 crc kubenswrapper[4764]: I0202 09:25:57.502756 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2a132428-772f-4d1a-bd11-bf9f69301bfa-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2a132428-772f-4d1a-bd11-bf9f69301bfa\") " pod="openstack/ceilometer-0" Feb 02 09:25:57 crc kubenswrapper[4764]: I0202 09:25:57.502847 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2a132428-772f-4d1a-bd11-bf9f69301bfa-scripts\") pod \"ceilometer-0\" (UID: \"2a132428-772f-4d1a-bd11-bf9f69301bfa\") " pod="openstack/ceilometer-0" Feb 02 09:25:57 crc kubenswrapper[4764]: I0202 09:25:57.502917 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a132428-772f-4d1a-bd11-bf9f69301bfa-config-data\") pod \"ceilometer-0\" (UID: \"2a132428-772f-4d1a-bd11-bf9f69301bfa\") " pod="openstack/ceilometer-0" Feb 02 09:25:57 crc kubenswrapper[4764]: I0202 09:25:57.503029 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4b8e0fd6-1949-4b4f-bbc5-b663c51f08c4-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"4b8e0fd6-1949-4b4f-bbc5-b663c51f08c4\") " pod="openstack/cinder-api-0" Feb 02 09:25:57 crc kubenswrapper[4764]: I0202 09:25:57.503110 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4b8e0fd6-1949-4b4f-bbc5-b663c51f08c4-etc-machine-id\") pod \"cinder-api-0\" (UID: \"4b8e0fd6-1949-4b4f-bbc5-b663c51f08c4\") " pod="openstack/cinder-api-0" Feb 02 09:25:57 crc kubenswrapper[4764]: I0202 09:25:57.503194 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4b8e0fd6-1949-4b4f-bbc5-b663c51f08c4-logs\") pod \"cinder-api-0\" (UID: \"4b8e0fd6-1949-4b4f-bbc5-b663c51f08c4\") " pod="openstack/cinder-api-0" Feb 02 09:25:57 crc kubenswrapper[4764]: I0202 09:25:57.503269 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4b8e0fd6-1949-4b4f-bbc5-b663c51f08c4-public-tls-certs\") pod \"cinder-api-0\" (UID: \"4b8e0fd6-1949-4b4f-bbc5-b663c51f08c4\") " pod="openstack/cinder-api-0" Feb 02 09:25:57 crc kubenswrapper[4764]: I0202 09:25:57.503339 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4b8e0fd6-1949-4b4f-bbc5-b663c51f08c4-scripts\") pod \"cinder-api-0\" (UID: \"4b8e0fd6-1949-4b4f-bbc5-b663c51f08c4\") " pod="openstack/cinder-api-0" Feb 02 09:25:57 crc kubenswrapper[4764]: I0202 09:25:57.503415 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4b8e0fd6-1949-4b4f-bbc5-b663c51f08c4-config-data-custom\") pod \"cinder-api-0\" (UID: \"4b8e0fd6-1949-4b4f-bbc5-b663c51f08c4\") " pod="openstack/cinder-api-0" Feb 02 09:25:57 crc kubenswrapper[4764]: I0202 09:25:57.503490 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a132428-772f-4d1a-bd11-bf9f69301bfa-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2a132428-772f-4d1a-bd11-bf9f69301bfa\") " pod="openstack/ceilometer-0" Feb 02 09:25:57 crc kubenswrapper[4764]: I0202 09:25:57.503557 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c74wc\" (UniqueName: \"kubernetes.io/projected/2a132428-772f-4d1a-bd11-bf9f69301bfa-kube-api-access-c74wc\") pod \"ceilometer-0\" (UID: \"2a132428-772f-4d1a-bd11-bf9f69301bfa\") " pod="openstack/ceilometer-0" Feb 02 09:25:57 crc kubenswrapper[4764]: I0202 09:25:57.504377 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2a132428-772f-4d1a-bd11-bf9f69301bfa-run-httpd\") pod \"ceilometer-0\" (UID: \"2a132428-772f-4d1a-bd11-bf9f69301bfa\") " pod="openstack/ceilometer-0" Feb 02 09:25:57 crc kubenswrapper[4764]: I0202 09:25:57.504429 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2a132428-772f-4d1a-bd11-bf9f69301bfa-log-httpd\") pod \"ceilometer-0\" (UID: \"2a132428-772f-4d1a-bd11-bf9f69301bfa\") " pod="openstack/ceilometer-0" Feb 02 09:25:57 crc kubenswrapper[4764]: I0202 09:25:57.504782 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4b8e0fd6-1949-4b4f-bbc5-b663c51f08c4-logs\") pod \"cinder-api-0\" (UID: \"4b8e0fd6-1949-4b4f-bbc5-b663c51f08c4\") " pod="openstack/cinder-api-0" Feb 02 09:25:57 crc kubenswrapper[4764]: I0202 09:25:57.506596 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4b8e0fd6-1949-4b4f-bbc5-b663c51f08c4-etc-machine-id\") pod \"cinder-api-0\" (UID: \"4b8e0fd6-1949-4b4f-bbc5-b663c51f08c4\") " pod="openstack/cinder-api-0" Feb 02 09:25:57 crc kubenswrapper[4764]: I0202 09:25:57.507760 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b8e0fd6-1949-4b4f-bbc5-b663c51f08c4-config-data\") pod \"cinder-api-0\" (UID: \"4b8e0fd6-1949-4b4f-bbc5-b663c51f08c4\") " pod="openstack/cinder-api-0" Feb 02 09:25:57 crc kubenswrapper[4764]: I0202 09:25:57.508264 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2a132428-772f-4d1a-bd11-bf9f69301bfa-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2a132428-772f-4d1a-bd11-bf9f69301bfa\") " pod="openstack/ceilometer-0" Feb 02 09:25:57 crc kubenswrapper[4764]: I0202 09:25:57.509179 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4b8e0fd6-1949-4b4f-bbc5-b663c51f08c4-config-data-custom\") pod \"cinder-api-0\" (UID: \"4b8e0fd6-1949-4b4f-bbc5-b663c51f08c4\") " pod="openstack/cinder-api-0" Feb 02 09:25:57 crc kubenswrapper[4764]: I0202 09:25:57.511357 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4b8e0fd6-1949-4b4f-bbc5-b663c51f08c4-public-tls-certs\") pod \"cinder-api-0\" (UID: \"4b8e0fd6-1949-4b4f-bbc5-b663c51f08c4\") " pod="openstack/cinder-api-0" Feb 02 09:25:57 crc kubenswrapper[4764]: I0202 09:25:57.518658 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2a132428-772f-4d1a-bd11-bf9f69301bfa-scripts\") pod \"ceilometer-0\" (UID: \"2a132428-772f-4d1a-bd11-bf9f69301bfa\") " pod="openstack/ceilometer-0" Feb 02 09:25:57 crc kubenswrapper[4764]: I0202 09:25:57.519317 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4b8e0fd6-1949-4b4f-bbc5-b663c51f08c4-scripts\") pod \"cinder-api-0\" (UID: \"4b8e0fd6-1949-4b4f-bbc5-b663c51f08c4\") " pod="openstack/cinder-api-0" Feb 02 09:25:57 crc kubenswrapper[4764]: I0202 09:25:57.519770 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a132428-772f-4d1a-bd11-bf9f69301bfa-config-data\") pod \"ceilometer-0\" (UID: \"2a132428-772f-4d1a-bd11-bf9f69301bfa\") " pod="openstack/ceilometer-0" Feb 02 09:25:57 crc kubenswrapper[4764]: I0202 09:25:57.519894 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a132428-772f-4d1a-bd11-bf9f69301bfa-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2a132428-772f-4d1a-bd11-bf9f69301bfa\") " pod="openstack/ceilometer-0" Feb 02 09:25:57 crc kubenswrapper[4764]: I0202 09:25:57.521316 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4b8e0fd6-1949-4b4f-bbc5-b663c51f08c4-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"4b8e0fd6-1949-4b4f-bbc5-b663c51f08c4\") " pod="openstack/cinder-api-0" Feb 02 09:25:57 crc kubenswrapper[4764]: I0202 09:25:57.521606 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b8e0fd6-1949-4b4f-bbc5-b663c51f08c4-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"4b8e0fd6-1949-4b4f-bbc5-b663c51f08c4\") " pod="openstack/cinder-api-0" Feb 02 09:25:57 crc kubenswrapper[4764]: I0202 09:25:57.526491 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c74wc\" (UniqueName: \"kubernetes.io/projected/2a132428-772f-4d1a-bd11-bf9f69301bfa-kube-api-access-c74wc\") pod \"ceilometer-0\" (UID: \"2a132428-772f-4d1a-bd11-bf9f69301bfa\") " pod="openstack/ceilometer-0" Feb 02 09:25:57 crc kubenswrapper[4764]: I0202 09:25:57.532742 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s9chv\" (UniqueName: \"kubernetes.io/projected/4b8e0fd6-1949-4b4f-bbc5-b663c51f08c4-kube-api-access-s9chv\") pod \"cinder-api-0\" (UID: \"4b8e0fd6-1949-4b4f-bbc5-b663c51f08c4\") " pod="openstack/cinder-api-0" Feb 02 09:25:57 crc kubenswrapper[4764]: I0202 09:25:57.655877 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 02 09:25:57 crc kubenswrapper[4764]: I0202 09:25:57.668497 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 02 09:25:57 crc kubenswrapper[4764]: I0202 09:25:57.842403 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6181d3f4-677f-40dd-b165-c409d72bcb7e" path="/var/lib/kubelet/pods/6181d3f4-677f-40dd-b165-c409d72bcb7e/volumes" Feb 02 09:25:57 crc kubenswrapper[4764]: I0202 09:25:57.843858 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a518caeb-6b91-4b71-a47f-0d71d965af2e" path="/var/lib/kubelet/pods/a518caeb-6b91-4b71-a47f-0d71d965af2e/volumes" Feb 02 09:25:58 crc kubenswrapper[4764]: I0202 09:25:58.043793 4764 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-7b6fbdbb96-fgr8m" podUID="40e640ad-6fb1-4c56-9b33-14fe28103b07" containerName="barbican-api" probeResult="failure" output="Get \"https://10.217.0.146:9311/healthcheck\": read tcp 10.217.0.2:55706->10.217.0.146:9311: read: connection reset by peer" Feb 02 09:25:58 crc kubenswrapper[4764]: I0202 09:25:58.043814 4764 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-7b6fbdbb96-fgr8m" podUID="40e640ad-6fb1-4c56-9b33-14fe28103b07" containerName="barbican-api-log" probeResult="failure" output="Get \"https://10.217.0.146:9311/healthcheck\": read tcp 10.217.0.2:55720->10.217.0.146:9311: read: connection reset by peer" Feb 02 09:25:58 crc kubenswrapper[4764]: I0202 09:25:58.149801 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 02 09:25:58 crc kubenswrapper[4764]: W0202 09:25:58.160229 4764 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2a132428_772f_4d1a_bd11_bf9f69301bfa.slice/crio-0d9e5b254ddcc1904aeffaf74ca61243210dbd9838a95eb26b14a4f6fccb0093 WatchSource:0}: Error finding container 0d9e5b254ddcc1904aeffaf74ca61243210dbd9838a95eb26b14a4f6fccb0093: Status 404 returned error can't find the container with id 0d9e5b254ddcc1904aeffaf74ca61243210dbd9838a95eb26b14a4f6fccb0093 Feb 02 09:25:58 crc kubenswrapper[4764]: I0202 09:25:58.185035 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2a132428-772f-4d1a-bd11-bf9f69301bfa","Type":"ContainerStarted","Data":"0d9e5b254ddcc1904aeffaf74ca61243210dbd9838a95eb26b14a4f6fccb0093"} Feb 02 09:25:58 crc kubenswrapper[4764]: I0202 09:25:58.190296 4764 generic.go:334] "Generic (PLEG): container finished" podID="40e640ad-6fb1-4c56-9b33-14fe28103b07" containerID="23ab8753e3ef9d74bf5c41dc1f2f62af8d159606e3323372c877898d35fe5001" exitCode=0 Feb 02 09:25:58 crc kubenswrapper[4764]: I0202 09:25:58.190349 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7b6fbdbb96-fgr8m" event={"ID":"40e640ad-6fb1-4c56-9b33-14fe28103b07","Type":"ContainerDied","Data":"23ab8753e3ef9d74bf5c41dc1f2f62af8d159606e3323372c877898d35fe5001"} Feb 02 09:25:58 crc kubenswrapper[4764]: I0202 09:25:58.285264 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Feb 02 09:25:58 crc kubenswrapper[4764]: I0202 09:25:58.449151 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7b6fbdbb96-fgr8m" Feb 02 09:25:58 crc kubenswrapper[4764]: I0202 09:25:58.523654 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/40e640ad-6fb1-4c56-9b33-14fe28103b07-logs\") pod \"40e640ad-6fb1-4c56-9b33-14fe28103b07\" (UID: \"40e640ad-6fb1-4c56-9b33-14fe28103b07\") " Feb 02 09:25:58 crc kubenswrapper[4764]: I0202 09:25:58.523713 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/40e640ad-6fb1-4c56-9b33-14fe28103b07-config-data\") pod \"40e640ad-6fb1-4c56-9b33-14fe28103b07\" (UID: \"40e640ad-6fb1-4c56-9b33-14fe28103b07\") " Feb 02 09:25:58 crc kubenswrapper[4764]: I0202 09:25:58.523765 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40e640ad-6fb1-4c56-9b33-14fe28103b07-combined-ca-bundle\") pod \"40e640ad-6fb1-4c56-9b33-14fe28103b07\" (UID: \"40e640ad-6fb1-4c56-9b33-14fe28103b07\") " Feb 02 09:25:58 crc kubenswrapper[4764]: I0202 09:25:58.523797 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9sxl9\" (UniqueName: \"kubernetes.io/projected/40e640ad-6fb1-4c56-9b33-14fe28103b07-kube-api-access-9sxl9\") pod \"40e640ad-6fb1-4c56-9b33-14fe28103b07\" (UID: \"40e640ad-6fb1-4c56-9b33-14fe28103b07\") " Feb 02 09:25:58 crc kubenswrapper[4764]: I0202 09:25:58.523833 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/40e640ad-6fb1-4c56-9b33-14fe28103b07-internal-tls-certs\") pod \"40e640ad-6fb1-4c56-9b33-14fe28103b07\" (UID: \"40e640ad-6fb1-4c56-9b33-14fe28103b07\") " Feb 02 09:25:58 crc kubenswrapper[4764]: I0202 09:25:58.523899 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/40e640ad-6fb1-4c56-9b33-14fe28103b07-config-data-custom\") pod \"40e640ad-6fb1-4c56-9b33-14fe28103b07\" (UID: \"40e640ad-6fb1-4c56-9b33-14fe28103b07\") " Feb 02 09:25:58 crc kubenswrapper[4764]: I0202 09:25:58.523945 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/40e640ad-6fb1-4c56-9b33-14fe28103b07-public-tls-certs\") pod \"40e640ad-6fb1-4c56-9b33-14fe28103b07\" (UID: \"40e640ad-6fb1-4c56-9b33-14fe28103b07\") " Feb 02 09:25:58 crc kubenswrapper[4764]: I0202 09:25:58.529157 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/40e640ad-6fb1-4c56-9b33-14fe28103b07-logs" (OuterVolumeSpecName: "logs") pod "40e640ad-6fb1-4c56-9b33-14fe28103b07" (UID: "40e640ad-6fb1-4c56-9b33-14fe28103b07"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 09:25:58 crc kubenswrapper[4764]: I0202 09:25:58.532183 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/40e640ad-6fb1-4c56-9b33-14fe28103b07-kube-api-access-9sxl9" (OuterVolumeSpecName: "kube-api-access-9sxl9") pod "40e640ad-6fb1-4c56-9b33-14fe28103b07" (UID: "40e640ad-6fb1-4c56-9b33-14fe28103b07"). InnerVolumeSpecName "kube-api-access-9sxl9". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:25:58 crc kubenswrapper[4764]: I0202 09:25:58.535870 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/40e640ad-6fb1-4c56-9b33-14fe28103b07-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "40e640ad-6fb1-4c56-9b33-14fe28103b07" (UID: "40e640ad-6fb1-4c56-9b33-14fe28103b07"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:25:58 crc kubenswrapper[4764]: I0202 09:25:58.553206 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/40e640ad-6fb1-4c56-9b33-14fe28103b07-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "40e640ad-6fb1-4c56-9b33-14fe28103b07" (UID: "40e640ad-6fb1-4c56-9b33-14fe28103b07"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:25:58 crc kubenswrapper[4764]: I0202 09:25:58.575566 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/40e640ad-6fb1-4c56-9b33-14fe28103b07-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "40e640ad-6fb1-4c56-9b33-14fe28103b07" (UID: "40e640ad-6fb1-4c56-9b33-14fe28103b07"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:25:58 crc kubenswrapper[4764]: I0202 09:25:58.581519 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/40e640ad-6fb1-4c56-9b33-14fe28103b07-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "40e640ad-6fb1-4c56-9b33-14fe28103b07" (UID: "40e640ad-6fb1-4c56-9b33-14fe28103b07"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:25:58 crc kubenswrapper[4764]: I0202 09:25:58.600704 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/40e640ad-6fb1-4c56-9b33-14fe28103b07-config-data" (OuterVolumeSpecName: "config-data") pod "40e640ad-6fb1-4c56-9b33-14fe28103b07" (UID: "40e640ad-6fb1-4c56-9b33-14fe28103b07"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:25:58 crc kubenswrapper[4764]: I0202 09:25:58.626195 4764 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/40e640ad-6fb1-4c56-9b33-14fe28103b07-logs\") on node \"crc\" DevicePath \"\"" Feb 02 09:25:58 crc kubenswrapper[4764]: I0202 09:25:58.626222 4764 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/40e640ad-6fb1-4c56-9b33-14fe28103b07-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 09:25:58 crc kubenswrapper[4764]: I0202 09:25:58.626233 4764 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40e640ad-6fb1-4c56-9b33-14fe28103b07-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 09:25:58 crc kubenswrapper[4764]: I0202 09:25:58.626246 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9sxl9\" (UniqueName: \"kubernetes.io/projected/40e640ad-6fb1-4c56-9b33-14fe28103b07-kube-api-access-9sxl9\") on node \"crc\" DevicePath \"\"" Feb 02 09:25:58 crc kubenswrapper[4764]: I0202 09:25:58.626255 4764 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/40e640ad-6fb1-4c56-9b33-14fe28103b07-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 02 09:25:58 crc kubenswrapper[4764]: I0202 09:25:58.626263 4764 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/40e640ad-6fb1-4c56-9b33-14fe28103b07-config-data-custom\") on node \"crc\" DevicePath \"\"" Feb 02 09:25:58 crc kubenswrapper[4764]: I0202 09:25:58.626270 4764 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/40e640ad-6fb1-4c56-9b33-14fe28103b07-public-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 02 09:25:59 crc kubenswrapper[4764]: I0202 09:25:59.203060 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7b6fbdbb96-fgr8m" event={"ID":"40e640ad-6fb1-4c56-9b33-14fe28103b07","Type":"ContainerDied","Data":"d67e68a11ee8449b0e97f4f27db9fd2f842d895b18942e639dfc4dc74232b8c0"} Feb 02 09:25:59 crc kubenswrapper[4764]: I0202 09:25:59.203364 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7b6fbdbb96-fgr8m" Feb 02 09:25:59 crc kubenswrapper[4764]: I0202 09:25:59.203378 4764 scope.go:117] "RemoveContainer" containerID="23ab8753e3ef9d74bf5c41dc1f2f62af8d159606e3323372c877898d35fe5001" Feb 02 09:25:59 crc kubenswrapper[4764]: I0202 09:25:59.207645 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"4b8e0fd6-1949-4b4f-bbc5-b663c51f08c4","Type":"ContainerStarted","Data":"6e0eadab2e8e98709c9fbc3eb71154933b5292da6a74f2c3a4a79d6c5f75714b"} Feb 02 09:25:59 crc kubenswrapper[4764]: I0202 09:25:59.207684 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"4b8e0fd6-1949-4b4f-bbc5-b663c51f08c4","Type":"ContainerStarted","Data":"2f9537fb3c1aa90233f64654b6c357eabcc8993a2a54e63422b1e10ac03ba546"} Feb 02 09:25:59 crc kubenswrapper[4764]: I0202 09:25:59.211820 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2a132428-772f-4d1a-bd11-bf9f69301bfa","Type":"ContainerStarted","Data":"3386c5db9eb0a841c11d8c70caac741e63d302e60b74f18c22e2bfb79426d277"} Feb 02 09:25:59 crc kubenswrapper[4764]: I0202 09:25:59.234219 4764 scope.go:117] "RemoveContainer" containerID="dc70e7817c8136c324242956b2cb32f6723b34276a53f37132eb9a7cf1641096" Feb 02 09:25:59 crc kubenswrapper[4764]: I0202 09:25:59.251069 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-7b6fbdbb96-fgr8m"] Feb 02 09:25:59 crc kubenswrapper[4764]: I0202 09:25:59.260163 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-7b6fbdbb96-fgr8m"] Feb 02 09:25:59 crc kubenswrapper[4764]: I0202 09:25:59.839693 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="40e640ad-6fb1-4c56-9b33-14fe28103b07" path="/var/lib/kubelet/pods/40e640ad-6fb1-4c56-9b33-14fe28103b07/volumes" Feb 02 09:26:00 crc kubenswrapper[4764]: I0202 09:26:00.240626 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"4b8e0fd6-1949-4b4f-bbc5-b663c51f08c4","Type":"ContainerStarted","Data":"f8179320a307d787f5fe3becb7538d92cb998314ec4b167272cd0114432f479d"} Feb 02 09:26:00 crc kubenswrapper[4764]: I0202 09:26:00.241717 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Feb 02 09:26:00 crc kubenswrapper[4764]: I0202 09:26:00.250875 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2a132428-772f-4d1a-bd11-bf9f69301bfa","Type":"ContainerStarted","Data":"54eef320d116977e546ffca1391a1214baff8b9d8ec4a69ff013e5e639a4a3a3"} Feb 02 09:26:00 crc kubenswrapper[4764]: I0202 09:26:00.250916 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2a132428-772f-4d1a-bd11-bf9f69301bfa","Type":"ContainerStarted","Data":"4d51a30a1a083b0634c344cfdb52ad8831f3eb816ca0a7bafe73b8536d9c6f8f"} Feb 02 09:26:00 crc kubenswrapper[4764]: I0202 09:26:00.254676 4764 generic.go:334] "Generic (PLEG): container finished" podID="364d30e9-d4ef-4c84-b82f-673b41ffe9ff" containerID="46ac0362dd33241a49276203ed445721439244f4f312c3c5389ac4126d87b43f" exitCode=0 Feb 02 09:26:00 crc kubenswrapper[4764]: I0202 09:26:00.254706 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-69b69cc889-m4fck" event={"ID":"364d30e9-d4ef-4c84-b82f-673b41ffe9ff","Type":"ContainerDied","Data":"46ac0362dd33241a49276203ed445721439244f4f312c3c5389ac4126d87b43f"} Feb 02 09:26:00 crc kubenswrapper[4764]: I0202 09:26:00.254721 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-69b69cc889-m4fck" event={"ID":"364d30e9-d4ef-4c84-b82f-673b41ffe9ff","Type":"ContainerDied","Data":"edb74c3a670b8b4d80206645680542a348eec13833c041852e92d949d0bffb54"} Feb 02 09:26:00 crc kubenswrapper[4764]: I0202 09:26:00.254731 4764 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="edb74c3a670b8b4d80206645680542a348eec13833c041852e92d949d0bffb54" Feb 02 09:26:00 crc kubenswrapper[4764]: I0202 09:26:00.268304 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-69b69cc889-m4fck" Feb 02 09:26:00 crc kubenswrapper[4764]: I0202 09:26:00.269104 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.269087721 podStartE2EDuration="3.269087721s" podCreationTimestamp="2026-02-02 09:25:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:26:00.26210027 +0000 UTC m=+1123.195824358" watchObservedRunningTime="2026-02-02 09:26:00.269087721 +0000 UTC m=+1123.202811799" Feb 02 09:26:00 crc kubenswrapper[4764]: I0202 09:26:00.357888 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/364d30e9-d4ef-4c84-b82f-673b41ffe9ff-httpd-config\") pod \"364d30e9-d4ef-4c84-b82f-673b41ffe9ff\" (UID: \"364d30e9-d4ef-4c84-b82f-673b41ffe9ff\") " Feb 02 09:26:00 crc kubenswrapper[4764]: I0202 09:26:00.357990 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/364d30e9-d4ef-4c84-b82f-673b41ffe9ff-ovndb-tls-certs\") pod \"364d30e9-d4ef-4c84-b82f-673b41ffe9ff\" (UID: \"364d30e9-d4ef-4c84-b82f-673b41ffe9ff\") " Feb 02 09:26:00 crc kubenswrapper[4764]: I0202 09:26:00.358057 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/364d30e9-d4ef-4c84-b82f-673b41ffe9ff-public-tls-certs\") pod \"364d30e9-d4ef-4c84-b82f-673b41ffe9ff\" (UID: \"364d30e9-d4ef-4c84-b82f-673b41ffe9ff\") " Feb 02 09:26:00 crc kubenswrapper[4764]: I0202 09:26:00.358169 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/364d30e9-d4ef-4c84-b82f-673b41ffe9ff-internal-tls-certs\") pod \"364d30e9-d4ef-4c84-b82f-673b41ffe9ff\" (UID: \"364d30e9-d4ef-4c84-b82f-673b41ffe9ff\") " Feb 02 09:26:00 crc kubenswrapper[4764]: I0202 09:26:00.358202 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/364d30e9-d4ef-4c84-b82f-673b41ffe9ff-combined-ca-bundle\") pod \"364d30e9-d4ef-4c84-b82f-673b41ffe9ff\" (UID: \"364d30e9-d4ef-4c84-b82f-673b41ffe9ff\") " Feb 02 09:26:00 crc kubenswrapper[4764]: I0202 09:26:00.358221 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/364d30e9-d4ef-4c84-b82f-673b41ffe9ff-config\") pod \"364d30e9-d4ef-4c84-b82f-673b41ffe9ff\" (UID: \"364d30e9-d4ef-4c84-b82f-673b41ffe9ff\") " Feb 02 09:26:00 crc kubenswrapper[4764]: I0202 09:26:00.358254 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hmggt\" (UniqueName: \"kubernetes.io/projected/364d30e9-d4ef-4c84-b82f-673b41ffe9ff-kube-api-access-hmggt\") pod \"364d30e9-d4ef-4c84-b82f-673b41ffe9ff\" (UID: \"364d30e9-d4ef-4c84-b82f-673b41ffe9ff\") " Feb 02 09:26:00 crc kubenswrapper[4764]: I0202 09:26:00.375989 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/364d30e9-d4ef-4c84-b82f-673b41ffe9ff-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "364d30e9-d4ef-4c84-b82f-673b41ffe9ff" (UID: "364d30e9-d4ef-4c84-b82f-673b41ffe9ff"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:26:00 crc kubenswrapper[4764]: I0202 09:26:00.376018 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/364d30e9-d4ef-4c84-b82f-673b41ffe9ff-kube-api-access-hmggt" (OuterVolumeSpecName: "kube-api-access-hmggt") pod "364d30e9-d4ef-4c84-b82f-673b41ffe9ff" (UID: "364d30e9-d4ef-4c84-b82f-673b41ffe9ff"). InnerVolumeSpecName "kube-api-access-hmggt". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:26:00 crc kubenswrapper[4764]: I0202 09:26:00.402014 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/364d30e9-d4ef-4c84-b82f-673b41ffe9ff-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "364d30e9-d4ef-4c84-b82f-673b41ffe9ff" (UID: "364d30e9-d4ef-4c84-b82f-673b41ffe9ff"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:26:00 crc kubenswrapper[4764]: I0202 09:26:00.411194 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/364d30e9-d4ef-4c84-b82f-673b41ffe9ff-config" (OuterVolumeSpecName: "config") pod "364d30e9-d4ef-4c84-b82f-673b41ffe9ff" (UID: "364d30e9-d4ef-4c84-b82f-673b41ffe9ff"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:26:00 crc kubenswrapper[4764]: I0202 09:26:00.420771 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/364d30e9-d4ef-4c84-b82f-673b41ffe9ff-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "364d30e9-d4ef-4c84-b82f-673b41ffe9ff" (UID: "364d30e9-d4ef-4c84-b82f-673b41ffe9ff"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:26:00 crc kubenswrapper[4764]: I0202 09:26:00.423644 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/364d30e9-d4ef-4c84-b82f-673b41ffe9ff-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "364d30e9-d4ef-4c84-b82f-673b41ffe9ff" (UID: "364d30e9-d4ef-4c84-b82f-673b41ffe9ff"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:26:00 crc kubenswrapper[4764]: I0202 09:26:00.442991 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/364d30e9-d4ef-4c84-b82f-673b41ffe9ff-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "364d30e9-d4ef-4c84-b82f-673b41ffe9ff" (UID: "364d30e9-d4ef-4c84-b82f-673b41ffe9ff"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:26:00 crc kubenswrapper[4764]: I0202 09:26:00.460129 4764 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/364d30e9-d4ef-4c84-b82f-673b41ffe9ff-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 02 09:26:00 crc kubenswrapper[4764]: I0202 09:26:00.460160 4764 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/364d30e9-d4ef-4c84-b82f-673b41ffe9ff-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 09:26:00 crc kubenswrapper[4764]: I0202 09:26:00.460169 4764 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/364d30e9-d4ef-4c84-b82f-673b41ffe9ff-config\") on node \"crc\" DevicePath \"\"" Feb 02 09:26:00 crc kubenswrapper[4764]: I0202 09:26:00.460181 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hmggt\" (UniqueName: \"kubernetes.io/projected/364d30e9-d4ef-4c84-b82f-673b41ffe9ff-kube-api-access-hmggt\") on node \"crc\" DevicePath \"\"" Feb 02 09:26:00 crc kubenswrapper[4764]: I0202 09:26:00.460191 4764 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/364d30e9-d4ef-4c84-b82f-673b41ffe9ff-httpd-config\") on node \"crc\" DevicePath \"\"" Feb 02 09:26:00 crc kubenswrapper[4764]: I0202 09:26:00.460199 4764 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/364d30e9-d4ef-4c84-b82f-673b41ffe9ff-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 02 09:26:00 crc kubenswrapper[4764]: I0202 09:26:00.460206 4764 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/364d30e9-d4ef-4c84-b82f-673b41ffe9ff-public-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 02 09:26:01 crc kubenswrapper[4764]: I0202 09:26:01.264607 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-69b69cc889-m4fck" Feb 02 09:26:01 crc kubenswrapper[4764]: I0202 09:26:01.324733 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-69b69cc889-m4fck"] Feb 02 09:26:01 crc kubenswrapper[4764]: I0202 09:26:01.330485 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-69b69cc889-m4fck"] Feb 02 09:26:01 crc kubenswrapper[4764]: I0202 09:26:01.696282 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-555cfdffc9-rw6ht" Feb 02 09:26:01 crc kubenswrapper[4764]: I0202 09:26:01.799722 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-74cf444645-8gljr"] Feb 02 09:26:01 crc kubenswrapper[4764]: I0202 09:26:01.799971 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-74cf444645-8gljr" podUID="64a8ff87-c26b-49c9-9a34-0b1833b6b7a5" containerName="dnsmasq-dns" containerID="cri-o://28b2d3a2705c93f075ce7d5742d6e8515c70e94f15bbe822bd302c82836394ae" gracePeriod=10 Feb 02 09:26:01 crc kubenswrapper[4764]: I0202 09:26:01.886604 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="364d30e9-d4ef-4c84-b82f-673b41ffe9ff" path="/var/lib/kubelet/pods/364d30e9-d4ef-4c84-b82f-673b41ffe9ff/volumes" Feb 02 09:26:02 crc kubenswrapper[4764]: I0202 09:26:02.240032 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Feb 02 09:26:02 crc kubenswrapper[4764]: I0202 09:26:02.304326 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2a132428-772f-4d1a-bd11-bf9f69301bfa","Type":"ContainerStarted","Data":"9c53f871c848ed0cc02a5e809afa4de8cb1fb273467711b34ff71e95cd140427"} Feb 02 09:26:02 crc kubenswrapper[4764]: I0202 09:26:02.305693 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 02 09:26:02 crc kubenswrapper[4764]: I0202 09:26:02.312443 4764 generic.go:334] "Generic (PLEG): container finished" podID="64a8ff87-c26b-49c9-9a34-0b1833b6b7a5" containerID="28b2d3a2705c93f075ce7d5742d6e8515c70e94f15bbe822bd302c82836394ae" exitCode=0 Feb 02 09:26:02 crc kubenswrapper[4764]: I0202 09:26:02.312486 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74cf444645-8gljr" event={"ID":"64a8ff87-c26b-49c9-9a34-0b1833b6b7a5","Type":"ContainerDied","Data":"28b2d3a2705c93f075ce7d5742d6e8515c70e94f15bbe822bd302c82836394ae"} Feb 02 09:26:02 crc kubenswrapper[4764]: I0202 09:26:02.339314 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 02 09:26:02 crc kubenswrapper[4764]: I0202 09:26:02.339543 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="d509e876-498c-4e04-95af-bdb84c334a99" containerName="cinder-scheduler" containerID="cri-o://c301ea8b2f199dad2ba71f47cdbeb2bc337210f442107f205bf2728ba0595afd" gracePeriod=30 Feb 02 09:26:02 crc kubenswrapper[4764]: I0202 09:26:02.339901 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="d509e876-498c-4e04-95af-bdb84c334a99" containerName="probe" containerID="cri-o://51619052573cea6bacf4e847a8df670a9036620cd4def95428f044400cefce9a" gracePeriod=30 Feb 02 09:26:02 crc kubenswrapper[4764]: I0202 09:26:02.342896 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.736359251 podStartE2EDuration="5.342876504s" podCreationTimestamp="2026-02-02 09:25:57 +0000 UTC" firstStartedPulling="2026-02-02 09:25:58.162671937 +0000 UTC m=+1121.096396025" lastFinishedPulling="2026-02-02 09:26:01.76918917 +0000 UTC m=+1124.702913278" observedRunningTime="2026-02-02 09:26:02.32841571 +0000 UTC m=+1125.262139788" watchObservedRunningTime="2026-02-02 09:26:02.342876504 +0000 UTC m=+1125.276600592" Feb 02 09:26:02 crc kubenswrapper[4764]: I0202 09:26:02.378707 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74cf444645-8gljr" Feb 02 09:26:02 crc kubenswrapper[4764]: I0202 09:26:02.496385 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7t727\" (UniqueName: \"kubernetes.io/projected/64a8ff87-c26b-49c9-9a34-0b1833b6b7a5-kube-api-access-7t727\") pod \"64a8ff87-c26b-49c9-9a34-0b1833b6b7a5\" (UID: \"64a8ff87-c26b-49c9-9a34-0b1833b6b7a5\") " Feb 02 09:26:02 crc kubenswrapper[4764]: I0202 09:26:02.496456 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/64a8ff87-c26b-49c9-9a34-0b1833b6b7a5-dns-svc\") pod \"64a8ff87-c26b-49c9-9a34-0b1833b6b7a5\" (UID: \"64a8ff87-c26b-49c9-9a34-0b1833b6b7a5\") " Feb 02 09:26:02 crc kubenswrapper[4764]: I0202 09:26:02.496543 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/64a8ff87-c26b-49c9-9a34-0b1833b6b7a5-config\") pod \"64a8ff87-c26b-49c9-9a34-0b1833b6b7a5\" (UID: \"64a8ff87-c26b-49c9-9a34-0b1833b6b7a5\") " Feb 02 09:26:02 crc kubenswrapper[4764]: I0202 09:26:02.496622 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/64a8ff87-c26b-49c9-9a34-0b1833b6b7a5-ovsdbserver-nb\") pod \"64a8ff87-c26b-49c9-9a34-0b1833b6b7a5\" (UID: \"64a8ff87-c26b-49c9-9a34-0b1833b6b7a5\") " Feb 02 09:26:02 crc kubenswrapper[4764]: I0202 09:26:02.496721 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/64a8ff87-c26b-49c9-9a34-0b1833b6b7a5-ovsdbserver-sb\") pod \"64a8ff87-c26b-49c9-9a34-0b1833b6b7a5\" (UID: \"64a8ff87-c26b-49c9-9a34-0b1833b6b7a5\") " Feb 02 09:26:02 crc kubenswrapper[4764]: I0202 09:26:02.501925 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/64a8ff87-c26b-49c9-9a34-0b1833b6b7a5-kube-api-access-7t727" (OuterVolumeSpecName: "kube-api-access-7t727") pod "64a8ff87-c26b-49c9-9a34-0b1833b6b7a5" (UID: "64a8ff87-c26b-49c9-9a34-0b1833b6b7a5"). InnerVolumeSpecName "kube-api-access-7t727". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:26:02 crc kubenswrapper[4764]: I0202 09:26:02.537392 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/64a8ff87-c26b-49c9-9a34-0b1833b6b7a5-config" (OuterVolumeSpecName: "config") pod "64a8ff87-c26b-49c9-9a34-0b1833b6b7a5" (UID: "64a8ff87-c26b-49c9-9a34-0b1833b6b7a5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:26:02 crc kubenswrapper[4764]: I0202 09:26:02.537803 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/64a8ff87-c26b-49c9-9a34-0b1833b6b7a5-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "64a8ff87-c26b-49c9-9a34-0b1833b6b7a5" (UID: "64a8ff87-c26b-49c9-9a34-0b1833b6b7a5"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:26:02 crc kubenswrapper[4764]: I0202 09:26:02.545050 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/64a8ff87-c26b-49c9-9a34-0b1833b6b7a5-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "64a8ff87-c26b-49c9-9a34-0b1833b6b7a5" (UID: "64a8ff87-c26b-49c9-9a34-0b1833b6b7a5"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:26:02 crc kubenswrapper[4764]: I0202 09:26:02.582420 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/64a8ff87-c26b-49c9-9a34-0b1833b6b7a5-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "64a8ff87-c26b-49c9-9a34-0b1833b6b7a5" (UID: "64a8ff87-c26b-49c9-9a34-0b1833b6b7a5"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:26:02 crc kubenswrapper[4764]: I0202 09:26:02.599000 4764 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/64a8ff87-c26b-49c9-9a34-0b1833b6b7a5-config\") on node \"crc\" DevicePath \"\"" Feb 02 09:26:02 crc kubenswrapper[4764]: I0202 09:26:02.599029 4764 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/64a8ff87-c26b-49c9-9a34-0b1833b6b7a5-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 02 09:26:02 crc kubenswrapper[4764]: I0202 09:26:02.599041 4764 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/64a8ff87-c26b-49c9-9a34-0b1833b6b7a5-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 02 09:26:02 crc kubenswrapper[4764]: I0202 09:26:02.599050 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7t727\" (UniqueName: \"kubernetes.io/projected/64a8ff87-c26b-49c9-9a34-0b1833b6b7a5-kube-api-access-7t727\") on node \"crc\" DevicePath \"\"" Feb 02 09:26:02 crc kubenswrapper[4764]: I0202 09:26:02.599059 4764 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/64a8ff87-c26b-49c9-9a34-0b1833b6b7a5-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 02 09:26:03 crc kubenswrapper[4764]: I0202 09:26:03.324297 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74cf444645-8gljr" Feb 02 09:26:03 crc kubenswrapper[4764]: I0202 09:26:03.326982 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74cf444645-8gljr" event={"ID":"64a8ff87-c26b-49c9-9a34-0b1833b6b7a5","Type":"ContainerDied","Data":"2faf3e8aa7d0587946beb4a728098bf586b0031811a48af2f1746583d014ddef"} Feb 02 09:26:03 crc kubenswrapper[4764]: I0202 09:26:03.327017 4764 scope.go:117] "RemoveContainer" containerID="28b2d3a2705c93f075ce7d5742d6e8515c70e94f15bbe822bd302c82836394ae" Feb 02 09:26:03 crc kubenswrapper[4764]: I0202 09:26:03.357087 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-74cf444645-8gljr"] Feb 02 09:26:03 crc kubenswrapper[4764]: I0202 09:26:03.376709 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-74cf444645-8gljr"] Feb 02 09:26:03 crc kubenswrapper[4764]: I0202 09:26:03.404667 4764 scope.go:117] "RemoveContainer" containerID="b516111137de647525a54a53b73b341f4425b981f9a2c917699bda63892f062c" Feb 02 09:26:03 crc kubenswrapper[4764]: I0202 09:26:03.835509 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="64a8ff87-c26b-49c9-9a34-0b1833b6b7a5" path="/var/lib/kubelet/pods/64a8ff87-c26b-49c9-9a34-0b1833b6b7a5/volumes" Feb 02 09:26:04 crc kubenswrapper[4764]: I0202 09:26:04.339703 4764 generic.go:334] "Generic (PLEG): container finished" podID="d509e876-498c-4e04-95af-bdb84c334a99" containerID="51619052573cea6bacf4e847a8df670a9036620cd4def95428f044400cefce9a" exitCode=0 Feb 02 09:26:04 crc kubenswrapper[4764]: I0202 09:26:04.339733 4764 generic.go:334] "Generic (PLEG): container finished" podID="d509e876-498c-4e04-95af-bdb84c334a99" containerID="c301ea8b2f199dad2ba71f47cdbeb2bc337210f442107f205bf2728ba0595afd" exitCode=0 Feb 02 09:26:04 crc kubenswrapper[4764]: I0202 09:26:04.339765 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"d509e876-498c-4e04-95af-bdb84c334a99","Type":"ContainerDied","Data":"51619052573cea6bacf4e847a8df670a9036620cd4def95428f044400cefce9a"} Feb 02 09:26:04 crc kubenswrapper[4764]: I0202 09:26:04.339791 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"d509e876-498c-4e04-95af-bdb84c334a99","Type":"ContainerDied","Data":"c301ea8b2f199dad2ba71f47cdbeb2bc337210f442107f205bf2728ba0595afd"} Feb 02 09:26:04 crc kubenswrapper[4764]: I0202 09:26:04.339803 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"d509e876-498c-4e04-95af-bdb84c334a99","Type":"ContainerDied","Data":"bc3465eeb4fca0708d8513f5a6cc31ab9f11387e9359a3b8c5d4c3cc5912bd7f"} Feb 02 09:26:04 crc kubenswrapper[4764]: I0202 09:26:04.339813 4764 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bc3465eeb4fca0708d8513f5a6cc31ab9f11387e9359a3b8c5d4c3cc5912bd7f" Feb 02 09:26:04 crc kubenswrapper[4764]: I0202 09:26:04.366380 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Feb 02 09:26:04 crc kubenswrapper[4764]: I0202 09:26:04.393558 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-65fcd88786-dggfp" Feb 02 09:26:04 crc kubenswrapper[4764]: I0202 09:26:04.415486 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-65fcd88786-dggfp" Feb 02 09:26:04 crc kubenswrapper[4764]: I0202 09:26:04.424148 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d509e876-498c-4e04-95af-bdb84c334a99-etc-machine-id\") pod \"d509e876-498c-4e04-95af-bdb84c334a99\" (UID: \"d509e876-498c-4e04-95af-bdb84c334a99\") " Feb 02 09:26:04 crc kubenswrapper[4764]: I0202 09:26:04.424182 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d509e876-498c-4e04-95af-bdb84c334a99-scripts\") pod \"d509e876-498c-4e04-95af-bdb84c334a99\" (UID: \"d509e876-498c-4e04-95af-bdb84c334a99\") " Feb 02 09:26:04 crc kubenswrapper[4764]: I0202 09:26:04.424285 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tjcrn\" (UniqueName: \"kubernetes.io/projected/d509e876-498c-4e04-95af-bdb84c334a99-kube-api-access-tjcrn\") pod \"d509e876-498c-4e04-95af-bdb84c334a99\" (UID: \"d509e876-498c-4e04-95af-bdb84c334a99\") " Feb 02 09:26:04 crc kubenswrapper[4764]: I0202 09:26:04.424311 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d509e876-498c-4e04-95af-bdb84c334a99-combined-ca-bundle\") pod \"d509e876-498c-4e04-95af-bdb84c334a99\" (UID: \"d509e876-498c-4e04-95af-bdb84c334a99\") " Feb 02 09:26:04 crc kubenswrapper[4764]: I0202 09:26:04.424393 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d509e876-498c-4e04-95af-bdb84c334a99-config-data\") pod \"d509e876-498c-4e04-95af-bdb84c334a99\" (UID: \"d509e876-498c-4e04-95af-bdb84c334a99\") " Feb 02 09:26:04 crc kubenswrapper[4764]: I0202 09:26:04.424479 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d509e876-498c-4e04-95af-bdb84c334a99-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "d509e876-498c-4e04-95af-bdb84c334a99" (UID: "d509e876-498c-4e04-95af-bdb84c334a99"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 09:26:04 crc kubenswrapper[4764]: I0202 09:26:04.424515 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d509e876-498c-4e04-95af-bdb84c334a99-config-data-custom\") pod \"d509e876-498c-4e04-95af-bdb84c334a99\" (UID: \"d509e876-498c-4e04-95af-bdb84c334a99\") " Feb 02 09:26:04 crc kubenswrapper[4764]: I0202 09:26:04.425899 4764 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d509e876-498c-4e04-95af-bdb84c334a99-etc-machine-id\") on node \"crc\" DevicePath \"\"" Feb 02 09:26:04 crc kubenswrapper[4764]: I0202 09:26:04.447047 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d509e876-498c-4e04-95af-bdb84c334a99-scripts" (OuterVolumeSpecName: "scripts") pod "d509e876-498c-4e04-95af-bdb84c334a99" (UID: "d509e876-498c-4e04-95af-bdb84c334a99"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:26:04 crc kubenswrapper[4764]: I0202 09:26:04.461225 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d509e876-498c-4e04-95af-bdb84c334a99-kube-api-access-tjcrn" (OuterVolumeSpecName: "kube-api-access-tjcrn") pod "d509e876-498c-4e04-95af-bdb84c334a99" (UID: "d509e876-498c-4e04-95af-bdb84c334a99"). InnerVolumeSpecName "kube-api-access-tjcrn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:26:04 crc kubenswrapper[4764]: I0202 09:26:04.461412 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d509e876-498c-4e04-95af-bdb84c334a99-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "d509e876-498c-4e04-95af-bdb84c334a99" (UID: "d509e876-498c-4e04-95af-bdb84c334a99"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:26:04 crc kubenswrapper[4764]: I0202 09:26:04.529067 4764 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d509e876-498c-4e04-95af-bdb84c334a99-config-data-custom\") on node \"crc\" DevicePath \"\"" Feb 02 09:26:04 crc kubenswrapper[4764]: I0202 09:26:04.529094 4764 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d509e876-498c-4e04-95af-bdb84c334a99-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 09:26:04 crc kubenswrapper[4764]: I0202 09:26:04.529105 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tjcrn\" (UniqueName: \"kubernetes.io/projected/d509e876-498c-4e04-95af-bdb84c334a99-kube-api-access-tjcrn\") on node \"crc\" DevicePath \"\"" Feb 02 09:26:04 crc kubenswrapper[4764]: I0202 09:26:04.533751 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d509e876-498c-4e04-95af-bdb84c334a99-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d509e876-498c-4e04-95af-bdb84c334a99" (UID: "d509e876-498c-4e04-95af-bdb84c334a99"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:26:04 crc kubenswrapper[4764]: I0202 09:26:04.607047 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d509e876-498c-4e04-95af-bdb84c334a99-config-data" (OuterVolumeSpecName: "config-data") pod "d509e876-498c-4e04-95af-bdb84c334a99" (UID: "d509e876-498c-4e04-95af-bdb84c334a99"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:26:04 crc kubenswrapper[4764]: I0202 09:26:04.630574 4764 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d509e876-498c-4e04-95af-bdb84c334a99-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 09:26:04 crc kubenswrapper[4764]: I0202 09:26:04.630609 4764 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d509e876-498c-4e04-95af-bdb84c334a99-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 09:26:04 crc kubenswrapper[4764]: I0202 09:26:04.653541 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-68d4474c54-98j2g"] Feb 02 09:26:04 crc kubenswrapper[4764]: E0202 09:26:04.654047 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d509e876-498c-4e04-95af-bdb84c334a99" containerName="cinder-scheduler" Feb 02 09:26:04 crc kubenswrapper[4764]: I0202 09:26:04.654847 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="d509e876-498c-4e04-95af-bdb84c334a99" containerName="cinder-scheduler" Feb 02 09:26:04 crc kubenswrapper[4764]: E0202 09:26:04.654967 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40e640ad-6fb1-4c56-9b33-14fe28103b07" containerName="barbican-api-log" Feb 02 09:26:04 crc kubenswrapper[4764]: I0202 09:26:04.655070 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="40e640ad-6fb1-4c56-9b33-14fe28103b07" containerName="barbican-api-log" Feb 02 09:26:04 crc kubenswrapper[4764]: E0202 09:26:04.655141 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64a8ff87-c26b-49c9-9a34-0b1833b6b7a5" containerName="dnsmasq-dns" Feb 02 09:26:04 crc kubenswrapper[4764]: I0202 09:26:04.655192 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="64a8ff87-c26b-49c9-9a34-0b1833b6b7a5" containerName="dnsmasq-dns" Feb 02 09:26:04 crc kubenswrapper[4764]: E0202 09:26:04.655249 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40e640ad-6fb1-4c56-9b33-14fe28103b07" containerName="barbican-api" Feb 02 09:26:04 crc kubenswrapper[4764]: I0202 09:26:04.655297 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="40e640ad-6fb1-4c56-9b33-14fe28103b07" containerName="barbican-api" Feb 02 09:26:04 crc kubenswrapper[4764]: E0202 09:26:04.655356 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d509e876-498c-4e04-95af-bdb84c334a99" containerName="probe" Feb 02 09:26:04 crc kubenswrapper[4764]: I0202 09:26:04.655404 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="d509e876-498c-4e04-95af-bdb84c334a99" containerName="probe" Feb 02 09:26:04 crc kubenswrapper[4764]: E0202 09:26:04.655475 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="364d30e9-d4ef-4c84-b82f-673b41ffe9ff" containerName="neutron-api" Feb 02 09:26:04 crc kubenswrapper[4764]: I0202 09:26:04.655540 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="364d30e9-d4ef-4c84-b82f-673b41ffe9ff" containerName="neutron-api" Feb 02 09:26:04 crc kubenswrapper[4764]: E0202 09:26:04.655617 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="364d30e9-d4ef-4c84-b82f-673b41ffe9ff" containerName="neutron-httpd" Feb 02 09:26:04 crc kubenswrapper[4764]: I0202 09:26:04.655679 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="364d30e9-d4ef-4c84-b82f-673b41ffe9ff" containerName="neutron-httpd" Feb 02 09:26:04 crc kubenswrapper[4764]: E0202 09:26:04.655738 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64a8ff87-c26b-49c9-9a34-0b1833b6b7a5" containerName="init" Feb 02 09:26:04 crc kubenswrapper[4764]: I0202 09:26:04.655847 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="64a8ff87-c26b-49c9-9a34-0b1833b6b7a5" containerName="init" Feb 02 09:26:04 crc kubenswrapper[4764]: I0202 09:26:04.656147 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="364d30e9-d4ef-4c84-b82f-673b41ffe9ff" containerName="neutron-api" Feb 02 09:26:04 crc kubenswrapper[4764]: I0202 09:26:04.656243 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="364d30e9-d4ef-4c84-b82f-673b41ffe9ff" containerName="neutron-httpd" Feb 02 09:26:04 crc kubenswrapper[4764]: I0202 09:26:04.656304 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="40e640ad-6fb1-4c56-9b33-14fe28103b07" containerName="barbican-api-log" Feb 02 09:26:04 crc kubenswrapper[4764]: I0202 09:26:04.656374 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="40e640ad-6fb1-4c56-9b33-14fe28103b07" containerName="barbican-api" Feb 02 09:26:04 crc kubenswrapper[4764]: I0202 09:26:04.656458 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="d509e876-498c-4e04-95af-bdb84c334a99" containerName="probe" Feb 02 09:26:04 crc kubenswrapper[4764]: I0202 09:26:04.656553 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="64a8ff87-c26b-49c9-9a34-0b1833b6b7a5" containerName="dnsmasq-dns" Feb 02 09:26:04 crc kubenswrapper[4764]: I0202 09:26:04.656637 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="d509e876-498c-4e04-95af-bdb84c334a99" containerName="cinder-scheduler" Feb 02 09:26:04 crc kubenswrapper[4764]: I0202 09:26:04.657568 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-68d4474c54-98j2g" Feb 02 09:26:04 crc kubenswrapper[4764]: I0202 09:26:04.672365 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-68d4474c54-98j2g"] Feb 02 09:26:04 crc kubenswrapper[4764]: I0202 09:26:04.732203 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f7a36a2c-9ade-4bb2-84d2-faa137207385-internal-tls-certs\") pod \"placement-68d4474c54-98j2g\" (UID: \"f7a36a2c-9ade-4bb2-84d2-faa137207385\") " pod="openstack/placement-68d4474c54-98j2g" Feb 02 09:26:04 crc kubenswrapper[4764]: I0202 09:26:04.732246 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7a36a2c-9ade-4bb2-84d2-faa137207385-combined-ca-bundle\") pod \"placement-68d4474c54-98j2g\" (UID: \"f7a36a2c-9ade-4bb2-84d2-faa137207385\") " pod="openstack/placement-68d4474c54-98j2g" Feb 02 09:26:04 crc kubenswrapper[4764]: I0202 09:26:04.732272 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7a36a2c-9ade-4bb2-84d2-faa137207385-scripts\") pod \"placement-68d4474c54-98j2g\" (UID: \"f7a36a2c-9ade-4bb2-84d2-faa137207385\") " pod="openstack/placement-68d4474c54-98j2g" Feb 02 09:26:04 crc kubenswrapper[4764]: I0202 09:26:04.732295 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f7a36a2c-9ade-4bb2-84d2-faa137207385-logs\") pod \"placement-68d4474c54-98j2g\" (UID: \"f7a36a2c-9ade-4bb2-84d2-faa137207385\") " pod="openstack/placement-68d4474c54-98j2g" Feb 02 09:26:04 crc kubenswrapper[4764]: I0202 09:26:04.732358 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k2hgp\" (UniqueName: \"kubernetes.io/projected/f7a36a2c-9ade-4bb2-84d2-faa137207385-kube-api-access-k2hgp\") pod \"placement-68d4474c54-98j2g\" (UID: \"f7a36a2c-9ade-4bb2-84d2-faa137207385\") " pod="openstack/placement-68d4474c54-98j2g" Feb 02 09:26:04 crc kubenswrapper[4764]: I0202 09:26:04.732425 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7a36a2c-9ade-4bb2-84d2-faa137207385-config-data\") pod \"placement-68d4474c54-98j2g\" (UID: \"f7a36a2c-9ade-4bb2-84d2-faa137207385\") " pod="openstack/placement-68d4474c54-98j2g" Feb 02 09:26:04 crc kubenswrapper[4764]: I0202 09:26:04.732450 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f7a36a2c-9ade-4bb2-84d2-faa137207385-public-tls-certs\") pod \"placement-68d4474c54-98j2g\" (UID: \"f7a36a2c-9ade-4bb2-84d2-faa137207385\") " pod="openstack/placement-68d4474c54-98j2g" Feb 02 09:26:04 crc kubenswrapper[4764]: I0202 09:26:04.833900 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k2hgp\" (UniqueName: \"kubernetes.io/projected/f7a36a2c-9ade-4bb2-84d2-faa137207385-kube-api-access-k2hgp\") pod \"placement-68d4474c54-98j2g\" (UID: \"f7a36a2c-9ade-4bb2-84d2-faa137207385\") " pod="openstack/placement-68d4474c54-98j2g" Feb 02 09:26:04 crc kubenswrapper[4764]: I0202 09:26:04.834039 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7a36a2c-9ade-4bb2-84d2-faa137207385-config-data\") pod \"placement-68d4474c54-98j2g\" (UID: \"f7a36a2c-9ade-4bb2-84d2-faa137207385\") " pod="openstack/placement-68d4474c54-98j2g" Feb 02 09:26:04 crc kubenswrapper[4764]: I0202 09:26:04.834073 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f7a36a2c-9ade-4bb2-84d2-faa137207385-public-tls-certs\") pod \"placement-68d4474c54-98j2g\" (UID: \"f7a36a2c-9ade-4bb2-84d2-faa137207385\") " pod="openstack/placement-68d4474c54-98j2g" Feb 02 09:26:04 crc kubenswrapper[4764]: I0202 09:26:04.834110 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f7a36a2c-9ade-4bb2-84d2-faa137207385-internal-tls-certs\") pod \"placement-68d4474c54-98j2g\" (UID: \"f7a36a2c-9ade-4bb2-84d2-faa137207385\") " pod="openstack/placement-68d4474c54-98j2g" Feb 02 09:26:04 crc kubenswrapper[4764]: I0202 09:26:04.834136 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7a36a2c-9ade-4bb2-84d2-faa137207385-combined-ca-bundle\") pod \"placement-68d4474c54-98j2g\" (UID: \"f7a36a2c-9ade-4bb2-84d2-faa137207385\") " pod="openstack/placement-68d4474c54-98j2g" Feb 02 09:26:04 crc kubenswrapper[4764]: I0202 09:26:04.834167 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7a36a2c-9ade-4bb2-84d2-faa137207385-scripts\") pod \"placement-68d4474c54-98j2g\" (UID: \"f7a36a2c-9ade-4bb2-84d2-faa137207385\") " pod="openstack/placement-68d4474c54-98j2g" Feb 02 09:26:04 crc kubenswrapper[4764]: I0202 09:26:04.834196 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f7a36a2c-9ade-4bb2-84d2-faa137207385-logs\") pod \"placement-68d4474c54-98j2g\" (UID: \"f7a36a2c-9ade-4bb2-84d2-faa137207385\") " pod="openstack/placement-68d4474c54-98j2g" Feb 02 09:26:04 crc kubenswrapper[4764]: I0202 09:26:04.834688 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f7a36a2c-9ade-4bb2-84d2-faa137207385-logs\") pod \"placement-68d4474c54-98j2g\" (UID: \"f7a36a2c-9ade-4bb2-84d2-faa137207385\") " pod="openstack/placement-68d4474c54-98j2g" Feb 02 09:26:04 crc kubenswrapper[4764]: I0202 09:26:04.841738 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7a36a2c-9ade-4bb2-84d2-faa137207385-scripts\") pod \"placement-68d4474c54-98j2g\" (UID: \"f7a36a2c-9ade-4bb2-84d2-faa137207385\") " pod="openstack/placement-68d4474c54-98j2g" Feb 02 09:26:04 crc kubenswrapper[4764]: I0202 09:26:04.842059 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7a36a2c-9ade-4bb2-84d2-faa137207385-combined-ca-bundle\") pod \"placement-68d4474c54-98j2g\" (UID: \"f7a36a2c-9ade-4bb2-84d2-faa137207385\") " pod="openstack/placement-68d4474c54-98j2g" Feb 02 09:26:04 crc kubenswrapper[4764]: I0202 09:26:04.848603 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7a36a2c-9ade-4bb2-84d2-faa137207385-config-data\") pod \"placement-68d4474c54-98j2g\" (UID: \"f7a36a2c-9ade-4bb2-84d2-faa137207385\") " pod="openstack/placement-68d4474c54-98j2g" Feb 02 09:26:04 crc kubenswrapper[4764]: I0202 09:26:04.850335 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f7a36a2c-9ade-4bb2-84d2-faa137207385-public-tls-certs\") pod \"placement-68d4474c54-98j2g\" (UID: \"f7a36a2c-9ade-4bb2-84d2-faa137207385\") " pod="openstack/placement-68d4474c54-98j2g" Feb 02 09:26:04 crc kubenswrapper[4764]: I0202 09:26:04.861084 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f7a36a2c-9ade-4bb2-84d2-faa137207385-internal-tls-certs\") pod \"placement-68d4474c54-98j2g\" (UID: \"f7a36a2c-9ade-4bb2-84d2-faa137207385\") " pod="openstack/placement-68d4474c54-98j2g" Feb 02 09:26:04 crc kubenswrapper[4764]: I0202 09:26:04.862291 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k2hgp\" (UniqueName: \"kubernetes.io/projected/f7a36a2c-9ade-4bb2-84d2-faa137207385-kube-api-access-k2hgp\") pod \"placement-68d4474c54-98j2g\" (UID: \"f7a36a2c-9ade-4bb2-84d2-faa137207385\") " pod="openstack/placement-68d4474c54-98j2g" Feb 02 09:26:04 crc kubenswrapper[4764]: I0202 09:26:04.976356 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-68d4474c54-98j2g" Feb 02 09:26:05 crc kubenswrapper[4764]: I0202 09:26:05.349449 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Feb 02 09:26:05 crc kubenswrapper[4764]: I0202 09:26:05.381945 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 02 09:26:05 crc kubenswrapper[4764]: I0202 09:26:05.390890 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 02 09:26:05 crc kubenswrapper[4764]: I0202 09:26:05.401851 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-68d4474c54-98j2g"] Feb 02 09:26:05 crc kubenswrapper[4764]: I0202 09:26:05.409083 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Feb 02 09:26:05 crc kubenswrapper[4764]: I0202 09:26:05.420833 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Feb 02 09:26:05 crc kubenswrapper[4764]: I0202 09:26:05.424653 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Feb 02 09:26:05 crc kubenswrapper[4764]: I0202 09:26:05.441856 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/10cfff03-5b66-41ef-948b-0bad4f6e9a91-scripts\") pod \"cinder-scheduler-0\" (UID: \"10cfff03-5b66-41ef-948b-0bad4f6e9a91\") " pod="openstack/cinder-scheduler-0" Feb 02 09:26:05 crc kubenswrapper[4764]: I0202 09:26:05.441951 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/10cfff03-5b66-41ef-948b-0bad4f6e9a91-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"10cfff03-5b66-41ef-948b-0bad4f6e9a91\") " pod="openstack/cinder-scheduler-0" Feb 02 09:26:05 crc kubenswrapper[4764]: I0202 09:26:05.442007 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8d5cb\" (UniqueName: \"kubernetes.io/projected/10cfff03-5b66-41ef-948b-0bad4f6e9a91-kube-api-access-8d5cb\") pod \"cinder-scheduler-0\" (UID: \"10cfff03-5b66-41ef-948b-0bad4f6e9a91\") " pod="openstack/cinder-scheduler-0" Feb 02 09:26:05 crc kubenswrapper[4764]: I0202 09:26:05.442071 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/10cfff03-5b66-41ef-948b-0bad4f6e9a91-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"10cfff03-5b66-41ef-948b-0bad4f6e9a91\") " pod="openstack/cinder-scheduler-0" Feb 02 09:26:05 crc kubenswrapper[4764]: I0202 09:26:05.442087 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10cfff03-5b66-41ef-948b-0bad4f6e9a91-config-data\") pod \"cinder-scheduler-0\" (UID: \"10cfff03-5b66-41ef-948b-0bad4f6e9a91\") " pod="openstack/cinder-scheduler-0" Feb 02 09:26:05 crc kubenswrapper[4764]: I0202 09:26:05.442115 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10cfff03-5b66-41ef-948b-0bad4f6e9a91-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"10cfff03-5b66-41ef-948b-0bad4f6e9a91\") " pod="openstack/cinder-scheduler-0" Feb 02 09:26:05 crc kubenswrapper[4764]: I0202 09:26:05.456095 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 02 09:26:05 crc kubenswrapper[4764]: I0202 09:26:05.543218 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/10cfff03-5b66-41ef-948b-0bad4f6e9a91-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"10cfff03-5b66-41ef-948b-0bad4f6e9a91\") " pod="openstack/cinder-scheduler-0" Feb 02 09:26:05 crc kubenswrapper[4764]: I0202 09:26:05.543287 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8d5cb\" (UniqueName: \"kubernetes.io/projected/10cfff03-5b66-41ef-948b-0bad4f6e9a91-kube-api-access-8d5cb\") pod \"cinder-scheduler-0\" (UID: \"10cfff03-5b66-41ef-948b-0bad4f6e9a91\") " pod="openstack/cinder-scheduler-0" Feb 02 09:26:05 crc kubenswrapper[4764]: I0202 09:26:05.543324 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/10cfff03-5b66-41ef-948b-0bad4f6e9a91-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"10cfff03-5b66-41ef-948b-0bad4f6e9a91\") " pod="openstack/cinder-scheduler-0" Feb 02 09:26:05 crc kubenswrapper[4764]: I0202 09:26:05.543362 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10cfff03-5b66-41ef-948b-0bad4f6e9a91-config-data\") pod \"cinder-scheduler-0\" (UID: \"10cfff03-5b66-41ef-948b-0bad4f6e9a91\") " pod="openstack/cinder-scheduler-0" Feb 02 09:26:05 crc kubenswrapper[4764]: I0202 09:26:05.543366 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/10cfff03-5b66-41ef-948b-0bad4f6e9a91-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"10cfff03-5b66-41ef-948b-0bad4f6e9a91\") " pod="openstack/cinder-scheduler-0" Feb 02 09:26:05 crc kubenswrapper[4764]: I0202 09:26:05.543391 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10cfff03-5b66-41ef-948b-0bad4f6e9a91-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"10cfff03-5b66-41ef-948b-0bad4f6e9a91\") " pod="openstack/cinder-scheduler-0" Feb 02 09:26:05 crc kubenswrapper[4764]: I0202 09:26:05.544591 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/10cfff03-5b66-41ef-948b-0bad4f6e9a91-scripts\") pod \"cinder-scheduler-0\" (UID: \"10cfff03-5b66-41ef-948b-0bad4f6e9a91\") " pod="openstack/cinder-scheduler-0" Feb 02 09:26:05 crc kubenswrapper[4764]: I0202 09:26:05.548278 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10cfff03-5b66-41ef-948b-0bad4f6e9a91-config-data\") pod \"cinder-scheduler-0\" (UID: \"10cfff03-5b66-41ef-948b-0bad4f6e9a91\") " pod="openstack/cinder-scheduler-0" Feb 02 09:26:05 crc kubenswrapper[4764]: I0202 09:26:05.548665 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/10cfff03-5b66-41ef-948b-0bad4f6e9a91-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"10cfff03-5b66-41ef-948b-0bad4f6e9a91\") " pod="openstack/cinder-scheduler-0" Feb 02 09:26:05 crc kubenswrapper[4764]: I0202 09:26:05.549012 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/10cfff03-5b66-41ef-948b-0bad4f6e9a91-scripts\") pod \"cinder-scheduler-0\" (UID: \"10cfff03-5b66-41ef-948b-0bad4f6e9a91\") " pod="openstack/cinder-scheduler-0" Feb 02 09:26:05 crc kubenswrapper[4764]: I0202 09:26:05.549859 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10cfff03-5b66-41ef-948b-0bad4f6e9a91-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"10cfff03-5b66-41ef-948b-0bad4f6e9a91\") " pod="openstack/cinder-scheduler-0" Feb 02 09:26:05 crc kubenswrapper[4764]: I0202 09:26:05.558318 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8d5cb\" (UniqueName: \"kubernetes.io/projected/10cfff03-5b66-41ef-948b-0bad4f6e9a91-kube-api-access-8d5cb\") pod \"cinder-scheduler-0\" (UID: \"10cfff03-5b66-41ef-948b-0bad4f6e9a91\") " pod="openstack/cinder-scheduler-0" Feb 02 09:26:05 crc kubenswrapper[4764]: I0202 09:26:05.587162 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Feb 02 09:26:05 crc kubenswrapper[4764]: I0202 09:26:05.834105 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d509e876-498c-4e04-95af-bdb84c334a99" path="/var/lib/kubelet/pods/d509e876-498c-4e04-95af-bdb84c334a99/volumes" Feb 02 09:26:05 crc kubenswrapper[4764]: I0202 09:26:05.994297 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 02 09:26:06 crc kubenswrapper[4764]: W0202 09:26:06.000077 4764 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod10cfff03_5b66_41ef_948b_0bad4f6e9a91.slice/crio-5a7144d9541e78f7b195ce4326a584b16c57535e205f5bb54a9a011847db057e WatchSource:0}: Error finding container 5a7144d9541e78f7b195ce4326a584b16c57535e205f5bb54a9a011847db057e: Status 404 returned error can't find the container with id 5a7144d9541e78f7b195ce4326a584b16c57535e205f5bb54a9a011847db057e Feb 02 09:26:06 crc kubenswrapper[4764]: I0202 09:26:06.369461 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-68d4474c54-98j2g" event={"ID":"f7a36a2c-9ade-4bb2-84d2-faa137207385","Type":"ContainerStarted","Data":"bf6c271cecad3f4bc0dd42829e10f86bf4d2fde6e3aabf35649a4ba8479b00d2"} Feb 02 09:26:06 crc kubenswrapper[4764]: I0202 09:26:06.369849 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-68d4474c54-98j2g" event={"ID":"f7a36a2c-9ade-4bb2-84d2-faa137207385","Type":"ContainerStarted","Data":"20f7e92160083ebdb7a0bde0d2e3dbb485fa41e8e1e8a7e3e43b106fc06a1d71"} Feb 02 09:26:06 crc kubenswrapper[4764]: I0202 09:26:06.369867 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-68d4474c54-98j2g" event={"ID":"f7a36a2c-9ade-4bb2-84d2-faa137207385","Type":"ContainerStarted","Data":"13eab1074cd3f56f9a8ba66b6da5bf323d2f564d79f0acaa3ea0775717b52256"} Feb 02 09:26:06 crc kubenswrapper[4764]: I0202 09:26:06.369892 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-68d4474c54-98j2g" Feb 02 09:26:06 crc kubenswrapper[4764]: I0202 09:26:06.370871 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"10cfff03-5b66-41ef-948b-0bad4f6e9a91","Type":"ContainerStarted","Data":"5a7144d9541e78f7b195ce4326a584b16c57535e205f5bb54a9a011847db057e"} Feb 02 09:26:06 crc kubenswrapper[4764]: I0202 09:26:06.389976 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-68d4474c54-98j2g" podStartSLOduration=2.389962884 podStartE2EDuration="2.389962884s" podCreationTimestamp="2026-02-02 09:26:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:26:06.389167513 +0000 UTC m=+1129.322891601" watchObservedRunningTime="2026-02-02 09:26:06.389962884 +0000 UTC m=+1129.323686972" Feb 02 09:26:07 crc kubenswrapper[4764]: I0202 09:26:07.385245 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"10cfff03-5b66-41ef-948b-0bad4f6e9a91","Type":"ContainerStarted","Data":"66330e7ed418fa934963957923bc47796c53d5f897bfe6cda99306799ba5c725"} Feb 02 09:26:07 crc kubenswrapper[4764]: I0202 09:26:07.385670 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"10cfff03-5b66-41ef-948b-0bad4f6e9a91","Type":"ContainerStarted","Data":"b020d91a96d0ed08efa06acf80a2864ebf22a404a625155cb8f5616c4e54b0a6"} Feb 02 09:26:07 crc kubenswrapper[4764]: I0202 09:26:07.385724 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-68d4474c54-98j2g" Feb 02 09:26:07 crc kubenswrapper[4764]: I0202 09:26:07.406990 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=2.4069611 podStartE2EDuration="2.4069611s" podCreationTimestamp="2026-02-02 09:26:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:26:07.403778513 +0000 UTC m=+1130.337502631" watchObservedRunningTime="2026-02-02 09:26:07.4069611 +0000 UTC m=+1130.340685228" Feb 02 09:26:09 crc kubenswrapper[4764]: I0202 09:26:09.208169 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-b47bfbff6-m265b" Feb 02 09:26:09 crc kubenswrapper[4764]: I0202 09:26:09.766461 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Feb 02 09:26:10 crc kubenswrapper[4764]: I0202 09:26:10.587854 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Feb 02 09:26:13 crc kubenswrapper[4764]: I0202 09:26:13.416820 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Feb 02 09:26:13 crc kubenswrapper[4764]: I0202 09:26:13.418091 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Feb 02 09:26:13 crc kubenswrapper[4764]: I0202 09:26:13.420041 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Feb 02 09:26:13 crc kubenswrapper[4764]: I0202 09:26:13.420889 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Feb 02 09:26:13 crc kubenswrapper[4764]: I0202 09:26:13.424326 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-cqx4q" Feb 02 09:26:13 crc kubenswrapper[4764]: I0202 09:26:13.440670 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Feb 02 09:26:13 crc kubenswrapper[4764]: I0202 09:26:13.490551 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zzk48\" (UniqueName: \"kubernetes.io/projected/54c115e4-eceb-49af-96ba-854f48802c73-kube-api-access-zzk48\") pod \"openstackclient\" (UID: \"54c115e4-eceb-49af-96ba-854f48802c73\") " pod="openstack/openstackclient" Feb 02 09:26:13 crc kubenswrapper[4764]: I0202 09:26:13.490637 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/54c115e4-eceb-49af-96ba-854f48802c73-openstack-config-secret\") pod \"openstackclient\" (UID: \"54c115e4-eceb-49af-96ba-854f48802c73\") " pod="openstack/openstackclient" Feb 02 09:26:13 crc kubenswrapper[4764]: I0202 09:26:13.490673 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/54c115e4-eceb-49af-96ba-854f48802c73-openstack-config\") pod \"openstackclient\" (UID: \"54c115e4-eceb-49af-96ba-854f48802c73\") " pod="openstack/openstackclient" Feb 02 09:26:13 crc kubenswrapper[4764]: I0202 09:26:13.490694 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54c115e4-eceb-49af-96ba-854f48802c73-combined-ca-bundle\") pod \"openstackclient\" (UID: \"54c115e4-eceb-49af-96ba-854f48802c73\") " pod="openstack/openstackclient" Feb 02 09:26:13 crc kubenswrapper[4764]: I0202 09:26:13.522661 4764 patch_prober.go:28] interesting pod/machine-config-daemon-4ndm4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 09:26:13 crc kubenswrapper[4764]: I0202 09:26:13.522717 4764 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 09:26:13 crc kubenswrapper[4764]: I0202 09:26:13.591896 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/54c115e4-eceb-49af-96ba-854f48802c73-openstack-config-secret\") pod \"openstackclient\" (UID: \"54c115e4-eceb-49af-96ba-854f48802c73\") " pod="openstack/openstackclient" Feb 02 09:26:13 crc kubenswrapper[4764]: I0202 09:26:13.592026 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/54c115e4-eceb-49af-96ba-854f48802c73-openstack-config\") pod \"openstackclient\" (UID: \"54c115e4-eceb-49af-96ba-854f48802c73\") " pod="openstack/openstackclient" Feb 02 09:26:13 crc kubenswrapper[4764]: I0202 09:26:13.592051 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54c115e4-eceb-49af-96ba-854f48802c73-combined-ca-bundle\") pod \"openstackclient\" (UID: \"54c115e4-eceb-49af-96ba-854f48802c73\") " pod="openstack/openstackclient" Feb 02 09:26:13 crc kubenswrapper[4764]: I0202 09:26:13.592134 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zzk48\" (UniqueName: \"kubernetes.io/projected/54c115e4-eceb-49af-96ba-854f48802c73-kube-api-access-zzk48\") pod \"openstackclient\" (UID: \"54c115e4-eceb-49af-96ba-854f48802c73\") " pod="openstack/openstackclient" Feb 02 09:26:13 crc kubenswrapper[4764]: I0202 09:26:13.593668 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/54c115e4-eceb-49af-96ba-854f48802c73-openstack-config\") pod \"openstackclient\" (UID: \"54c115e4-eceb-49af-96ba-854f48802c73\") " pod="openstack/openstackclient" Feb 02 09:26:13 crc kubenswrapper[4764]: I0202 09:26:13.599876 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54c115e4-eceb-49af-96ba-854f48802c73-combined-ca-bundle\") pod \"openstackclient\" (UID: \"54c115e4-eceb-49af-96ba-854f48802c73\") " pod="openstack/openstackclient" Feb 02 09:26:13 crc kubenswrapper[4764]: I0202 09:26:13.602238 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/54c115e4-eceb-49af-96ba-854f48802c73-openstack-config-secret\") pod \"openstackclient\" (UID: \"54c115e4-eceb-49af-96ba-854f48802c73\") " pod="openstack/openstackclient" Feb 02 09:26:13 crc kubenswrapper[4764]: I0202 09:26:13.607100 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zzk48\" (UniqueName: \"kubernetes.io/projected/54c115e4-eceb-49af-96ba-854f48802c73-kube-api-access-zzk48\") pod \"openstackclient\" (UID: \"54c115e4-eceb-49af-96ba-854f48802c73\") " pod="openstack/openstackclient" Feb 02 09:26:13 crc kubenswrapper[4764]: I0202 09:26:13.735118 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Feb 02 09:26:14 crc kubenswrapper[4764]: I0202 09:26:14.210791 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Feb 02 09:26:14 crc kubenswrapper[4764]: I0202 09:26:14.442397 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"54c115e4-eceb-49af-96ba-854f48802c73","Type":"ContainerStarted","Data":"cac6208abd372ba45a915c61c40fcc2a0539cf97aa92c8dd7cf4a34c899e8d8f"} Feb 02 09:26:15 crc kubenswrapper[4764]: I0202 09:26:15.821331 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Feb 02 09:26:24 crc kubenswrapper[4764]: I0202 09:26:24.533403 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"54c115e4-eceb-49af-96ba-854f48802c73","Type":"ContainerStarted","Data":"17c9822b9287b7b099bc3f00044933baaf833f2de41192a0cef37ab59016fac5"} Feb 02 09:26:24 crc kubenswrapper[4764]: I0202 09:26:24.554536 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.235191402 podStartE2EDuration="11.554520178s" podCreationTimestamp="2026-02-02 09:26:13 +0000 UTC" firstStartedPulling="2026-02-02 09:26:14.214724212 +0000 UTC m=+1137.148448300" lastFinishedPulling="2026-02-02 09:26:23.534052988 +0000 UTC m=+1146.467777076" observedRunningTime="2026-02-02 09:26:24.550757355 +0000 UTC m=+1147.484481443" watchObservedRunningTime="2026-02-02 09:26:24.554520178 +0000 UTC m=+1147.488244266" Feb 02 09:26:24 crc kubenswrapper[4764]: I0202 09:26:24.902281 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 02 09:26:24 crc kubenswrapper[4764]: I0202 09:26:24.902747 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2a132428-772f-4d1a-bd11-bf9f69301bfa" containerName="ceilometer-central-agent" containerID="cri-o://3386c5db9eb0a841c11d8c70caac741e63d302e60b74f18c22e2bfb79426d277" gracePeriod=30 Feb 02 09:26:24 crc kubenswrapper[4764]: I0202 09:26:24.902834 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2a132428-772f-4d1a-bd11-bf9f69301bfa" containerName="sg-core" containerID="cri-o://54eef320d116977e546ffca1391a1214baff8b9d8ec4a69ff013e5e639a4a3a3" gracePeriod=30 Feb 02 09:26:24 crc kubenswrapper[4764]: I0202 09:26:24.902867 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2a132428-772f-4d1a-bd11-bf9f69301bfa" containerName="ceilometer-notification-agent" containerID="cri-o://4d51a30a1a083b0634c344cfdb52ad8831f3eb816ca0a7bafe73b8536d9c6f8f" gracePeriod=30 Feb 02 09:26:24 crc kubenswrapper[4764]: I0202 09:26:24.903123 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2a132428-772f-4d1a-bd11-bf9f69301bfa" containerName="proxy-httpd" containerID="cri-o://9c53f871c848ed0cc02a5e809afa4de8cb1fb273467711b34ff71e95cd140427" gracePeriod=30 Feb 02 09:26:24 crc kubenswrapper[4764]: I0202 09:26:24.928008 4764 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="2a132428-772f-4d1a-bd11-bf9f69301bfa" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 502" Feb 02 09:26:25 crc kubenswrapper[4764]: I0202 09:26:25.421312 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-6b9f774cdf-j8pdr" Feb 02 09:26:25 crc kubenswrapper[4764]: I0202 09:26:25.504421 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-7bd869cccb-wnw6g"] Feb 02 09:26:25 crc kubenswrapper[4764]: I0202 09:26:25.505719 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-7bd869cccb-wnw6g" podUID="d63e0dd6-3a9b-4ac4-8a81-a0a50fc4c2bb" containerName="neutron-api" containerID="cri-o://b85d54c7bce23ff6e47f6ec753906bcc21fe69a94f2b437e86386fa7b0a474fc" gracePeriod=30 Feb 02 09:26:25 crc kubenswrapper[4764]: I0202 09:26:25.507718 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-7bd869cccb-wnw6g" podUID="d63e0dd6-3a9b-4ac4-8a81-a0a50fc4c2bb" containerName="neutron-httpd" containerID="cri-o://7db2f812a41d2d8d5ff08564eb909dcecd25bc8f721bb2a9e01ad5ae0a5c26a5" gracePeriod=30 Feb 02 09:26:25 crc kubenswrapper[4764]: I0202 09:26:25.545808 4764 generic.go:334] "Generic (PLEG): container finished" podID="2a132428-772f-4d1a-bd11-bf9f69301bfa" containerID="9c53f871c848ed0cc02a5e809afa4de8cb1fb273467711b34ff71e95cd140427" exitCode=0 Feb 02 09:26:25 crc kubenswrapper[4764]: I0202 09:26:25.545843 4764 generic.go:334] "Generic (PLEG): container finished" podID="2a132428-772f-4d1a-bd11-bf9f69301bfa" containerID="54eef320d116977e546ffca1391a1214baff8b9d8ec4a69ff013e5e639a4a3a3" exitCode=2 Feb 02 09:26:25 crc kubenswrapper[4764]: I0202 09:26:25.545851 4764 generic.go:334] "Generic (PLEG): container finished" podID="2a132428-772f-4d1a-bd11-bf9f69301bfa" containerID="3386c5db9eb0a841c11d8c70caac741e63d302e60b74f18c22e2bfb79426d277" exitCode=0 Feb 02 09:26:25 crc kubenswrapper[4764]: I0202 09:26:25.546394 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2a132428-772f-4d1a-bd11-bf9f69301bfa","Type":"ContainerDied","Data":"9c53f871c848ed0cc02a5e809afa4de8cb1fb273467711b34ff71e95cd140427"} Feb 02 09:26:25 crc kubenswrapper[4764]: I0202 09:26:25.546454 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2a132428-772f-4d1a-bd11-bf9f69301bfa","Type":"ContainerDied","Data":"54eef320d116977e546ffca1391a1214baff8b9d8ec4a69ff013e5e639a4a3a3"} Feb 02 09:26:25 crc kubenswrapper[4764]: I0202 09:26:25.546466 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2a132428-772f-4d1a-bd11-bf9f69301bfa","Type":"ContainerDied","Data":"3386c5db9eb0a841c11d8c70caac741e63d302e60b74f18c22e2bfb79426d277"} Feb 02 09:26:26 crc kubenswrapper[4764]: I0202 09:26:26.556325 4764 generic.go:334] "Generic (PLEG): container finished" podID="d63e0dd6-3a9b-4ac4-8a81-a0a50fc4c2bb" containerID="7db2f812a41d2d8d5ff08564eb909dcecd25bc8f721bb2a9e01ad5ae0a5c26a5" exitCode=0 Feb 02 09:26:26 crc kubenswrapper[4764]: I0202 09:26:26.556417 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7bd869cccb-wnw6g" event={"ID":"d63e0dd6-3a9b-4ac4-8a81-a0a50fc4c2bb","Type":"ContainerDied","Data":"7db2f812a41d2d8d5ff08564eb909dcecd25bc8f721bb2a9e01ad5ae0a5c26a5"} Feb 02 09:26:27 crc kubenswrapper[4764]: I0202 09:26:27.449557 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 02 09:26:27 crc kubenswrapper[4764]: I0202 09:26:27.544696 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a132428-772f-4d1a-bd11-bf9f69301bfa-combined-ca-bundle\") pod \"2a132428-772f-4d1a-bd11-bf9f69301bfa\" (UID: \"2a132428-772f-4d1a-bd11-bf9f69301bfa\") " Feb 02 09:26:27 crc kubenswrapper[4764]: I0202 09:26:27.544804 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2a132428-772f-4d1a-bd11-bf9f69301bfa-run-httpd\") pod \"2a132428-772f-4d1a-bd11-bf9f69301bfa\" (UID: \"2a132428-772f-4d1a-bd11-bf9f69301bfa\") " Feb 02 09:26:27 crc kubenswrapper[4764]: I0202 09:26:27.545343 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2a132428-772f-4d1a-bd11-bf9f69301bfa-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "2a132428-772f-4d1a-bd11-bf9f69301bfa" (UID: "2a132428-772f-4d1a-bd11-bf9f69301bfa"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 09:26:27 crc kubenswrapper[4764]: I0202 09:26:27.545416 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2a132428-772f-4d1a-bd11-bf9f69301bfa-scripts\") pod \"2a132428-772f-4d1a-bd11-bf9f69301bfa\" (UID: \"2a132428-772f-4d1a-bd11-bf9f69301bfa\") " Feb 02 09:26:27 crc kubenswrapper[4764]: I0202 09:26:27.545814 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c74wc\" (UniqueName: \"kubernetes.io/projected/2a132428-772f-4d1a-bd11-bf9f69301bfa-kube-api-access-c74wc\") pod \"2a132428-772f-4d1a-bd11-bf9f69301bfa\" (UID: \"2a132428-772f-4d1a-bd11-bf9f69301bfa\") " Feb 02 09:26:27 crc kubenswrapper[4764]: I0202 09:26:27.545880 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2a132428-772f-4d1a-bd11-bf9f69301bfa-sg-core-conf-yaml\") pod \"2a132428-772f-4d1a-bd11-bf9f69301bfa\" (UID: \"2a132428-772f-4d1a-bd11-bf9f69301bfa\") " Feb 02 09:26:27 crc kubenswrapper[4764]: I0202 09:26:27.545901 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2a132428-772f-4d1a-bd11-bf9f69301bfa-log-httpd\") pod \"2a132428-772f-4d1a-bd11-bf9f69301bfa\" (UID: \"2a132428-772f-4d1a-bd11-bf9f69301bfa\") " Feb 02 09:26:27 crc kubenswrapper[4764]: I0202 09:26:27.545923 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a132428-772f-4d1a-bd11-bf9f69301bfa-config-data\") pod \"2a132428-772f-4d1a-bd11-bf9f69301bfa\" (UID: \"2a132428-772f-4d1a-bd11-bf9f69301bfa\") " Feb 02 09:26:27 crc kubenswrapper[4764]: I0202 09:26:27.546277 4764 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2a132428-772f-4d1a-bd11-bf9f69301bfa-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 02 09:26:27 crc kubenswrapper[4764]: I0202 09:26:27.553375 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a132428-772f-4d1a-bd11-bf9f69301bfa-scripts" (OuterVolumeSpecName: "scripts") pod "2a132428-772f-4d1a-bd11-bf9f69301bfa" (UID: "2a132428-772f-4d1a-bd11-bf9f69301bfa"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:26:27 crc kubenswrapper[4764]: I0202 09:26:27.558551 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2a132428-772f-4d1a-bd11-bf9f69301bfa-kube-api-access-c74wc" (OuterVolumeSpecName: "kube-api-access-c74wc") pod "2a132428-772f-4d1a-bd11-bf9f69301bfa" (UID: "2a132428-772f-4d1a-bd11-bf9f69301bfa"). InnerVolumeSpecName "kube-api-access-c74wc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:26:27 crc kubenswrapper[4764]: I0202 09:26:27.574131 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2a132428-772f-4d1a-bd11-bf9f69301bfa-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "2a132428-772f-4d1a-bd11-bf9f69301bfa" (UID: "2a132428-772f-4d1a-bd11-bf9f69301bfa"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 09:26:27 crc kubenswrapper[4764]: I0202 09:26:27.601835 4764 generic.go:334] "Generic (PLEG): container finished" podID="2a132428-772f-4d1a-bd11-bf9f69301bfa" containerID="4d51a30a1a083b0634c344cfdb52ad8831f3eb816ca0a7bafe73b8536d9c6f8f" exitCode=0 Feb 02 09:26:27 crc kubenswrapper[4764]: I0202 09:26:27.601878 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2a132428-772f-4d1a-bd11-bf9f69301bfa","Type":"ContainerDied","Data":"4d51a30a1a083b0634c344cfdb52ad8831f3eb816ca0a7bafe73b8536d9c6f8f"} Feb 02 09:26:27 crc kubenswrapper[4764]: I0202 09:26:27.601903 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2a132428-772f-4d1a-bd11-bf9f69301bfa","Type":"ContainerDied","Data":"0d9e5b254ddcc1904aeffaf74ca61243210dbd9838a95eb26b14a4f6fccb0093"} Feb 02 09:26:27 crc kubenswrapper[4764]: I0202 09:26:27.601920 4764 scope.go:117] "RemoveContainer" containerID="9c53f871c848ed0cc02a5e809afa4de8cb1fb273467711b34ff71e95cd140427" Feb 02 09:26:27 crc kubenswrapper[4764]: I0202 09:26:27.602046 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 02 09:26:27 crc kubenswrapper[4764]: I0202 09:26:27.608206 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a132428-772f-4d1a-bd11-bf9f69301bfa-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "2a132428-772f-4d1a-bd11-bf9f69301bfa" (UID: "2a132428-772f-4d1a-bd11-bf9f69301bfa"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:26:27 crc kubenswrapper[4764]: I0202 09:26:27.647679 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c74wc\" (UniqueName: \"kubernetes.io/projected/2a132428-772f-4d1a-bd11-bf9f69301bfa-kube-api-access-c74wc\") on node \"crc\" DevicePath \"\"" Feb 02 09:26:27 crc kubenswrapper[4764]: I0202 09:26:27.647708 4764 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2a132428-772f-4d1a-bd11-bf9f69301bfa-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 02 09:26:27 crc kubenswrapper[4764]: I0202 09:26:27.647717 4764 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2a132428-772f-4d1a-bd11-bf9f69301bfa-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 02 09:26:27 crc kubenswrapper[4764]: I0202 09:26:27.647725 4764 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2a132428-772f-4d1a-bd11-bf9f69301bfa-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 09:26:27 crc kubenswrapper[4764]: I0202 09:26:27.651669 4764 scope.go:117] "RemoveContainer" containerID="54eef320d116977e546ffca1391a1214baff8b9d8ec4a69ff013e5e639a4a3a3" Feb 02 09:26:27 crc kubenswrapper[4764]: I0202 09:26:27.677159 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a132428-772f-4d1a-bd11-bf9f69301bfa-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2a132428-772f-4d1a-bd11-bf9f69301bfa" (UID: "2a132428-772f-4d1a-bd11-bf9f69301bfa"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:26:27 crc kubenswrapper[4764]: I0202 09:26:27.688612 4764 scope.go:117] "RemoveContainer" containerID="4d51a30a1a083b0634c344cfdb52ad8831f3eb816ca0a7bafe73b8536d9c6f8f" Feb 02 09:26:27 crc kubenswrapper[4764]: I0202 09:26:27.707639 4764 scope.go:117] "RemoveContainer" containerID="3386c5db9eb0a841c11d8c70caac741e63d302e60b74f18c22e2bfb79426d277" Feb 02 09:26:27 crc kubenswrapper[4764]: I0202 09:26:27.716556 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a132428-772f-4d1a-bd11-bf9f69301bfa-config-data" (OuterVolumeSpecName: "config-data") pod "2a132428-772f-4d1a-bd11-bf9f69301bfa" (UID: "2a132428-772f-4d1a-bd11-bf9f69301bfa"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:26:27 crc kubenswrapper[4764]: I0202 09:26:27.734304 4764 scope.go:117] "RemoveContainer" containerID="9c53f871c848ed0cc02a5e809afa4de8cb1fb273467711b34ff71e95cd140427" Feb 02 09:26:27 crc kubenswrapper[4764]: E0202 09:26:27.734742 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9c53f871c848ed0cc02a5e809afa4de8cb1fb273467711b34ff71e95cd140427\": container with ID starting with 9c53f871c848ed0cc02a5e809afa4de8cb1fb273467711b34ff71e95cd140427 not found: ID does not exist" containerID="9c53f871c848ed0cc02a5e809afa4de8cb1fb273467711b34ff71e95cd140427" Feb 02 09:26:27 crc kubenswrapper[4764]: I0202 09:26:27.734852 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9c53f871c848ed0cc02a5e809afa4de8cb1fb273467711b34ff71e95cd140427"} err="failed to get container status \"9c53f871c848ed0cc02a5e809afa4de8cb1fb273467711b34ff71e95cd140427\": rpc error: code = NotFound desc = could not find container \"9c53f871c848ed0cc02a5e809afa4de8cb1fb273467711b34ff71e95cd140427\": container with ID starting with 9c53f871c848ed0cc02a5e809afa4de8cb1fb273467711b34ff71e95cd140427 not found: ID does not exist" Feb 02 09:26:27 crc kubenswrapper[4764]: I0202 09:26:27.734986 4764 scope.go:117] "RemoveContainer" containerID="54eef320d116977e546ffca1391a1214baff8b9d8ec4a69ff013e5e639a4a3a3" Feb 02 09:26:27 crc kubenswrapper[4764]: E0202 09:26:27.735586 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"54eef320d116977e546ffca1391a1214baff8b9d8ec4a69ff013e5e639a4a3a3\": container with ID starting with 54eef320d116977e546ffca1391a1214baff8b9d8ec4a69ff013e5e639a4a3a3 not found: ID does not exist" containerID="54eef320d116977e546ffca1391a1214baff8b9d8ec4a69ff013e5e639a4a3a3" Feb 02 09:26:27 crc kubenswrapper[4764]: I0202 09:26:27.735678 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"54eef320d116977e546ffca1391a1214baff8b9d8ec4a69ff013e5e639a4a3a3"} err="failed to get container status \"54eef320d116977e546ffca1391a1214baff8b9d8ec4a69ff013e5e639a4a3a3\": rpc error: code = NotFound desc = could not find container \"54eef320d116977e546ffca1391a1214baff8b9d8ec4a69ff013e5e639a4a3a3\": container with ID starting with 54eef320d116977e546ffca1391a1214baff8b9d8ec4a69ff013e5e639a4a3a3 not found: ID does not exist" Feb 02 09:26:27 crc kubenswrapper[4764]: I0202 09:26:27.735742 4764 scope.go:117] "RemoveContainer" containerID="4d51a30a1a083b0634c344cfdb52ad8831f3eb816ca0a7bafe73b8536d9c6f8f" Feb 02 09:26:27 crc kubenswrapper[4764]: E0202 09:26:27.736063 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4d51a30a1a083b0634c344cfdb52ad8831f3eb816ca0a7bafe73b8536d9c6f8f\": container with ID starting with 4d51a30a1a083b0634c344cfdb52ad8831f3eb816ca0a7bafe73b8536d9c6f8f not found: ID does not exist" containerID="4d51a30a1a083b0634c344cfdb52ad8831f3eb816ca0a7bafe73b8536d9c6f8f" Feb 02 09:26:27 crc kubenswrapper[4764]: I0202 09:26:27.736108 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4d51a30a1a083b0634c344cfdb52ad8831f3eb816ca0a7bafe73b8536d9c6f8f"} err="failed to get container status \"4d51a30a1a083b0634c344cfdb52ad8831f3eb816ca0a7bafe73b8536d9c6f8f\": rpc error: code = NotFound desc = could not find container \"4d51a30a1a083b0634c344cfdb52ad8831f3eb816ca0a7bafe73b8536d9c6f8f\": container with ID starting with 4d51a30a1a083b0634c344cfdb52ad8831f3eb816ca0a7bafe73b8536d9c6f8f not found: ID does not exist" Feb 02 09:26:27 crc kubenswrapper[4764]: I0202 09:26:27.736135 4764 scope.go:117] "RemoveContainer" containerID="3386c5db9eb0a841c11d8c70caac741e63d302e60b74f18c22e2bfb79426d277" Feb 02 09:26:27 crc kubenswrapper[4764]: E0202 09:26:27.736481 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3386c5db9eb0a841c11d8c70caac741e63d302e60b74f18c22e2bfb79426d277\": container with ID starting with 3386c5db9eb0a841c11d8c70caac741e63d302e60b74f18c22e2bfb79426d277 not found: ID does not exist" containerID="3386c5db9eb0a841c11d8c70caac741e63d302e60b74f18c22e2bfb79426d277" Feb 02 09:26:27 crc kubenswrapper[4764]: I0202 09:26:27.736587 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3386c5db9eb0a841c11d8c70caac741e63d302e60b74f18c22e2bfb79426d277"} err="failed to get container status \"3386c5db9eb0a841c11d8c70caac741e63d302e60b74f18c22e2bfb79426d277\": rpc error: code = NotFound desc = could not find container \"3386c5db9eb0a841c11d8c70caac741e63d302e60b74f18c22e2bfb79426d277\": container with ID starting with 3386c5db9eb0a841c11d8c70caac741e63d302e60b74f18c22e2bfb79426d277 not found: ID does not exist" Feb 02 09:26:27 crc kubenswrapper[4764]: I0202 09:26:27.748436 4764 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a132428-772f-4d1a-bd11-bf9f69301bfa-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 09:26:27 crc kubenswrapper[4764]: I0202 09:26:27.748702 4764 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a132428-772f-4d1a-bd11-bf9f69301bfa-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 09:26:27 crc kubenswrapper[4764]: I0202 09:26:27.921923 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 02 09:26:27 crc kubenswrapper[4764]: I0202 09:26:27.944395 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 02 09:26:27 crc kubenswrapper[4764]: I0202 09:26:27.958409 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 02 09:26:27 crc kubenswrapper[4764]: E0202 09:26:27.958740 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a132428-772f-4d1a-bd11-bf9f69301bfa" containerName="proxy-httpd" Feb 02 09:26:27 crc kubenswrapper[4764]: I0202 09:26:27.958756 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a132428-772f-4d1a-bd11-bf9f69301bfa" containerName="proxy-httpd" Feb 02 09:26:27 crc kubenswrapper[4764]: E0202 09:26:27.958769 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a132428-772f-4d1a-bd11-bf9f69301bfa" containerName="ceilometer-notification-agent" Feb 02 09:26:27 crc kubenswrapper[4764]: I0202 09:26:27.958775 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a132428-772f-4d1a-bd11-bf9f69301bfa" containerName="ceilometer-notification-agent" Feb 02 09:26:27 crc kubenswrapper[4764]: E0202 09:26:27.958787 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a132428-772f-4d1a-bd11-bf9f69301bfa" containerName="sg-core" Feb 02 09:26:27 crc kubenswrapper[4764]: I0202 09:26:27.958793 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a132428-772f-4d1a-bd11-bf9f69301bfa" containerName="sg-core" Feb 02 09:26:27 crc kubenswrapper[4764]: E0202 09:26:27.958801 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a132428-772f-4d1a-bd11-bf9f69301bfa" containerName="ceilometer-central-agent" Feb 02 09:26:27 crc kubenswrapper[4764]: I0202 09:26:27.958807 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a132428-772f-4d1a-bd11-bf9f69301bfa" containerName="ceilometer-central-agent" Feb 02 09:26:27 crc kubenswrapper[4764]: I0202 09:26:27.958990 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a132428-772f-4d1a-bd11-bf9f69301bfa" containerName="ceilometer-central-agent" Feb 02 09:26:27 crc kubenswrapper[4764]: I0202 09:26:27.959000 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a132428-772f-4d1a-bd11-bf9f69301bfa" containerName="ceilometer-notification-agent" Feb 02 09:26:27 crc kubenswrapper[4764]: I0202 09:26:27.959010 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a132428-772f-4d1a-bd11-bf9f69301bfa" containerName="sg-core" Feb 02 09:26:27 crc kubenswrapper[4764]: I0202 09:26:27.959021 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a132428-772f-4d1a-bd11-bf9f69301bfa" containerName="proxy-httpd" Feb 02 09:26:27 crc kubenswrapper[4764]: I0202 09:26:27.960372 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 02 09:26:27 crc kubenswrapper[4764]: I0202 09:26:27.966386 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 02 09:26:27 crc kubenswrapper[4764]: I0202 09:26:27.966411 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 02 09:26:27 crc kubenswrapper[4764]: I0202 09:26:27.970778 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 02 09:26:28 crc kubenswrapper[4764]: I0202 09:26:28.054339 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-68v2m\" (UniqueName: \"kubernetes.io/projected/812ecbf4-31e2-49db-98ec-82992886557b-kube-api-access-68v2m\") pod \"ceilometer-0\" (UID: \"812ecbf4-31e2-49db-98ec-82992886557b\") " pod="openstack/ceilometer-0" Feb 02 09:26:28 crc kubenswrapper[4764]: I0202 09:26:28.054412 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/812ecbf4-31e2-49db-98ec-82992886557b-run-httpd\") pod \"ceilometer-0\" (UID: \"812ecbf4-31e2-49db-98ec-82992886557b\") " pod="openstack/ceilometer-0" Feb 02 09:26:28 crc kubenswrapper[4764]: I0202 09:26:28.054438 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/812ecbf4-31e2-49db-98ec-82992886557b-config-data\") pod \"ceilometer-0\" (UID: \"812ecbf4-31e2-49db-98ec-82992886557b\") " pod="openstack/ceilometer-0" Feb 02 09:26:28 crc kubenswrapper[4764]: I0202 09:26:28.054457 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/812ecbf4-31e2-49db-98ec-82992886557b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"812ecbf4-31e2-49db-98ec-82992886557b\") " pod="openstack/ceilometer-0" Feb 02 09:26:28 crc kubenswrapper[4764]: I0202 09:26:28.054474 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/812ecbf4-31e2-49db-98ec-82992886557b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"812ecbf4-31e2-49db-98ec-82992886557b\") " pod="openstack/ceilometer-0" Feb 02 09:26:28 crc kubenswrapper[4764]: I0202 09:26:28.054497 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/812ecbf4-31e2-49db-98ec-82992886557b-scripts\") pod \"ceilometer-0\" (UID: \"812ecbf4-31e2-49db-98ec-82992886557b\") " pod="openstack/ceilometer-0" Feb 02 09:26:28 crc kubenswrapper[4764]: I0202 09:26:28.054548 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/812ecbf4-31e2-49db-98ec-82992886557b-log-httpd\") pod \"ceilometer-0\" (UID: \"812ecbf4-31e2-49db-98ec-82992886557b\") " pod="openstack/ceilometer-0" Feb 02 09:26:28 crc kubenswrapper[4764]: I0202 09:26:28.155494 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-68v2m\" (UniqueName: \"kubernetes.io/projected/812ecbf4-31e2-49db-98ec-82992886557b-kube-api-access-68v2m\") pod \"ceilometer-0\" (UID: \"812ecbf4-31e2-49db-98ec-82992886557b\") " pod="openstack/ceilometer-0" Feb 02 09:26:28 crc kubenswrapper[4764]: I0202 09:26:28.155553 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/812ecbf4-31e2-49db-98ec-82992886557b-run-httpd\") pod \"ceilometer-0\" (UID: \"812ecbf4-31e2-49db-98ec-82992886557b\") " pod="openstack/ceilometer-0" Feb 02 09:26:28 crc kubenswrapper[4764]: I0202 09:26:28.155575 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/812ecbf4-31e2-49db-98ec-82992886557b-config-data\") pod \"ceilometer-0\" (UID: \"812ecbf4-31e2-49db-98ec-82992886557b\") " pod="openstack/ceilometer-0" Feb 02 09:26:28 crc kubenswrapper[4764]: I0202 09:26:28.155593 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/812ecbf4-31e2-49db-98ec-82992886557b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"812ecbf4-31e2-49db-98ec-82992886557b\") " pod="openstack/ceilometer-0" Feb 02 09:26:28 crc kubenswrapper[4764]: I0202 09:26:28.155612 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/812ecbf4-31e2-49db-98ec-82992886557b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"812ecbf4-31e2-49db-98ec-82992886557b\") " pod="openstack/ceilometer-0" Feb 02 09:26:28 crc kubenswrapper[4764]: I0202 09:26:28.155634 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/812ecbf4-31e2-49db-98ec-82992886557b-scripts\") pod \"ceilometer-0\" (UID: \"812ecbf4-31e2-49db-98ec-82992886557b\") " pod="openstack/ceilometer-0" Feb 02 09:26:28 crc kubenswrapper[4764]: I0202 09:26:28.155659 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/812ecbf4-31e2-49db-98ec-82992886557b-log-httpd\") pod \"ceilometer-0\" (UID: \"812ecbf4-31e2-49db-98ec-82992886557b\") " pod="openstack/ceilometer-0" Feb 02 09:26:28 crc kubenswrapper[4764]: I0202 09:26:28.156198 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/812ecbf4-31e2-49db-98ec-82992886557b-run-httpd\") pod \"ceilometer-0\" (UID: \"812ecbf4-31e2-49db-98ec-82992886557b\") " pod="openstack/ceilometer-0" Feb 02 09:26:28 crc kubenswrapper[4764]: I0202 09:26:28.156213 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/812ecbf4-31e2-49db-98ec-82992886557b-log-httpd\") pod \"ceilometer-0\" (UID: \"812ecbf4-31e2-49db-98ec-82992886557b\") " pod="openstack/ceilometer-0" Feb 02 09:26:28 crc kubenswrapper[4764]: I0202 09:26:28.161283 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/812ecbf4-31e2-49db-98ec-82992886557b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"812ecbf4-31e2-49db-98ec-82992886557b\") " pod="openstack/ceilometer-0" Feb 02 09:26:28 crc kubenswrapper[4764]: I0202 09:26:28.161914 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/812ecbf4-31e2-49db-98ec-82992886557b-config-data\") pod \"ceilometer-0\" (UID: \"812ecbf4-31e2-49db-98ec-82992886557b\") " pod="openstack/ceilometer-0" Feb 02 09:26:28 crc kubenswrapper[4764]: I0202 09:26:28.167838 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/812ecbf4-31e2-49db-98ec-82992886557b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"812ecbf4-31e2-49db-98ec-82992886557b\") " pod="openstack/ceilometer-0" Feb 02 09:26:28 crc kubenswrapper[4764]: I0202 09:26:28.174691 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/812ecbf4-31e2-49db-98ec-82992886557b-scripts\") pod \"ceilometer-0\" (UID: \"812ecbf4-31e2-49db-98ec-82992886557b\") " pod="openstack/ceilometer-0" Feb 02 09:26:28 crc kubenswrapper[4764]: I0202 09:26:28.188834 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-68v2m\" (UniqueName: \"kubernetes.io/projected/812ecbf4-31e2-49db-98ec-82992886557b-kube-api-access-68v2m\") pod \"ceilometer-0\" (UID: \"812ecbf4-31e2-49db-98ec-82992886557b\") " pod="openstack/ceilometer-0" Feb 02 09:26:28 crc kubenswrapper[4764]: I0202 09:26:28.274954 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 02 09:26:28 crc kubenswrapper[4764]: I0202 09:26:28.882735 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 02 09:26:28 crc kubenswrapper[4764]: W0202 09:26:28.884808 4764 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod812ecbf4_31e2_49db_98ec_82992886557b.slice/crio-33d2297d59543380d47825935ce2864e6316b0fb01e207fd04c5274f270e6ee4 WatchSource:0}: Error finding container 33d2297d59543380d47825935ce2864e6316b0fb01e207fd04c5274f270e6ee4: Status 404 returned error can't find the container with id 33d2297d59543380d47825935ce2864e6316b0fb01e207fd04c5274f270e6ee4 Feb 02 09:26:29 crc kubenswrapper[4764]: I0202 09:26:29.173076 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 02 09:26:29 crc kubenswrapper[4764]: I0202 09:26:29.626653 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"812ecbf4-31e2-49db-98ec-82992886557b","Type":"ContainerStarted","Data":"b442ad160351ea8a9f086f5433501745e485e752e6379d55db1141934f356a62"} Feb 02 09:26:29 crc kubenswrapper[4764]: I0202 09:26:29.626693 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"812ecbf4-31e2-49db-98ec-82992886557b","Type":"ContainerStarted","Data":"33d2297d59543380d47825935ce2864e6316b0fb01e207fd04c5274f270e6ee4"} Feb 02 09:26:29 crc kubenswrapper[4764]: I0202 09:26:29.836621 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2a132428-772f-4d1a-bd11-bf9f69301bfa" path="/var/lib/kubelet/pods/2a132428-772f-4d1a-bd11-bf9f69301bfa/volumes" Feb 02 09:26:30 crc kubenswrapper[4764]: I0202 09:26:30.634791 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"812ecbf4-31e2-49db-98ec-82992886557b","Type":"ContainerStarted","Data":"ae959cfdce12c0cd337f49be9499023ed6ef970512d3aa412b6eb7fc974091e2"} Feb 02 09:26:31 crc kubenswrapper[4764]: I0202 09:26:31.666151 4764 generic.go:334] "Generic (PLEG): container finished" podID="d63e0dd6-3a9b-4ac4-8a81-a0a50fc4c2bb" containerID="b85d54c7bce23ff6e47f6ec753906bcc21fe69a94f2b437e86386fa7b0a474fc" exitCode=0 Feb 02 09:26:31 crc kubenswrapper[4764]: I0202 09:26:31.666238 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7bd869cccb-wnw6g" event={"ID":"d63e0dd6-3a9b-4ac4-8a81-a0a50fc4c2bb","Type":"ContainerDied","Data":"b85d54c7bce23ff6e47f6ec753906bcc21fe69a94f2b437e86386fa7b0a474fc"} Feb 02 09:26:31 crc kubenswrapper[4764]: I0202 09:26:31.670603 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"812ecbf4-31e2-49db-98ec-82992886557b","Type":"ContainerStarted","Data":"7cfbe64ae6de76de4b3bd49d5f588b9950147f47366928c962f5fb60ebfbdd1d"} Feb 02 09:26:31 crc kubenswrapper[4764]: I0202 09:26:31.937055 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7bd869cccb-wnw6g" Feb 02 09:26:32 crc kubenswrapper[4764]: I0202 09:26:32.022209 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kk9dj\" (UniqueName: \"kubernetes.io/projected/d63e0dd6-3a9b-4ac4-8a81-a0a50fc4c2bb-kube-api-access-kk9dj\") pod \"d63e0dd6-3a9b-4ac4-8a81-a0a50fc4c2bb\" (UID: \"d63e0dd6-3a9b-4ac4-8a81-a0a50fc4c2bb\") " Feb 02 09:26:32 crc kubenswrapper[4764]: I0202 09:26:32.022248 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d63e0dd6-3a9b-4ac4-8a81-a0a50fc4c2bb-combined-ca-bundle\") pod \"d63e0dd6-3a9b-4ac4-8a81-a0a50fc4c2bb\" (UID: \"d63e0dd6-3a9b-4ac4-8a81-a0a50fc4c2bb\") " Feb 02 09:26:32 crc kubenswrapper[4764]: I0202 09:26:32.022438 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d63e0dd6-3a9b-4ac4-8a81-a0a50fc4c2bb-ovndb-tls-certs\") pod \"d63e0dd6-3a9b-4ac4-8a81-a0a50fc4c2bb\" (UID: \"d63e0dd6-3a9b-4ac4-8a81-a0a50fc4c2bb\") " Feb 02 09:26:32 crc kubenswrapper[4764]: I0202 09:26:32.022480 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/d63e0dd6-3a9b-4ac4-8a81-a0a50fc4c2bb-config\") pod \"d63e0dd6-3a9b-4ac4-8a81-a0a50fc4c2bb\" (UID: \"d63e0dd6-3a9b-4ac4-8a81-a0a50fc4c2bb\") " Feb 02 09:26:32 crc kubenswrapper[4764]: I0202 09:26:32.022499 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/d63e0dd6-3a9b-4ac4-8a81-a0a50fc4c2bb-httpd-config\") pod \"d63e0dd6-3a9b-4ac4-8a81-a0a50fc4c2bb\" (UID: \"d63e0dd6-3a9b-4ac4-8a81-a0a50fc4c2bb\") " Feb 02 09:26:32 crc kubenswrapper[4764]: I0202 09:26:32.029703 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d63e0dd6-3a9b-4ac4-8a81-a0a50fc4c2bb-kube-api-access-kk9dj" (OuterVolumeSpecName: "kube-api-access-kk9dj") pod "d63e0dd6-3a9b-4ac4-8a81-a0a50fc4c2bb" (UID: "d63e0dd6-3a9b-4ac4-8a81-a0a50fc4c2bb"). InnerVolumeSpecName "kube-api-access-kk9dj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:26:32 crc kubenswrapper[4764]: I0202 09:26:32.031746 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d63e0dd6-3a9b-4ac4-8a81-a0a50fc4c2bb-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "d63e0dd6-3a9b-4ac4-8a81-a0a50fc4c2bb" (UID: "d63e0dd6-3a9b-4ac4-8a81-a0a50fc4c2bb"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:26:32 crc kubenswrapper[4764]: I0202 09:26:32.098989 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d63e0dd6-3a9b-4ac4-8a81-a0a50fc4c2bb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d63e0dd6-3a9b-4ac4-8a81-a0a50fc4c2bb" (UID: "d63e0dd6-3a9b-4ac4-8a81-a0a50fc4c2bb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:26:32 crc kubenswrapper[4764]: I0202 09:26:32.131379 4764 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/d63e0dd6-3a9b-4ac4-8a81-a0a50fc4c2bb-httpd-config\") on node \"crc\" DevicePath \"\"" Feb 02 09:26:32 crc kubenswrapper[4764]: I0202 09:26:32.131421 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kk9dj\" (UniqueName: \"kubernetes.io/projected/d63e0dd6-3a9b-4ac4-8a81-a0a50fc4c2bb-kube-api-access-kk9dj\") on node \"crc\" DevicePath \"\"" Feb 02 09:26:32 crc kubenswrapper[4764]: I0202 09:26:32.131434 4764 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d63e0dd6-3a9b-4ac4-8a81-a0a50fc4c2bb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 09:26:32 crc kubenswrapper[4764]: I0202 09:26:32.146057 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d63e0dd6-3a9b-4ac4-8a81-a0a50fc4c2bb-config" (OuterVolumeSpecName: "config") pod "d63e0dd6-3a9b-4ac4-8a81-a0a50fc4c2bb" (UID: "d63e0dd6-3a9b-4ac4-8a81-a0a50fc4c2bb"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:26:32 crc kubenswrapper[4764]: I0202 09:26:32.149635 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d63e0dd6-3a9b-4ac4-8a81-a0a50fc4c2bb-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "d63e0dd6-3a9b-4ac4-8a81-a0a50fc4c2bb" (UID: "d63e0dd6-3a9b-4ac4-8a81-a0a50fc4c2bb"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:26:32 crc kubenswrapper[4764]: I0202 09:26:32.233199 4764 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d63e0dd6-3a9b-4ac4-8a81-a0a50fc4c2bb-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 02 09:26:32 crc kubenswrapper[4764]: I0202 09:26:32.233522 4764 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/d63e0dd6-3a9b-4ac4-8a81-a0a50fc4c2bb-config\") on node \"crc\" DevicePath \"\"" Feb 02 09:26:32 crc kubenswrapper[4764]: I0202 09:26:32.680669 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7bd869cccb-wnw6g" event={"ID":"d63e0dd6-3a9b-4ac4-8a81-a0a50fc4c2bb","Type":"ContainerDied","Data":"e4a0856af1974d6db7b83e8050802367153fe4f357d12e03c204b4e1cdb41a7b"} Feb 02 09:26:32 crc kubenswrapper[4764]: I0202 09:26:32.680741 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7bd869cccb-wnw6g" Feb 02 09:26:32 crc kubenswrapper[4764]: I0202 09:26:32.681433 4764 scope.go:117] "RemoveContainer" containerID="7db2f812a41d2d8d5ff08564eb909dcecd25bc8f721bb2a9e01ad5ae0a5c26a5" Feb 02 09:26:32 crc kubenswrapper[4764]: I0202 09:26:32.700815 4764 scope.go:117] "RemoveContainer" containerID="b85d54c7bce23ff6e47f6ec753906bcc21fe69a94f2b437e86386fa7b0a474fc" Feb 02 09:26:32 crc kubenswrapper[4764]: I0202 09:26:32.732863 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-7bd869cccb-wnw6g"] Feb 02 09:26:32 crc kubenswrapper[4764]: I0202 09:26:32.740960 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-7bd869cccb-wnw6g"] Feb 02 09:26:33 crc kubenswrapper[4764]: I0202 09:26:33.694739 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"812ecbf4-31e2-49db-98ec-82992886557b","Type":"ContainerStarted","Data":"1d3bb6fd23c8071400e1bfce6e4a9329c92de571ae3ec99411fc07ea41ded3ed"} Feb 02 09:26:33 crc kubenswrapper[4764]: I0202 09:26:33.694896 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="812ecbf4-31e2-49db-98ec-82992886557b" containerName="ceilometer-central-agent" containerID="cri-o://b442ad160351ea8a9f086f5433501745e485e752e6379d55db1141934f356a62" gracePeriod=30 Feb 02 09:26:33 crc kubenswrapper[4764]: I0202 09:26:33.695227 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 02 09:26:33 crc kubenswrapper[4764]: I0202 09:26:33.695540 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="812ecbf4-31e2-49db-98ec-82992886557b" containerName="sg-core" containerID="cri-o://7cfbe64ae6de76de4b3bd49d5f588b9950147f47366928c962f5fb60ebfbdd1d" gracePeriod=30 Feb 02 09:26:33 crc kubenswrapper[4764]: I0202 09:26:33.695622 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="812ecbf4-31e2-49db-98ec-82992886557b" containerName="proxy-httpd" containerID="cri-o://1d3bb6fd23c8071400e1bfce6e4a9329c92de571ae3ec99411fc07ea41ded3ed" gracePeriod=30 Feb 02 09:26:33 crc kubenswrapper[4764]: I0202 09:26:33.695698 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="812ecbf4-31e2-49db-98ec-82992886557b" containerName="ceilometer-notification-agent" containerID="cri-o://ae959cfdce12c0cd337f49be9499023ed6ef970512d3aa412b6eb7fc974091e2" gracePeriod=30 Feb 02 09:26:33 crc kubenswrapper[4764]: I0202 09:26:33.733700 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.8823608099999998 podStartE2EDuration="6.733683694s" podCreationTimestamp="2026-02-02 09:26:27 +0000 UTC" firstStartedPulling="2026-02-02 09:26:28.887296033 +0000 UTC m=+1151.821020121" lastFinishedPulling="2026-02-02 09:26:32.738618897 +0000 UTC m=+1155.672343005" observedRunningTime="2026-02-02 09:26:33.726507809 +0000 UTC m=+1156.660231897" watchObservedRunningTime="2026-02-02 09:26:33.733683694 +0000 UTC m=+1156.667407772" Feb 02 09:26:33 crc kubenswrapper[4764]: I0202 09:26:33.857544 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d63e0dd6-3a9b-4ac4-8a81-a0a50fc4c2bb" path="/var/lib/kubelet/pods/d63e0dd6-3a9b-4ac4-8a81-a0a50fc4c2bb/volumes" Feb 02 09:26:34 crc kubenswrapper[4764]: I0202 09:26:34.717630 4764 generic.go:334] "Generic (PLEG): container finished" podID="812ecbf4-31e2-49db-98ec-82992886557b" containerID="1d3bb6fd23c8071400e1bfce6e4a9329c92de571ae3ec99411fc07ea41ded3ed" exitCode=0 Feb 02 09:26:34 crc kubenswrapper[4764]: I0202 09:26:34.717666 4764 generic.go:334] "Generic (PLEG): container finished" podID="812ecbf4-31e2-49db-98ec-82992886557b" containerID="7cfbe64ae6de76de4b3bd49d5f588b9950147f47366928c962f5fb60ebfbdd1d" exitCode=2 Feb 02 09:26:34 crc kubenswrapper[4764]: I0202 09:26:34.717674 4764 generic.go:334] "Generic (PLEG): container finished" podID="812ecbf4-31e2-49db-98ec-82992886557b" containerID="ae959cfdce12c0cd337f49be9499023ed6ef970512d3aa412b6eb7fc974091e2" exitCode=0 Feb 02 09:26:34 crc kubenswrapper[4764]: I0202 09:26:34.717697 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"812ecbf4-31e2-49db-98ec-82992886557b","Type":"ContainerDied","Data":"1d3bb6fd23c8071400e1bfce6e4a9329c92de571ae3ec99411fc07ea41ded3ed"} Feb 02 09:26:34 crc kubenswrapper[4764]: I0202 09:26:34.717723 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"812ecbf4-31e2-49db-98ec-82992886557b","Type":"ContainerDied","Data":"7cfbe64ae6de76de4b3bd49d5f588b9950147f47366928c962f5fb60ebfbdd1d"} Feb 02 09:26:34 crc kubenswrapper[4764]: I0202 09:26:34.717733 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"812ecbf4-31e2-49db-98ec-82992886557b","Type":"ContainerDied","Data":"ae959cfdce12c0cd337f49be9499023ed6ef970512d3aa412b6eb7fc974091e2"} Feb 02 09:26:36 crc kubenswrapper[4764]: I0202 09:26:36.328289 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-68d4474c54-98j2g" Feb 02 09:26:36 crc kubenswrapper[4764]: I0202 09:26:36.641216 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-68d4474c54-98j2g" Feb 02 09:26:36 crc kubenswrapper[4764]: I0202 09:26:36.704053 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-65fcd88786-dggfp"] Feb 02 09:26:36 crc kubenswrapper[4764]: I0202 09:26:36.704437 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-65fcd88786-dggfp" podUID="dedd7b1f-7ec4-4a7a-a703-7e228f7767b1" containerName="placement-log" containerID="cri-o://990d29535b531c195dc668ed0f9c16b9592d389ae625fafcb51d55c1040391d2" gracePeriod=30 Feb 02 09:26:36 crc kubenswrapper[4764]: I0202 09:26:36.704885 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-65fcd88786-dggfp" podUID="dedd7b1f-7ec4-4a7a-a703-7e228f7767b1" containerName="placement-api" containerID="cri-o://a3f865f4635480487ff622f30424f2b9a0ebc2e202ceb9b00303e7ebcdad3716" gracePeriod=30 Feb 02 09:26:37 crc kubenswrapper[4764]: I0202 09:26:37.010108 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-8ms85"] Feb 02 09:26:37 crc kubenswrapper[4764]: E0202 09:26:37.016048 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d63e0dd6-3a9b-4ac4-8a81-a0a50fc4c2bb" containerName="neutron-api" Feb 02 09:26:37 crc kubenswrapper[4764]: I0202 09:26:37.016081 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="d63e0dd6-3a9b-4ac4-8a81-a0a50fc4c2bb" containerName="neutron-api" Feb 02 09:26:37 crc kubenswrapper[4764]: E0202 09:26:37.016094 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d63e0dd6-3a9b-4ac4-8a81-a0a50fc4c2bb" containerName="neutron-httpd" Feb 02 09:26:37 crc kubenswrapper[4764]: I0202 09:26:37.016102 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="d63e0dd6-3a9b-4ac4-8a81-a0a50fc4c2bb" containerName="neutron-httpd" Feb 02 09:26:37 crc kubenswrapper[4764]: I0202 09:26:37.016312 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="d63e0dd6-3a9b-4ac4-8a81-a0a50fc4c2bb" containerName="neutron-httpd" Feb 02 09:26:37 crc kubenswrapper[4764]: I0202 09:26:37.016335 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="d63e0dd6-3a9b-4ac4-8a81-a0a50fc4c2bb" containerName="neutron-api" Feb 02 09:26:37 crc kubenswrapper[4764]: I0202 09:26:37.016849 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-8ms85" Feb 02 09:26:37 crc kubenswrapper[4764]: I0202 09:26:37.043632 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-8ms85"] Feb 02 09:26:37 crc kubenswrapper[4764]: I0202 09:26:37.110093 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/64d8d838-0afe-4f39-ae93-027fea76ced1-operator-scripts\") pod \"nova-api-db-create-8ms85\" (UID: \"64d8d838-0afe-4f39-ae93-027fea76ced1\") " pod="openstack/nova-api-db-create-8ms85" Feb 02 09:26:37 crc kubenswrapper[4764]: I0202 09:26:37.110194 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dvv25\" (UniqueName: \"kubernetes.io/projected/64d8d838-0afe-4f39-ae93-027fea76ced1-kube-api-access-dvv25\") pod \"nova-api-db-create-8ms85\" (UID: \"64d8d838-0afe-4f39-ae93-027fea76ced1\") " pod="openstack/nova-api-db-create-8ms85" Feb 02 09:26:37 crc kubenswrapper[4764]: I0202 09:26:37.114685 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-scmkd"] Feb 02 09:26:37 crc kubenswrapper[4764]: I0202 09:26:37.116754 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-scmkd" Feb 02 09:26:37 crc kubenswrapper[4764]: I0202 09:26:37.208155 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-91bb-account-create-update-gg7c4"] Feb 02 09:26:37 crc kubenswrapper[4764]: I0202 09:26:37.209204 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-91bb-account-create-update-gg7c4" Feb 02 09:26:37 crc kubenswrapper[4764]: I0202 09:26:37.212230 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Feb 02 09:26:37 crc kubenswrapper[4764]: I0202 09:26:37.217357 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/64d8d838-0afe-4f39-ae93-027fea76ced1-operator-scripts\") pod \"nova-api-db-create-8ms85\" (UID: \"64d8d838-0afe-4f39-ae93-027fea76ced1\") " pod="openstack/nova-api-db-create-8ms85" Feb 02 09:26:37 crc kubenswrapper[4764]: I0202 09:26:37.217430 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dvv25\" (UniqueName: \"kubernetes.io/projected/64d8d838-0afe-4f39-ae93-027fea76ced1-kube-api-access-dvv25\") pod \"nova-api-db-create-8ms85\" (UID: \"64d8d838-0afe-4f39-ae93-027fea76ced1\") " pod="openstack/nova-api-db-create-8ms85" Feb 02 09:26:37 crc kubenswrapper[4764]: I0202 09:26:37.217508 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2ccd54a4-fbee-4ad6-b0eb-9d22fdb1eebe-operator-scripts\") pod \"nova-cell0-db-create-scmkd\" (UID: \"2ccd54a4-fbee-4ad6-b0eb-9d22fdb1eebe\") " pod="openstack/nova-cell0-db-create-scmkd" Feb 02 09:26:37 crc kubenswrapper[4764]: I0202 09:26:37.217561 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-db9gm\" (UniqueName: \"kubernetes.io/projected/2ccd54a4-fbee-4ad6-b0eb-9d22fdb1eebe-kube-api-access-db9gm\") pod \"nova-cell0-db-create-scmkd\" (UID: \"2ccd54a4-fbee-4ad6-b0eb-9d22fdb1eebe\") " pod="openstack/nova-cell0-db-create-scmkd" Feb 02 09:26:37 crc kubenswrapper[4764]: I0202 09:26:37.218041 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/64d8d838-0afe-4f39-ae93-027fea76ced1-operator-scripts\") pod \"nova-api-db-create-8ms85\" (UID: \"64d8d838-0afe-4f39-ae93-027fea76ced1\") " pod="openstack/nova-api-db-create-8ms85" Feb 02 09:26:37 crc kubenswrapper[4764]: I0202 09:26:37.219646 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-91bb-account-create-update-gg7c4"] Feb 02 09:26:37 crc kubenswrapper[4764]: I0202 09:26:37.230007 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-scmkd"] Feb 02 09:26:37 crc kubenswrapper[4764]: I0202 09:26:37.260465 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dvv25\" (UniqueName: \"kubernetes.io/projected/64d8d838-0afe-4f39-ae93-027fea76ced1-kube-api-access-dvv25\") pod \"nova-api-db-create-8ms85\" (UID: \"64d8d838-0afe-4f39-ae93-027fea76ced1\") " pod="openstack/nova-api-db-create-8ms85" Feb 02 09:26:37 crc kubenswrapper[4764]: I0202 09:26:37.312995 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-448kc"] Feb 02 09:26:37 crc kubenswrapper[4764]: I0202 09:26:37.314180 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-448kc" Feb 02 09:26:37 crc kubenswrapper[4764]: I0202 09:26:37.318612 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2ccd54a4-fbee-4ad6-b0eb-9d22fdb1eebe-operator-scripts\") pod \"nova-cell0-db-create-scmkd\" (UID: \"2ccd54a4-fbee-4ad6-b0eb-9d22fdb1eebe\") " pod="openstack/nova-cell0-db-create-scmkd" Feb 02 09:26:37 crc kubenswrapper[4764]: I0202 09:26:37.318664 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-db9gm\" (UniqueName: \"kubernetes.io/projected/2ccd54a4-fbee-4ad6-b0eb-9d22fdb1eebe-kube-api-access-db9gm\") pod \"nova-cell0-db-create-scmkd\" (UID: \"2ccd54a4-fbee-4ad6-b0eb-9d22fdb1eebe\") " pod="openstack/nova-cell0-db-create-scmkd" Feb 02 09:26:37 crc kubenswrapper[4764]: I0202 09:26:37.318775 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3ee0de70-5bfb-4b6e-974c-601071027218-operator-scripts\") pod \"nova-api-91bb-account-create-update-gg7c4\" (UID: \"3ee0de70-5bfb-4b6e-974c-601071027218\") " pod="openstack/nova-api-91bb-account-create-update-gg7c4" Feb 02 09:26:37 crc kubenswrapper[4764]: I0202 09:26:37.318811 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7xsvj\" (UniqueName: \"kubernetes.io/projected/3ee0de70-5bfb-4b6e-974c-601071027218-kube-api-access-7xsvj\") pod \"nova-api-91bb-account-create-update-gg7c4\" (UID: \"3ee0de70-5bfb-4b6e-974c-601071027218\") " pod="openstack/nova-api-91bb-account-create-update-gg7c4" Feb 02 09:26:37 crc kubenswrapper[4764]: I0202 09:26:37.319498 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2ccd54a4-fbee-4ad6-b0eb-9d22fdb1eebe-operator-scripts\") pod \"nova-cell0-db-create-scmkd\" (UID: \"2ccd54a4-fbee-4ad6-b0eb-9d22fdb1eebe\") " pod="openstack/nova-cell0-db-create-scmkd" Feb 02 09:26:37 crc kubenswrapper[4764]: I0202 09:26:37.326575 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-448kc"] Feb 02 09:26:37 crc kubenswrapper[4764]: I0202 09:26:37.339305 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-8ms85" Feb 02 09:26:37 crc kubenswrapper[4764]: I0202 09:26:37.358806 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-db9gm\" (UniqueName: \"kubernetes.io/projected/2ccd54a4-fbee-4ad6-b0eb-9d22fdb1eebe-kube-api-access-db9gm\") pod \"nova-cell0-db-create-scmkd\" (UID: \"2ccd54a4-fbee-4ad6-b0eb-9d22fdb1eebe\") " pod="openstack/nova-cell0-db-create-scmkd" Feb 02 09:26:37 crc kubenswrapper[4764]: I0202 09:26:37.420051 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3ee0de70-5bfb-4b6e-974c-601071027218-operator-scripts\") pod \"nova-api-91bb-account-create-update-gg7c4\" (UID: \"3ee0de70-5bfb-4b6e-974c-601071027218\") " pod="openstack/nova-api-91bb-account-create-update-gg7c4" Feb 02 09:26:37 crc kubenswrapper[4764]: I0202 09:26:37.420261 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7xsvj\" (UniqueName: \"kubernetes.io/projected/3ee0de70-5bfb-4b6e-974c-601071027218-kube-api-access-7xsvj\") pod \"nova-api-91bb-account-create-update-gg7c4\" (UID: \"3ee0de70-5bfb-4b6e-974c-601071027218\") " pod="openstack/nova-api-91bb-account-create-update-gg7c4" Feb 02 09:26:37 crc kubenswrapper[4764]: I0202 09:26:37.420385 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d492e01f-38f0-4c76-b2aa-a177760a757f-operator-scripts\") pod \"nova-cell1-db-create-448kc\" (UID: \"d492e01f-38f0-4c76-b2aa-a177760a757f\") " pod="openstack/nova-cell1-db-create-448kc" Feb 02 09:26:37 crc kubenswrapper[4764]: I0202 09:26:37.420567 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-626gj\" (UniqueName: \"kubernetes.io/projected/d492e01f-38f0-4c76-b2aa-a177760a757f-kube-api-access-626gj\") pod \"nova-cell1-db-create-448kc\" (UID: \"d492e01f-38f0-4c76-b2aa-a177760a757f\") " pod="openstack/nova-cell1-db-create-448kc" Feb 02 09:26:37 crc kubenswrapper[4764]: I0202 09:26:37.421450 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3ee0de70-5bfb-4b6e-974c-601071027218-operator-scripts\") pod \"nova-api-91bb-account-create-update-gg7c4\" (UID: \"3ee0de70-5bfb-4b6e-974c-601071027218\") " pod="openstack/nova-api-91bb-account-create-update-gg7c4" Feb 02 09:26:37 crc kubenswrapper[4764]: I0202 09:26:37.423149 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-10df-account-create-update-gb8cd"] Feb 02 09:26:37 crc kubenswrapper[4764]: I0202 09:26:37.424205 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-10df-account-create-update-gb8cd" Feb 02 09:26:37 crc kubenswrapper[4764]: I0202 09:26:37.462244 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-scmkd" Feb 02 09:26:37 crc kubenswrapper[4764]: I0202 09:26:37.463391 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Feb 02 09:26:37 crc kubenswrapper[4764]: I0202 09:26:37.466628 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7xsvj\" (UniqueName: \"kubernetes.io/projected/3ee0de70-5bfb-4b6e-974c-601071027218-kube-api-access-7xsvj\") pod \"nova-api-91bb-account-create-update-gg7c4\" (UID: \"3ee0de70-5bfb-4b6e-974c-601071027218\") " pod="openstack/nova-api-91bb-account-create-update-gg7c4" Feb 02 09:26:37 crc kubenswrapper[4764]: I0202 09:26:37.520691 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-10df-account-create-update-gb8cd"] Feb 02 09:26:37 crc kubenswrapper[4764]: I0202 09:26:37.522233 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-626gj\" (UniqueName: \"kubernetes.io/projected/d492e01f-38f0-4c76-b2aa-a177760a757f-kube-api-access-626gj\") pod \"nova-cell1-db-create-448kc\" (UID: \"d492e01f-38f0-4c76-b2aa-a177760a757f\") " pod="openstack/nova-cell1-db-create-448kc" Feb 02 09:26:37 crc kubenswrapper[4764]: I0202 09:26:37.522322 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d492e01f-38f0-4c76-b2aa-a177760a757f-operator-scripts\") pod \"nova-cell1-db-create-448kc\" (UID: \"d492e01f-38f0-4c76-b2aa-a177760a757f\") " pod="openstack/nova-cell1-db-create-448kc" Feb 02 09:26:37 crc kubenswrapper[4764]: I0202 09:26:37.522389 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2t6cp\" (UniqueName: \"kubernetes.io/projected/902c771b-8da4-4c38-911e-c8b531d076a8-kube-api-access-2t6cp\") pod \"nova-cell0-10df-account-create-update-gb8cd\" (UID: \"902c771b-8da4-4c38-911e-c8b531d076a8\") " pod="openstack/nova-cell0-10df-account-create-update-gb8cd" Feb 02 09:26:37 crc kubenswrapper[4764]: I0202 09:26:37.522453 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/902c771b-8da4-4c38-911e-c8b531d076a8-operator-scripts\") pod \"nova-cell0-10df-account-create-update-gb8cd\" (UID: \"902c771b-8da4-4c38-911e-c8b531d076a8\") " pod="openstack/nova-cell0-10df-account-create-update-gb8cd" Feb 02 09:26:37 crc kubenswrapper[4764]: I0202 09:26:37.525591 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d492e01f-38f0-4c76-b2aa-a177760a757f-operator-scripts\") pod \"nova-cell1-db-create-448kc\" (UID: \"d492e01f-38f0-4c76-b2aa-a177760a757f\") " pod="openstack/nova-cell1-db-create-448kc" Feb 02 09:26:37 crc kubenswrapper[4764]: I0202 09:26:37.529951 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-91bb-account-create-update-gg7c4" Feb 02 09:26:37 crc kubenswrapper[4764]: I0202 09:26:37.550737 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-626gj\" (UniqueName: \"kubernetes.io/projected/d492e01f-38f0-4c76-b2aa-a177760a757f-kube-api-access-626gj\") pod \"nova-cell1-db-create-448kc\" (UID: \"d492e01f-38f0-4c76-b2aa-a177760a757f\") " pod="openstack/nova-cell1-db-create-448kc" Feb 02 09:26:37 crc kubenswrapper[4764]: I0202 09:26:37.618748 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-39d3-account-create-update-mtxtf"] Feb 02 09:26:37 crc kubenswrapper[4764]: I0202 09:26:37.620126 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-39d3-account-create-update-mtxtf" Feb 02 09:26:37 crc kubenswrapper[4764]: I0202 09:26:37.625277 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Feb 02 09:26:37 crc kubenswrapper[4764]: I0202 09:26:37.626340 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2t6cp\" (UniqueName: \"kubernetes.io/projected/902c771b-8da4-4c38-911e-c8b531d076a8-kube-api-access-2t6cp\") pod \"nova-cell0-10df-account-create-update-gb8cd\" (UID: \"902c771b-8da4-4c38-911e-c8b531d076a8\") " pod="openstack/nova-cell0-10df-account-create-update-gb8cd" Feb 02 09:26:37 crc kubenswrapper[4764]: I0202 09:26:37.626726 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/902c771b-8da4-4c38-911e-c8b531d076a8-operator-scripts\") pod \"nova-cell0-10df-account-create-update-gb8cd\" (UID: \"902c771b-8da4-4c38-911e-c8b531d076a8\") " pod="openstack/nova-cell0-10df-account-create-update-gb8cd" Feb 02 09:26:37 crc kubenswrapper[4764]: I0202 09:26:37.627412 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/902c771b-8da4-4c38-911e-c8b531d076a8-operator-scripts\") pod \"nova-cell0-10df-account-create-update-gb8cd\" (UID: \"902c771b-8da4-4c38-911e-c8b531d076a8\") " pod="openstack/nova-cell0-10df-account-create-update-gb8cd" Feb 02 09:26:37 crc kubenswrapper[4764]: I0202 09:26:37.637499 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-448kc" Feb 02 09:26:37 crc kubenswrapper[4764]: I0202 09:26:37.641188 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-39d3-account-create-update-mtxtf"] Feb 02 09:26:37 crc kubenswrapper[4764]: I0202 09:26:37.662904 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2t6cp\" (UniqueName: \"kubernetes.io/projected/902c771b-8da4-4c38-911e-c8b531d076a8-kube-api-access-2t6cp\") pod \"nova-cell0-10df-account-create-update-gb8cd\" (UID: \"902c771b-8da4-4c38-911e-c8b531d076a8\") " pod="openstack/nova-cell0-10df-account-create-update-gb8cd" Feb 02 09:26:37 crc kubenswrapper[4764]: I0202 09:26:37.728956 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/afbe3865-2077-45c5-8a94-3d0a6774d8e2-operator-scripts\") pod \"nova-cell1-39d3-account-create-update-mtxtf\" (UID: \"afbe3865-2077-45c5-8a94-3d0a6774d8e2\") " pod="openstack/nova-cell1-39d3-account-create-update-mtxtf" Feb 02 09:26:37 crc kubenswrapper[4764]: I0202 09:26:37.729028 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vqvwk\" (UniqueName: \"kubernetes.io/projected/afbe3865-2077-45c5-8a94-3d0a6774d8e2-kube-api-access-vqvwk\") pod \"nova-cell1-39d3-account-create-update-mtxtf\" (UID: \"afbe3865-2077-45c5-8a94-3d0a6774d8e2\") " pod="openstack/nova-cell1-39d3-account-create-update-mtxtf" Feb 02 09:26:37 crc kubenswrapper[4764]: I0202 09:26:37.769788 4764 generic.go:334] "Generic (PLEG): container finished" podID="dedd7b1f-7ec4-4a7a-a703-7e228f7767b1" containerID="990d29535b531c195dc668ed0f9c16b9592d389ae625fafcb51d55c1040391d2" exitCode=143 Feb 02 09:26:37 crc kubenswrapper[4764]: I0202 09:26:37.769832 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-65fcd88786-dggfp" event={"ID":"dedd7b1f-7ec4-4a7a-a703-7e228f7767b1","Type":"ContainerDied","Data":"990d29535b531c195dc668ed0f9c16b9592d389ae625fafcb51d55c1040391d2"} Feb 02 09:26:37 crc kubenswrapper[4764]: I0202 09:26:37.830538 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/afbe3865-2077-45c5-8a94-3d0a6774d8e2-operator-scripts\") pod \"nova-cell1-39d3-account-create-update-mtxtf\" (UID: \"afbe3865-2077-45c5-8a94-3d0a6774d8e2\") " pod="openstack/nova-cell1-39d3-account-create-update-mtxtf" Feb 02 09:26:37 crc kubenswrapper[4764]: I0202 09:26:37.830950 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vqvwk\" (UniqueName: \"kubernetes.io/projected/afbe3865-2077-45c5-8a94-3d0a6774d8e2-kube-api-access-vqvwk\") pod \"nova-cell1-39d3-account-create-update-mtxtf\" (UID: \"afbe3865-2077-45c5-8a94-3d0a6774d8e2\") " pod="openstack/nova-cell1-39d3-account-create-update-mtxtf" Feb 02 09:26:37 crc kubenswrapper[4764]: I0202 09:26:37.832484 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-10df-account-create-update-gb8cd" Feb 02 09:26:37 crc kubenswrapper[4764]: I0202 09:26:37.834087 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/afbe3865-2077-45c5-8a94-3d0a6774d8e2-operator-scripts\") pod \"nova-cell1-39d3-account-create-update-mtxtf\" (UID: \"afbe3865-2077-45c5-8a94-3d0a6774d8e2\") " pod="openstack/nova-cell1-39d3-account-create-update-mtxtf" Feb 02 09:26:37 crc kubenswrapper[4764]: I0202 09:26:37.880307 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vqvwk\" (UniqueName: \"kubernetes.io/projected/afbe3865-2077-45c5-8a94-3d0a6774d8e2-kube-api-access-vqvwk\") pod \"nova-cell1-39d3-account-create-update-mtxtf\" (UID: \"afbe3865-2077-45c5-8a94-3d0a6774d8e2\") " pod="openstack/nova-cell1-39d3-account-create-update-mtxtf" Feb 02 09:26:37 crc kubenswrapper[4764]: I0202 09:26:37.981449 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-39d3-account-create-update-mtxtf" Feb 02 09:26:38 crc kubenswrapper[4764]: I0202 09:26:38.108999 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-8ms85"] Feb 02 09:26:38 crc kubenswrapper[4764]: I0202 09:26:38.241851 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-91bb-account-create-update-gg7c4"] Feb 02 09:26:38 crc kubenswrapper[4764]: I0202 09:26:38.407969 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-scmkd"] Feb 02 09:26:38 crc kubenswrapper[4764]: I0202 09:26:38.434087 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-448kc"] Feb 02 09:26:38 crc kubenswrapper[4764]: W0202 09:26:38.443729 4764 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2ccd54a4_fbee_4ad6_b0eb_9d22fdb1eebe.slice/crio-2f9ce56e59d7af521e668975b6e7e1dd45cd838d0b03e98772350698aa9429d2 WatchSource:0}: Error finding container 2f9ce56e59d7af521e668975b6e7e1dd45cd838d0b03e98772350698aa9429d2: Status 404 returned error can't find the container with id 2f9ce56e59d7af521e668975b6e7e1dd45cd838d0b03e98772350698aa9429d2 Feb 02 09:26:38 crc kubenswrapper[4764]: I0202 09:26:38.596708 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-10df-account-create-update-gb8cd"] Feb 02 09:26:38 crc kubenswrapper[4764]: I0202 09:26:38.786957 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-91bb-account-create-update-gg7c4" event={"ID":"3ee0de70-5bfb-4b6e-974c-601071027218","Type":"ContainerStarted","Data":"08540cf734e603fff5d5dcb78a4f068507e9eb3462471ce00d9d5cfaf8926d7d"} Feb 02 09:26:38 crc kubenswrapper[4764]: I0202 09:26:38.796622 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-91bb-account-create-update-gg7c4" event={"ID":"3ee0de70-5bfb-4b6e-974c-601071027218","Type":"ContainerStarted","Data":"55784b5be0c377291e9b80c756aa5a694cc7c6e084aa45e3e9b3be27f5a828c4"} Feb 02 09:26:38 crc kubenswrapper[4764]: I0202 09:26:38.797281 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-39d3-account-create-update-mtxtf"] Feb 02 09:26:38 crc kubenswrapper[4764]: I0202 09:26:38.798211 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-448kc" event={"ID":"d492e01f-38f0-4c76-b2aa-a177760a757f","Type":"ContainerStarted","Data":"5db23b665031218d72259cf4646733ed63a4bd51ab8afc325096c5343d72c977"} Feb 02 09:26:38 crc kubenswrapper[4764]: I0202 09:26:38.798250 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-448kc" event={"ID":"d492e01f-38f0-4c76-b2aa-a177760a757f","Type":"ContainerStarted","Data":"2e49a4f4d15e0f9f4d44b65d60eddbba800762c03a9848cdc615a75922f93b74"} Feb 02 09:26:38 crc kubenswrapper[4764]: I0202 09:26:38.807951 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-91bb-account-create-update-gg7c4" podStartSLOduration=1.807919015 podStartE2EDuration="1.807919015s" podCreationTimestamp="2026-02-02 09:26:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:26:38.804388509 +0000 UTC m=+1161.738112597" watchObservedRunningTime="2026-02-02 09:26:38.807919015 +0000 UTC m=+1161.741643103" Feb 02 09:26:38 crc kubenswrapper[4764]: I0202 09:26:38.826302 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-scmkd" event={"ID":"2ccd54a4-fbee-4ad6-b0eb-9d22fdb1eebe","Type":"ContainerStarted","Data":"f3f1a2804c26e261d58dd4afe0a1d559aee3da232f720be90cb093d535a10712"} Feb 02 09:26:38 crc kubenswrapper[4764]: I0202 09:26:38.851947 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-scmkd" event={"ID":"2ccd54a4-fbee-4ad6-b0eb-9d22fdb1eebe","Type":"ContainerStarted","Data":"2f9ce56e59d7af521e668975b6e7e1dd45cd838d0b03e98772350698aa9429d2"} Feb 02 09:26:38 crc kubenswrapper[4764]: I0202 09:26:38.851987 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-8ms85" event={"ID":"64d8d838-0afe-4f39-ae93-027fea76ced1","Type":"ContainerStarted","Data":"2afeea0c0b612522569ddb9572b7ff6d5cf8c43140c52fc6c0f8275474893368"} Feb 02 09:26:38 crc kubenswrapper[4764]: I0202 09:26:38.852000 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-8ms85" event={"ID":"64d8d838-0afe-4f39-ae93-027fea76ced1","Type":"ContainerStarted","Data":"1ffe35fcf31d18c8926d39776d3bc1fa1b1dde90860fc181303ed3a28f1b6416"} Feb 02 09:26:38 crc kubenswrapper[4764]: I0202 09:26:38.852011 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-10df-account-create-update-gb8cd" event={"ID":"902c771b-8da4-4c38-911e-c8b531d076a8","Type":"ContainerStarted","Data":"3e478671319cdce58bea9f612fda89e04a4085302fd2bf4576d05402661d0d03"} Feb 02 09:26:38 crc kubenswrapper[4764]: I0202 09:26:38.936174 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-db-create-448kc" podStartSLOduration=1.93615194 podStartE2EDuration="1.93615194s" podCreationTimestamp="2026-02-02 09:26:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:26:38.837187553 +0000 UTC m=+1161.770911641" watchObservedRunningTime="2026-02-02 09:26:38.93615194 +0000 UTC m=+1161.869876028" Feb 02 09:26:38 crc kubenswrapper[4764]: I0202 09:26:38.941814 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-db-create-scmkd" podStartSLOduration=1.941799594 podStartE2EDuration="1.941799594s" podCreationTimestamp="2026-02-02 09:26:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:26:38.85580362 +0000 UTC m=+1161.789527708" watchObservedRunningTime="2026-02-02 09:26:38.941799594 +0000 UTC m=+1161.875523682" Feb 02 09:26:38 crc kubenswrapper[4764]: I0202 09:26:38.954541 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-db-create-8ms85" podStartSLOduration=2.954529421 podStartE2EDuration="2.954529421s" podCreationTimestamp="2026-02-02 09:26:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:26:38.891292128 +0000 UTC m=+1161.825016236" watchObservedRunningTime="2026-02-02 09:26:38.954529421 +0000 UTC m=+1161.888253509" Feb 02 09:26:39 crc kubenswrapper[4764]: I0202 09:26:39.729354 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 02 09:26:39 crc kubenswrapper[4764]: I0202 09:26:39.775495 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/812ecbf4-31e2-49db-98ec-82992886557b-run-httpd\") pod \"812ecbf4-31e2-49db-98ec-82992886557b\" (UID: \"812ecbf4-31e2-49db-98ec-82992886557b\") " Feb 02 09:26:39 crc kubenswrapper[4764]: I0202 09:26:39.775547 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/812ecbf4-31e2-49db-98ec-82992886557b-sg-core-conf-yaml\") pod \"812ecbf4-31e2-49db-98ec-82992886557b\" (UID: \"812ecbf4-31e2-49db-98ec-82992886557b\") " Feb 02 09:26:39 crc kubenswrapper[4764]: I0202 09:26:39.775572 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/812ecbf4-31e2-49db-98ec-82992886557b-log-httpd\") pod \"812ecbf4-31e2-49db-98ec-82992886557b\" (UID: \"812ecbf4-31e2-49db-98ec-82992886557b\") " Feb 02 09:26:39 crc kubenswrapper[4764]: I0202 09:26:39.775656 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-68v2m\" (UniqueName: \"kubernetes.io/projected/812ecbf4-31e2-49db-98ec-82992886557b-kube-api-access-68v2m\") pod \"812ecbf4-31e2-49db-98ec-82992886557b\" (UID: \"812ecbf4-31e2-49db-98ec-82992886557b\") " Feb 02 09:26:39 crc kubenswrapper[4764]: I0202 09:26:39.775691 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/812ecbf4-31e2-49db-98ec-82992886557b-scripts\") pod \"812ecbf4-31e2-49db-98ec-82992886557b\" (UID: \"812ecbf4-31e2-49db-98ec-82992886557b\") " Feb 02 09:26:39 crc kubenswrapper[4764]: I0202 09:26:39.775730 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/812ecbf4-31e2-49db-98ec-82992886557b-combined-ca-bundle\") pod \"812ecbf4-31e2-49db-98ec-82992886557b\" (UID: \"812ecbf4-31e2-49db-98ec-82992886557b\") " Feb 02 09:26:39 crc kubenswrapper[4764]: I0202 09:26:39.775798 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/812ecbf4-31e2-49db-98ec-82992886557b-config-data\") pod \"812ecbf4-31e2-49db-98ec-82992886557b\" (UID: \"812ecbf4-31e2-49db-98ec-82992886557b\") " Feb 02 09:26:39 crc kubenswrapper[4764]: I0202 09:26:39.789948 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/812ecbf4-31e2-49db-98ec-82992886557b-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "812ecbf4-31e2-49db-98ec-82992886557b" (UID: "812ecbf4-31e2-49db-98ec-82992886557b"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 09:26:39 crc kubenswrapper[4764]: I0202 09:26:39.796589 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/812ecbf4-31e2-49db-98ec-82992886557b-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "812ecbf4-31e2-49db-98ec-82992886557b" (UID: "812ecbf4-31e2-49db-98ec-82992886557b"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 09:26:39 crc kubenswrapper[4764]: I0202 09:26:39.800510 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/812ecbf4-31e2-49db-98ec-82992886557b-kube-api-access-68v2m" (OuterVolumeSpecName: "kube-api-access-68v2m") pod "812ecbf4-31e2-49db-98ec-82992886557b" (UID: "812ecbf4-31e2-49db-98ec-82992886557b"). InnerVolumeSpecName "kube-api-access-68v2m". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:26:39 crc kubenswrapper[4764]: I0202 09:26:39.804915 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/812ecbf4-31e2-49db-98ec-82992886557b-scripts" (OuterVolumeSpecName: "scripts") pod "812ecbf4-31e2-49db-98ec-82992886557b" (UID: "812ecbf4-31e2-49db-98ec-82992886557b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:26:39 crc kubenswrapper[4764]: I0202 09:26:39.817986 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/812ecbf4-31e2-49db-98ec-82992886557b-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "812ecbf4-31e2-49db-98ec-82992886557b" (UID: "812ecbf4-31e2-49db-98ec-82992886557b"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:26:39 crc kubenswrapper[4764]: I0202 09:26:39.857608 4764 generic.go:334] "Generic (PLEG): container finished" podID="3ee0de70-5bfb-4b6e-974c-601071027218" containerID="08540cf734e603fff5d5dcb78a4f068507e9eb3462471ce00d9d5cfaf8926d7d" exitCode=0 Feb 02 09:26:39 crc kubenswrapper[4764]: I0202 09:26:39.862243 4764 generic.go:334] "Generic (PLEG): container finished" podID="afbe3865-2077-45c5-8a94-3d0a6774d8e2" containerID="e7f0299d2bd51585893ffb5380d548111de8ffb0dd7db4b97a394a0df7f69d58" exitCode=0 Feb 02 09:26:39 crc kubenswrapper[4764]: I0202 09:26:39.864190 4764 generic.go:334] "Generic (PLEG): container finished" podID="d492e01f-38f0-4c76-b2aa-a177760a757f" containerID="5db23b665031218d72259cf4646733ed63a4bd51ab8afc325096c5343d72c977" exitCode=0 Feb 02 09:26:39 crc kubenswrapper[4764]: I0202 09:26:39.867172 4764 generic.go:334] "Generic (PLEG): container finished" podID="2ccd54a4-fbee-4ad6-b0eb-9d22fdb1eebe" containerID="f3f1a2804c26e261d58dd4afe0a1d559aee3da232f720be90cb093d535a10712" exitCode=0 Feb 02 09:26:39 crc kubenswrapper[4764]: I0202 09:26:39.876569 4764 generic.go:334] "Generic (PLEG): container finished" podID="812ecbf4-31e2-49db-98ec-82992886557b" containerID="b442ad160351ea8a9f086f5433501745e485e752e6379d55db1141934f356a62" exitCode=0 Feb 02 09:26:39 crc kubenswrapper[4764]: I0202 09:26:39.876717 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 02 09:26:39 crc kubenswrapper[4764]: I0202 09:26:39.878197 4764 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/812ecbf4-31e2-49db-98ec-82992886557b-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 02 09:26:39 crc kubenswrapper[4764]: I0202 09:26:39.879393 4764 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/812ecbf4-31e2-49db-98ec-82992886557b-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 02 09:26:39 crc kubenswrapper[4764]: I0202 09:26:39.879469 4764 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/812ecbf4-31e2-49db-98ec-82992886557b-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 02 09:26:39 crc kubenswrapper[4764]: I0202 09:26:39.879525 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-68v2m\" (UniqueName: \"kubernetes.io/projected/812ecbf4-31e2-49db-98ec-82992886557b-kube-api-access-68v2m\") on node \"crc\" DevicePath \"\"" Feb 02 09:26:39 crc kubenswrapper[4764]: I0202 09:26:39.879602 4764 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/812ecbf4-31e2-49db-98ec-82992886557b-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 09:26:39 crc kubenswrapper[4764]: I0202 09:26:39.883836 4764 generic.go:334] "Generic (PLEG): container finished" podID="64d8d838-0afe-4f39-ae93-027fea76ced1" containerID="2afeea0c0b612522569ddb9572b7ff6d5cf8c43140c52fc6c0f8275474893368" exitCode=0 Feb 02 09:26:39 crc kubenswrapper[4764]: I0202 09:26:39.895172 4764 generic.go:334] "Generic (PLEG): container finished" podID="902c771b-8da4-4c38-911e-c8b531d076a8" containerID="5e0a384e8eeb871467c9fe900e3e0979df01eba7e41cc26e321f6df83fb5eae8" exitCode=0 Feb 02 09:26:39 crc kubenswrapper[4764]: I0202 09:26:39.897611 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/812ecbf4-31e2-49db-98ec-82992886557b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "812ecbf4-31e2-49db-98ec-82992886557b" (UID: "812ecbf4-31e2-49db-98ec-82992886557b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:26:39 crc kubenswrapper[4764]: I0202 09:26:39.902144 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/812ecbf4-31e2-49db-98ec-82992886557b-config-data" (OuterVolumeSpecName: "config-data") pod "812ecbf4-31e2-49db-98ec-82992886557b" (UID: "812ecbf4-31e2-49db-98ec-82992886557b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:26:39 crc kubenswrapper[4764]: I0202 09:26:39.981377 4764 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/812ecbf4-31e2-49db-98ec-82992886557b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 09:26:39 crc kubenswrapper[4764]: I0202 09:26:39.981403 4764 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/812ecbf4-31e2-49db-98ec-82992886557b-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 09:26:40 crc kubenswrapper[4764]: I0202 09:26:40.010158 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-91bb-account-create-update-gg7c4" event={"ID":"3ee0de70-5bfb-4b6e-974c-601071027218","Type":"ContainerDied","Data":"08540cf734e603fff5d5dcb78a4f068507e9eb3462471ce00d9d5cfaf8926d7d"} Feb 02 09:26:40 crc kubenswrapper[4764]: I0202 09:26:40.010210 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-39d3-account-create-update-mtxtf" event={"ID":"afbe3865-2077-45c5-8a94-3d0a6774d8e2","Type":"ContainerDied","Data":"e7f0299d2bd51585893ffb5380d548111de8ffb0dd7db4b97a394a0df7f69d58"} Feb 02 09:26:40 crc kubenswrapper[4764]: I0202 09:26:40.010230 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-39d3-account-create-update-mtxtf" event={"ID":"afbe3865-2077-45c5-8a94-3d0a6774d8e2","Type":"ContainerStarted","Data":"005e2dbe434c241d182f43904437136b4ea21e54961283408aac9bda495dd3fc"} Feb 02 09:26:40 crc kubenswrapper[4764]: I0202 09:26:40.010240 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-448kc" event={"ID":"d492e01f-38f0-4c76-b2aa-a177760a757f","Type":"ContainerDied","Data":"5db23b665031218d72259cf4646733ed63a4bd51ab8afc325096c5343d72c977"} Feb 02 09:26:40 crc kubenswrapper[4764]: I0202 09:26:40.010252 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-scmkd" event={"ID":"2ccd54a4-fbee-4ad6-b0eb-9d22fdb1eebe","Type":"ContainerDied","Data":"f3f1a2804c26e261d58dd4afe0a1d559aee3da232f720be90cb093d535a10712"} Feb 02 09:26:40 crc kubenswrapper[4764]: I0202 09:26:40.010265 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"812ecbf4-31e2-49db-98ec-82992886557b","Type":"ContainerDied","Data":"b442ad160351ea8a9f086f5433501745e485e752e6379d55db1141934f356a62"} Feb 02 09:26:40 crc kubenswrapper[4764]: I0202 09:26:40.010279 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"812ecbf4-31e2-49db-98ec-82992886557b","Type":"ContainerDied","Data":"33d2297d59543380d47825935ce2864e6316b0fb01e207fd04c5274f270e6ee4"} Feb 02 09:26:40 crc kubenswrapper[4764]: I0202 09:26:40.010290 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-8ms85" event={"ID":"64d8d838-0afe-4f39-ae93-027fea76ced1","Type":"ContainerDied","Data":"2afeea0c0b612522569ddb9572b7ff6d5cf8c43140c52fc6c0f8275474893368"} Feb 02 09:26:40 crc kubenswrapper[4764]: I0202 09:26:40.010302 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-10df-account-create-update-gb8cd" event={"ID":"902c771b-8da4-4c38-911e-c8b531d076a8","Type":"ContainerDied","Data":"5e0a384e8eeb871467c9fe900e3e0979df01eba7e41cc26e321f6df83fb5eae8"} Feb 02 09:26:40 crc kubenswrapper[4764]: I0202 09:26:40.010330 4764 scope.go:117] "RemoveContainer" containerID="1d3bb6fd23c8071400e1bfce6e4a9329c92de571ae3ec99411fc07ea41ded3ed" Feb 02 09:26:40 crc kubenswrapper[4764]: I0202 09:26:40.064999 4764 scope.go:117] "RemoveContainer" containerID="7cfbe64ae6de76de4b3bd49d5f588b9950147f47366928c962f5fb60ebfbdd1d" Feb 02 09:26:40 crc kubenswrapper[4764]: I0202 09:26:40.081994 4764 scope.go:117] "RemoveContainer" containerID="ae959cfdce12c0cd337f49be9499023ed6ef970512d3aa412b6eb7fc974091e2" Feb 02 09:26:40 crc kubenswrapper[4764]: I0202 09:26:40.099679 4764 scope.go:117] "RemoveContainer" containerID="b442ad160351ea8a9f086f5433501745e485e752e6379d55db1141934f356a62" Feb 02 09:26:40 crc kubenswrapper[4764]: I0202 09:26:40.125002 4764 scope.go:117] "RemoveContainer" containerID="1d3bb6fd23c8071400e1bfce6e4a9329c92de571ae3ec99411fc07ea41ded3ed" Feb 02 09:26:40 crc kubenswrapper[4764]: E0202 09:26:40.125422 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1d3bb6fd23c8071400e1bfce6e4a9329c92de571ae3ec99411fc07ea41ded3ed\": container with ID starting with 1d3bb6fd23c8071400e1bfce6e4a9329c92de571ae3ec99411fc07ea41ded3ed not found: ID does not exist" containerID="1d3bb6fd23c8071400e1bfce6e4a9329c92de571ae3ec99411fc07ea41ded3ed" Feb 02 09:26:40 crc kubenswrapper[4764]: I0202 09:26:40.125450 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d3bb6fd23c8071400e1bfce6e4a9329c92de571ae3ec99411fc07ea41ded3ed"} err="failed to get container status \"1d3bb6fd23c8071400e1bfce6e4a9329c92de571ae3ec99411fc07ea41ded3ed\": rpc error: code = NotFound desc = could not find container \"1d3bb6fd23c8071400e1bfce6e4a9329c92de571ae3ec99411fc07ea41ded3ed\": container with ID starting with 1d3bb6fd23c8071400e1bfce6e4a9329c92de571ae3ec99411fc07ea41ded3ed not found: ID does not exist" Feb 02 09:26:40 crc kubenswrapper[4764]: I0202 09:26:40.125470 4764 scope.go:117] "RemoveContainer" containerID="7cfbe64ae6de76de4b3bd49d5f588b9950147f47366928c962f5fb60ebfbdd1d" Feb 02 09:26:40 crc kubenswrapper[4764]: E0202 09:26:40.125915 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7cfbe64ae6de76de4b3bd49d5f588b9950147f47366928c962f5fb60ebfbdd1d\": container with ID starting with 7cfbe64ae6de76de4b3bd49d5f588b9950147f47366928c962f5fb60ebfbdd1d not found: ID does not exist" containerID="7cfbe64ae6de76de4b3bd49d5f588b9950147f47366928c962f5fb60ebfbdd1d" Feb 02 09:26:40 crc kubenswrapper[4764]: I0202 09:26:40.125959 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7cfbe64ae6de76de4b3bd49d5f588b9950147f47366928c962f5fb60ebfbdd1d"} err="failed to get container status \"7cfbe64ae6de76de4b3bd49d5f588b9950147f47366928c962f5fb60ebfbdd1d\": rpc error: code = NotFound desc = could not find container \"7cfbe64ae6de76de4b3bd49d5f588b9950147f47366928c962f5fb60ebfbdd1d\": container with ID starting with 7cfbe64ae6de76de4b3bd49d5f588b9950147f47366928c962f5fb60ebfbdd1d not found: ID does not exist" Feb 02 09:26:40 crc kubenswrapper[4764]: I0202 09:26:40.125973 4764 scope.go:117] "RemoveContainer" containerID="ae959cfdce12c0cd337f49be9499023ed6ef970512d3aa412b6eb7fc974091e2" Feb 02 09:26:40 crc kubenswrapper[4764]: E0202 09:26:40.126175 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ae959cfdce12c0cd337f49be9499023ed6ef970512d3aa412b6eb7fc974091e2\": container with ID starting with ae959cfdce12c0cd337f49be9499023ed6ef970512d3aa412b6eb7fc974091e2 not found: ID does not exist" containerID="ae959cfdce12c0cd337f49be9499023ed6ef970512d3aa412b6eb7fc974091e2" Feb 02 09:26:40 crc kubenswrapper[4764]: I0202 09:26:40.126196 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ae959cfdce12c0cd337f49be9499023ed6ef970512d3aa412b6eb7fc974091e2"} err="failed to get container status \"ae959cfdce12c0cd337f49be9499023ed6ef970512d3aa412b6eb7fc974091e2\": rpc error: code = NotFound desc = could not find container \"ae959cfdce12c0cd337f49be9499023ed6ef970512d3aa412b6eb7fc974091e2\": container with ID starting with ae959cfdce12c0cd337f49be9499023ed6ef970512d3aa412b6eb7fc974091e2 not found: ID does not exist" Feb 02 09:26:40 crc kubenswrapper[4764]: I0202 09:26:40.126208 4764 scope.go:117] "RemoveContainer" containerID="b442ad160351ea8a9f086f5433501745e485e752e6379d55db1141934f356a62" Feb 02 09:26:40 crc kubenswrapper[4764]: E0202 09:26:40.126380 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b442ad160351ea8a9f086f5433501745e485e752e6379d55db1141934f356a62\": container with ID starting with b442ad160351ea8a9f086f5433501745e485e752e6379d55db1141934f356a62 not found: ID does not exist" containerID="b442ad160351ea8a9f086f5433501745e485e752e6379d55db1141934f356a62" Feb 02 09:26:40 crc kubenswrapper[4764]: I0202 09:26:40.126393 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b442ad160351ea8a9f086f5433501745e485e752e6379d55db1141934f356a62"} err="failed to get container status \"b442ad160351ea8a9f086f5433501745e485e752e6379d55db1141934f356a62\": rpc error: code = NotFound desc = could not find container \"b442ad160351ea8a9f086f5433501745e485e752e6379d55db1141934f356a62\": container with ID starting with b442ad160351ea8a9f086f5433501745e485e752e6379d55db1141934f356a62 not found: ID does not exist" Feb 02 09:26:40 crc kubenswrapper[4764]: I0202 09:26:40.219626 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 02 09:26:40 crc kubenswrapper[4764]: I0202 09:26:40.231271 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 02 09:26:40 crc kubenswrapper[4764]: I0202 09:26:40.250655 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 02 09:26:40 crc kubenswrapper[4764]: E0202 09:26:40.251000 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="812ecbf4-31e2-49db-98ec-82992886557b" containerName="ceilometer-central-agent" Feb 02 09:26:40 crc kubenswrapper[4764]: I0202 09:26:40.251012 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="812ecbf4-31e2-49db-98ec-82992886557b" containerName="ceilometer-central-agent" Feb 02 09:26:40 crc kubenswrapper[4764]: E0202 09:26:40.251033 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="812ecbf4-31e2-49db-98ec-82992886557b" containerName="ceilometer-notification-agent" Feb 02 09:26:40 crc kubenswrapper[4764]: I0202 09:26:40.251038 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="812ecbf4-31e2-49db-98ec-82992886557b" containerName="ceilometer-notification-agent" Feb 02 09:26:40 crc kubenswrapper[4764]: E0202 09:26:40.251056 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="812ecbf4-31e2-49db-98ec-82992886557b" containerName="proxy-httpd" Feb 02 09:26:40 crc kubenswrapper[4764]: I0202 09:26:40.251062 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="812ecbf4-31e2-49db-98ec-82992886557b" containerName="proxy-httpd" Feb 02 09:26:40 crc kubenswrapper[4764]: E0202 09:26:40.251075 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="812ecbf4-31e2-49db-98ec-82992886557b" containerName="sg-core" Feb 02 09:26:40 crc kubenswrapper[4764]: I0202 09:26:40.251080 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="812ecbf4-31e2-49db-98ec-82992886557b" containerName="sg-core" Feb 02 09:26:40 crc kubenswrapper[4764]: I0202 09:26:40.251228 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="812ecbf4-31e2-49db-98ec-82992886557b" containerName="ceilometer-central-agent" Feb 02 09:26:40 crc kubenswrapper[4764]: I0202 09:26:40.251243 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="812ecbf4-31e2-49db-98ec-82992886557b" containerName="ceilometer-notification-agent" Feb 02 09:26:40 crc kubenswrapper[4764]: I0202 09:26:40.251254 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="812ecbf4-31e2-49db-98ec-82992886557b" containerName="proxy-httpd" Feb 02 09:26:40 crc kubenswrapper[4764]: I0202 09:26:40.251260 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="812ecbf4-31e2-49db-98ec-82992886557b" containerName="sg-core" Feb 02 09:26:40 crc kubenswrapper[4764]: I0202 09:26:40.252583 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 02 09:26:40 crc kubenswrapper[4764]: I0202 09:26:40.257180 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 02 09:26:40 crc kubenswrapper[4764]: I0202 09:26:40.257370 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 02 09:26:40 crc kubenswrapper[4764]: I0202 09:26:40.275249 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 02 09:26:40 crc kubenswrapper[4764]: I0202 09:26:40.286458 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5911a7f4-32cf-4b23-a056-baf957ebf6f6-scripts\") pod \"ceilometer-0\" (UID: \"5911a7f4-32cf-4b23-a056-baf957ebf6f6\") " pod="openstack/ceilometer-0" Feb 02 09:26:40 crc kubenswrapper[4764]: I0202 09:26:40.286507 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5911a7f4-32cf-4b23-a056-baf957ebf6f6-log-httpd\") pod \"ceilometer-0\" (UID: \"5911a7f4-32cf-4b23-a056-baf957ebf6f6\") " pod="openstack/ceilometer-0" Feb 02 09:26:40 crc kubenswrapper[4764]: I0202 09:26:40.286566 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5911a7f4-32cf-4b23-a056-baf957ebf6f6-run-httpd\") pod \"ceilometer-0\" (UID: \"5911a7f4-32cf-4b23-a056-baf957ebf6f6\") " pod="openstack/ceilometer-0" Feb 02 09:26:40 crc kubenswrapper[4764]: I0202 09:26:40.286595 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hrpnk\" (UniqueName: \"kubernetes.io/projected/5911a7f4-32cf-4b23-a056-baf957ebf6f6-kube-api-access-hrpnk\") pod \"ceilometer-0\" (UID: \"5911a7f4-32cf-4b23-a056-baf957ebf6f6\") " pod="openstack/ceilometer-0" Feb 02 09:26:40 crc kubenswrapper[4764]: I0202 09:26:40.286632 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5911a7f4-32cf-4b23-a056-baf957ebf6f6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5911a7f4-32cf-4b23-a056-baf957ebf6f6\") " pod="openstack/ceilometer-0" Feb 02 09:26:40 crc kubenswrapper[4764]: I0202 09:26:40.286649 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5911a7f4-32cf-4b23-a056-baf957ebf6f6-config-data\") pod \"ceilometer-0\" (UID: \"5911a7f4-32cf-4b23-a056-baf957ebf6f6\") " pod="openstack/ceilometer-0" Feb 02 09:26:40 crc kubenswrapper[4764]: I0202 09:26:40.286666 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5911a7f4-32cf-4b23-a056-baf957ebf6f6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5911a7f4-32cf-4b23-a056-baf957ebf6f6\") " pod="openstack/ceilometer-0" Feb 02 09:26:40 crc kubenswrapper[4764]: I0202 09:26:40.388922 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5911a7f4-32cf-4b23-a056-baf957ebf6f6-scripts\") pod \"ceilometer-0\" (UID: \"5911a7f4-32cf-4b23-a056-baf957ebf6f6\") " pod="openstack/ceilometer-0" Feb 02 09:26:40 crc kubenswrapper[4764]: I0202 09:26:40.389214 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5911a7f4-32cf-4b23-a056-baf957ebf6f6-log-httpd\") pod \"ceilometer-0\" (UID: \"5911a7f4-32cf-4b23-a056-baf957ebf6f6\") " pod="openstack/ceilometer-0" Feb 02 09:26:40 crc kubenswrapper[4764]: I0202 09:26:40.389272 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5911a7f4-32cf-4b23-a056-baf957ebf6f6-run-httpd\") pod \"ceilometer-0\" (UID: \"5911a7f4-32cf-4b23-a056-baf957ebf6f6\") " pod="openstack/ceilometer-0" Feb 02 09:26:40 crc kubenswrapper[4764]: I0202 09:26:40.389302 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hrpnk\" (UniqueName: \"kubernetes.io/projected/5911a7f4-32cf-4b23-a056-baf957ebf6f6-kube-api-access-hrpnk\") pod \"ceilometer-0\" (UID: \"5911a7f4-32cf-4b23-a056-baf957ebf6f6\") " pod="openstack/ceilometer-0" Feb 02 09:26:40 crc kubenswrapper[4764]: I0202 09:26:40.389343 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5911a7f4-32cf-4b23-a056-baf957ebf6f6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5911a7f4-32cf-4b23-a056-baf957ebf6f6\") " pod="openstack/ceilometer-0" Feb 02 09:26:40 crc kubenswrapper[4764]: I0202 09:26:40.389360 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5911a7f4-32cf-4b23-a056-baf957ebf6f6-config-data\") pod \"ceilometer-0\" (UID: \"5911a7f4-32cf-4b23-a056-baf957ebf6f6\") " pod="openstack/ceilometer-0" Feb 02 09:26:40 crc kubenswrapper[4764]: I0202 09:26:40.389378 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5911a7f4-32cf-4b23-a056-baf957ebf6f6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5911a7f4-32cf-4b23-a056-baf957ebf6f6\") " pod="openstack/ceilometer-0" Feb 02 09:26:40 crc kubenswrapper[4764]: I0202 09:26:40.389673 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5911a7f4-32cf-4b23-a056-baf957ebf6f6-log-httpd\") pod \"ceilometer-0\" (UID: \"5911a7f4-32cf-4b23-a056-baf957ebf6f6\") " pod="openstack/ceilometer-0" Feb 02 09:26:40 crc kubenswrapper[4764]: I0202 09:26:40.389723 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5911a7f4-32cf-4b23-a056-baf957ebf6f6-run-httpd\") pod \"ceilometer-0\" (UID: \"5911a7f4-32cf-4b23-a056-baf957ebf6f6\") " pod="openstack/ceilometer-0" Feb 02 09:26:40 crc kubenswrapper[4764]: I0202 09:26:40.393123 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5911a7f4-32cf-4b23-a056-baf957ebf6f6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5911a7f4-32cf-4b23-a056-baf957ebf6f6\") " pod="openstack/ceilometer-0" Feb 02 09:26:40 crc kubenswrapper[4764]: I0202 09:26:40.394663 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5911a7f4-32cf-4b23-a056-baf957ebf6f6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5911a7f4-32cf-4b23-a056-baf957ebf6f6\") " pod="openstack/ceilometer-0" Feb 02 09:26:40 crc kubenswrapper[4764]: I0202 09:26:40.396511 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5911a7f4-32cf-4b23-a056-baf957ebf6f6-config-data\") pod \"ceilometer-0\" (UID: \"5911a7f4-32cf-4b23-a056-baf957ebf6f6\") " pod="openstack/ceilometer-0" Feb 02 09:26:40 crc kubenswrapper[4764]: I0202 09:26:40.397535 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5911a7f4-32cf-4b23-a056-baf957ebf6f6-scripts\") pod \"ceilometer-0\" (UID: \"5911a7f4-32cf-4b23-a056-baf957ebf6f6\") " pod="openstack/ceilometer-0" Feb 02 09:26:40 crc kubenswrapper[4764]: I0202 09:26:40.398052 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-65fcd88786-dggfp" Feb 02 09:26:40 crc kubenswrapper[4764]: I0202 09:26:40.405327 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hrpnk\" (UniqueName: \"kubernetes.io/projected/5911a7f4-32cf-4b23-a056-baf957ebf6f6-kube-api-access-hrpnk\") pod \"ceilometer-0\" (UID: \"5911a7f4-32cf-4b23-a056-baf957ebf6f6\") " pod="openstack/ceilometer-0" Feb 02 09:26:40 crc kubenswrapper[4764]: I0202 09:26:40.490831 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jvkzr\" (UniqueName: \"kubernetes.io/projected/dedd7b1f-7ec4-4a7a-a703-7e228f7767b1-kube-api-access-jvkzr\") pod \"dedd7b1f-7ec4-4a7a-a703-7e228f7767b1\" (UID: \"dedd7b1f-7ec4-4a7a-a703-7e228f7767b1\") " Feb 02 09:26:40 crc kubenswrapper[4764]: I0202 09:26:40.490943 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/dedd7b1f-7ec4-4a7a-a703-7e228f7767b1-public-tls-certs\") pod \"dedd7b1f-7ec4-4a7a-a703-7e228f7767b1\" (UID: \"dedd7b1f-7ec4-4a7a-a703-7e228f7767b1\") " Feb 02 09:26:40 crc kubenswrapper[4764]: I0202 09:26:40.490976 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dedd7b1f-7ec4-4a7a-a703-7e228f7767b1-config-data\") pod \"dedd7b1f-7ec4-4a7a-a703-7e228f7767b1\" (UID: \"dedd7b1f-7ec4-4a7a-a703-7e228f7767b1\") " Feb 02 09:26:40 crc kubenswrapper[4764]: I0202 09:26:40.491040 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dedd7b1f-7ec4-4a7a-a703-7e228f7767b1-logs\") pod \"dedd7b1f-7ec4-4a7a-a703-7e228f7767b1\" (UID: \"dedd7b1f-7ec4-4a7a-a703-7e228f7767b1\") " Feb 02 09:26:40 crc kubenswrapper[4764]: I0202 09:26:40.491124 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dedd7b1f-7ec4-4a7a-a703-7e228f7767b1-scripts\") pod \"dedd7b1f-7ec4-4a7a-a703-7e228f7767b1\" (UID: \"dedd7b1f-7ec4-4a7a-a703-7e228f7767b1\") " Feb 02 09:26:40 crc kubenswrapper[4764]: I0202 09:26:40.491162 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/dedd7b1f-7ec4-4a7a-a703-7e228f7767b1-internal-tls-certs\") pod \"dedd7b1f-7ec4-4a7a-a703-7e228f7767b1\" (UID: \"dedd7b1f-7ec4-4a7a-a703-7e228f7767b1\") " Feb 02 09:26:40 crc kubenswrapper[4764]: I0202 09:26:40.491213 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dedd7b1f-7ec4-4a7a-a703-7e228f7767b1-combined-ca-bundle\") pod \"dedd7b1f-7ec4-4a7a-a703-7e228f7767b1\" (UID: \"dedd7b1f-7ec4-4a7a-a703-7e228f7767b1\") " Feb 02 09:26:40 crc kubenswrapper[4764]: I0202 09:26:40.491561 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dedd7b1f-7ec4-4a7a-a703-7e228f7767b1-logs" (OuterVolumeSpecName: "logs") pod "dedd7b1f-7ec4-4a7a-a703-7e228f7767b1" (UID: "dedd7b1f-7ec4-4a7a-a703-7e228f7767b1"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 09:26:40 crc kubenswrapper[4764]: I0202 09:26:40.493660 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dedd7b1f-7ec4-4a7a-a703-7e228f7767b1-kube-api-access-jvkzr" (OuterVolumeSpecName: "kube-api-access-jvkzr") pod "dedd7b1f-7ec4-4a7a-a703-7e228f7767b1" (UID: "dedd7b1f-7ec4-4a7a-a703-7e228f7767b1"). InnerVolumeSpecName "kube-api-access-jvkzr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:26:40 crc kubenswrapper[4764]: I0202 09:26:40.495544 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dedd7b1f-7ec4-4a7a-a703-7e228f7767b1-scripts" (OuterVolumeSpecName: "scripts") pod "dedd7b1f-7ec4-4a7a-a703-7e228f7767b1" (UID: "dedd7b1f-7ec4-4a7a-a703-7e228f7767b1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:26:40 crc kubenswrapper[4764]: I0202 09:26:40.534659 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dedd7b1f-7ec4-4a7a-a703-7e228f7767b1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "dedd7b1f-7ec4-4a7a-a703-7e228f7767b1" (UID: "dedd7b1f-7ec4-4a7a-a703-7e228f7767b1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:26:40 crc kubenswrapper[4764]: I0202 09:26:40.536019 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dedd7b1f-7ec4-4a7a-a703-7e228f7767b1-config-data" (OuterVolumeSpecName: "config-data") pod "dedd7b1f-7ec4-4a7a-a703-7e228f7767b1" (UID: "dedd7b1f-7ec4-4a7a-a703-7e228f7767b1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:26:40 crc kubenswrapper[4764]: I0202 09:26:40.579626 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dedd7b1f-7ec4-4a7a-a703-7e228f7767b1-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "dedd7b1f-7ec4-4a7a-a703-7e228f7767b1" (UID: "dedd7b1f-7ec4-4a7a-a703-7e228f7767b1"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:26:40 crc kubenswrapper[4764]: I0202 09:26:40.591322 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 02 09:26:40 crc kubenswrapper[4764]: I0202 09:26:40.592739 4764 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dedd7b1f-7ec4-4a7a-a703-7e228f7767b1-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 09:26:40 crc kubenswrapper[4764]: I0202 09:26:40.592772 4764 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/dedd7b1f-7ec4-4a7a-a703-7e228f7767b1-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 02 09:26:40 crc kubenswrapper[4764]: I0202 09:26:40.592787 4764 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dedd7b1f-7ec4-4a7a-a703-7e228f7767b1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 09:26:40 crc kubenswrapper[4764]: I0202 09:26:40.592796 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jvkzr\" (UniqueName: \"kubernetes.io/projected/dedd7b1f-7ec4-4a7a-a703-7e228f7767b1-kube-api-access-jvkzr\") on node \"crc\" DevicePath \"\"" Feb 02 09:26:40 crc kubenswrapper[4764]: I0202 09:26:40.592805 4764 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dedd7b1f-7ec4-4a7a-a703-7e228f7767b1-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 09:26:40 crc kubenswrapper[4764]: I0202 09:26:40.592813 4764 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dedd7b1f-7ec4-4a7a-a703-7e228f7767b1-logs\") on node \"crc\" DevicePath \"\"" Feb 02 09:26:40 crc kubenswrapper[4764]: I0202 09:26:40.593144 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dedd7b1f-7ec4-4a7a-a703-7e228f7767b1-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "dedd7b1f-7ec4-4a7a-a703-7e228f7767b1" (UID: "dedd7b1f-7ec4-4a7a-a703-7e228f7767b1"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:26:40 crc kubenswrapper[4764]: I0202 09:26:40.693847 4764 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/dedd7b1f-7ec4-4a7a-a703-7e228f7767b1-public-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 02 09:26:40 crc kubenswrapper[4764]: I0202 09:26:40.904993 4764 generic.go:334] "Generic (PLEG): container finished" podID="dedd7b1f-7ec4-4a7a-a703-7e228f7767b1" containerID="a3f865f4635480487ff622f30424f2b9a0ebc2e202ceb9b00303e7ebcdad3716" exitCode=0 Feb 02 09:26:40 crc kubenswrapper[4764]: I0202 09:26:40.905056 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-65fcd88786-dggfp" Feb 02 09:26:40 crc kubenswrapper[4764]: I0202 09:26:40.905082 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-65fcd88786-dggfp" event={"ID":"dedd7b1f-7ec4-4a7a-a703-7e228f7767b1","Type":"ContainerDied","Data":"a3f865f4635480487ff622f30424f2b9a0ebc2e202ceb9b00303e7ebcdad3716"} Feb 02 09:26:40 crc kubenswrapper[4764]: I0202 09:26:40.905358 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-65fcd88786-dggfp" event={"ID":"dedd7b1f-7ec4-4a7a-a703-7e228f7767b1","Type":"ContainerDied","Data":"d7df5c1af9c56a5971f2d2cda7bbcd4f2d4a1b32db5eeeb2b19911d615045fd5"} Feb 02 09:26:40 crc kubenswrapper[4764]: I0202 09:26:40.905376 4764 scope.go:117] "RemoveContainer" containerID="a3f865f4635480487ff622f30424f2b9a0ebc2e202ceb9b00303e7ebcdad3716" Feb 02 09:26:40 crc kubenswrapper[4764]: I0202 09:26:40.944887 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-65fcd88786-dggfp"] Feb 02 09:26:40 crc kubenswrapper[4764]: I0202 09:26:40.949659 4764 scope.go:117] "RemoveContainer" containerID="990d29535b531c195dc668ed0f9c16b9592d389ae625fafcb51d55c1040391d2" Feb 02 09:26:40 crc kubenswrapper[4764]: I0202 09:26:40.950400 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-65fcd88786-dggfp"] Feb 02 09:26:41 crc kubenswrapper[4764]: I0202 09:26:41.020683 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 02 09:26:41 crc kubenswrapper[4764]: I0202 09:26:41.040982 4764 scope.go:117] "RemoveContainer" containerID="a3f865f4635480487ff622f30424f2b9a0ebc2e202ceb9b00303e7ebcdad3716" Feb 02 09:26:41 crc kubenswrapper[4764]: E0202 09:26:41.041399 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a3f865f4635480487ff622f30424f2b9a0ebc2e202ceb9b00303e7ebcdad3716\": container with ID starting with a3f865f4635480487ff622f30424f2b9a0ebc2e202ceb9b00303e7ebcdad3716 not found: ID does not exist" containerID="a3f865f4635480487ff622f30424f2b9a0ebc2e202ceb9b00303e7ebcdad3716" Feb 02 09:26:41 crc kubenswrapper[4764]: I0202 09:26:41.041430 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a3f865f4635480487ff622f30424f2b9a0ebc2e202ceb9b00303e7ebcdad3716"} err="failed to get container status \"a3f865f4635480487ff622f30424f2b9a0ebc2e202ceb9b00303e7ebcdad3716\": rpc error: code = NotFound desc = could not find container \"a3f865f4635480487ff622f30424f2b9a0ebc2e202ceb9b00303e7ebcdad3716\": container with ID starting with a3f865f4635480487ff622f30424f2b9a0ebc2e202ceb9b00303e7ebcdad3716 not found: ID does not exist" Feb 02 09:26:41 crc kubenswrapper[4764]: I0202 09:26:41.041450 4764 scope.go:117] "RemoveContainer" containerID="990d29535b531c195dc668ed0f9c16b9592d389ae625fafcb51d55c1040391d2" Feb 02 09:26:41 crc kubenswrapper[4764]: E0202 09:26:41.041691 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"990d29535b531c195dc668ed0f9c16b9592d389ae625fafcb51d55c1040391d2\": container with ID starting with 990d29535b531c195dc668ed0f9c16b9592d389ae625fafcb51d55c1040391d2 not found: ID does not exist" containerID="990d29535b531c195dc668ed0f9c16b9592d389ae625fafcb51d55c1040391d2" Feb 02 09:26:41 crc kubenswrapper[4764]: I0202 09:26:41.041717 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"990d29535b531c195dc668ed0f9c16b9592d389ae625fafcb51d55c1040391d2"} err="failed to get container status \"990d29535b531c195dc668ed0f9c16b9592d389ae625fafcb51d55c1040391d2\": rpc error: code = NotFound desc = could not find container \"990d29535b531c195dc668ed0f9c16b9592d389ae625fafcb51d55c1040391d2\": container with ID starting with 990d29535b531c195dc668ed0f9c16b9592d389ae625fafcb51d55c1040391d2 not found: ID does not exist" Feb 02 09:26:41 crc kubenswrapper[4764]: I0202 09:26:41.290457 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-scmkd" Feb 02 09:26:41 crc kubenswrapper[4764]: I0202 09:26:41.414084 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-db9gm\" (UniqueName: \"kubernetes.io/projected/2ccd54a4-fbee-4ad6-b0eb-9d22fdb1eebe-kube-api-access-db9gm\") pod \"2ccd54a4-fbee-4ad6-b0eb-9d22fdb1eebe\" (UID: \"2ccd54a4-fbee-4ad6-b0eb-9d22fdb1eebe\") " Feb 02 09:26:41 crc kubenswrapper[4764]: I0202 09:26:41.414227 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2ccd54a4-fbee-4ad6-b0eb-9d22fdb1eebe-operator-scripts\") pod \"2ccd54a4-fbee-4ad6-b0eb-9d22fdb1eebe\" (UID: \"2ccd54a4-fbee-4ad6-b0eb-9d22fdb1eebe\") " Feb 02 09:26:41 crc kubenswrapper[4764]: I0202 09:26:41.415389 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2ccd54a4-fbee-4ad6-b0eb-9d22fdb1eebe-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2ccd54a4-fbee-4ad6-b0eb-9d22fdb1eebe" (UID: "2ccd54a4-fbee-4ad6-b0eb-9d22fdb1eebe"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:26:41 crc kubenswrapper[4764]: I0202 09:26:41.480860 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2ccd54a4-fbee-4ad6-b0eb-9d22fdb1eebe-kube-api-access-db9gm" (OuterVolumeSpecName: "kube-api-access-db9gm") pod "2ccd54a4-fbee-4ad6-b0eb-9d22fdb1eebe" (UID: "2ccd54a4-fbee-4ad6-b0eb-9d22fdb1eebe"). InnerVolumeSpecName "kube-api-access-db9gm". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:26:41 crc kubenswrapper[4764]: I0202 09:26:41.516255 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-db9gm\" (UniqueName: \"kubernetes.io/projected/2ccd54a4-fbee-4ad6-b0eb-9d22fdb1eebe-kube-api-access-db9gm\") on node \"crc\" DevicePath \"\"" Feb 02 09:26:41 crc kubenswrapper[4764]: I0202 09:26:41.516579 4764 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2ccd54a4-fbee-4ad6-b0eb-9d22fdb1eebe-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 09:26:41 crc kubenswrapper[4764]: I0202 09:26:41.521470 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-448kc" Feb 02 09:26:41 crc kubenswrapper[4764]: I0202 09:26:41.539281 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-8ms85" Feb 02 09:26:41 crc kubenswrapper[4764]: I0202 09:26:41.549730 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-91bb-account-create-update-gg7c4" Feb 02 09:26:41 crc kubenswrapper[4764]: I0202 09:26:41.564545 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-39d3-account-create-update-mtxtf" Feb 02 09:26:41 crc kubenswrapper[4764]: I0202 09:26:41.567068 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-10df-account-create-update-gb8cd" Feb 02 09:26:41 crc kubenswrapper[4764]: I0202 09:26:41.617805 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7xsvj\" (UniqueName: \"kubernetes.io/projected/3ee0de70-5bfb-4b6e-974c-601071027218-kube-api-access-7xsvj\") pod \"3ee0de70-5bfb-4b6e-974c-601071027218\" (UID: \"3ee0de70-5bfb-4b6e-974c-601071027218\") " Feb 02 09:26:41 crc kubenswrapper[4764]: I0202 09:26:41.617949 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d492e01f-38f0-4c76-b2aa-a177760a757f-operator-scripts\") pod \"d492e01f-38f0-4c76-b2aa-a177760a757f\" (UID: \"d492e01f-38f0-4c76-b2aa-a177760a757f\") " Feb 02 09:26:41 crc kubenswrapper[4764]: I0202 09:26:41.618054 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3ee0de70-5bfb-4b6e-974c-601071027218-operator-scripts\") pod \"3ee0de70-5bfb-4b6e-974c-601071027218\" (UID: \"3ee0de70-5bfb-4b6e-974c-601071027218\") " Feb 02 09:26:41 crc kubenswrapper[4764]: I0202 09:26:41.618145 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2t6cp\" (UniqueName: \"kubernetes.io/projected/902c771b-8da4-4c38-911e-c8b531d076a8-kube-api-access-2t6cp\") pod \"902c771b-8da4-4c38-911e-c8b531d076a8\" (UID: \"902c771b-8da4-4c38-911e-c8b531d076a8\") " Feb 02 09:26:41 crc kubenswrapper[4764]: I0202 09:26:41.618225 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vqvwk\" (UniqueName: \"kubernetes.io/projected/afbe3865-2077-45c5-8a94-3d0a6774d8e2-kube-api-access-vqvwk\") pod \"afbe3865-2077-45c5-8a94-3d0a6774d8e2\" (UID: \"afbe3865-2077-45c5-8a94-3d0a6774d8e2\") " Feb 02 09:26:41 crc kubenswrapper[4764]: I0202 09:26:41.618305 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/902c771b-8da4-4c38-911e-c8b531d076a8-operator-scripts\") pod \"902c771b-8da4-4c38-911e-c8b531d076a8\" (UID: \"902c771b-8da4-4c38-911e-c8b531d076a8\") " Feb 02 09:26:41 crc kubenswrapper[4764]: I0202 09:26:41.618404 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-626gj\" (UniqueName: \"kubernetes.io/projected/d492e01f-38f0-4c76-b2aa-a177760a757f-kube-api-access-626gj\") pod \"d492e01f-38f0-4c76-b2aa-a177760a757f\" (UID: \"d492e01f-38f0-4c76-b2aa-a177760a757f\") " Feb 02 09:26:41 crc kubenswrapper[4764]: I0202 09:26:41.618488 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/afbe3865-2077-45c5-8a94-3d0a6774d8e2-operator-scripts\") pod \"afbe3865-2077-45c5-8a94-3d0a6774d8e2\" (UID: \"afbe3865-2077-45c5-8a94-3d0a6774d8e2\") " Feb 02 09:26:41 crc kubenswrapper[4764]: I0202 09:26:41.618568 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/64d8d838-0afe-4f39-ae93-027fea76ced1-operator-scripts\") pod \"64d8d838-0afe-4f39-ae93-027fea76ced1\" (UID: \"64d8d838-0afe-4f39-ae93-027fea76ced1\") " Feb 02 09:26:41 crc kubenswrapper[4764]: I0202 09:26:41.618724 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dvv25\" (UniqueName: \"kubernetes.io/projected/64d8d838-0afe-4f39-ae93-027fea76ced1-kube-api-access-dvv25\") pod \"64d8d838-0afe-4f39-ae93-027fea76ced1\" (UID: \"64d8d838-0afe-4f39-ae93-027fea76ced1\") " Feb 02 09:26:41 crc kubenswrapper[4764]: I0202 09:26:41.620334 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3ee0de70-5bfb-4b6e-974c-601071027218-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "3ee0de70-5bfb-4b6e-974c-601071027218" (UID: "3ee0de70-5bfb-4b6e-974c-601071027218"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:26:41 crc kubenswrapper[4764]: I0202 09:26:41.624583 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/902c771b-8da4-4c38-911e-c8b531d076a8-kube-api-access-2t6cp" (OuterVolumeSpecName: "kube-api-access-2t6cp") pod "902c771b-8da4-4c38-911e-c8b531d076a8" (UID: "902c771b-8da4-4c38-911e-c8b531d076a8"). InnerVolumeSpecName "kube-api-access-2t6cp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:26:41 crc kubenswrapper[4764]: I0202 09:26:41.626624 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/902c771b-8da4-4c38-911e-c8b531d076a8-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "902c771b-8da4-4c38-911e-c8b531d076a8" (UID: "902c771b-8da4-4c38-911e-c8b531d076a8"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:26:41 crc kubenswrapper[4764]: I0202 09:26:41.627013 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/64d8d838-0afe-4f39-ae93-027fea76ced1-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "64d8d838-0afe-4f39-ae93-027fea76ced1" (UID: "64d8d838-0afe-4f39-ae93-027fea76ced1"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:26:41 crc kubenswrapper[4764]: I0202 09:26:41.630265 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d492e01f-38f0-4c76-b2aa-a177760a757f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d492e01f-38f0-4c76-b2aa-a177760a757f" (UID: "d492e01f-38f0-4c76-b2aa-a177760a757f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:26:41 crc kubenswrapper[4764]: I0202 09:26:41.630979 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/afbe3865-2077-45c5-8a94-3d0a6774d8e2-kube-api-access-vqvwk" (OuterVolumeSpecName: "kube-api-access-vqvwk") pod "afbe3865-2077-45c5-8a94-3d0a6774d8e2" (UID: "afbe3865-2077-45c5-8a94-3d0a6774d8e2"). InnerVolumeSpecName "kube-api-access-vqvwk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:26:41 crc kubenswrapper[4764]: I0202 09:26:41.631269 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d492e01f-38f0-4c76-b2aa-a177760a757f-kube-api-access-626gj" (OuterVolumeSpecName: "kube-api-access-626gj") pod "d492e01f-38f0-4c76-b2aa-a177760a757f" (UID: "d492e01f-38f0-4c76-b2aa-a177760a757f"). InnerVolumeSpecName "kube-api-access-626gj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:26:41 crc kubenswrapper[4764]: I0202 09:26:41.632551 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/afbe3865-2077-45c5-8a94-3d0a6774d8e2-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "afbe3865-2077-45c5-8a94-3d0a6774d8e2" (UID: "afbe3865-2077-45c5-8a94-3d0a6774d8e2"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:26:41 crc kubenswrapper[4764]: I0202 09:26:41.637359 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/64d8d838-0afe-4f39-ae93-027fea76ced1-kube-api-access-dvv25" (OuterVolumeSpecName: "kube-api-access-dvv25") pod "64d8d838-0afe-4f39-ae93-027fea76ced1" (UID: "64d8d838-0afe-4f39-ae93-027fea76ced1"). InnerVolumeSpecName "kube-api-access-dvv25". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:26:41 crc kubenswrapper[4764]: I0202 09:26:41.637455 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ee0de70-5bfb-4b6e-974c-601071027218-kube-api-access-7xsvj" (OuterVolumeSpecName: "kube-api-access-7xsvj") pod "3ee0de70-5bfb-4b6e-974c-601071027218" (UID: "3ee0de70-5bfb-4b6e-974c-601071027218"). InnerVolumeSpecName "kube-api-access-7xsvj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:26:41 crc kubenswrapper[4764]: I0202 09:26:41.720798 4764 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3ee0de70-5bfb-4b6e-974c-601071027218-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 09:26:41 crc kubenswrapper[4764]: I0202 09:26:41.720822 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2t6cp\" (UniqueName: \"kubernetes.io/projected/902c771b-8da4-4c38-911e-c8b531d076a8-kube-api-access-2t6cp\") on node \"crc\" DevicePath \"\"" Feb 02 09:26:41 crc kubenswrapper[4764]: I0202 09:26:41.720835 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vqvwk\" (UniqueName: \"kubernetes.io/projected/afbe3865-2077-45c5-8a94-3d0a6774d8e2-kube-api-access-vqvwk\") on node \"crc\" DevicePath \"\"" Feb 02 09:26:41 crc kubenswrapper[4764]: I0202 09:26:41.720845 4764 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/902c771b-8da4-4c38-911e-c8b531d076a8-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 09:26:41 crc kubenswrapper[4764]: I0202 09:26:41.720854 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-626gj\" (UniqueName: \"kubernetes.io/projected/d492e01f-38f0-4c76-b2aa-a177760a757f-kube-api-access-626gj\") on node \"crc\" DevicePath \"\"" Feb 02 09:26:41 crc kubenswrapper[4764]: I0202 09:26:41.720866 4764 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/afbe3865-2077-45c5-8a94-3d0a6774d8e2-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 09:26:41 crc kubenswrapper[4764]: I0202 09:26:41.720878 4764 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/64d8d838-0afe-4f39-ae93-027fea76ced1-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 09:26:41 crc kubenswrapper[4764]: I0202 09:26:41.720890 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dvv25\" (UniqueName: \"kubernetes.io/projected/64d8d838-0afe-4f39-ae93-027fea76ced1-kube-api-access-dvv25\") on node \"crc\" DevicePath \"\"" Feb 02 09:26:41 crc kubenswrapper[4764]: I0202 09:26:41.720900 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7xsvj\" (UniqueName: \"kubernetes.io/projected/3ee0de70-5bfb-4b6e-974c-601071027218-kube-api-access-7xsvj\") on node \"crc\" DevicePath \"\"" Feb 02 09:26:41 crc kubenswrapper[4764]: I0202 09:26:41.720910 4764 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d492e01f-38f0-4c76-b2aa-a177760a757f-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 09:26:41 crc kubenswrapper[4764]: I0202 09:26:41.834440 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="812ecbf4-31e2-49db-98ec-82992886557b" path="/var/lib/kubelet/pods/812ecbf4-31e2-49db-98ec-82992886557b/volumes" Feb 02 09:26:41 crc kubenswrapper[4764]: I0202 09:26:41.835495 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dedd7b1f-7ec4-4a7a-a703-7e228f7767b1" path="/var/lib/kubelet/pods/dedd7b1f-7ec4-4a7a-a703-7e228f7767b1/volumes" Feb 02 09:26:41 crc kubenswrapper[4764]: I0202 09:26:41.915603 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-448kc" event={"ID":"d492e01f-38f0-4c76-b2aa-a177760a757f","Type":"ContainerDied","Data":"2e49a4f4d15e0f9f4d44b65d60eddbba800762c03a9848cdc615a75922f93b74"} Feb 02 09:26:41 crc kubenswrapper[4764]: I0202 09:26:41.915647 4764 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2e49a4f4d15e0f9f4d44b65d60eddbba800762c03a9848cdc615a75922f93b74" Feb 02 09:26:41 crc kubenswrapper[4764]: I0202 09:26:41.915699 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-448kc" Feb 02 09:26:41 crc kubenswrapper[4764]: I0202 09:26:41.918507 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-scmkd" Feb 02 09:26:41 crc kubenswrapper[4764]: I0202 09:26:41.918490 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-scmkd" event={"ID":"2ccd54a4-fbee-4ad6-b0eb-9d22fdb1eebe","Type":"ContainerDied","Data":"2f9ce56e59d7af521e668975b6e7e1dd45cd838d0b03e98772350698aa9429d2"} Feb 02 09:26:41 crc kubenswrapper[4764]: I0202 09:26:41.918625 4764 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2f9ce56e59d7af521e668975b6e7e1dd45cd838d0b03e98772350698aa9429d2" Feb 02 09:26:41 crc kubenswrapper[4764]: I0202 09:26:41.922417 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-8ms85" event={"ID":"64d8d838-0afe-4f39-ae93-027fea76ced1","Type":"ContainerDied","Data":"1ffe35fcf31d18c8926d39776d3bc1fa1b1dde90860fc181303ed3a28f1b6416"} Feb 02 09:26:41 crc kubenswrapper[4764]: I0202 09:26:41.922441 4764 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1ffe35fcf31d18c8926d39776d3bc1fa1b1dde90860fc181303ed3a28f1b6416" Feb 02 09:26:41 crc kubenswrapper[4764]: I0202 09:26:41.922487 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-8ms85" Feb 02 09:26:41 crc kubenswrapper[4764]: I0202 09:26:41.925138 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5911a7f4-32cf-4b23-a056-baf957ebf6f6","Type":"ContainerStarted","Data":"0f3f96d6aac3c120e6c0d11b39195a3860921f2a1e5f90e9ceaef528007dda43"} Feb 02 09:26:41 crc kubenswrapper[4764]: I0202 09:26:41.925170 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5911a7f4-32cf-4b23-a056-baf957ebf6f6","Type":"ContainerStarted","Data":"5929d0fef49bb9e5d0fca76029da779077a8c42391a4c2df5550f28ffd1b1e99"} Feb 02 09:26:41 crc kubenswrapper[4764]: I0202 09:26:41.927711 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-10df-account-create-update-gb8cd" event={"ID":"902c771b-8da4-4c38-911e-c8b531d076a8","Type":"ContainerDied","Data":"3e478671319cdce58bea9f612fda89e04a4085302fd2bf4576d05402661d0d03"} Feb 02 09:26:41 crc kubenswrapper[4764]: I0202 09:26:41.927735 4764 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3e478671319cdce58bea9f612fda89e04a4085302fd2bf4576d05402661d0d03" Feb 02 09:26:41 crc kubenswrapper[4764]: I0202 09:26:41.927876 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-10df-account-create-update-gb8cd" Feb 02 09:26:41 crc kubenswrapper[4764]: I0202 09:26:41.930118 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-91bb-account-create-update-gg7c4" event={"ID":"3ee0de70-5bfb-4b6e-974c-601071027218","Type":"ContainerDied","Data":"55784b5be0c377291e9b80c756aa5a694cc7c6e084aa45e3e9b3be27f5a828c4"} Feb 02 09:26:41 crc kubenswrapper[4764]: I0202 09:26:41.930139 4764 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="55784b5be0c377291e9b80c756aa5a694cc7c6e084aa45e3e9b3be27f5a828c4" Feb 02 09:26:41 crc kubenswrapper[4764]: I0202 09:26:41.930174 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-91bb-account-create-update-gg7c4" Feb 02 09:26:41 crc kubenswrapper[4764]: I0202 09:26:41.937016 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-39d3-account-create-update-mtxtf" event={"ID":"afbe3865-2077-45c5-8a94-3d0a6774d8e2","Type":"ContainerDied","Data":"005e2dbe434c241d182f43904437136b4ea21e54961283408aac9bda495dd3fc"} Feb 02 09:26:41 crc kubenswrapper[4764]: I0202 09:26:41.937083 4764 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="005e2dbe434c241d182f43904437136b4ea21e54961283408aac9bda495dd3fc" Feb 02 09:26:41 crc kubenswrapper[4764]: I0202 09:26:41.937124 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-39d3-account-create-update-mtxtf" Feb 02 09:26:42 crc kubenswrapper[4764]: I0202 09:26:42.946375 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5911a7f4-32cf-4b23-a056-baf957ebf6f6","Type":"ContainerStarted","Data":"f4fd960565a5d7dfb5dfb7231467bc7bfee873667a0216fafd3ae631cc625310"} Feb 02 09:26:42 crc kubenswrapper[4764]: I0202 09:26:42.946605 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5911a7f4-32cf-4b23-a056-baf957ebf6f6","Type":"ContainerStarted","Data":"a2bd00ba81e1bcf19805161498046cf03eab3deedb3ae7e8f62e0a2cb91e2cb2"} Feb 02 09:26:43 crc kubenswrapper[4764]: I0202 09:26:43.522757 4764 patch_prober.go:28] interesting pod/machine-config-daemon-4ndm4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 09:26:43 crc kubenswrapper[4764]: I0202 09:26:43.523148 4764 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 09:26:43 crc kubenswrapper[4764]: I0202 09:26:43.523201 4764 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" Feb 02 09:26:43 crc kubenswrapper[4764]: I0202 09:26:43.523816 4764 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"bbb9c6cc86c729fedc84038f3af8b700a433083aa4652e832aca9d3d773be9fd"} pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 02 09:26:43 crc kubenswrapper[4764]: I0202 09:26:43.523886 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" containerName="machine-config-daemon" containerID="cri-o://bbb9c6cc86c729fedc84038f3af8b700a433083aa4652e832aca9d3d773be9fd" gracePeriod=600 Feb 02 09:26:43 crc kubenswrapper[4764]: I0202 09:26:43.967782 4764 generic.go:334] "Generic (PLEG): container finished" podID="d192f670-9f9d-4539-9641-e4bed73acdd4" containerID="bbb9c6cc86c729fedc84038f3af8b700a433083aa4652e832aca9d3d773be9fd" exitCode=0 Feb 02 09:26:43 crc kubenswrapper[4764]: I0202 09:26:43.968152 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" event={"ID":"d192f670-9f9d-4539-9641-e4bed73acdd4","Type":"ContainerDied","Data":"bbb9c6cc86c729fedc84038f3af8b700a433083aa4652e832aca9d3d773be9fd"} Feb 02 09:26:43 crc kubenswrapper[4764]: I0202 09:26:43.968181 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" event={"ID":"d192f670-9f9d-4539-9641-e4bed73acdd4","Type":"ContainerStarted","Data":"7e0e1c40ccad424f2dce09dc2a68260202cf2abc147b119996f40fbb4f3a453b"} Feb 02 09:26:43 crc kubenswrapper[4764]: I0202 09:26:43.968204 4764 scope.go:117] "RemoveContainer" containerID="a4c68bd9cbe36bd80883876d850d402edf3b17e0660281a9f5527447a8167a49" Feb 02 09:26:45 crc kubenswrapper[4764]: I0202 09:26:45.986968 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5911a7f4-32cf-4b23-a056-baf957ebf6f6","Type":"ContainerStarted","Data":"258ed1f7f66da1abdfd97048e694ee26e556d7ffa1afdd538daf64f5bf2505de"} Feb 02 09:26:45 crc kubenswrapper[4764]: I0202 09:26:45.988466 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 02 09:26:46 crc kubenswrapper[4764]: I0202 09:26:46.020104 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.795555673 podStartE2EDuration="6.020084139s" podCreationTimestamp="2026-02-02 09:26:40 +0000 UTC" firstStartedPulling="2026-02-02 09:26:41.05383944 +0000 UTC m=+1163.987563528" lastFinishedPulling="2026-02-02 09:26:45.278367866 +0000 UTC m=+1168.212091994" observedRunningTime="2026-02-02 09:26:46.011214537 +0000 UTC m=+1168.944938625" watchObservedRunningTime="2026-02-02 09:26:46.020084139 +0000 UTC m=+1168.953808237" Feb 02 09:26:47 crc kubenswrapper[4764]: I0202 09:26:47.739688 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-bnkbg"] Feb 02 09:26:47 crc kubenswrapper[4764]: E0202 09:26:47.744392 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="afbe3865-2077-45c5-8a94-3d0a6774d8e2" containerName="mariadb-account-create-update" Feb 02 09:26:47 crc kubenswrapper[4764]: I0202 09:26:47.744570 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="afbe3865-2077-45c5-8a94-3d0a6774d8e2" containerName="mariadb-account-create-update" Feb 02 09:26:47 crc kubenswrapper[4764]: E0202 09:26:47.744657 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ee0de70-5bfb-4b6e-974c-601071027218" containerName="mariadb-account-create-update" Feb 02 09:26:47 crc kubenswrapper[4764]: I0202 09:26:47.744728 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ee0de70-5bfb-4b6e-974c-601071027218" containerName="mariadb-account-create-update" Feb 02 09:26:47 crc kubenswrapper[4764]: E0202 09:26:47.744803 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dedd7b1f-7ec4-4a7a-a703-7e228f7767b1" containerName="placement-log" Feb 02 09:26:47 crc kubenswrapper[4764]: I0202 09:26:47.744856 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="dedd7b1f-7ec4-4a7a-a703-7e228f7767b1" containerName="placement-log" Feb 02 09:26:47 crc kubenswrapper[4764]: E0202 09:26:47.744924 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d492e01f-38f0-4c76-b2aa-a177760a757f" containerName="mariadb-database-create" Feb 02 09:26:47 crc kubenswrapper[4764]: I0202 09:26:47.744993 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="d492e01f-38f0-4c76-b2aa-a177760a757f" containerName="mariadb-database-create" Feb 02 09:26:47 crc kubenswrapper[4764]: E0202 09:26:47.745058 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="902c771b-8da4-4c38-911e-c8b531d076a8" containerName="mariadb-account-create-update" Feb 02 09:26:47 crc kubenswrapper[4764]: I0202 09:26:47.745116 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="902c771b-8da4-4c38-911e-c8b531d076a8" containerName="mariadb-account-create-update" Feb 02 09:26:47 crc kubenswrapper[4764]: E0202 09:26:47.745178 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ccd54a4-fbee-4ad6-b0eb-9d22fdb1eebe" containerName="mariadb-database-create" Feb 02 09:26:47 crc kubenswrapper[4764]: I0202 09:26:47.745227 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ccd54a4-fbee-4ad6-b0eb-9d22fdb1eebe" containerName="mariadb-database-create" Feb 02 09:26:47 crc kubenswrapper[4764]: E0202 09:26:47.745282 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64d8d838-0afe-4f39-ae93-027fea76ced1" containerName="mariadb-database-create" Feb 02 09:26:47 crc kubenswrapper[4764]: I0202 09:26:47.745333 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="64d8d838-0afe-4f39-ae93-027fea76ced1" containerName="mariadb-database-create" Feb 02 09:26:47 crc kubenswrapper[4764]: E0202 09:26:47.745394 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dedd7b1f-7ec4-4a7a-a703-7e228f7767b1" containerName="placement-api" Feb 02 09:26:47 crc kubenswrapper[4764]: I0202 09:26:47.745467 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="dedd7b1f-7ec4-4a7a-a703-7e228f7767b1" containerName="placement-api" Feb 02 09:26:47 crc kubenswrapper[4764]: I0202 09:26:47.746005 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="dedd7b1f-7ec4-4a7a-a703-7e228f7767b1" containerName="placement-api" Feb 02 09:26:47 crc kubenswrapper[4764]: I0202 09:26:47.746099 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="902c771b-8da4-4c38-911e-c8b531d076a8" containerName="mariadb-account-create-update" Feb 02 09:26:47 crc kubenswrapper[4764]: I0202 09:26:47.746174 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ccd54a4-fbee-4ad6-b0eb-9d22fdb1eebe" containerName="mariadb-database-create" Feb 02 09:26:47 crc kubenswrapper[4764]: I0202 09:26:47.746235 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ee0de70-5bfb-4b6e-974c-601071027218" containerName="mariadb-account-create-update" Feb 02 09:26:47 crc kubenswrapper[4764]: I0202 09:26:47.746286 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="64d8d838-0afe-4f39-ae93-027fea76ced1" containerName="mariadb-database-create" Feb 02 09:26:47 crc kubenswrapper[4764]: I0202 09:26:47.746350 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="dedd7b1f-7ec4-4a7a-a703-7e228f7767b1" containerName="placement-log" Feb 02 09:26:47 crc kubenswrapper[4764]: I0202 09:26:47.746432 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="afbe3865-2077-45c5-8a94-3d0a6774d8e2" containerName="mariadb-account-create-update" Feb 02 09:26:47 crc kubenswrapper[4764]: I0202 09:26:47.746492 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="d492e01f-38f0-4c76-b2aa-a177760a757f" containerName="mariadb-database-create" Feb 02 09:26:47 crc kubenswrapper[4764]: I0202 09:26:47.747288 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-bnkbg" Feb 02 09:26:47 crc kubenswrapper[4764]: I0202 09:26:47.750361 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Feb 02 09:26:47 crc kubenswrapper[4764]: I0202 09:26:47.750646 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-c8w6h" Feb 02 09:26:47 crc kubenswrapper[4764]: I0202 09:26:47.750707 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Feb 02 09:26:47 crc kubenswrapper[4764]: I0202 09:26:47.794989 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-bnkbg"] Feb 02 09:26:47 crc kubenswrapper[4764]: I0202 09:26:47.830530 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4rlr2\" (UniqueName: \"kubernetes.io/projected/40a1e8e4-e9f7-4108-858e-ff65c0becc8d-kube-api-access-4rlr2\") pod \"nova-cell0-conductor-db-sync-bnkbg\" (UID: \"40a1e8e4-e9f7-4108-858e-ff65c0becc8d\") " pod="openstack/nova-cell0-conductor-db-sync-bnkbg" Feb 02 09:26:47 crc kubenswrapper[4764]: I0202 09:26:47.830664 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40a1e8e4-e9f7-4108-858e-ff65c0becc8d-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-bnkbg\" (UID: \"40a1e8e4-e9f7-4108-858e-ff65c0becc8d\") " pod="openstack/nova-cell0-conductor-db-sync-bnkbg" Feb 02 09:26:47 crc kubenswrapper[4764]: I0202 09:26:47.830692 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/40a1e8e4-e9f7-4108-858e-ff65c0becc8d-scripts\") pod \"nova-cell0-conductor-db-sync-bnkbg\" (UID: \"40a1e8e4-e9f7-4108-858e-ff65c0becc8d\") " pod="openstack/nova-cell0-conductor-db-sync-bnkbg" Feb 02 09:26:47 crc kubenswrapper[4764]: I0202 09:26:47.830807 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/40a1e8e4-e9f7-4108-858e-ff65c0becc8d-config-data\") pod \"nova-cell0-conductor-db-sync-bnkbg\" (UID: \"40a1e8e4-e9f7-4108-858e-ff65c0becc8d\") " pod="openstack/nova-cell0-conductor-db-sync-bnkbg" Feb 02 09:26:47 crc kubenswrapper[4764]: I0202 09:26:47.933852 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/40a1e8e4-e9f7-4108-858e-ff65c0becc8d-config-data\") pod \"nova-cell0-conductor-db-sync-bnkbg\" (UID: \"40a1e8e4-e9f7-4108-858e-ff65c0becc8d\") " pod="openstack/nova-cell0-conductor-db-sync-bnkbg" Feb 02 09:26:47 crc kubenswrapper[4764]: I0202 09:26:47.933940 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4rlr2\" (UniqueName: \"kubernetes.io/projected/40a1e8e4-e9f7-4108-858e-ff65c0becc8d-kube-api-access-4rlr2\") pod \"nova-cell0-conductor-db-sync-bnkbg\" (UID: \"40a1e8e4-e9f7-4108-858e-ff65c0becc8d\") " pod="openstack/nova-cell0-conductor-db-sync-bnkbg" Feb 02 09:26:47 crc kubenswrapper[4764]: I0202 09:26:47.934053 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/40a1e8e4-e9f7-4108-858e-ff65c0becc8d-scripts\") pod \"nova-cell0-conductor-db-sync-bnkbg\" (UID: \"40a1e8e4-e9f7-4108-858e-ff65c0becc8d\") " pod="openstack/nova-cell0-conductor-db-sync-bnkbg" Feb 02 09:26:47 crc kubenswrapper[4764]: I0202 09:26:47.934075 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40a1e8e4-e9f7-4108-858e-ff65c0becc8d-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-bnkbg\" (UID: \"40a1e8e4-e9f7-4108-858e-ff65c0becc8d\") " pod="openstack/nova-cell0-conductor-db-sync-bnkbg" Feb 02 09:26:47 crc kubenswrapper[4764]: I0202 09:26:47.945007 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/40a1e8e4-e9f7-4108-858e-ff65c0becc8d-scripts\") pod \"nova-cell0-conductor-db-sync-bnkbg\" (UID: \"40a1e8e4-e9f7-4108-858e-ff65c0becc8d\") " pod="openstack/nova-cell0-conductor-db-sync-bnkbg" Feb 02 09:26:47 crc kubenswrapper[4764]: I0202 09:26:47.948707 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/40a1e8e4-e9f7-4108-858e-ff65c0becc8d-config-data\") pod \"nova-cell0-conductor-db-sync-bnkbg\" (UID: \"40a1e8e4-e9f7-4108-858e-ff65c0becc8d\") " pod="openstack/nova-cell0-conductor-db-sync-bnkbg" Feb 02 09:26:47 crc kubenswrapper[4764]: I0202 09:26:47.949285 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40a1e8e4-e9f7-4108-858e-ff65c0becc8d-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-bnkbg\" (UID: \"40a1e8e4-e9f7-4108-858e-ff65c0becc8d\") " pod="openstack/nova-cell0-conductor-db-sync-bnkbg" Feb 02 09:26:47 crc kubenswrapper[4764]: I0202 09:26:47.958441 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4rlr2\" (UniqueName: \"kubernetes.io/projected/40a1e8e4-e9f7-4108-858e-ff65c0becc8d-kube-api-access-4rlr2\") pod \"nova-cell0-conductor-db-sync-bnkbg\" (UID: \"40a1e8e4-e9f7-4108-858e-ff65c0becc8d\") " pod="openstack/nova-cell0-conductor-db-sync-bnkbg" Feb 02 09:26:48 crc kubenswrapper[4764]: I0202 09:26:48.069726 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-bnkbg" Feb 02 09:26:48 crc kubenswrapper[4764]: I0202 09:26:48.533981 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-bnkbg"] Feb 02 09:26:49 crc kubenswrapper[4764]: I0202 09:26:49.025819 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-bnkbg" event={"ID":"40a1e8e4-e9f7-4108-858e-ff65c0becc8d","Type":"ContainerStarted","Data":"75715139c8d8e9a0d7ba4f4a038438ad5ba7edf924be8083c76b8cde502eac94"} Feb 02 09:26:51 crc kubenswrapper[4764]: I0202 09:26:51.076805 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 02 09:26:51 crc kubenswrapper[4764]: I0202 09:26:51.077339 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5911a7f4-32cf-4b23-a056-baf957ebf6f6" containerName="ceilometer-central-agent" containerID="cri-o://0f3f96d6aac3c120e6c0d11b39195a3860921f2a1e5f90e9ceaef528007dda43" gracePeriod=30 Feb 02 09:26:51 crc kubenswrapper[4764]: I0202 09:26:51.077399 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5911a7f4-32cf-4b23-a056-baf957ebf6f6" containerName="proxy-httpd" containerID="cri-o://258ed1f7f66da1abdfd97048e694ee26e556d7ffa1afdd538daf64f5bf2505de" gracePeriod=30 Feb 02 09:26:51 crc kubenswrapper[4764]: I0202 09:26:51.077456 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5911a7f4-32cf-4b23-a056-baf957ebf6f6" containerName="ceilometer-notification-agent" containerID="cri-o://a2bd00ba81e1bcf19805161498046cf03eab3deedb3ae7e8f62e0a2cb91e2cb2" gracePeriod=30 Feb 02 09:26:51 crc kubenswrapper[4764]: I0202 09:26:51.077625 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5911a7f4-32cf-4b23-a056-baf957ebf6f6" containerName="sg-core" containerID="cri-o://f4fd960565a5d7dfb5dfb7231467bc7bfee873667a0216fafd3ae631cc625310" gracePeriod=30 Feb 02 09:26:52 crc kubenswrapper[4764]: I0202 09:26:52.048523 4764 generic.go:334] "Generic (PLEG): container finished" podID="5911a7f4-32cf-4b23-a056-baf957ebf6f6" containerID="258ed1f7f66da1abdfd97048e694ee26e556d7ffa1afdd538daf64f5bf2505de" exitCode=0 Feb 02 09:26:52 crc kubenswrapper[4764]: I0202 09:26:52.048781 4764 generic.go:334] "Generic (PLEG): container finished" podID="5911a7f4-32cf-4b23-a056-baf957ebf6f6" containerID="f4fd960565a5d7dfb5dfb7231467bc7bfee873667a0216fafd3ae631cc625310" exitCode=2 Feb 02 09:26:52 crc kubenswrapper[4764]: I0202 09:26:52.048788 4764 generic.go:334] "Generic (PLEG): container finished" podID="5911a7f4-32cf-4b23-a056-baf957ebf6f6" containerID="a2bd00ba81e1bcf19805161498046cf03eab3deedb3ae7e8f62e0a2cb91e2cb2" exitCode=0 Feb 02 09:26:52 crc kubenswrapper[4764]: I0202 09:26:52.048795 4764 generic.go:334] "Generic (PLEG): container finished" podID="5911a7f4-32cf-4b23-a056-baf957ebf6f6" containerID="0f3f96d6aac3c120e6c0d11b39195a3860921f2a1e5f90e9ceaef528007dda43" exitCode=0 Feb 02 09:26:52 crc kubenswrapper[4764]: I0202 09:26:52.048813 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5911a7f4-32cf-4b23-a056-baf957ebf6f6","Type":"ContainerDied","Data":"258ed1f7f66da1abdfd97048e694ee26e556d7ffa1afdd538daf64f5bf2505de"} Feb 02 09:26:52 crc kubenswrapper[4764]: I0202 09:26:52.048837 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5911a7f4-32cf-4b23-a056-baf957ebf6f6","Type":"ContainerDied","Data":"f4fd960565a5d7dfb5dfb7231467bc7bfee873667a0216fafd3ae631cc625310"} Feb 02 09:26:52 crc kubenswrapper[4764]: I0202 09:26:52.048846 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5911a7f4-32cf-4b23-a056-baf957ebf6f6","Type":"ContainerDied","Data":"a2bd00ba81e1bcf19805161498046cf03eab3deedb3ae7e8f62e0a2cb91e2cb2"} Feb 02 09:26:52 crc kubenswrapper[4764]: I0202 09:26:52.048855 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5911a7f4-32cf-4b23-a056-baf957ebf6f6","Type":"ContainerDied","Data":"0f3f96d6aac3c120e6c0d11b39195a3860921f2a1e5f90e9ceaef528007dda43"} Feb 02 09:26:55 crc kubenswrapper[4764]: I0202 09:26:55.754901 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 02 09:26:55 crc kubenswrapper[4764]: I0202 09:26:55.862488 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5911a7f4-32cf-4b23-a056-baf957ebf6f6-log-httpd\") pod \"5911a7f4-32cf-4b23-a056-baf957ebf6f6\" (UID: \"5911a7f4-32cf-4b23-a056-baf957ebf6f6\") " Feb 02 09:26:55 crc kubenswrapper[4764]: I0202 09:26:55.862864 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hrpnk\" (UniqueName: \"kubernetes.io/projected/5911a7f4-32cf-4b23-a056-baf957ebf6f6-kube-api-access-hrpnk\") pod \"5911a7f4-32cf-4b23-a056-baf957ebf6f6\" (UID: \"5911a7f4-32cf-4b23-a056-baf957ebf6f6\") " Feb 02 09:26:55 crc kubenswrapper[4764]: I0202 09:26:55.862894 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5911a7f4-32cf-4b23-a056-baf957ebf6f6-run-httpd\") pod \"5911a7f4-32cf-4b23-a056-baf957ebf6f6\" (UID: \"5911a7f4-32cf-4b23-a056-baf957ebf6f6\") " Feb 02 09:26:55 crc kubenswrapper[4764]: I0202 09:26:55.862921 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5911a7f4-32cf-4b23-a056-baf957ebf6f6-scripts\") pod \"5911a7f4-32cf-4b23-a056-baf957ebf6f6\" (UID: \"5911a7f4-32cf-4b23-a056-baf957ebf6f6\") " Feb 02 09:26:55 crc kubenswrapper[4764]: I0202 09:26:55.862973 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5911a7f4-32cf-4b23-a056-baf957ebf6f6-config-data\") pod \"5911a7f4-32cf-4b23-a056-baf957ebf6f6\" (UID: \"5911a7f4-32cf-4b23-a056-baf957ebf6f6\") " Feb 02 09:26:55 crc kubenswrapper[4764]: I0202 09:26:55.863086 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5911a7f4-32cf-4b23-a056-baf957ebf6f6-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "5911a7f4-32cf-4b23-a056-baf957ebf6f6" (UID: "5911a7f4-32cf-4b23-a056-baf957ebf6f6"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 09:26:55 crc kubenswrapper[4764]: I0202 09:26:55.863142 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5911a7f4-32cf-4b23-a056-baf957ebf6f6-sg-core-conf-yaml\") pod \"5911a7f4-32cf-4b23-a056-baf957ebf6f6\" (UID: \"5911a7f4-32cf-4b23-a056-baf957ebf6f6\") " Feb 02 09:26:55 crc kubenswrapper[4764]: I0202 09:26:55.863163 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5911a7f4-32cf-4b23-a056-baf957ebf6f6-combined-ca-bundle\") pod \"5911a7f4-32cf-4b23-a056-baf957ebf6f6\" (UID: \"5911a7f4-32cf-4b23-a056-baf957ebf6f6\") " Feb 02 09:26:55 crc kubenswrapper[4764]: I0202 09:26:55.863315 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5911a7f4-32cf-4b23-a056-baf957ebf6f6-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "5911a7f4-32cf-4b23-a056-baf957ebf6f6" (UID: "5911a7f4-32cf-4b23-a056-baf957ebf6f6"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 09:26:55 crc kubenswrapper[4764]: I0202 09:26:55.863498 4764 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5911a7f4-32cf-4b23-a056-baf957ebf6f6-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 02 09:26:55 crc kubenswrapper[4764]: I0202 09:26:55.863515 4764 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5911a7f4-32cf-4b23-a056-baf957ebf6f6-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 02 09:26:55 crc kubenswrapper[4764]: I0202 09:26:55.869478 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5911a7f4-32cf-4b23-a056-baf957ebf6f6-scripts" (OuterVolumeSpecName: "scripts") pod "5911a7f4-32cf-4b23-a056-baf957ebf6f6" (UID: "5911a7f4-32cf-4b23-a056-baf957ebf6f6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:26:55 crc kubenswrapper[4764]: I0202 09:26:55.869666 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5911a7f4-32cf-4b23-a056-baf957ebf6f6-kube-api-access-hrpnk" (OuterVolumeSpecName: "kube-api-access-hrpnk") pod "5911a7f4-32cf-4b23-a056-baf957ebf6f6" (UID: "5911a7f4-32cf-4b23-a056-baf957ebf6f6"). InnerVolumeSpecName "kube-api-access-hrpnk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:26:55 crc kubenswrapper[4764]: I0202 09:26:55.894371 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5911a7f4-32cf-4b23-a056-baf957ebf6f6-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "5911a7f4-32cf-4b23-a056-baf957ebf6f6" (UID: "5911a7f4-32cf-4b23-a056-baf957ebf6f6"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:26:55 crc kubenswrapper[4764]: I0202 09:26:55.921830 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5911a7f4-32cf-4b23-a056-baf957ebf6f6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5911a7f4-32cf-4b23-a056-baf957ebf6f6" (UID: "5911a7f4-32cf-4b23-a056-baf957ebf6f6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:26:55 crc kubenswrapper[4764]: I0202 09:26:55.939093 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5911a7f4-32cf-4b23-a056-baf957ebf6f6-config-data" (OuterVolumeSpecName: "config-data") pod "5911a7f4-32cf-4b23-a056-baf957ebf6f6" (UID: "5911a7f4-32cf-4b23-a056-baf957ebf6f6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:26:55 crc kubenswrapper[4764]: I0202 09:26:55.964815 4764 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5911a7f4-32cf-4b23-a056-baf957ebf6f6-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 02 09:26:55 crc kubenswrapper[4764]: I0202 09:26:55.964840 4764 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5911a7f4-32cf-4b23-a056-baf957ebf6f6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 09:26:55 crc kubenswrapper[4764]: I0202 09:26:55.964850 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hrpnk\" (UniqueName: \"kubernetes.io/projected/5911a7f4-32cf-4b23-a056-baf957ebf6f6-kube-api-access-hrpnk\") on node \"crc\" DevicePath \"\"" Feb 02 09:26:55 crc kubenswrapper[4764]: I0202 09:26:55.964861 4764 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5911a7f4-32cf-4b23-a056-baf957ebf6f6-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 09:26:55 crc kubenswrapper[4764]: I0202 09:26:55.964870 4764 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5911a7f4-32cf-4b23-a056-baf957ebf6f6-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 09:26:56 crc kubenswrapper[4764]: I0202 09:26:56.086313 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-bnkbg" event={"ID":"40a1e8e4-e9f7-4108-858e-ff65c0becc8d","Type":"ContainerStarted","Data":"7fbdf340596d9e601a0c1ec9ec2dbca538a4261351156b563a3e08f1a090259f"} Feb 02 09:26:56 crc kubenswrapper[4764]: I0202 09:26:56.088857 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5911a7f4-32cf-4b23-a056-baf957ebf6f6","Type":"ContainerDied","Data":"5929d0fef49bb9e5d0fca76029da779077a8c42391a4c2df5550f28ffd1b1e99"} Feb 02 09:26:56 crc kubenswrapper[4764]: I0202 09:26:56.088889 4764 scope.go:117] "RemoveContainer" containerID="258ed1f7f66da1abdfd97048e694ee26e556d7ffa1afdd538daf64f5bf2505de" Feb 02 09:26:56 crc kubenswrapper[4764]: I0202 09:26:56.088907 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 02 09:26:56 crc kubenswrapper[4764]: I0202 09:26:56.108425 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-bnkbg" podStartSLOduration=2.110767045 podStartE2EDuration="9.108407031s" podCreationTimestamp="2026-02-02 09:26:47 +0000 UTC" firstStartedPulling="2026-02-02 09:26:48.545596933 +0000 UTC m=+1171.479321021" lastFinishedPulling="2026-02-02 09:26:55.543236919 +0000 UTC m=+1178.476961007" observedRunningTime="2026-02-02 09:26:56.100837555 +0000 UTC m=+1179.034561643" watchObservedRunningTime="2026-02-02 09:26:56.108407031 +0000 UTC m=+1179.042131119" Feb 02 09:26:56 crc kubenswrapper[4764]: I0202 09:26:56.114603 4764 scope.go:117] "RemoveContainer" containerID="f4fd960565a5d7dfb5dfb7231467bc7bfee873667a0216fafd3ae631cc625310" Feb 02 09:26:56 crc kubenswrapper[4764]: I0202 09:26:56.139468 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 02 09:26:56 crc kubenswrapper[4764]: I0202 09:26:56.163196 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 02 09:26:56 crc kubenswrapper[4764]: I0202 09:26:56.171250 4764 scope.go:117] "RemoveContainer" containerID="a2bd00ba81e1bcf19805161498046cf03eab3deedb3ae7e8f62e0a2cb91e2cb2" Feb 02 09:26:56 crc kubenswrapper[4764]: E0202 09:26:56.177797 4764 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5911a7f4_32cf_4b23_a056_baf957ebf6f6.slice/crio-5929d0fef49bb9e5d0fca76029da779077a8c42391a4c2df5550f28ffd1b1e99\": RecentStats: unable to find data in memory cache]" Feb 02 09:26:56 crc kubenswrapper[4764]: I0202 09:26:56.190439 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 02 09:26:56 crc kubenswrapper[4764]: E0202 09:26:56.190796 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5911a7f4-32cf-4b23-a056-baf957ebf6f6" containerName="ceilometer-notification-agent" Feb 02 09:26:56 crc kubenswrapper[4764]: I0202 09:26:56.190889 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="5911a7f4-32cf-4b23-a056-baf957ebf6f6" containerName="ceilometer-notification-agent" Feb 02 09:26:56 crc kubenswrapper[4764]: E0202 09:26:56.190970 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5911a7f4-32cf-4b23-a056-baf957ebf6f6" containerName="proxy-httpd" Feb 02 09:26:56 crc kubenswrapper[4764]: I0202 09:26:56.191022 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="5911a7f4-32cf-4b23-a056-baf957ebf6f6" containerName="proxy-httpd" Feb 02 09:26:56 crc kubenswrapper[4764]: E0202 09:26:56.191078 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5911a7f4-32cf-4b23-a056-baf957ebf6f6" containerName="sg-core" Feb 02 09:26:56 crc kubenswrapper[4764]: I0202 09:26:56.191199 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="5911a7f4-32cf-4b23-a056-baf957ebf6f6" containerName="sg-core" Feb 02 09:26:56 crc kubenswrapper[4764]: E0202 09:26:56.191261 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5911a7f4-32cf-4b23-a056-baf957ebf6f6" containerName="ceilometer-central-agent" Feb 02 09:26:56 crc kubenswrapper[4764]: I0202 09:26:56.191318 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="5911a7f4-32cf-4b23-a056-baf957ebf6f6" containerName="ceilometer-central-agent" Feb 02 09:26:56 crc kubenswrapper[4764]: I0202 09:26:56.191583 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="5911a7f4-32cf-4b23-a056-baf957ebf6f6" containerName="ceilometer-notification-agent" Feb 02 09:26:56 crc kubenswrapper[4764]: I0202 09:26:56.191669 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="5911a7f4-32cf-4b23-a056-baf957ebf6f6" containerName="sg-core" Feb 02 09:26:56 crc kubenswrapper[4764]: I0202 09:26:56.191755 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="5911a7f4-32cf-4b23-a056-baf957ebf6f6" containerName="proxy-httpd" Feb 02 09:26:56 crc kubenswrapper[4764]: I0202 09:26:56.191833 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="5911a7f4-32cf-4b23-a056-baf957ebf6f6" containerName="ceilometer-central-agent" Feb 02 09:26:56 crc kubenswrapper[4764]: I0202 09:26:56.193132 4764 scope.go:117] "RemoveContainer" containerID="0f3f96d6aac3c120e6c0d11b39195a3860921f2a1e5f90e9ceaef528007dda43" Feb 02 09:26:56 crc kubenswrapper[4764]: I0202 09:26:56.194616 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 02 09:26:56 crc kubenswrapper[4764]: I0202 09:26:56.196524 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 02 09:26:56 crc kubenswrapper[4764]: I0202 09:26:56.197420 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 02 09:26:56 crc kubenswrapper[4764]: I0202 09:26:56.202486 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 02 09:26:56 crc kubenswrapper[4764]: I0202 09:26:56.372767 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3cad43ae-e284-446f-9dbc-ba224b11aac4-log-httpd\") pod \"ceilometer-0\" (UID: \"3cad43ae-e284-446f-9dbc-ba224b11aac4\") " pod="openstack/ceilometer-0" Feb 02 09:26:56 crc kubenswrapper[4764]: I0202 09:26:56.372813 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3cad43ae-e284-446f-9dbc-ba224b11aac4-scripts\") pod \"ceilometer-0\" (UID: \"3cad43ae-e284-446f-9dbc-ba224b11aac4\") " pod="openstack/ceilometer-0" Feb 02 09:26:56 crc kubenswrapper[4764]: I0202 09:26:56.372880 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6v4xn\" (UniqueName: \"kubernetes.io/projected/3cad43ae-e284-446f-9dbc-ba224b11aac4-kube-api-access-6v4xn\") pod \"ceilometer-0\" (UID: \"3cad43ae-e284-446f-9dbc-ba224b11aac4\") " pod="openstack/ceilometer-0" Feb 02 09:26:56 crc kubenswrapper[4764]: I0202 09:26:56.372904 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cad43ae-e284-446f-9dbc-ba224b11aac4-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3cad43ae-e284-446f-9dbc-ba224b11aac4\") " pod="openstack/ceilometer-0" Feb 02 09:26:56 crc kubenswrapper[4764]: I0202 09:26:56.372949 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3cad43ae-e284-446f-9dbc-ba224b11aac4-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3cad43ae-e284-446f-9dbc-ba224b11aac4\") " pod="openstack/ceilometer-0" Feb 02 09:26:56 crc kubenswrapper[4764]: I0202 09:26:56.372975 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3cad43ae-e284-446f-9dbc-ba224b11aac4-run-httpd\") pod \"ceilometer-0\" (UID: \"3cad43ae-e284-446f-9dbc-ba224b11aac4\") " pod="openstack/ceilometer-0" Feb 02 09:26:56 crc kubenswrapper[4764]: I0202 09:26:56.372996 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3cad43ae-e284-446f-9dbc-ba224b11aac4-config-data\") pod \"ceilometer-0\" (UID: \"3cad43ae-e284-446f-9dbc-ba224b11aac4\") " pod="openstack/ceilometer-0" Feb 02 09:26:56 crc kubenswrapper[4764]: I0202 09:26:56.474734 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3cad43ae-e284-446f-9dbc-ba224b11aac4-log-httpd\") pod \"ceilometer-0\" (UID: \"3cad43ae-e284-446f-9dbc-ba224b11aac4\") " pod="openstack/ceilometer-0" Feb 02 09:26:56 crc kubenswrapper[4764]: I0202 09:26:56.475711 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3cad43ae-e284-446f-9dbc-ba224b11aac4-scripts\") pod \"ceilometer-0\" (UID: \"3cad43ae-e284-446f-9dbc-ba224b11aac4\") " pod="openstack/ceilometer-0" Feb 02 09:26:56 crc kubenswrapper[4764]: I0202 09:26:56.475900 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6v4xn\" (UniqueName: \"kubernetes.io/projected/3cad43ae-e284-446f-9dbc-ba224b11aac4-kube-api-access-6v4xn\") pod \"ceilometer-0\" (UID: \"3cad43ae-e284-446f-9dbc-ba224b11aac4\") " pod="openstack/ceilometer-0" Feb 02 09:26:56 crc kubenswrapper[4764]: I0202 09:26:56.476041 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cad43ae-e284-446f-9dbc-ba224b11aac4-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3cad43ae-e284-446f-9dbc-ba224b11aac4\") " pod="openstack/ceilometer-0" Feb 02 09:26:56 crc kubenswrapper[4764]: I0202 09:26:56.476153 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3cad43ae-e284-446f-9dbc-ba224b11aac4-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3cad43ae-e284-446f-9dbc-ba224b11aac4\") " pod="openstack/ceilometer-0" Feb 02 09:26:56 crc kubenswrapper[4764]: I0202 09:26:56.476250 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3cad43ae-e284-446f-9dbc-ba224b11aac4-run-httpd\") pod \"ceilometer-0\" (UID: \"3cad43ae-e284-446f-9dbc-ba224b11aac4\") " pod="openstack/ceilometer-0" Feb 02 09:26:56 crc kubenswrapper[4764]: I0202 09:26:56.476667 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3cad43ae-e284-446f-9dbc-ba224b11aac4-config-data\") pod \"ceilometer-0\" (UID: \"3cad43ae-e284-446f-9dbc-ba224b11aac4\") " pod="openstack/ceilometer-0" Feb 02 09:26:56 crc kubenswrapper[4764]: I0202 09:26:56.475667 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3cad43ae-e284-446f-9dbc-ba224b11aac4-log-httpd\") pod \"ceilometer-0\" (UID: \"3cad43ae-e284-446f-9dbc-ba224b11aac4\") " pod="openstack/ceilometer-0" Feb 02 09:26:56 crc kubenswrapper[4764]: I0202 09:26:56.476827 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3cad43ae-e284-446f-9dbc-ba224b11aac4-run-httpd\") pod \"ceilometer-0\" (UID: \"3cad43ae-e284-446f-9dbc-ba224b11aac4\") " pod="openstack/ceilometer-0" Feb 02 09:26:56 crc kubenswrapper[4764]: I0202 09:26:56.485513 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3cad43ae-e284-446f-9dbc-ba224b11aac4-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3cad43ae-e284-446f-9dbc-ba224b11aac4\") " pod="openstack/ceilometer-0" Feb 02 09:26:56 crc kubenswrapper[4764]: I0202 09:26:56.485737 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3cad43ae-e284-446f-9dbc-ba224b11aac4-scripts\") pod \"ceilometer-0\" (UID: \"3cad43ae-e284-446f-9dbc-ba224b11aac4\") " pod="openstack/ceilometer-0" Feb 02 09:26:56 crc kubenswrapper[4764]: I0202 09:26:56.485822 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cad43ae-e284-446f-9dbc-ba224b11aac4-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3cad43ae-e284-446f-9dbc-ba224b11aac4\") " pod="openstack/ceilometer-0" Feb 02 09:26:56 crc kubenswrapper[4764]: I0202 09:26:56.486347 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3cad43ae-e284-446f-9dbc-ba224b11aac4-config-data\") pod \"ceilometer-0\" (UID: \"3cad43ae-e284-446f-9dbc-ba224b11aac4\") " pod="openstack/ceilometer-0" Feb 02 09:26:56 crc kubenswrapper[4764]: I0202 09:26:56.493348 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6v4xn\" (UniqueName: \"kubernetes.io/projected/3cad43ae-e284-446f-9dbc-ba224b11aac4-kube-api-access-6v4xn\") pod \"ceilometer-0\" (UID: \"3cad43ae-e284-446f-9dbc-ba224b11aac4\") " pod="openstack/ceilometer-0" Feb 02 09:26:56 crc kubenswrapper[4764]: I0202 09:26:56.512325 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 02 09:26:56 crc kubenswrapper[4764]: I0202 09:26:56.997948 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 02 09:26:57 crc kubenswrapper[4764]: I0202 09:26:57.099231 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3cad43ae-e284-446f-9dbc-ba224b11aac4","Type":"ContainerStarted","Data":"39fc7654d7dd07e41524b35a271817faa467c6b3dcc8ab16623b678f54955d5a"} Feb 02 09:26:57 crc kubenswrapper[4764]: I0202 09:26:57.841344 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5911a7f4-32cf-4b23-a056-baf957ebf6f6" path="/var/lib/kubelet/pods/5911a7f4-32cf-4b23-a056-baf957ebf6f6/volumes" Feb 02 09:26:58 crc kubenswrapper[4764]: I0202 09:26:58.110472 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3cad43ae-e284-446f-9dbc-ba224b11aac4","Type":"ContainerStarted","Data":"f0e10976314a43245db7f1bfe6ae5503a29b455557f59bd621d463a0219b2bb4"} Feb 02 09:26:59 crc kubenswrapper[4764]: I0202 09:26:59.120318 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3cad43ae-e284-446f-9dbc-ba224b11aac4","Type":"ContainerStarted","Data":"ec37269bba1eeedbf80feeb128bd400885ce3747aa038d326f942e91b66f9cc1"} Feb 02 09:26:59 crc kubenswrapper[4764]: I0202 09:26:59.120907 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3cad43ae-e284-446f-9dbc-ba224b11aac4","Type":"ContainerStarted","Data":"52d9f8634f508399e6f41017e512c93e0cc403c3b39ea0c39e366455b8ee787c"} Feb 02 09:27:01 crc kubenswrapper[4764]: I0202 09:27:01.139181 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3cad43ae-e284-446f-9dbc-ba224b11aac4","Type":"ContainerStarted","Data":"9ba9869e46c55b3868c63cd7c0e91be1c9eec3f6b90f2c1a868d1065a651680a"} Feb 02 09:27:01 crc kubenswrapper[4764]: I0202 09:27:01.139904 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 02 09:27:08 crc kubenswrapper[4764]: I0202 09:27:08.214781 4764 generic.go:334] "Generic (PLEG): container finished" podID="40a1e8e4-e9f7-4108-858e-ff65c0becc8d" containerID="7fbdf340596d9e601a0c1ec9ec2dbca538a4261351156b563a3e08f1a090259f" exitCode=0 Feb 02 09:27:08 crc kubenswrapper[4764]: I0202 09:27:08.214909 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-bnkbg" event={"ID":"40a1e8e4-e9f7-4108-858e-ff65c0becc8d","Type":"ContainerDied","Data":"7fbdf340596d9e601a0c1ec9ec2dbca538a4261351156b563a3e08f1a090259f"} Feb 02 09:27:08 crc kubenswrapper[4764]: I0202 09:27:08.243484 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=8.495827882 podStartE2EDuration="12.243463972s" podCreationTimestamp="2026-02-02 09:26:56 +0000 UTC" firstStartedPulling="2026-02-02 09:26:56.993526502 +0000 UTC m=+1179.927250590" lastFinishedPulling="2026-02-02 09:27:00.741162582 +0000 UTC m=+1183.674886680" observedRunningTime="2026-02-02 09:27:01.173132434 +0000 UTC m=+1184.106856522" watchObservedRunningTime="2026-02-02 09:27:08.243463972 +0000 UTC m=+1191.177188070" Feb 02 09:27:09 crc kubenswrapper[4764]: I0202 09:27:09.592367 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-bnkbg" Feb 02 09:27:09 crc kubenswrapper[4764]: I0202 09:27:09.706156 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/40a1e8e4-e9f7-4108-858e-ff65c0becc8d-config-data\") pod \"40a1e8e4-e9f7-4108-858e-ff65c0becc8d\" (UID: \"40a1e8e4-e9f7-4108-858e-ff65c0becc8d\") " Feb 02 09:27:09 crc kubenswrapper[4764]: I0202 09:27:09.706278 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4rlr2\" (UniqueName: \"kubernetes.io/projected/40a1e8e4-e9f7-4108-858e-ff65c0becc8d-kube-api-access-4rlr2\") pod \"40a1e8e4-e9f7-4108-858e-ff65c0becc8d\" (UID: \"40a1e8e4-e9f7-4108-858e-ff65c0becc8d\") " Feb 02 09:27:09 crc kubenswrapper[4764]: I0202 09:27:09.706434 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40a1e8e4-e9f7-4108-858e-ff65c0becc8d-combined-ca-bundle\") pod \"40a1e8e4-e9f7-4108-858e-ff65c0becc8d\" (UID: \"40a1e8e4-e9f7-4108-858e-ff65c0becc8d\") " Feb 02 09:27:09 crc kubenswrapper[4764]: I0202 09:27:09.707375 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/40a1e8e4-e9f7-4108-858e-ff65c0becc8d-scripts\") pod \"40a1e8e4-e9f7-4108-858e-ff65c0becc8d\" (UID: \"40a1e8e4-e9f7-4108-858e-ff65c0becc8d\") " Feb 02 09:27:09 crc kubenswrapper[4764]: I0202 09:27:09.717551 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/40a1e8e4-e9f7-4108-858e-ff65c0becc8d-kube-api-access-4rlr2" (OuterVolumeSpecName: "kube-api-access-4rlr2") pod "40a1e8e4-e9f7-4108-858e-ff65c0becc8d" (UID: "40a1e8e4-e9f7-4108-858e-ff65c0becc8d"). InnerVolumeSpecName "kube-api-access-4rlr2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:27:09 crc kubenswrapper[4764]: I0202 09:27:09.730151 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/40a1e8e4-e9f7-4108-858e-ff65c0becc8d-scripts" (OuterVolumeSpecName: "scripts") pod "40a1e8e4-e9f7-4108-858e-ff65c0becc8d" (UID: "40a1e8e4-e9f7-4108-858e-ff65c0becc8d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:27:09 crc kubenswrapper[4764]: I0202 09:27:09.747100 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/40a1e8e4-e9f7-4108-858e-ff65c0becc8d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "40a1e8e4-e9f7-4108-858e-ff65c0becc8d" (UID: "40a1e8e4-e9f7-4108-858e-ff65c0becc8d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:27:09 crc kubenswrapper[4764]: I0202 09:27:09.799150 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/40a1e8e4-e9f7-4108-858e-ff65c0becc8d-config-data" (OuterVolumeSpecName: "config-data") pod "40a1e8e4-e9f7-4108-858e-ff65c0becc8d" (UID: "40a1e8e4-e9f7-4108-858e-ff65c0becc8d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:27:09 crc kubenswrapper[4764]: I0202 09:27:09.810368 4764 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40a1e8e4-e9f7-4108-858e-ff65c0becc8d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 09:27:09 crc kubenswrapper[4764]: I0202 09:27:09.810400 4764 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/40a1e8e4-e9f7-4108-858e-ff65c0becc8d-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 09:27:09 crc kubenswrapper[4764]: I0202 09:27:09.810409 4764 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/40a1e8e4-e9f7-4108-858e-ff65c0becc8d-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 09:27:09 crc kubenswrapper[4764]: I0202 09:27:09.810417 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4rlr2\" (UniqueName: \"kubernetes.io/projected/40a1e8e4-e9f7-4108-858e-ff65c0becc8d-kube-api-access-4rlr2\") on node \"crc\" DevicePath \"\"" Feb 02 09:27:10 crc kubenswrapper[4764]: I0202 09:27:10.234577 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-bnkbg" event={"ID":"40a1e8e4-e9f7-4108-858e-ff65c0becc8d","Type":"ContainerDied","Data":"75715139c8d8e9a0d7ba4f4a038438ad5ba7edf924be8083c76b8cde502eac94"} Feb 02 09:27:10 crc kubenswrapper[4764]: I0202 09:27:10.234619 4764 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="75715139c8d8e9a0d7ba4f4a038438ad5ba7edf924be8083c76b8cde502eac94" Feb 02 09:27:10 crc kubenswrapper[4764]: I0202 09:27:10.234645 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-bnkbg" Feb 02 09:27:10 crc kubenswrapper[4764]: I0202 09:27:10.363942 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Feb 02 09:27:10 crc kubenswrapper[4764]: E0202 09:27:10.364495 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40a1e8e4-e9f7-4108-858e-ff65c0becc8d" containerName="nova-cell0-conductor-db-sync" Feb 02 09:27:10 crc kubenswrapper[4764]: I0202 09:27:10.364511 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="40a1e8e4-e9f7-4108-858e-ff65c0becc8d" containerName="nova-cell0-conductor-db-sync" Feb 02 09:27:10 crc kubenswrapper[4764]: I0202 09:27:10.364701 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="40a1e8e4-e9f7-4108-858e-ff65c0becc8d" containerName="nova-cell0-conductor-db-sync" Feb 02 09:27:10 crc kubenswrapper[4764]: I0202 09:27:10.365251 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Feb 02 09:27:10 crc kubenswrapper[4764]: I0202 09:27:10.369612 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-c8w6h" Feb 02 09:27:10 crc kubenswrapper[4764]: I0202 09:27:10.371076 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Feb 02 09:27:10 crc kubenswrapper[4764]: I0202 09:27:10.379902 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Feb 02 09:27:10 crc kubenswrapper[4764]: I0202 09:27:10.421863 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9ww9q\" (UniqueName: \"kubernetes.io/projected/82f0b9ff-5694-4e86-acb1-6a8ef7b5b97a-kube-api-access-9ww9q\") pod \"nova-cell0-conductor-0\" (UID: \"82f0b9ff-5694-4e86-acb1-6a8ef7b5b97a\") " pod="openstack/nova-cell0-conductor-0" Feb 02 09:27:10 crc kubenswrapper[4764]: I0202 09:27:10.422115 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82f0b9ff-5694-4e86-acb1-6a8ef7b5b97a-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"82f0b9ff-5694-4e86-acb1-6a8ef7b5b97a\") " pod="openstack/nova-cell0-conductor-0" Feb 02 09:27:10 crc kubenswrapper[4764]: I0202 09:27:10.422258 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82f0b9ff-5694-4e86-acb1-6a8ef7b5b97a-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"82f0b9ff-5694-4e86-acb1-6a8ef7b5b97a\") " pod="openstack/nova-cell0-conductor-0" Feb 02 09:27:10 crc kubenswrapper[4764]: I0202 09:27:10.524214 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9ww9q\" (UniqueName: \"kubernetes.io/projected/82f0b9ff-5694-4e86-acb1-6a8ef7b5b97a-kube-api-access-9ww9q\") pod \"nova-cell0-conductor-0\" (UID: \"82f0b9ff-5694-4e86-acb1-6a8ef7b5b97a\") " pod="openstack/nova-cell0-conductor-0" Feb 02 09:27:10 crc kubenswrapper[4764]: I0202 09:27:10.524265 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82f0b9ff-5694-4e86-acb1-6a8ef7b5b97a-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"82f0b9ff-5694-4e86-acb1-6a8ef7b5b97a\") " pod="openstack/nova-cell0-conductor-0" Feb 02 09:27:10 crc kubenswrapper[4764]: I0202 09:27:10.524322 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82f0b9ff-5694-4e86-acb1-6a8ef7b5b97a-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"82f0b9ff-5694-4e86-acb1-6a8ef7b5b97a\") " pod="openstack/nova-cell0-conductor-0" Feb 02 09:27:10 crc kubenswrapper[4764]: I0202 09:27:10.528401 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82f0b9ff-5694-4e86-acb1-6a8ef7b5b97a-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"82f0b9ff-5694-4e86-acb1-6a8ef7b5b97a\") " pod="openstack/nova-cell0-conductor-0" Feb 02 09:27:10 crc kubenswrapper[4764]: I0202 09:27:10.528627 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82f0b9ff-5694-4e86-acb1-6a8ef7b5b97a-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"82f0b9ff-5694-4e86-acb1-6a8ef7b5b97a\") " pod="openstack/nova-cell0-conductor-0" Feb 02 09:27:10 crc kubenswrapper[4764]: I0202 09:27:10.543537 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9ww9q\" (UniqueName: \"kubernetes.io/projected/82f0b9ff-5694-4e86-acb1-6a8ef7b5b97a-kube-api-access-9ww9q\") pod \"nova-cell0-conductor-0\" (UID: \"82f0b9ff-5694-4e86-acb1-6a8ef7b5b97a\") " pod="openstack/nova-cell0-conductor-0" Feb 02 09:27:10 crc kubenswrapper[4764]: I0202 09:27:10.678990 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Feb 02 09:27:11 crc kubenswrapper[4764]: I0202 09:27:11.501624 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Feb 02 09:27:11 crc kubenswrapper[4764]: W0202 09:27:11.502167 4764 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod82f0b9ff_5694_4e86_acb1_6a8ef7b5b97a.slice/crio-580f0fda02904dd7b649eb69655ecb68d688427efd8df16cb565b19a8863141f WatchSource:0}: Error finding container 580f0fda02904dd7b649eb69655ecb68d688427efd8df16cb565b19a8863141f: Status 404 returned error can't find the container with id 580f0fda02904dd7b649eb69655ecb68d688427efd8df16cb565b19a8863141f Feb 02 09:27:12 crc kubenswrapper[4764]: I0202 09:27:12.262639 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"82f0b9ff-5694-4e86-acb1-6a8ef7b5b97a","Type":"ContainerStarted","Data":"98e878ae897de28765558723e1028a2c146f64938f66755c1d0e7fbd7707884a"} Feb 02 09:27:12 crc kubenswrapper[4764]: I0202 09:27:12.263041 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Feb 02 09:27:12 crc kubenswrapper[4764]: I0202 09:27:12.263060 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"82f0b9ff-5694-4e86-acb1-6a8ef7b5b97a","Type":"ContainerStarted","Data":"580f0fda02904dd7b649eb69655ecb68d688427efd8df16cb565b19a8863141f"} Feb 02 09:27:12 crc kubenswrapper[4764]: I0202 09:27:12.295335 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.2953121579999998 podStartE2EDuration="2.295312158s" podCreationTimestamp="2026-02-02 09:27:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:27:12.284879774 +0000 UTC m=+1195.218603902" watchObservedRunningTime="2026-02-02 09:27:12.295312158 +0000 UTC m=+1195.229036276" Feb 02 09:27:20 crc kubenswrapper[4764]: I0202 09:27:20.706887 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Feb 02 09:27:21 crc kubenswrapper[4764]: I0202 09:27:21.187102 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-zfgz6"] Feb 02 09:27:21 crc kubenswrapper[4764]: I0202 09:27:21.188330 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-zfgz6" Feb 02 09:27:21 crc kubenswrapper[4764]: I0202 09:27:21.191319 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Feb 02 09:27:21 crc kubenswrapper[4764]: I0202 09:27:21.192077 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Feb 02 09:27:21 crc kubenswrapper[4764]: I0202 09:27:21.207562 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-zfgz6"] Feb 02 09:27:21 crc kubenswrapper[4764]: I0202 09:27:21.231816 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6408cd2c-fa48-4c79-a6dd-558008245a0f-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-zfgz6\" (UID: \"6408cd2c-fa48-4c79-a6dd-558008245a0f\") " pod="openstack/nova-cell0-cell-mapping-zfgz6" Feb 02 09:27:21 crc kubenswrapper[4764]: I0202 09:27:21.231883 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6408cd2c-fa48-4c79-a6dd-558008245a0f-config-data\") pod \"nova-cell0-cell-mapping-zfgz6\" (UID: \"6408cd2c-fa48-4c79-a6dd-558008245a0f\") " pod="openstack/nova-cell0-cell-mapping-zfgz6" Feb 02 09:27:21 crc kubenswrapper[4764]: I0202 09:27:21.232052 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6408cd2c-fa48-4c79-a6dd-558008245a0f-scripts\") pod \"nova-cell0-cell-mapping-zfgz6\" (UID: \"6408cd2c-fa48-4c79-a6dd-558008245a0f\") " pod="openstack/nova-cell0-cell-mapping-zfgz6" Feb 02 09:27:21 crc kubenswrapper[4764]: I0202 09:27:21.232089 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9dtjd\" (UniqueName: \"kubernetes.io/projected/6408cd2c-fa48-4c79-a6dd-558008245a0f-kube-api-access-9dtjd\") pod \"nova-cell0-cell-mapping-zfgz6\" (UID: \"6408cd2c-fa48-4c79-a6dd-558008245a0f\") " pod="openstack/nova-cell0-cell-mapping-zfgz6" Feb 02 09:27:21 crc kubenswrapper[4764]: I0202 09:27:21.338039 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6408cd2c-fa48-4c79-a6dd-558008245a0f-config-data\") pod \"nova-cell0-cell-mapping-zfgz6\" (UID: \"6408cd2c-fa48-4c79-a6dd-558008245a0f\") " pod="openstack/nova-cell0-cell-mapping-zfgz6" Feb 02 09:27:21 crc kubenswrapper[4764]: I0202 09:27:21.338110 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6408cd2c-fa48-4c79-a6dd-558008245a0f-scripts\") pod \"nova-cell0-cell-mapping-zfgz6\" (UID: \"6408cd2c-fa48-4c79-a6dd-558008245a0f\") " pod="openstack/nova-cell0-cell-mapping-zfgz6" Feb 02 09:27:21 crc kubenswrapper[4764]: I0202 09:27:21.338136 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9dtjd\" (UniqueName: \"kubernetes.io/projected/6408cd2c-fa48-4c79-a6dd-558008245a0f-kube-api-access-9dtjd\") pod \"nova-cell0-cell-mapping-zfgz6\" (UID: \"6408cd2c-fa48-4c79-a6dd-558008245a0f\") " pod="openstack/nova-cell0-cell-mapping-zfgz6" Feb 02 09:27:21 crc kubenswrapper[4764]: I0202 09:27:21.338236 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6408cd2c-fa48-4c79-a6dd-558008245a0f-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-zfgz6\" (UID: \"6408cd2c-fa48-4c79-a6dd-558008245a0f\") " pod="openstack/nova-cell0-cell-mapping-zfgz6" Feb 02 09:27:21 crc kubenswrapper[4764]: I0202 09:27:21.343979 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6408cd2c-fa48-4c79-a6dd-558008245a0f-config-data\") pod \"nova-cell0-cell-mapping-zfgz6\" (UID: \"6408cd2c-fa48-4c79-a6dd-558008245a0f\") " pod="openstack/nova-cell0-cell-mapping-zfgz6" Feb 02 09:27:21 crc kubenswrapper[4764]: I0202 09:27:21.349474 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6408cd2c-fa48-4c79-a6dd-558008245a0f-scripts\") pod \"nova-cell0-cell-mapping-zfgz6\" (UID: \"6408cd2c-fa48-4c79-a6dd-558008245a0f\") " pod="openstack/nova-cell0-cell-mapping-zfgz6" Feb 02 09:27:21 crc kubenswrapper[4764]: I0202 09:27:21.349913 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6408cd2c-fa48-4c79-a6dd-558008245a0f-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-zfgz6\" (UID: \"6408cd2c-fa48-4c79-a6dd-558008245a0f\") " pod="openstack/nova-cell0-cell-mapping-zfgz6" Feb 02 09:27:21 crc kubenswrapper[4764]: I0202 09:27:21.355177 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Feb 02 09:27:21 crc kubenswrapper[4764]: I0202 09:27:21.356578 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 02 09:27:21 crc kubenswrapper[4764]: I0202 09:27:21.362066 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Feb 02 09:27:21 crc kubenswrapper[4764]: I0202 09:27:21.375857 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 02 09:27:21 crc kubenswrapper[4764]: I0202 09:27:21.381247 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9dtjd\" (UniqueName: \"kubernetes.io/projected/6408cd2c-fa48-4c79-a6dd-558008245a0f-kube-api-access-9dtjd\") pod \"nova-cell0-cell-mapping-zfgz6\" (UID: \"6408cd2c-fa48-4c79-a6dd-558008245a0f\") " pod="openstack/nova-cell0-cell-mapping-zfgz6" Feb 02 09:27:21 crc kubenswrapper[4764]: I0202 09:27:21.441061 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ae215e9-dc8e-4274-873d-1641cd67c6c1-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"7ae215e9-dc8e-4274-873d-1641cd67c6c1\") " pod="openstack/nova-api-0" Feb 02 09:27:21 crc kubenswrapper[4764]: I0202 09:27:21.441109 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ae215e9-dc8e-4274-873d-1641cd67c6c1-config-data\") pod \"nova-api-0\" (UID: \"7ae215e9-dc8e-4274-873d-1641cd67c6c1\") " pod="openstack/nova-api-0" Feb 02 09:27:21 crc kubenswrapper[4764]: I0202 09:27:21.441234 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7ae215e9-dc8e-4274-873d-1641cd67c6c1-logs\") pod \"nova-api-0\" (UID: \"7ae215e9-dc8e-4274-873d-1641cd67c6c1\") " pod="openstack/nova-api-0" Feb 02 09:27:21 crc kubenswrapper[4764]: I0202 09:27:21.441262 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-trvls\" (UniqueName: \"kubernetes.io/projected/7ae215e9-dc8e-4274-873d-1641cd67c6c1-kube-api-access-trvls\") pod \"nova-api-0\" (UID: \"7ae215e9-dc8e-4274-873d-1641cd67c6c1\") " pod="openstack/nova-api-0" Feb 02 09:27:21 crc kubenswrapper[4764]: I0202 09:27:21.487406 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Feb 02 09:27:21 crc kubenswrapper[4764]: I0202 09:27:21.488440 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 02 09:27:21 crc kubenswrapper[4764]: I0202 09:27:21.494129 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Feb 02 09:27:21 crc kubenswrapper[4764]: I0202 09:27:21.511365 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-zfgz6" Feb 02 09:27:21 crc kubenswrapper[4764]: I0202 09:27:21.517724 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Feb 02 09:27:21 crc kubenswrapper[4764]: I0202 09:27:21.542562 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab56b2c6-5132-488a-bdde-b4bf3cd38b6d-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"ab56b2c6-5132-488a-bdde-b4bf3cd38b6d\") " pod="openstack/nova-scheduler-0" Feb 02 09:27:21 crc kubenswrapper[4764]: I0202 09:27:21.542628 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab56b2c6-5132-488a-bdde-b4bf3cd38b6d-config-data\") pod \"nova-scheduler-0\" (UID: \"ab56b2c6-5132-488a-bdde-b4bf3cd38b6d\") " pod="openstack/nova-scheduler-0" Feb 02 09:27:21 crc kubenswrapper[4764]: I0202 09:27:21.542669 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7ae215e9-dc8e-4274-873d-1641cd67c6c1-logs\") pod \"nova-api-0\" (UID: \"7ae215e9-dc8e-4274-873d-1641cd67c6c1\") " pod="openstack/nova-api-0" Feb 02 09:27:21 crc kubenswrapper[4764]: I0202 09:27:21.542707 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-trvls\" (UniqueName: \"kubernetes.io/projected/7ae215e9-dc8e-4274-873d-1641cd67c6c1-kube-api-access-trvls\") pod \"nova-api-0\" (UID: \"7ae215e9-dc8e-4274-873d-1641cd67c6c1\") " pod="openstack/nova-api-0" Feb 02 09:27:21 crc kubenswrapper[4764]: I0202 09:27:21.542731 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ae215e9-dc8e-4274-873d-1641cd67c6c1-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"7ae215e9-dc8e-4274-873d-1641cd67c6c1\") " pod="openstack/nova-api-0" Feb 02 09:27:21 crc kubenswrapper[4764]: I0202 09:27:21.542747 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ae215e9-dc8e-4274-873d-1641cd67c6c1-config-data\") pod \"nova-api-0\" (UID: \"7ae215e9-dc8e-4274-873d-1641cd67c6c1\") " pod="openstack/nova-api-0" Feb 02 09:27:21 crc kubenswrapper[4764]: I0202 09:27:21.542769 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4df6s\" (UniqueName: \"kubernetes.io/projected/ab56b2c6-5132-488a-bdde-b4bf3cd38b6d-kube-api-access-4df6s\") pod \"nova-scheduler-0\" (UID: \"ab56b2c6-5132-488a-bdde-b4bf3cd38b6d\") " pod="openstack/nova-scheduler-0" Feb 02 09:27:21 crc kubenswrapper[4764]: I0202 09:27:21.544527 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7ae215e9-dc8e-4274-873d-1641cd67c6c1-logs\") pod \"nova-api-0\" (UID: \"7ae215e9-dc8e-4274-873d-1641cd67c6c1\") " pod="openstack/nova-api-0" Feb 02 09:27:21 crc kubenswrapper[4764]: I0202 09:27:21.557660 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ae215e9-dc8e-4274-873d-1641cd67c6c1-config-data\") pod \"nova-api-0\" (UID: \"7ae215e9-dc8e-4274-873d-1641cd67c6c1\") " pod="openstack/nova-api-0" Feb 02 09:27:21 crc kubenswrapper[4764]: I0202 09:27:21.558770 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ae215e9-dc8e-4274-873d-1641cd67c6c1-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"7ae215e9-dc8e-4274-873d-1641cd67c6c1\") " pod="openstack/nova-api-0" Feb 02 09:27:21 crc kubenswrapper[4764]: I0202 09:27:21.582527 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-trvls\" (UniqueName: \"kubernetes.io/projected/7ae215e9-dc8e-4274-873d-1641cd67c6c1-kube-api-access-trvls\") pod \"nova-api-0\" (UID: \"7ae215e9-dc8e-4274-873d-1641cd67c6c1\") " pod="openstack/nova-api-0" Feb 02 09:27:21 crc kubenswrapper[4764]: I0202 09:27:21.643868 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4df6s\" (UniqueName: \"kubernetes.io/projected/ab56b2c6-5132-488a-bdde-b4bf3cd38b6d-kube-api-access-4df6s\") pod \"nova-scheduler-0\" (UID: \"ab56b2c6-5132-488a-bdde-b4bf3cd38b6d\") " pod="openstack/nova-scheduler-0" Feb 02 09:27:21 crc kubenswrapper[4764]: I0202 09:27:21.643980 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab56b2c6-5132-488a-bdde-b4bf3cd38b6d-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"ab56b2c6-5132-488a-bdde-b4bf3cd38b6d\") " pod="openstack/nova-scheduler-0" Feb 02 09:27:21 crc kubenswrapper[4764]: I0202 09:27:21.644023 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab56b2c6-5132-488a-bdde-b4bf3cd38b6d-config-data\") pod \"nova-scheduler-0\" (UID: \"ab56b2c6-5132-488a-bdde-b4bf3cd38b6d\") " pod="openstack/nova-scheduler-0" Feb 02 09:27:21 crc kubenswrapper[4764]: I0202 09:27:21.648242 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab56b2c6-5132-488a-bdde-b4bf3cd38b6d-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"ab56b2c6-5132-488a-bdde-b4bf3cd38b6d\") " pod="openstack/nova-scheduler-0" Feb 02 09:27:21 crc kubenswrapper[4764]: I0202 09:27:21.649347 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab56b2c6-5132-488a-bdde-b4bf3cd38b6d-config-data\") pod \"nova-scheduler-0\" (UID: \"ab56b2c6-5132-488a-bdde-b4bf3cd38b6d\") " pod="openstack/nova-scheduler-0" Feb 02 09:27:21 crc kubenswrapper[4764]: I0202 09:27:21.660695 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 02 09:27:21 crc kubenswrapper[4764]: I0202 09:27:21.661805 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Feb 02 09:27:21 crc kubenswrapper[4764]: I0202 09:27:21.665718 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Feb 02 09:27:21 crc kubenswrapper[4764]: I0202 09:27:21.692030 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4df6s\" (UniqueName: \"kubernetes.io/projected/ab56b2c6-5132-488a-bdde-b4bf3cd38b6d-kube-api-access-4df6s\") pod \"nova-scheduler-0\" (UID: \"ab56b2c6-5132-488a-bdde-b4bf3cd38b6d\") " pod="openstack/nova-scheduler-0" Feb 02 09:27:21 crc kubenswrapper[4764]: I0202 09:27:21.728383 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 02 09:27:21 crc kubenswrapper[4764]: I0202 09:27:21.746033 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51b710f2-c87f-4d7f-8eda-a3f091f39f1d-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"51b710f2-c87f-4d7f-8eda-a3f091f39f1d\") " pod="openstack/nova-cell1-novncproxy-0" Feb 02 09:27:21 crc kubenswrapper[4764]: I0202 09:27:21.746119 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8999j\" (UniqueName: \"kubernetes.io/projected/51b710f2-c87f-4d7f-8eda-a3f091f39f1d-kube-api-access-8999j\") pod \"nova-cell1-novncproxy-0\" (UID: \"51b710f2-c87f-4d7f-8eda-a3f091f39f1d\") " pod="openstack/nova-cell1-novncproxy-0" Feb 02 09:27:21 crc kubenswrapper[4764]: I0202 09:27:21.746176 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51b710f2-c87f-4d7f-8eda-a3f091f39f1d-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"51b710f2-c87f-4d7f-8eda-a3f091f39f1d\") " pod="openstack/nova-cell1-novncproxy-0" Feb 02 09:27:21 crc kubenswrapper[4764]: I0202 09:27:21.750831 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 02 09:27:21 crc kubenswrapper[4764]: I0202 09:27:21.777505 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Feb 02 09:27:21 crc kubenswrapper[4764]: I0202 09:27:21.779496 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 02 09:27:21 crc kubenswrapper[4764]: I0202 09:27:21.791456 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Feb 02 09:27:21 crc kubenswrapper[4764]: I0202 09:27:21.809394 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 02 09:27:21 crc kubenswrapper[4764]: I0202 09:27:21.847783 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51b710f2-c87f-4d7f-8eda-a3f091f39f1d-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"51b710f2-c87f-4d7f-8eda-a3f091f39f1d\") " pod="openstack/nova-cell1-novncproxy-0" Feb 02 09:27:21 crc kubenswrapper[4764]: I0202 09:27:21.847898 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mhvvf\" (UniqueName: \"kubernetes.io/projected/c5ef67bf-81bb-4b1c-8d9c-f78d7039bc2d-kube-api-access-mhvvf\") pod \"nova-metadata-0\" (UID: \"c5ef67bf-81bb-4b1c-8d9c-f78d7039bc2d\") " pod="openstack/nova-metadata-0" Feb 02 09:27:21 crc kubenswrapper[4764]: I0202 09:27:21.851301 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51b710f2-c87f-4d7f-8eda-a3f091f39f1d-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"51b710f2-c87f-4d7f-8eda-a3f091f39f1d\") " pod="openstack/nova-cell1-novncproxy-0" Feb 02 09:27:21 crc kubenswrapper[4764]: I0202 09:27:21.851771 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c5ef67bf-81bb-4b1c-8d9c-f78d7039bc2d-logs\") pod \"nova-metadata-0\" (UID: \"c5ef67bf-81bb-4b1c-8d9c-f78d7039bc2d\") " pod="openstack/nova-metadata-0" Feb 02 09:27:21 crc kubenswrapper[4764]: I0202 09:27:21.851857 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5ef67bf-81bb-4b1c-8d9c-f78d7039bc2d-config-data\") pod \"nova-metadata-0\" (UID: \"c5ef67bf-81bb-4b1c-8d9c-f78d7039bc2d\") " pod="openstack/nova-metadata-0" Feb 02 09:27:21 crc kubenswrapper[4764]: I0202 09:27:21.851961 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8999j\" (UniqueName: \"kubernetes.io/projected/51b710f2-c87f-4d7f-8eda-a3f091f39f1d-kube-api-access-8999j\") pod \"nova-cell1-novncproxy-0\" (UID: \"51b710f2-c87f-4d7f-8eda-a3f091f39f1d\") " pod="openstack/nova-cell1-novncproxy-0" Feb 02 09:27:21 crc kubenswrapper[4764]: I0202 09:27:21.852186 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5ef67bf-81bb-4b1c-8d9c-f78d7039bc2d-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"c5ef67bf-81bb-4b1c-8d9c-f78d7039bc2d\") " pod="openstack/nova-metadata-0" Feb 02 09:27:21 crc kubenswrapper[4764]: I0202 09:27:21.866533 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51b710f2-c87f-4d7f-8eda-a3f091f39f1d-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"51b710f2-c87f-4d7f-8eda-a3f091f39f1d\") " pod="openstack/nova-cell1-novncproxy-0" Feb 02 09:27:21 crc kubenswrapper[4764]: I0202 09:27:21.877783 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51b710f2-c87f-4d7f-8eda-a3f091f39f1d-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"51b710f2-c87f-4d7f-8eda-a3f091f39f1d\") " pod="openstack/nova-cell1-novncproxy-0" Feb 02 09:27:21 crc kubenswrapper[4764]: I0202 09:27:21.906972 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 02 09:27:21 crc kubenswrapper[4764]: I0202 09:27:21.920581 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8999j\" (UniqueName: \"kubernetes.io/projected/51b710f2-c87f-4d7f-8eda-a3f091f39f1d-kube-api-access-8999j\") pod \"nova-cell1-novncproxy-0\" (UID: \"51b710f2-c87f-4d7f-8eda-a3f091f39f1d\") " pod="openstack/nova-cell1-novncproxy-0" Feb 02 09:27:21 crc kubenswrapper[4764]: I0202 09:27:21.953998 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mhvvf\" (UniqueName: \"kubernetes.io/projected/c5ef67bf-81bb-4b1c-8d9c-f78d7039bc2d-kube-api-access-mhvvf\") pod \"nova-metadata-0\" (UID: \"c5ef67bf-81bb-4b1c-8d9c-f78d7039bc2d\") " pod="openstack/nova-metadata-0" Feb 02 09:27:21 crc kubenswrapper[4764]: I0202 09:27:21.954068 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c5ef67bf-81bb-4b1c-8d9c-f78d7039bc2d-logs\") pod \"nova-metadata-0\" (UID: \"c5ef67bf-81bb-4b1c-8d9c-f78d7039bc2d\") " pod="openstack/nova-metadata-0" Feb 02 09:27:21 crc kubenswrapper[4764]: I0202 09:27:21.954098 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5ef67bf-81bb-4b1c-8d9c-f78d7039bc2d-config-data\") pod \"nova-metadata-0\" (UID: \"c5ef67bf-81bb-4b1c-8d9c-f78d7039bc2d\") " pod="openstack/nova-metadata-0" Feb 02 09:27:21 crc kubenswrapper[4764]: I0202 09:27:21.954139 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5ef67bf-81bb-4b1c-8d9c-f78d7039bc2d-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"c5ef67bf-81bb-4b1c-8d9c-f78d7039bc2d\") " pod="openstack/nova-metadata-0" Feb 02 09:27:21 crc kubenswrapper[4764]: I0202 09:27:21.977059 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c5ef67bf-81bb-4b1c-8d9c-f78d7039bc2d-logs\") pod \"nova-metadata-0\" (UID: \"c5ef67bf-81bb-4b1c-8d9c-f78d7039bc2d\") " pod="openstack/nova-metadata-0" Feb 02 09:27:21 crc kubenswrapper[4764]: I0202 09:27:21.991850 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5ef67bf-81bb-4b1c-8d9c-f78d7039bc2d-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"c5ef67bf-81bb-4b1c-8d9c-f78d7039bc2d\") " pod="openstack/nova-metadata-0" Feb 02 09:27:22 crc kubenswrapper[4764]: I0202 09:27:21.999594 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5ef67bf-81bb-4b1c-8d9c-f78d7039bc2d-config-data\") pod \"nova-metadata-0\" (UID: \"c5ef67bf-81bb-4b1c-8d9c-f78d7039bc2d\") " pod="openstack/nova-metadata-0" Feb 02 09:27:22 crc kubenswrapper[4764]: I0202 09:27:22.005364 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Feb 02 09:27:22 crc kubenswrapper[4764]: I0202 09:27:22.058414 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-65b8795f57-z4kh6"] Feb 02 09:27:22 crc kubenswrapper[4764]: I0202 09:27:22.061736 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-65b8795f57-z4kh6" Feb 02 09:27:22 crc kubenswrapper[4764]: I0202 09:27:22.121419 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mhvvf\" (UniqueName: \"kubernetes.io/projected/c5ef67bf-81bb-4b1c-8d9c-f78d7039bc2d-kube-api-access-mhvvf\") pod \"nova-metadata-0\" (UID: \"c5ef67bf-81bb-4b1c-8d9c-f78d7039bc2d\") " pod="openstack/nova-metadata-0" Feb 02 09:27:22 crc kubenswrapper[4764]: I0202 09:27:22.123705 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-65b8795f57-z4kh6"] Feb 02 09:27:22 crc kubenswrapper[4764]: I0202 09:27:22.124127 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 02 09:27:22 crc kubenswrapper[4764]: I0202 09:27:22.186362 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/39a2a1fd-e483-46ec-807d-7a8b63f4d403-ovsdbserver-sb\") pod \"dnsmasq-dns-65b8795f57-z4kh6\" (UID: \"39a2a1fd-e483-46ec-807d-7a8b63f4d403\") " pod="openstack/dnsmasq-dns-65b8795f57-z4kh6" Feb 02 09:27:22 crc kubenswrapper[4764]: I0202 09:27:22.186607 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/39a2a1fd-e483-46ec-807d-7a8b63f4d403-dns-svc\") pod \"dnsmasq-dns-65b8795f57-z4kh6\" (UID: \"39a2a1fd-e483-46ec-807d-7a8b63f4d403\") " pod="openstack/dnsmasq-dns-65b8795f57-z4kh6" Feb 02 09:27:22 crc kubenswrapper[4764]: I0202 09:27:22.186626 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/39a2a1fd-e483-46ec-807d-7a8b63f4d403-config\") pod \"dnsmasq-dns-65b8795f57-z4kh6\" (UID: \"39a2a1fd-e483-46ec-807d-7a8b63f4d403\") " pod="openstack/dnsmasq-dns-65b8795f57-z4kh6" Feb 02 09:27:22 crc kubenswrapper[4764]: I0202 09:27:22.186655 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/39a2a1fd-e483-46ec-807d-7a8b63f4d403-ovsdbserver-nb\") pod \"dnsmasq-dns-65b8795f57-z4kh6\" (UID: \"39a2a1fd-e483-46ec-807d-7a8b63f4d403\") " pod="openstack/dnsmasq-dns-65b8795f57-z4kh6" Feb 02 09:27:22 crc kubenswrapper[4764]: I0202 09:27:22.186695 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2lhpm\" (UniqueName: \"kubernetes.io/projected/39a2a1fd-e483-46ec-807d-7a8b63f4d403-kube-api-access-2lhpm\") pod \"dnsmasq-dns-65b8795f57-z4kh6\" (UID: \"39a2a1fd-e483-46ec-807d-7a8b63f4d403\") " pod="openstack/dnsmasq-dns-65b8795f57-z4kh6" Feb 02 09:27:22 crc kubenswrapper[4764]: I0202 09:27:22.297388 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/39a2a1fd-e483-46ec-807d-7a8b63f4d403-ovsdbserver-sb\") pod \"dnsmasq-dns-65b8795f57-z4kh6\" (UID: \"39a2a1fd-e483-46ec-807d-7a8b63f4d403\") " pod="openstack/dnsmasq-dns-65b8795f57-z4kh6" Feb 02 09:27:22 crc kubenswrapper[4764]: I0202 09:27:22.297607 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/39a2a1fd-e483-46ec-807d-7a8b63f4d403-dns-svc\") pod \"dnsmasq-dns-65b8795f57-z4kh6\" (UID: \"39a2a1fd-e483-46ec-807d-7a8b63f4d403\") " pod="openstack/dnsmasq-dns-65b8795f57-z4kh6" Feb 02 09:27:22 crc kubenswrapper[4764]: I0202 09:27:22.297691 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/39a2a1fd-e483-46ec-807d-7a8b63f4d403-config\") pod \"dnsmasq-dns-65b8795f57-z4kh6\" (UID: \"39a2a1fd-e483-46ec-807d-7a8b63f4d403\") " pod="openstack/dnsmasq-dns-65b8795f57-z4kh6" Feb 02 09:27:22 crc kubenswrapper[4764]: I0202 09:27:22.297773 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/39a2a1fd-e483-46ec-807d-7a8b63f4d403-ovsdbserver-nb\") pod \"dnsmasq-dns-65b8795f57-z4kh6\" (UID: \"39a2a1fd-e483-46ec-807d-7a8b63f4d403\") " pod="openstack/dnsmasq-dns-65b8795f57-z4kh6" Feb 02 09:27:22 crc kubenswrapper[4764]: I0202 09:27:22.297878 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2lhpm\" (UniqueName: \"kubernetes.io/projected/39a2a1fd-e483-46ec-807d-7a8b63f4d403-kube-api-access-2lhpm\") pod \"dnsmasq-dns-65b8795f57-z4kh6\" (UID: \"39a2a1fd-e483-46ec-807d-7a8b63f4d403\") " pod="openstack/dnsmasq-dns-65b8795f57-z4kh6" Feb 02 09:27:22 crc kubenswrapper[4764]: I0202 09:27:22.299341 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/39a2a1fd-e483-46ec-807d-7a8b63f4d403-dns-svc\") pod \"dnsmasq-dns-65b8795f57-z4kh6\" (UID: \"39a2a1fd-e483-46ec-807d-7a8b63f4d403\") " pod="openstack/dnsmasq-dns-65b8795f57-z4kh6" Feb 02 09:27:22 crc kubenswrapper[4764]: I0202 09:27:22.299393 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/39a2a1fd-e483-46ec-807d-7a8b63f4d403-config\") pod \"dnsmasq-dns-65b8795f57-z4kh6\" (UID: \"39a2a1fd-e483-46ec-807d-7a8b63f4d403\") " pod="openstack/dnsmasq-dns-65b8795f57-z4kh6" Feb 02 09:27:22 crc kubenswrapper[4764]: I0202 09:27:22.299515 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/39a2a1fd-e483-46ec-807d-7a8b63f4d403-ovsdbserver-nb\") pod \"dnsmasq-dns-65b8795f57-z4kh6\" (UID: \"39a2a1fd-e483-46ec-807d-7a8b63f4d403\") " pod="openstack/dnsmasq-dns-65b8795f57-z4kh6" Feb 02 09:27:22 crc kubenswrapper[4764]: I0202 09:27:22.299876 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/39a2a1fd-e483-46ec-807d-7a8b63f4d403-ovsdbserver-sb\") pod \"dnsmasq-dns-65b8795f57-z4kh6\" (UID: \"39a2a1fd-e483-46ec-807d-7a8b63f4d403\") " pod="openstack/dnsmasq-dns-65b8795f57-z4kh6" Feb 02 09:27:22 crc kubenswrapper[4764]: I0202 09:27:22.326596 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2lhpm\" (UniqueName: \"kubernetes.io/projected/39a2a1fd-e483-46ec-807d-7a8b63f4d403-kube-api-access-2lhpm\") pod \"dnsmasq-dns-65b8795f57-z4kh6\" (UID: \"39a2a1fd-e483-46ec-807d-7a8b63f4d403\") " pod="openstack/dnsmasq-dns-65b8795f57-z4kh6" Feb 02 09:27:22 crc kubenswrapper[4764]: I0202 09:27:22.500608 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-65b8795f57-z4kh6" Feb 02 09:27:22 crc kubenswrapper[4764]: I0202 09:27:22.513847 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-zfgz6"] Feb 02 09:27:22 crc kubenswrapper[4764]: I0202 09:27:22.719316 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 02 09:27:22 crc kubenswrapper[4764]: I0202 09:27:22.775037 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 02 09:27:22 crc kubenswrapper[4764]: I0202 09:27:22.869739 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 02 09:27:22 crc kubenswrapper[4764]: W0202 09:27:22.878264 4764 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc5ef67bf_81bb_4b1c_8d9c_f78d7039bc2d.slice/crio-b62e3ac70835936cae628863e0bce7b076ff2cf5320bc6b6ad1bb9b13871bd1b WatchSource:0}: Error finding container b62e3ac70835936cae628863e0bce7b076ff2cf5320bc6b6ad1bb9b13871bd1b: Status 404 returned error can't find the container with id b62e3ac70835936cae628863e0bce7b076ff2cf5320bc6b6ad1bb9b13871bd1b Feb 02 09:27:22 crc kubenswrapper[4764]: I0202 09:27:22.911056 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Feb 02 09:27:22 crc kubenswrapper[4764]: W0202 09:27:22.918453 4764 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podab56b2c6_5132_488a_bdde_b4bf3cd38b6d.slice/crio-65e121a8e165a425ccad9bf86df6ca33926b1fa84b544811507623c15b912aed WatchSource:0}: Error finding container 65e121a8e165a425ccad9bf86df6ca33926b1fa84b544811507623c15b912aed: Status 404 returned error can't find the container with id 65e121a8e165a425ccad9bf86df6ca33926b1fa84b544811507623c15b912aed Feb 02 09:27:23 crc kubenswrapper[4764]: I0202 09:27:23.045535 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-9zzqm"] Feb 02 09:27:23 crc kubenswrapper[4764]: I0202 09:27:23.046541 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-9zzqm" Feb 02 09:27:23 crc kubenswrapper[4764]: I0202 09:27:23.048122 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Feb 02 09:27:23 crc kubenswrapper[4764]: I0202 09:27:23.050348 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Feb 02 09:27:23 crc kubenswrapper[4764]: I0202 09:27:23.064187 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-9zzqm"] Feb 02 09:27:23 crc kubenswrapper[4764]: I0202 09:27:23.108774 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-65b8795f57-z4kh6"] Feb 02 09:27:23 crc kubenswrapper[4764]: I0202 09:27:23.125068 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91e8d389-437f-4eed-b6b1-2b64802f9bc9-config-data\") pod \"nova-cell1-conductor-db-sync-9zzqm\" (UID: \"91e8d389-437f-4eed-b6b1-2b64802f9bc9\") " pod="openstack/nova-cell1-conductor-db-sync-9zzqm" Feb 02 09:27:23 crc kubenswrapper[4764]: I0202 09:27:23.125124 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j7hlw\" (UniqueName: \"kubernetes.io/projected/91e8d389-437f-4eed-b6b1-2b64802f9bc9-kube-api-access-j7hlw\") pod \"nova-cell1-conductor-db-sync-9zzqm\" (UID: \"91e8d389-437f-4eed-b6b1-2b64802f9bc9\") " pod="openstack/nova-cell1-conductor-db-sync-9zzqm" Feb 02 09:27:23 crc kubenswrapper[4764]: I0202 09:27:23.125162 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91e8d389-437f-4eed-b6b1-2b64802f9bc9-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-9zzqm\" (UID: \"91e8d389-437f-4eed-b6b1-2b64802f9bc9\") " pod="openstack/nova-cell1-conductor-db-sync-9zzqm" Feb 02 09:27:23 crc kubenswrapper[4764]: I0202 09:27:23.125186 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/91e8d389-437f-4eed-b6b1-2b64802f9bc9-scripts\") pod \"nova-cell1-conductor-db-sync-9zzqm\" (UID: \"91e8d389-437f-4eed-b6b1-2b64802f9bc9\") " pod="openstack/nova-cell1-conductor-db-sync-9zzqm" Feb 02 09:27:23 crc kubenswrapper[4764]: I0202 09:27:23.226587 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j7hlw\" (UniqueName: \"kubernetes.io/projected/91e8d389-437f-4eed-b6b1-2b64802f9bc9-kube-api-access-j7hlw\") pod \"nova-cell1-conductor-db-sync-9zzqm\" (UID: \"91e8d389-437f-4eed-b6b1-2b64802f9bc9\") " pod="openstack/nova-cell1-conductor-db-sync-9zzqm" Feb 02 09:27:23 crc kubenswrapper[4764]: I0202 09:27:23.226644 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91e8d389-437f-4eed-b6b1-2b64802f9bc9-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-9zzqm\" (UID: \"91e8d389-437f-4eed-b6b1-2b64802f9bc9\") " pod="openstack/nova-cell1-conductor-db-sync-9zzqm" Feb 02 09:27:23 crc kubenswrapper[4764]: I0202 09:27:23.226667 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/91e8d389-437f-4eed-b6b1-2b64802f9bc9-scripts\") pod \"nova-cell1-conductor-db-sync-9zzqm\" (UID: \"91e8d389-437f-4eed-b6b1-2b64802f9bc9\") " pod="openstack/nova-cell1-conductor-db-sync-9zzqm" Feb 02 09:27:23 crc kubenswrapper[4764]: I0202 09:27:23.226768 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91e8d389-437f-4eed-b6b1-2b64802f9bc9-config-data\") pod \"nova-cell1-conductor-db-sync-9zzqm\" (UID: \"91e8d389-437f-4eed-b6b1-2b64802f9bc9\") " pod="openstack/nova-cell1-conductor-db-sync-9zzqm" Feb 02 09:27:23 crc kubenswrapper[4764]: I0202 09:27:23.230836 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/91e8d389-437f-4eed-b6b1-2b64802f9bc9-scripts\") pod \"nova-cell1-conductor-db-sync-9zzqm\" (UID: \"91e8d389-437f-4eed-b6b1-2b64802f9bc9\") " pod="openstack/nova-cell1-conductor-db-sync-9zzqm" Feb 02 09:27:23 crc kubenswrapper[4764]: I0202 09:27:23.233221 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91e8d389-437f-4eed-b6b1-2b64802f9bc9-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-9zzqm\" (UID: \"91e8d389-437f-4eed-b6b1-2b64802f9bc9\") " pod="openstack/nova-cell1-conductor-db-sync-9zzqm" Feb 02 09:27:23 crc kubenswrapper[4764]: I0202 09:27:23.234002 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91e8d389-437f-4eed-b6b1-2b64802f9bc9-config-data\") pod \"nova-cell1-conductor-db-sync-9zzqm\" (UID: \"91e8d389-437f-4eed-b6b1-2b64802f9bc9\") " pod="openstack/nova-cell1-conductor-db-sync-9zzqm" Feb 02 09:27:23 crc kubenswrapper[4764]: I0202 09:27:23.246654 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j7hlw\" (UniqueName: \"kubernetes.io/projected/91e8d389-437f-4eed-b6b1-2b64802f9bc9-kube-api-access-j7hlw\") pod \"nova-cell1-conductor-db-sync-9zzqm\" (UID: \"91e8d389-437f-4eed-b6b1-2b64802f9bc9\") " pod="openstack/nova-cell1-conductor-db-sync-9zzqm" Feb 02 09:27:23 crc kubenswrapper[4764]: I0202 09:27:23.366132 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-9zzqm" Feb 02 09:27:23 crc kubenswrapper[4764]: I0202 09:27:23.405971 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7ae215e9-dc8e-4274-873d-1641cd67c6c1","Type":"ContainerStarted","Data":"cd1add0181318b42ce8be7f09a6f35238bed8d7ff7688f486b36f99f37598193"} Feb 02 09:27:23 crc kubenswrapper[4764]: I0202 09:27:23.421274 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-65b8795f57-z4kh6" event={"ID":"39a2a1fd-e483-46ec-807d-7a8b63f4d403","Type":"ContainerStarted","Data":"53dec1d4f771e36dacae86b0a9391c587b248da8664f3f08dd707ad478ced95d"} Feb 02 09:27:23 crc kubenswrapper[4764]: I0202 09:27:23.421317 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-65b8795f57-z4kh6" event={"ID":"39a2a1fd-e483-46ec-807d-7a8b63f4d403","Type":"ContainerStarted","Data":"3ecf47785943539ce4dde17602b85eabb4236aa1a8b1df77f23f50a2db37f417"} Feb 02 09:27:23 crc kubenswrapper[4764]: I0202 09:27:23.429242 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"ab56b2c6-5132-488a-bdde-b4bf3cd38b6d","Type":"ContainerStarted","Data":"65e121a8e165a425ccad9bf86df6ca33926b1fa84b544811507623c15b912aed"} Feb 02 09:27:23 crc kubenswrapper[4764]: I0202 09:27:23.430506 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c5ef67bf-81bb-4b1c-8d9c-f78d7039bc2d","Type":"ContainerStarted","Data":"b62e3ac70835936cae628863e0bce7b076ff2cf5320bc6b6ad1bb9b13871bd1b"} Feb 02 09:27:23 crc kubenswrapper[4764]: I0202 09:27:23.431690 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"51b710f2-c87f-4d7f-8eda-a3f091f39f1d","Type":"ContainerStarted","Data":"c84b66c7939231229aeb51b3e319c065cd437adcd46c7f28c5b2260ca2693737"} Feb 02 09:27:23 crc kubenswrapper[4764]: I0202 09:27:23.450019 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-zfgz6" event={"ID":"6408cd2c-fa48-4c79-a6dd-558008245a0f","Type":"ContainerStarted","Data":"0567ac860308c61ce0ead95bb217f119d788a67221475b28b7de06c2f202c35a"} Feb 02 09:27:23 crc kubenswrapper[4764]: I0202 09:27:23.450063 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-zfgz6" event={"ID":"6408cd2c-fa48-4c79-a6dd-558008245a0f","Type":"ContainerStarted","Data":"f4150bd6d4b45e25745599e54d91595b40d88919b2c5fb004faa7c241a9ad90e"} Feb 02 09:27:23 crc kubenswrapper[4764]: I0202 09:27:23.479797 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-zfgz6" podStartSLOduration=2.479771553 podStartE2EDuration="2.479771553s" podCreationTimestamp="2026-02-02 09:27:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:27:23.465689819 +0000 UTC m=+1206.399413907" watchObservedRunningTime="2026-02-02 09:27:23.479771553 +0000 UTC m=+1206.413495641" Feb 02 09:27:23 crc kubenswrapper[4764]: I0202 09:27:23.943601 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-9zzqm"] Feb 02 09:27:24 crc kubenswrapper[4764]: I0202 09:27:24.466732 4764 generic.go:334] "Generic (PLEG): container finished" podID="39a2a1fd-e483-46ec-807d-7a8b63f4d403" containerID="53dec1d4f771e36dacae86b0a9391c587b248da8664f3f08dd707ad478ced95d" exitCode=0 Feb 02 09:27:24 crc kubenswrapper[4764]: I0202 09:27:24.468058 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-65b8795f57-z4kh6" event={"ID":"39a2a1fd-e483-46ec-807d-7a8b63f4d403","Type":"ContainerDied","Data":"53dec1d4f771e36dacae86b0a9391c587b248da8664f3f08dd707ad478ced95d"} Feb 02 09:27:24 crc kubenswrapper[4764]: I0202 09:27:24.468107 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-65b8795f57-z4kh6" Feb 02 09:27:24 crc kubenswrapper[4764]: I0202 09:27:24.468119 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-65b8795f57-z4kh6" event={"ID":"39a2a1fd-e483-46ec-807d-7a8b63f4d403","Type":"ContainerStarted","Data":"a1b11461d74d94922fb1108c376a1ebcd01bd6511e487c221b592c97deda502d"} Feb 02 09:27:24 crc kubenswrapper[4764]: I0202 09:27:24.476900 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-9zzqm" event={"ID":"91e8d389-437f-4eed-b6b1-2b64802f9bc9","Type":"ContainerStarted","Data":"f26b51118aaf57ee0b36ac05fa18a18fb719c48c913ef9f25bad65b10713ba4a"} Feb 02 09:27:24 crc kubenswrapper[4764]: I0202 09:27:24.476928 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-9zzqm" event={"ID":"91e8d389-437f-4eed-b6b1-2b64802f9bc9","Type":"ContainerStarted","Data":"ca23690127dab495bbd267b79e78f32157ab0f7861187b6a7d6e4e5d1fef3363"} Feb 02 09:27:24 crc kubenswrapper[4764]: I0202 09:27:24.508258 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-65b8795f57-z4kh6" podStartSLOduration=3.508242171 podStartE2EDuration="3.508242171s" podCreationTimestamp="2026-02-02 09:27:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:27:24.49025552 +0000 UTC m=+1207.423979608" watchObservedRunningTime="2026-02-02 09:27:24.508242171 +0000 UTC m=+1207.441966259" Feb 02 09:27:25 crc kubenswrapper[4764]: I0202 09:27:25.380498 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-9zzqm" podStartSLOduration=2.380480981 podStartE2EDuration="2.380480981s" podCreationTimestamp="2026-02-02 09:27:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:27:24.516449234 +0000 UTC m=+1207.450173322" watchObservedRunningTime="2026-02-02 09:27:25.380480981 +0000 UTC m=+1208.314205059" Feb 02 09:27:25 crc kubenswrapper[4764]: I0202 09:27:25.381769 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Feb 02 09:27:25 crc kubenswrapper[4764]: I0202 09:27:25.400960 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 02 09:27:26 crc kubenswrapper[4764]: I0202 09:27:26.505808 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"ab56b2c6-5132-488a-bdde-b4bf3cd38b6d","Type":"ContainerStarted","Data":"b0f942aaab42d58433c36cefea216b2c0d65fcc55f664d23674f0b4d1b63671c"} Feb 02 09:27:26 crc kubenswrapper[4764]: I0202 09:27:26.507954 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c5ef67bf-81bb-4b1c-8d9c-f78d7039bc2d","Type":"ContainerStarted","Data":"6fa19856a6bfc97c13f585574269ad4991d26862353374772268d42bb73c334e"} Feb 02 09:27:26 crc kubenswrapper[4764]: I0202 09:27:26.509204 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"51b710f2-c87f-4d7f-8eda-a3f091f39f1d","Type":"ContainerStarted","Data":"7ec1f78b5f30dc980d60bb86d95602550ea60909e3f0b5adad9b61f7efadbcb0"} Feb 02 09:27:26 crc kubenswrapper[4764]: I0202 09:27:26.509314 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="51b710f2-c87f-4d7f-8eda-a3f091f39f1d" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://7ec1f78b5f30dc980d60bb86d95602550ea60909e3f0b5adad9b61f7efadbcb0" gracePeriod=30 Feb 02 09:27:26 crc kubenswrapper[4764]: I0202 09:27:26.516288 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7ae215e9-dc8e-4274-873d-1641cd67c6c1","Type":"ContainerStarted","Data":"2e8a49f04e50df3a7f183dd6086b65d007fe49e9d3efa236ddd67bf433ae7409"} Feb 02 09:27:26 crc kubenswrapper[4764]: I0202 09:27:26.530516 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.36707724 podStartE2EDuration="5.53050157s" podCreationTimestamp="2026-02-02 09:27:21 +0000 UTC" firstStartedPulling="2026-02-02 09:27:22.920754018 +0000 UTC m=+1205.854478106" lastFinishedPulling="2026-02-02 09:27:26.084178348 +0000 UTC m=+1209.017902436" observedRunningTime="2026-02-02 09:27:26.525175235 +0000 UTC m=+1209.458899313" watchObservedRunningTime="2026-02-02 09:27:26.53050157 +0000 UTC m=+1209.464225658" Feb 02 09:27:26 crc kubenswrapper[4764]: I0202 09:27:26.531430 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Feb 02 09:27:26 crc kubenswrapper[4764]: I0202 09:27:26.544488 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.245854078 podStartE2EDuration="5.544474181s" podCreationTimestamp="2026-02-02 09:27:21 +0000 UTC" firstStartedPulling="2026-02-02 09:27:22.785609456 +0000 UTC m=+1205.719333544" lastFinishedPulling="2026-02-02 09:27:26.084229559 +0000 UTC m=+1209.017953647" observedRunningTime="2026-02-02 09:27:26.537501651 +0000 UTC m=+1209.471225739" watchObservedRunningTime="2026-02-02 09:27:26.544474181 +0000 UTC m=+1209.478198269" Feb 02 09:27:26 crc kubenswrapper[4764]: I0202 09:27:26.810031 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Feb 02 09:27:27 crc kubenswrapper[4764]: I0202 09:27:27.006522 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Feb 02 09:27:27 crc kubenswrapper[4764]: I0202 09:27:27.530319 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7ae215e9-dc8e-4274-873d-1641cd67c6c1","Type":"ContainerStarted","Data":"fa1c95d637a3c5dfe63e9ee6d8afba0bf9bc07979fd8f0b0d00c7b5cbf36b0d6"} Feb 02 09:27:27 crc kubenswrapper[4764]: I0202 09:27:27.532856 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="c5ef67bf-81bb-4b1c-8d9c-f78d7039bc2d" containerName="nova-metadata-log" containerID="cri-o://6fa19856a6bfc97c13f585574269ad4991d26862353374772268d42bb73c334e" gracePeriod=30 Feb 02 09:27:27 crc kubenswrapper[4764]: I0202 09:27:27.533639 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="c5ef67bf-81bb-4b1c-8d9c-f78d7039bc2d" containerName="nova-metadata-metadata" containerID="cri-o://08480e622e3fb3194d62be0f6bd9155fa1709fdde2a64822d1a6bb57e302fa43" gracePeriod=30 Feb 02 09:27:27 crc kubenswrapper[4764]: I0202 09:27:27.533618 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c5ef67bf-81bb-4b1c-8d9c-f78d7039bc2d","Type":"ContainerStarted","Data":"08480e622e3fb3194d62be0f6bd9155fa1709fdde2a64822d1a6bb57e302fa43"} Feb 02 09:27:27 crc kubenswrapper[4764]: I0202 09:27:27.571705 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.201022847 podStartE2EDuration="6.571678704s" podCreationTimestamp="2026-02-02 09:27:21 +0000 UTC" firstStartedPulling="2026-02-02 09:27:22.716856072 +0000 UTC m=+1205.650580160" lastFinishedPulling="2026-02-02 09:27:26.087511909 +0000 UTC m=+1209.021236017" observedRunningTime="2026-02-02 09:27:27.560336955 +0000 UTC m=+1210.494061043" watchObservedRunningTime="2026-02-02 09:27:27.571678704 +0000 UTC m=+1210.505402832" Feb 02 09:27:27 crc kubenswrapper[4764]: I0202 09:27:27.605793 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.410444114 podStartE2EDuration="6.605771993s" podCreationTimestamp="2026-02-02 09:27:21 +0000 UTC" firstStartedPulling="2026-02-02 09:27:22.880714287 +0000 UTC m=+1205.814438375" lastFinishedPulling="2026-02-02 09:27:26.076042156 +0000 UTC m=+1209.009766254" observedRunningTime="2026-02-02 09:27:27.584663748 +0000 UTC m=+1210.518387846" watchObservedRunningTime="2026-02-02 09:27:27.605771993 +0000 UTC m=+1210.539496091" Feb 02 09:27:28 crc kubenswrapper[4764]: I0202 09:27:28.107350 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 02 09:27:28 crc kubenswrapper[4764]: I0202 09:27:28.138886 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mhvvf\" (UniqueName: \"kubernetes.io/projected/c5ef67bf-81bb-4b1c-8d9c-f78d7039bc2d-kube-api-access-mhvvf\") pod \"c5ef67bf-81bb-4b1c-8d9c-f78d7039bc2d\" (UID: \"c5ef67bf-81bb-4b1c-8d9c-f78d7039bc2d\") " Feb 02 09:27:28 crc kubenswrapper[4764]: I0202 09:27:28.138946 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5ef67bf-81bb-4b1c-8d9c-f78d7039bc2d-config-data\") pod \"c5ef67bf-81bb-4b1c-8d9c-f78d7039bc2d\" (UID: \"c5ef67bf-81bb-4b1c-8d9c-f78d7039bc2d\") " Feb 02 09:27:28 crc kubenswrapper[4764]: I0202 09:27:28.138972 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5ef67bf-81bb-4b1c-8d9c-f78d7039bc2d-combined-ca-bundle\") pod \"c5ef67bf-81bb-4b1c-8d9c-f78d7039bc2d\" (UID: \"c5ef67bf-81bb-4b1c-8d9c-f78d7039bc2d\") " Feb 02 09:27:28 crc kubenswrapper[4764]: I0202 09:27:28.139091 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c5ef67bf-81bb-4b1c-8d9c-f78d7039bc2d-logs\") pod \"c5ef67bf-81bb-4b1c-8d9c-f78d7039bc2d\" (UID: \"c5ef67bf-81bb-4b1c-8d9c-f78d7039bc2d\") " Feb 02 09:27:28 crc kubenswrapper[4764]: I0202 09:27:28.139778 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c5ef67bf-81bb-4b1c-8d9c-f78d7039bc2d-logs" (OuterVolumeSpecName: "logs") pod "c5ef67bf-81bb-4b1c-8d9c-f78d7039bc2d" (UID: "c5ef67bf-81bb-4b1c-8d9c-f78d7039bc2d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 09:27:28 crc kubenswrapper[4764]: I0202 09:27:28.174271 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c5ef67bf-81bb-4b1c-8d9c-f78d7039bc2d-kube-api-access-mhvvf" (OuterVolumeSpecName: "kube-api-access-mhvvf") pod "c5ef67bf-81bb-4b1c-8d9c-f78d7039bc2d" (UID: "c5ef67bf-81bb-4b1c-8d9c-f78d7039bc2d"). InnerVolumeSpecName "kube-api-access-mhvvf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:27:28 crc kubenswrapper[4764]: I0202 09:27:28.190804 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5ef67bf-81bb-4b1c-8d9c-f78d7039bc2d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c5ef67bf-81bb-4b1c-8d9c-f78d7039bc2d" (UID: "c5ef67bf-81bb-4b1c-8d9c-f78d7039bc2d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:27:28 crc kubenswrapper[4764]: I0202 09:27:28.238510 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5ef67bf-81bb-4b1c-8d9c-f78d7039bc2d-config-data" (OuterVolumeSpecName: "config-data") pod "c5ef67bf-81bb-4b1c-8d9c-f78d7039bc2d" (UID: "c5ef67bf-81bb-4b1c-8d9c-f78d7039bc2d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:27:28 crc kubenswrapper[4764]: I0202 09:27:28.240693 4764 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c5ef67bf-81bb-4b1c-8d9c-f78d7039bc2d-logs\") on node \"crc\" DevicePath \"\"" Feb 02 09:27:28 crc kubenswrapper[4764]: I0202 09:27:28.240836 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mhvvf\" (UniqueName: \"kubernetes.io/projected/c5ef67bf-81bb-4b1c-8d9c-f78d7039bc2d-kube-api-access-mhvvf\") on node \"crc\" DevicePath \"\"" Feb 02 09:27:28 crc kubenswrapper[4764]: I0202 09:27:28.240927 4764 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5ef67bf-81bb-4b1c-8d9c-f78d7039bc2d-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 09:27:28 crc kubenswrapper[4764]: I0202 09:27:28.241071 4764 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5ef67bf-81bb-4b1c-8d9c-f78d7039bc2d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 09:27:28 crc kubenswrapper[4764]: I0202 09:27:28.540013 4764 generic.go:334] "Generic (PLEG): container finished" podID="c5ef67bf-81bb-4b1c-8d9c-f78d7039bc2d" containerID="08480e622e3fb3194d62be0f6bd9155fa1709fdde2a64822d1a6bb57e302fa43" exitCode=0 Feb 02 09:27:28 crc kubenswrapper[4764]: I0202 09:27:28.540059 4764 generic.go:334] "Generic (PLEG): container finished" podID="c5ef67bf-81bb-4b1c-8d9c-f78d7039bc2d" containerID="6fa19856a6bfc97c13f585574269ad4991d26862353374772268d42bb73c334e" exitCode=143 Feb 02 09:27:28 crc kubenswrapper[4764]: I0202 09:27:28.541125 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 02 09:27:28 crc kubenswrapper[4764]: I0202 09:27:28.542110 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c5ef67bf-81bb-4b1c-8d9c-f78d7039bc2d","Type":"ContainerDied","Data":"08480e622e3fb3194d62be0f6bd9155fa1709fdde2a64822d1a6bb57e302fa43"} Feb 02 09:27:28 crc kubenswrapper[4764]: I0202 09:27:28.542161 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c5ef67bf-81bb-4b1c-8d9c-f78d7039bc2d","Type":"ContainerDied","Data":"6fa19856a6bfc97c13f585574269ad4991d26862353374772268d42bb73c334e"} Feb 02 09:27:28 crc kubenswrapper[4764]: I0202 09:27:28.542173 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c5ef67bf-81bb-4b1c-8d9c-f78d7039bc2d","Type":"ContainerDied","Data":"b62e3ac70835936cae628863e0bce7b076ff2cf5320bc6b6ad1bb9b13871bd1b"} Feb 02 09:27:28 crc kubenswrapper[4764]: I0202 09:27:28.542187 4764 scope.go:117] "RemoveContainer" containerID="08480e622e3fb3194d62be0f6bd9155fa1709fdde2a64822d1a6bb57e302fa43" Feb 02 09:27:28 crc kubenswrapper[4764]: I0202 09:27:28.585091 4764 scope.go:117] "RemoveContainer" containerID="6fa19856a6bfc97c13f585574269ad4991d26862353374772268d42bb73c334e" Feb 02 09:27:28 crc kubenswrapper[4764]: I0202 09:27:28.596699 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Feb 02 09:27:28 crc kubenswrapper[4764]: I0202 09:27:28.612662 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Feb 02 09:27:28 crc kubenswrapper[4764]: I0202 09:27:28.619689 4764 scope.go:117] "RemoveContainer" containerID="08480e622e3fb3194d62be0f6bd9155fa1709fdde2a64822d1a6bb57e302fa43" Feb 02 09:27:28 crc kubenswrapper[4764]: E0202 09:27:28.620288 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"08480e622e3fb3194d62be0f6bd9155fa1709fdde2a64822d1a6bb57e302fa43\": container with ID starting with 08480e622e3fb3194d62be0f6bd9155fa1709fdde2a64822d1a6bb57e302fa43 not found: ID does not exist" containerID="08480e622e3fb3194d62be0f6bd9155fa1709fdde2a64822d1a6bb57e302fa43" Feb 02 09:27:28 crc kubenswrapper[4764]: I0202 09:27:28.620317 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"08480e622e3fb3194d62be0f6bd9155fa1709fdde2a64822d1a6bb57e302fa43"} err="failed to get container status \"08480e622e3fb3194d62be0f6bd9155fa1709fdde2a64822d1a6bb57e302fa43\": rpc error: code = NotFound desc = could not find container \"08480e622e3fb3194d62be0f6bd9155fa1709fdde2a64822d1a6bb57e302fa43\": container with ID starting with 08480e622e3fb3194d62be0f6bd9155fa1709fdde2a64822d1a6bb57e302fa43 not found: ID does not exist" Feb 02 09:27:28 crc kubenswrapper[4764]: I0202 09:27:28.620339 4764 scope.go:117] "RemoveContainer" containerID="6fa19856a6bfc97c13f585574269ad4991d26862353374772268d42bb73c334e" Feb 02 09:27:28 crc kubenswrapper[4764]: E0202 09:27:28.622123 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6fa19856a6bfc97c13f585574269ad4991d26862353374772268d42bb73c334e\": container with ID starting with 6fa19856a6bfc97c13f585574269ad4991d26862353374772268d42bb73c334e not found: ID does not exist" containerID="6fa19856a6bfc97c13f585574269ad4991d26862353374772268d42bb73c334e" Feb 02 09:27:28 crc kubenswrapper[4764]: I0202 09:27:28.622177 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6fa19856a6bfc97c13f585574269ad4991d26862353374772268d42bb73c334e"} err="failed to get container status \"6fa19856a6bfc97c13f585574269ad4991d26862353374772268d42bb73c334e\": rpc error: code = NotFound desc = could not find container \"6fa19856a6bfc97c13f585574269ad4991d26862353374772268d42bb73c334e\": container with ID starting with 6fa19856a6bfc97c13f585574269ad4991d26862353374772268d42bb73c334e not found: ID does not exist" Feb 02 09:27:28 crc kubenswrapper[4764]: I0202 09:27:28.622202 4764 scope.go:117] "RemoveContainer" containerID="08480e622e3fb3194d62be0f6bd9155fa1709fdde2a64822d1a6bb57e302fa43" Feb 02 09:27:28 crc kubenswrapper[4764]: I0202 09:27:28.622656 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"08480e622e3fb3194d62be0f6bd9155fa1709fdde2a64822d1a6bb57e302fa43"} err="failed to get container status \"08480e622e3fb3194d62be0f6bd9155fa1709fdde2a64822d1a6bb57e302fa43\": rpc error: code = NotFound desc = could not find container \"08480e622e3fb3194d62be0f6bd9155fa1709fdde2a64822d1a6bb57e302fa43\": container with ID starting with 08480e622e3fb3194d62be0f6bd9155fa1709fdde2a64822d1a6bb57e302fa43 not found: ID does not exist" Feb 02 09:27:28 crc kubenswrapper[4764]: I0202 09:27:28.622691 4764 scope.go:117] "RemoveContainer" containerID="6fa19856a6bfc97c13f585574269ad4991d26862353374772268d42bb73c334e" Feb 02 09:27:28 crc kubenswrapper[4764]: I0202 09:27:28.623092 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6fa19856a6bfc97c13f585574269ad4991d26862353374772268d42bb73c334e"} err="failed to get container status \"6fa19856a6bfc97c13f585574269ad4991d26862353374772268d42bb73c334e\": rpc error: code = NotFound desc = could not find container \"6fa19856a6bfc97c13f585574269ad4991d26862353374772268d42bb73c334e\": container with ID starting with 6fa19856a6bfc97c13f585574269ad4991d26862353374772268d42bb73c334e not found: ID does not exist" Feb 02 09:27:28 crc kubenswrapper[4764]: I0202 09:27:28.625034 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Feb 02 09:27:28 crc kubenswrapper[4764]: E0202 09:27:28.625404 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5ef67bf-81bb-4b1c-8d9c-f78d7039bc2d" containerName="nova-metadata-metadata" Feb 02 09:27:28 crc kubenswrapper[4764]: I0202 09:27:28.625421 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5ef67bf-81bb-4b1c-8d9c-f78d7039bc2d" containerName="nova-metadata-metadata" Feb 02 09:27:28 crc kubenswrapper[4764]: E0202 09:27:28.625439 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5ef67bf-81bb-4b1c-8d9c-f78d7039bc2d" containerName="nova-metadata-log" Feb 02 09:27:28 crc kubenswrapper[4764]: I0202 09:27:28.625445 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5ef67bf-81bb-4b1c-8d9c-f78d7039bc2d" containerName="nova-metadata-log" Feb 02 09:27:28 crc kubenswrapper[4764]: I0202 09:27:28.625643 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="c5ef67bf-81bb-4b1c-8d9c-f78d7039bc2d" containerName="nova-metadata-log" Feb 02 09:27:28 crc kubenswrapper[4764]: I0202 09:27:28.625688 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="c5ef67bf-81bb-4b1c-8d9c-f78d7039bc2d" containerName="nova-metadata-metadata" Feb 02 09:27:28 crc kubenswrapper[4764]: I0202 09:27:28.626927 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 02 09:27:28 crc kubenswrapper[4764]: I0202 09:27:28.637011 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Feb 02 09:27:28 crc kubenswrapper[4764]: I0202 09:27:28.637249 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Feb 02 09:27:28 crc kubenswrapper[4764]: I0202 09:27:28.648201 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 02 09:27:28 crc kubenswrapper[4764]: I0202 09:27:28.751538 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/df690cc6-0a88-44db-8ee5-2e967e51fa59-logs\") pod \"nova-metadata-0\" (UID: \"df690cc6-0a88-44db-8ee5-2e967e51fa59\") " pod="openstack/nova-metadata-0" Feb 02 09:27:28 crc kubenswrapper[4764]: I0202 09:27:28.751608 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/df690cc6-0a88-44db-8ee5-2e967e51fa59-config-data\") pod \"nova-metadata-0\" (UID: \"df690cc6-0a88-44db-8ee5-2e967e51fa59\") " pod="openstack/nova-metadata-0" Feb 02 09:27:28 crc kubenswrapper[4764]: I0202 09:27:28.751721 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/df690cc6-0a88-44db-8ee5-2e967e51fa59-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"df690cc6-0a88-44db-8ee5-2e967e51fa59\") " pod="openstack/nova-metadata-0" Feb 02 09:27:28 crc kubenswrapper[4764]: I0202 09:27:28.751765 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rhj77\" (UniqueName: \"kubernetes.io/projected/df690cc6-0a88-44db-8ee5-2e967e51fa59-kube-api-access-rhj77\") pod \"nova-metadata-0\" (UID: \"df690cc6-0a88-44db-8ee5-2e967e51fa59\") " pod="openstack/nova-metadata-0" Feb 02 09:27:28 crc kubenswrapper[4764]: I0202 09:27:28.752081 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df690cc6-0a88-44db-8ee5-2e967e51fa59-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"df690cc6-0a88-44db-8ee5-2e967e51fa59\") " pod="openstack/nova-metadata-0" Feb 02 09:27:28 crc kubenswrapper[4764]: I0202 09:27:28.853611 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/df690cc6-0a88-44db-8ee5-2e967e51fa59-logs\") pod \"nova-metadata-0\" (UID: \"df690cc6-0a88-44db-8ee5-2e967e51fa59\") " pod="openstack/nova-metadata-0" Feb 02 09:27:28 crc kubenswrapper[4764]: I0202 09:27:28.853991 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/df690cc6-0a88-44db-8ee5-2e967e51fa59-config-data\") pod \"nova-metadata-0\" (UID: \"df690cc6-0a88-44db-8ee5-2e967e51fa59\") " pod="openstack/nova-metadata-0" Feb 02 09:27:28 crc kubenswrapper[4764]: I0202 09:27:28.854045 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/df690cc6-0a88-44db-8ee5-2e967e51fa59-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"df690cc6-0a88-44db-8ee5-2e967e51fa59\") " pod="openstack/nova-metadata-0" Feb 02 09:27:28 crc kubenswrapper[4764]: I0202 09:27:28.854066 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rhj77\" (UniqueName: \"kubernetes.io/projected/df690cc6-0a88-44db-8ee5-2e967e51fa59-kube-api-access-rhj77\") pod \"nova-metadata-0\" (UID: \"df690cc6-0a88-44db-8ee5-2e967e51fa59\") " pod="openstack/nova-metadata-0" Feb 02 09:27:28 crc kubenswrapper[4764]: I0202 09:27:28.854162 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df690cc6-0a88-44db-8ee5-2e967e51fa59-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"df690cc6-0a88-44db-8ee5-2e967e51fa59\") " pod="openstack/nova-metadata-0" Feb 02 09:27:28 crc kubenswrapper[4764]: I0202 09:27:28.854605 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/df690cc6-0a88-44db-8ee5-2e967e51fa59-logs\") pod \"nova-metadata-0\" (UID: \"df690cc6-0a88-44db-8ee5-2e967e51fa59\") " pod="openstack/nova-metadata-0" Feb 02 09:27:28 crc kubenswrapper[4764]: I0202 09:27:28.857341 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/df690cc6-0a88-44db-8ee5-2e967e51fa59-config-data\") pod \"nova-metadata-0\" (UID: \"df690cc6-0a88-44db-8ee5-2e967e51fa59\") " pod="openstack/nova-metadata-0" Feb 02 09:27:28 crc kubenswrapper[4764]: I0202 09:27:28.857675 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/df690cc6-0a88-44db-8ee5-2e967e51fa59-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"df690cc6-0a88-44db-8ee5-2e967e51fa59\") " pod="openstack/nova-metadata-0" Feb 02 09:27:28 crc kubenswrapper[4764]: I0202 09:27:28.858579 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df690cc6-0a88-44db-8ee5-2e967e51fa59-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"df690cc6-0a88-44db-8ee5-2e967e51fa59\") " pod="openstack/nova-metadata-0" Feb 02 09:27:28 crc kubenswrapper[4764]: I0202 09:27:28.872835 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rhj77\" (UniqueName: \"kubernetes.io/projected/df690cc6-0a88-44db-8ee5-2e967e51fa59-kube-api-access-rhj77\") pod \"nova-metadata-0\" (UID: \"df690cc6-0a88-44db-8ee5-2e967e51fa59\") " pod="openstack/nova-metadata-0" Feb 02 09:27:28 crc kubenswrapper[4764]: I0202 09:27:28.950867 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 02 09:27:29 crc kubenswrapper[4764]: I0202 09:27:29.383234 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 02 09:27:29 crc kubenswrapper[4764]: I0202 09:27:29.383435 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="26d7a0b3-ea1e-417b-b417-69ae84bd5d24" containerName="kube-state-metrics" containerID="cri-o://1cb1bd29c445477024eea879da1cceda1244dd5e3b7e54856e664ed0cfe5ec68" gracePeriod=30 Feb 02 09:27:29 crc kubenswrapper[4764]: I0202 09:27:29.496990 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 02 09:27:29 crc kubenswrapper[4764]: I0202 09:27:29.568218 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"df690cc6-0a88-44db-8ee5-2e967e51fa59","Type":"ContainerStarted","Data":"99cfde25138a8798419ae8d9b93eec6d84a6be236156fd1c75437ac9be189a4f"} Feb 02 09:27:29 crc kubenswrapper[4764]: I0202 09:27:29.576440 4764 generic.go:334] "Generic (PLEG): container finished" podID="26d7a0b3-ea1e-417b-b417-69ae84bd5d24" containerID="1cb1bd29c445477024eea879da1cceda1244dd5e3b7e54856e664ed0cfe5ec68" exitCode=2 Feb 02 09:27:29 crc kubenswrapper[4764]: I0202 09:27:29.576484 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"26d7a0b3-ea1e-417b-b417-69ae84bd5d24","Type":"ContainerDied","Data":"1cb1bd29c445477024eea879da1cceda1244dd5e3b7e54856e664ed0cfe5ec68"} Feb 02 09:27:29 crc kubenswrapper[4764]: I0202 09:27:29.835531 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c5ef67bf-81bb-4b1c-8d9c-f78d7039bc2d" path="/var/lib/kubelet/pods/c5ef67bf-81bb-4b1c-8d9c-f78d7039bc2d/volumes" Feb 02 09:27:30 crc kubenswrapper[4764]: I0202 09:27:30.425138 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Feb 02 09:27:30 crc kubenswrapper[4764]: I0202 09:27:30.482824 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-98v7s\" (UniqueName: \"kubernetes.io/projected/26d7a0b3-ea1e-417b-b417-69ae84bd5d24-kube-api-access-98v7s\") pod \"26d7a0b3-ea1e-417b-b417-69ae84bd5d24\" (UID: \"26d7a0b3-ea1e-417b-b417-69ae84bd5d24\") " Feb 02 09:27:30 crc kubenswrapper[4764]: I0202 09:27:30.488693 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/26d7a0b3-ea1e-417b-b417-69ae84bd5d24-kube-api-access-98v7s" (OuterVolumeSpecName: "kube-api-access-98v7s") pod "26d7a0b3-ea1e-417b-b417-69ae84bd5d24" (UID: "26d7a0b3-ea1e-417b-b417-69ae84bd5d24"). InnerVolumeSpecName "kube-api-access-98v7s". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:27:30 crc kubenswrapper[4764]: I0202 09:27:30.589524 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-98v7s\" (UniqueName: \"kubernetes.io/projected/26d7a0b3-ea1e-417b-b417-69ae84bd5d24-kube-api-access-98v7s\") on node \"crc\" DevicePath \"\"" Feb 02 09:27:30 crc kubenswrapper[4764]: I0202 09:27:30.598258 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"df690cc6-0a88-44db-8ee5-2e967e51fa59","Type":"ContainerStarted","Data":"7f8ca67a7e4d45d98682b4584bb5c5eb834a5c546cc932ac81924f956a9e8f0e"} Feb 02 09:27:30 crc kubenswrapper[4764]: I0202 09:27:30.598317 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"df690cc6-0a88-44db-8ee5-2e967e51fa59","Type":"ContainerStarted","Data":"0a9a3f3ea0066a3c068850c336c289ba30f34f30340cbb09b8b3c548baca6b36"} Feb 02 09:27:30 crc kubenswrapper[4764]: I0202 09:27:30.600375 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"26d7a0b3-ea1e-417b-b417-69ae84bd5d24","Type":"ContainerDied","Data":"d8af482c5008789457daff07c4f5a0cf5cfc3e09895ed5dd3c018b07a566b275"} Feb 02 09:27:30 crc kubenswrapper[4764]: I0202 09:27:30.600441 4764 scope.go:117] "RemoveContainer" containerID="1cb1bd29c445477024eea879da1cceda1244dd5e3b7e54856e664ed0cfe5ec68" Feb 02 09:27:30 crc kubenswrapper[4764]: I0202 09:27:30.600523 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Feb 02 09:27:30 crc kubenswrapper[4764]: I0202 09:27:30.628034 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.628014974 podStartE2EDuration="2.628014974s" podCreationTimestamp="2026-02-02 09:27:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:27:30.616555642 +0000 UTC m=+1213.550279730" watchObservedRunningTime="2026-02-02 09:27:30.628014974 +0000 UTC m=+1213.561739062" Feb 02 09:27:30 crc kubenswrapper[4764]: I0202 09:27:30.668239 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 02 09:27:30 crc kubenswrapper[4764]: I0202 09:27:30.680055 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 02 09:27:30 crc kubenswrapper[4764]: I0202 09:27:30.697862 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Feb 02 09:27:30 crc kubenswrapper[4764]: E0202 09:27:30.698508 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26d7a0b3-ea1e-417b-b417-69ae84bd5d24" containerName="kube-state-metrics" Feb 02 09:27:30 crc kubenswrapper[4764]: I0202 09:27:30.698532 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="26d7a0b3-ea1e-417b-b417-69ae84bd5d24" containerName="kube-state-metrics" Feb 02 09:27:30 crc kubenswrapper[4764]: I0202 09:27:30.698842 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="26d7a0b3-ea1e-417b-b417-69ae84bd5d24" containerName="kube-state-metrics" Feb 02 09:27:30 crc kubenswrapper[4764]: I0202 09:27:30.705733 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Feb 02 09:27:30 crc kubenswrapper[4764]: I0202 09:27:30.708710 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 02 09:27:30 crc kubenswrapper[4764]: I0202 09:27:30.711169 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Feb 02 09:27:30 crc kubenswrapper[4764]: I0202 09:27:30.711388 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Feb 02 09:27:30 crc kubenswrapper[4764]: I0202 09:27:30.782624 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 02 09:27:30 crc kubenswrapper[4764]: I0202 09:27:30.782915 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3cad43ae-e284-446f-9dbc-ba224b11aac4" containerName="ceilometer-central-agent" containerID="cri-o://f0e10976314a43245db7f1bfe6ae5503a29b455557f59bd621d463a0219b2bb4" gracePeriod=30 Feb 02 09:27:30 crc kubenswrapper[4764]: I0202 09:27:30.783030 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3cad43ae-e284-446f-9dbc-ba224b11aac4" containerName="sg-core" containerID="cri-o://ec37269bba1eeedbf80feeb128bd400885ce3747aa038d326f942e91b66f9cc1" gracePeriod=30 Feb 02 09:27:30 crc kubenswrapper[4764]: I0202 09:27:30.783048 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3cad43ae-e284-446f-9dbc-ba224b11aac4" containerName="ceilometer-notification-agent" containerID="cri-o://52d9f8634f508399e6f41017e512c93e0cc403c3b39ea0c39e366455b8ee787c" gracePeriod=30 Feb 02 09:27:30 crc kubenswrapper[4764]: I0202 09:27:30.783146 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3cad43ae-e284-446f-9dbc-ba224b11aac4" containerName="proxy-httpd" containerID="cri-o://9ba9869e46c55b3868c63cd7c0e91be1c9eec3f6b90f2c1a868d1065a651680a" gracePeriod=30 Feb 02 09:27:30 crc kubenswrapper[4764]: I0202 09:27:30.797580 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/2fce6f0b-cb36-4ff1-ab73-35c40d709502-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"2fce6f0b-cb36-4ff1-ab73-35c40d709502\") " pod="openstack/kube-state-metrics-0" Feb 02 09:27:30 crc kubenswrapper[4764]: I0202 09:27:30.797674 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/2fce6f0b-cb36-4ff1-ab73-35c40d709502-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"2fce6f0b-cb36-4ff1-ab73-35c40d709502\") " pod="openstack/kube-state-metrics-0" Feb 02 09:27:30 crc kubenswrapper[4764]: I0202 09:27:30.797713 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tkj62\" (UniqueName: \"kubernetes.io/projected/2fce6f0b-cb36-4ff1-ab73-35c40d709502-kube-api-access-tkj62\") pod \"kube-state-metrics-0\" (UID: \"2fce6f0b-cb36-4ff1-ab73-35c40d709502\") " pod="openstack/kube-state-metrics-0" Feb 02 09:27:30 crc kubenswrapper[4764]: I0202 09:27:30.797803 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2fce6f0b-cb36-4ff1-ab73-35c40d709502-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"2fce6f0b-cb36-4ff1-ab73-35c40d709502\") " pod="openstack/kube-state-metrics-0" Feb 02 09:27:30 crc kubenswrapper[4764]: I0202 09:27:30.899823 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2fce6f0b-cb36-4ff1-ab73-35c40d709502-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"2fce6f0b-cb36-4ff1-ab73-35c40d709502\") " pod="openstack/kube-state-metrics-0" Feb 02 09:27:30 crc kubenswrapper[4764]: I0202 09:27:30.900301 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/2fce6f0b-cb36-4ff1-ab73-35c40d709502-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"2fce6f0b-cb36-4ff1-ab73-35c40d709502\") " pod="openstack/kube-state-metrics-0" Feb 02 09:27:30 crc kubenswrapper[4764]: I0202 09:27:30.900357 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/2fce6f0b-cb36-4ff1-ab73-35c40d709502-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"2fce6f0b-cb36-4ff1-ab73-35c40d709502\") " pod="openstack/kube-state-metrics-0" Feb 02 09:27:30 crc kubenswrapper[4764]: I0202 09:27:30.900377 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tkj62\" (UniqueName: \"kubernetes.io/projected/2fce6f0b-cb36-4ff1-ab73-35c40d709502-kube-api-access-tkj62\") pod \"kube-state-metrics-0\" (UID: \"2fce6f0b-cb36-4ff1-ab73-35c40d709502\") " pod="openstack/kube-state-metrics-0" Feb 02 09:27:30 crc kubenswrapper[4764]: I0202 09:27:30.905465 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/2fce6f0b-cb36-4ff1-ab73-35c40d709502-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"2fce6f0b-cb36-4ff1-ab73-35c40d709502\") " pod="openstack/kube-state-metrics-0" Feb 02 09:27:30 crc kubenswrapper[4764]: I0202 09:27:30.908203 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/2fce6f0b-cb36-4ff1-ab73-35c40d709502-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"2fce6f0b-cb36-4ff1-ab73-35c40d709502\") " pod="openstack/kube-state-metrics-0" Feb 02 09:27:30 crc kubenswrapper[4764]: I0202 09:27:30.908329 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2fce6f0b-cb36-4ff1-ab73-35c40d709502-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"2fce6f0b-cb36-4ff1-ab73-35c40d709502\") " pod="openstack/kube-state-metrics-0" Feb 02 09:27:30 crc kubenswrapper[4764]: I0202 09:27:30.920326 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tkj62\" (UniqueName: \"kubernetes.io/projected/2fce6f0b-cb36-4ff1-ab73-35c40d709502-kube-api-access-tkj62\") pod \"kube-state-metrics-0\" (UID: \"2fce6f0b-cb36-4ff1-ab73-35c40d709502\") " pod="openstack/kube-state-metrics-0" Feb 02 09:27:31 crc kubenswrapper[4764]: I0202 09:27:31.034040 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Feb 02 09:27:31 crc kubenswrapper[4764]: I0202 09:27:31.593233 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 02 09:27:31 crc kubenswrapper[4764]: I0202 09:27:31.620640 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"2fce6f0b-cb36-4ff1-ab73-35c40d709502","Type":"ContainerStarted","Data":"0f0b92e098d2ac98663f22c105470732c9e17fb2be9adaffbdeaefc2adbf3729"} Feb 02 09:27:31 crc kubenswrapper[4764]: I0202 09:27:31.622881 4764 generic.go:334] "Generic (PLEG): container finished" podID="6408cd2c-fa48-4c79-a6dd-558008245a0f" containerID="0567ac860308c61ce0ead95bb217f119d788a67221475b28b7de06c2f202c35a" exitCode=0 Feb 02 09:27:31 crc kubenswrapper[4764]: I0202 09:27:31.622953 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-zfgz6" event={"ID":"6408cd2c-fa48-4c79-a6dd-558008245a0f","Type":"ContainerDied","Data":"0567ac860308c61ce0ead95bb217f119d788a67221475b28b7de06c2f202c35a"} Feb 02 09:27:31 crc kubenswrapper[4764]: I0202 09:27:31.630188 4764 generic.go:334] "Generic (PLEG): container finished" podID="3cad43ae-e284-446f-9dbc-ba224b11aac4" containerID="9ba9869e46c55b3868c63cd7c0e91be1c9eec3f6b90f2c1a868d1065a651680a" exitCode=0 Feb 02 09:27:31 crc kubenswrapper[4764]: I0202 09:27:31.630230 4764 generic.go:334] "Generic (PLEG): container finished" podID="3cad43ae-e284-446f-9dbc-ba224b11aac4" containerID="ec37269bba1eeedbf80feeb128bd400885ce3747aa038d326f942e91b66f9cc1" exitCode=2 Feb 02 09:27:31 crc kubenswrapper[4764]: I0202 09:27:31.630240 4764 generic.go:334] "Generic (PLEG): container finished" podID="3cad43ae-e284-446f-9dbc-ba224b11aac4" containerID="f0e10976314a43245db7f1bfe6ae5503a29b455557f59bd621d463a0219b2bb4" exitCode=0 Feb 02 09:27:31 crc kubenswrapper[4764]: I0202 09:27:31.630264 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3cad43ae-e284-446f-9dbc-ba224b11aac4","Type":"ContainerDied","Data":"9ba9869e46c55b3868c63cd7c0e91be1c9eec3f6b90f2c1a868d1065a651680a"} Feb 02 09:27:31 crc kubenswrapper[4764]: I0202 09:27:31.630291 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3cad43ae-e284-446f-9dbc-ba224b11aac4","Type":"ContainerDied","Data":"ec37269bba1eeedbf80feeb128bd400885ce3747aa038d326f942e91b66f9cc1"} Feb 02 09:27:31 crc kubenswrapper[4764]: I0202 09:27:31.630303 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3cad43ae-e284-446f-9dbc-ba224b11aac4","Type":"ContainerDied","Data":"f0e10976314a43245db7f1bfe6ae5503a29b455557f59bd621d463a0219b2bb4"} Feb 02 09:27:31 crc kubenswrapper[4764]: I0202 09:27:31.752142 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 02 09:27:31 crc kubenswrapper[4764]: I0202 09:27:31.752203 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 02 09:27:31 crc kubenswrapper[4764]: I0202 09:27:31.809668 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Feb 02 09:27:31 crc kubenswrapper[4764]: I0202 09:27:31.837857 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="26d7a0b3-ea1e-417b-b417-69ae84bd5d24" path="/var/lib/kubelet/pods/26d7a0b3-ea1e-417b-b417-69ae84bd5d24/volumes" Feb 02 09:27:31 crc kubenswrapper[4764]: I0202 09:27:31.838436 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Feb 02 09:27:32 crc kubenswrapper[4764]: I0202 09:27:32.503149 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-65b8795f57-z4kh6" Feb 02 09:27:32 crc kubenswrapper[4764]: I0202 09:27:32.619205 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-555cfdffc9-rw6ht"] Feb 02 09:27:32 crc kubenswrapper[4764]: I0202 09:27:32.619429 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-555cfdffc9-rw6ht" podUID="2c66c255-9454-496c-baf3-5c331d5a2f16" containerName="dnsmasq-dns" containerID="cri-o://5ce5a71993fab6383785294c33c51d3ada6114b29ce3fe0de247942ba4acefce" gracePeriod=10 Feb 02 09:27:32 crc kubenswrapper[4764]: I0202 09:27:32.648528 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"2fce6f0b-cb36-4ff1-ab73-35c40d709502","Type":"ContainerStarted","Data":"fe9a9d1d913670cadf2ff089fa89e30618e99ccb7eee21765d62e2181e5967c2"} Feb 02 09:27:32 crc kubenswrapper[4764]: I0202 09:27:32.650820 4764 generic.go:334] "Generic (PLEG): container finished" podID="91e8d389-437f-4eed-b6b1-2b64802f9bc9" containerID="f26b51118aaf57ee0b36ac05fa18a18fb719c48c913ef9f25bad65b10713ba4a" exitCode=0 Feb 02 09:27:32 crc kubenswrapper[4764]: I0202 09:27:32.650856 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-9zzqm" event={"ID":"91e8d389-437f-4eed-b6b1-2b64802f9bc9","Type":"ContainerDied","Data":"f26b51118aaf57ee0b36ac05fa18a18fb719c48c913ef9f25bad65b10713ba4a"} Feb 02 09:27:32 crc kubenswrapper[4764]: I0202 09:27:32.684651 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.335408084 podStartE2EDuration="2.684635241s" podCreationTimestamp="2026-02-02 09:27:30 +0000 UTC" firstStartedPulling="2026-02-02 09:27:31.605082161 +0000 UTC m=+1214.538806249" lastFinishedPulling="2026-02-02 09:27:31.954309318 +0000 UTC m=+1214.888033406" observedRunningTime="2026-02-02 09:27:32.678470643 +0000 UTC m=+1215.612194731" watchObservedRunningTime="2026-02-02 09:27:32.684635241 +0000 UTC m=+1215.618359329" Feb 02 09:27:32 crc kubenswrapper[4764]: I0202 09:27:32.751483 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Feb 02 09:27:32 crc kubenswrapper[4764]: I0202 09:27:32.838775 4764 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="7ae215e9-dc8e-4274-873d-1641cd67c6c1" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.172:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 02 09:27:32 crc kubenswrapper[4764]: I0202 09:27:32.838836 4764 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="7ae215e9-dc8e-4274-873d-1641cd67c6c1" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.172:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 02 09:27:33 crc kubenswrapper[4764]: I0202 09:27:33.085845 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-zfgz6" Feb 02 09:27:33 crc kubenswrapper[4764]: I0202 09:27:33.156002 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6408cd2c-fa48-4c79-a6dd-558008245a0f-combined-ca-bundle\") pod \"6408cd2c-fa48-4c79-a6dd-558008245a0f\" (UID: \"6408cd2c-fa48-4c79-a6dd-558008245a0f\") " Feb 02 09:27:33 crc kubenswrapper[4764]: I0202 09:27:33.156237 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6408cd2c-fa48-4c79-a6dd-558008245a0f-config-data\") pod \"6408cd2c-fa48-4c79-a6dd-558008245a0f\" (UID: \"6408cd2c-fa48-4c79-a6dd-558008245a0f\") " Feb 02 09:27:33 crc kubenswrapper[4764]: I0202 09:27:33.156290 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9dtjd\" (UniqueName: \"kubernetes.io/projected/6408cd2c-fa48-4c79-a6dd-558008245a0f-kube-api-access-9dtjd\") pod \"6408cd2c-fa48-4c79-a6dd-558008245a0f\" (UID: \"6408cd2c-fa48-4c79-a6dd-558008245a0f\") " Feb 02 09:27:33 crc kubenswrapper[4764]: I0202 09:27:33.156326 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6408cd2c-fa48-4c79-a6dd-558008245a0f-scripts\") pod \"6408cd2c-fa48-4c79-a6dd-558008245a0f\" (UID: \"6408cd2c-fa48-4c79-a6dd-558008245a0f\") " Feb 02 09:27:33 crc kubenswrapper[4764]: I0202 09:27:33.172821 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6408cd2c-fa48-4c79-a6dd-558008245a0f-kube-api-access-9dtjd" (OuterVolumeSpecName: "kube-api-access-9dtjd") pod "6408cd2c-fa48-4c79-a6dd-558008245a0f" (UID: "6408cd2c-fa48-4c79-a6dd-558008245a0f"). InnerVolumeSpecName "kube-api-access-9dtjd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:27:33 crc kubenswrapper[4764]: I0202 09:27:33.177218 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6408cd2c-fa48-4c79-a6dd-558008245a0f-scripts" (OuterVolumeSpecName: "scripts") pod "6408cd2c-fa48-4c79-a6dd-558008245a0f" (UID: "6408cd2c-fa48-4c79-a6dd-558008245a0f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:27:33 crc kubenswrapper[4764]: I0202 09:27:33.208117 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6408cd2c-fa48-4c79-a6dd-558008245a0f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6408cd2c-fa48-4c79-a6dd-558008245a0f" (UID: "6408cd2c-fa48-4c79-a6dd-558008245a0f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:27:33 crc kubenswrapper[4764]: I0202 09:27:33.218363 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6408cd2c-fa48-4c79-a6dd-558008245a0f-config-data" (OuterVolumeSpecName: "config-data") pod "6408cd2c-fa48-4c79-a6dd-558008245a0f" (UID: "6408cd2c-fa48-4c79-a6dd-558008245a0f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:27:33 crc kubenswrapper[4764]: I0202 09:27:33.257338 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-555cfdffc9-rw6ht" Feb 02 09:27:33 crc kubenswrapper[4764]: I0202 09:27:33.259611 4764 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6408cd2c-fa48-4c79-a6dd-558008245a0f-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 09:27:33 crc kubenswrapper[4764]: I0202 09:27:33.259638 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9dtjd\" (UniqueName: \"kubernetes.io/projected/6408cd2c-fa48-4c79-a6dd-558008245a0f-kube-api-access-9dtjd\") on node \"crc\" DevicePath \"\"" Feb 02 09:27:33 crc kubenswrapper[4764]: I0202 09:27:33.259648 4764 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6408cd2c-fa48-4c79-a6dd-558008245a0f-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 09:27:33 crc kubenswrapper[4764]: I0202 09:27:33.259657 4764 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6408cd2c-fa48-4c79-a6dd-558008245a0f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 09:27:33 crc kubenswrapper[4764]: I0202 09:27:33.362994 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2c66c255-9454-496c-baf3-5c331d5a2f16-ovsdbserver-nb\") pod \"2c66c255-9454-496c-baf3-5c331d5a2f16\" (UID: \"2c66c255-9454-496c-baf3-5c331d5a2f16\") " Feb 02 09:27:33 crc kubenswrapper[4764]: I0202 09:27:33.363050 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2c66c255-9454-496c-baf3-5c331d5a2f16-config\") pod \"2c66c255-9454-496c-baf3-5c331d5a2f16\" (UID: \"2c66c255-9454-496c-baf3-5c331d5a2f16\") " Feb 02 09:27:33 crc kubenswrapper[4764]: I0202 09:27:33.363104 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9wzfx\" (UniqueName: \"kubernetes.io/projected/2c66c255-9454-496c-baf3-5c331d5a2f16-kube-api-access-9wzfx\") pod \"2c66c255-9454-496c-baf3-5c331d5a2f16\" (UID: \"2c66c255-9454-496c-baf3-5c331d5a2f16\") " Feb 02 09:27:33 crc kubenswrapper[4764]: I0202 09:27:33.363269 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2c66c255-9454-496c-baf3-5c331d5a2f16-dns-svc\") pod \"2c66c255-9454-496c-baf3-5c331d5a2f16\" (UID: \"2c66c255-9454-496c-baf3-5c331d5a2f16\") " Feb 02 09:27:33 crc kubenswrapper[4764]: I0202 09:27:33.363311 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2c66c255-9454-496c-baf3-5c331d5a2f16-ovsdbserver-sb\") pod \"2c66c255-9454-496c-baf3-5c331d5a2f16\" (UID: \"2c66c255-9454-496c-baf3-5c331d5a2f16\") " Feb 02 09:27:33 crc kubenswrapper[4764]: I0202 09:27:33.378124 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2c66c255-9454-496c-baf3-5c331d5a2f16-kube-api-access-9wzfx" (OuterVolumeSpecName: "kube-api-access-9wzfx") pod "2c66c255-9454-496c-baf3-5c331d5a2f16" (UID: "2c66c255-9454-496c-baf3-5c331d5a2f16"). InnerVolumeSpecName "kube-api-access-9wzfx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:27:33 crc kubenswrapper[4764]: I0202 09:27:33.439361 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2c66c255-9454-496c-baf3-5c331d5a2f16-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "2c66c255-9454-496c-baf3-5c331d5a2f16" (UID: "2c66c255-9454-496c-baf3-5c331d5a2f16"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:27:33 crc kubenswrapper[4764]: I0202 09:27:33.440317 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2c66c255-9454-496c-baf3-5c331d5a2f16-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "2c66c255-9454-496c-baf3-5c331d5a2f16" (UID: "2c66c255-9454-496c-baf3-5c331d5a2f16"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:27:33 crc kubenswrapper[4764]: I0202 09:27:33.441455 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2c66c255-9454-496c-baf3-5c331d5a2f16-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "2c66c255-9454-496c-baf3-5c331d5a2f16" (UID: "2c66c255-9454-496c-baf3-5c331d5a2f16"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:27:33 crc kubenswrapper[4764]: I0202 09:27:33.443193 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2c66c255-9454-496c-baf3-5c331d5a2f16-config" (OuterVolumeSpecName: "config") pod "2c66c255-9454-496c-baf3-5c331d5a2f16" (UID: "2c66c255-9454-496c-baf3-5c331d5a2f16"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:27:33 crc kubenswrapper[4764]: I0202 09:27:33.466036 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9wzfx\" (UniqueName: \"kubernetes.io/projected/2c66c255-9454-496c-baf3-5c331d5a2f16-kube-api-access-9wzfx\") on node \"crc\" DevicePath \"\"" Feb 02 09:27:33 crc kubenswrapper[4764]: I0202 09:27:33.472271 4764 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2c66c255-9454-496c-baf3-5c331d5a2f16-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 02 09:27:33 crc kubenswrapper[4764]: I0202 09:27:33.472418 4764 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2c66c255-9454-496c-baf3-5c331d5a2f16-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 02 09:27:33 crc kubenswrapper[4764]: I0202 09:27:33.472477 4764 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2c66c255-9454-496c-baf3-5c331d5a2f16-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 02 09:27:33 crc kubenswrapper[4764]: I0202 09:27:33.472538 4764 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2c66c255-9454-496c-baf3-5c331d5a2f16-config\") on node \"crc\" DevicePath \"\"" Feb 02 09:27:33 crc kubenswrapper[4764]: I0202 09:27:33.659827 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-zfgz6" Feb 02 09:27:33 crc kubenswrapper[4764]: I0202 09:27:33.659822 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-zfgz6" event={"ID":"6408cd2c-fa48-4c79-a6dd-558008245a0f","Type":"ContainerDied","Data":"f4150bd6d4b45e25745599e54d91595b40d88919b2c5fb004faa7c241a9ad90e"} Feb 02 09:27:33 crc kubenswrapper[4764]: I0202 09:27:33.659967 4764 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f4150bd6d4b45e25745599e54d91595b40d88919b2c5fb004faa7c241a9ad90e" Feb 02 09:27:33 crc kubenswrapper[4764]: I0202 09:27:33.662585 4764 generic.go:334] "Generic (PLEG): container finished" podID="2c66c255-9454-496c-baf3-5c331d5a2f16" containerID="5ce5a71993fab6383785294c33c51d3ada6114b29ce3fe0de247942ba4acefce" exitCode=0 Feb 02 09:27:33 crc kubenswrapper[4764]: I0202 09:27:33.663498 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-555cfdffc9-rw6ht" Feb 02 09:27:33 crc kubenswrapper[4764]: I0202 09:27:33.678834 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-555cfdffc9-rw6ht" event={"ID":"2c66c255-9454-496c-baf3-5c331d5a2f16","Type":"ContainerDied","Data":"5ce5a71993fab6383785294c33c51d3ada6114b29ce3fe0de247942ba4acefce"} Feb 02 09:27:33 crc kubenswrapper[4764]: I0202 09:27:33.678911 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Feb 02 09:27:33 crc kubenswrapper[4764]: I0202 09:27:33.678927 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-555cfdffc9-rw6ht" event={"ID":"2c66c255-9454-496c-baf3-5c331d5a2f16","Type":"ContainerDied","Data":"e42ba78f4103f80a77aff2863838250ea8c45442eba5684c8498bdffd7fdaaa5"} Feb 02 09:27:33 crc kubenswrapper[4764]: I0202 09:27:33.678972 4764 scope.go:117] "RemoveContainer" containerID="5ce5a71993fab6383785294c33c51d3ada6114b29ce3fe0de247942ba4acefce" Feb 02 09:27:33 crc kubenswrapper[4764]: I0202 09:27:33.714399 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-555cfdffc9-rw6ht"] Feb 02 09:27:33 crc kubenswrapper[4764]: I0202 09:27:33.724343 4764 scope.go:117] "RemoveContainer" containerID="94f0f40f3100f7ed2dd2d67e2a6da28fe26a61d8d19d2128470d3558ac0876cf" Feb 02 09:27:33 crc kubenswrapper[4764]: I0202 09:27:33.725714 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-555cfdffc9-rw6ht"] Feb 02 09:27:33 crc kubenswrapper[4764]: I0202 09:27:33.841858 4764 scope.go:117] "RemoveContainer" containerID="5ce5a71993fab6383785294c33c51d3ada6114b29ce3fe0de247942ba4acefce" Feb 02 09:27:33 crc kubenswrapper[4764]: E0202 09:27:33.844218 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5ce5a71993fab6383785294c33c51d3ada6114b29ce3fe0de247942ba4acefce\": container with ID starting with 5ce5a71993fab6383785294c33c51d3ada6114b29ce3fe0de247942ba4acefce not found: ID does not exist" containerID="5ce5a71993fab6383785294c33c51d3ada6114b29ce3fe0de247942ba4acefce" Feb 02 09:27:33 crc kubenswrapper[4764]: I0202 09:27:33.844441 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5ce5a71993fab6383785294c33c51d3ada6114b29ce3fe0de247942ba4acefce"} err="failed to get container status \"5ce5a71993fab6383785294c33c51d3ada6114b29ce3fe0de247942ba4acefce\": rpc error: code = NotFound desc = could not find container \"5ce5a71993fab6383785294c33c51d3ada6114b29ce3fe0de247942ba4acefce\": container with ID starting with 5ce5a71993fab6383785294c33c51d3ada6114b29ce3fe0de247942ba4acefce not found: ID does not exist" Feb 02 09:27:33 crc kubenswrapper[4764]: I0202 09:27:33.844484 4764 scope.go:117] "RemoveContainer" containerID="94f0f40f3100f7ed2dd2d67e2a6da28fe26a61d8d19d2128470d3558ac0876cf" Feb 02 09:27:33 crc kubenswrapper[4764]: I0202 09:27:33.847036 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2c66c255-9454-496c-baf3-5c331d5a2f16" path="/var/lib/kubelet/pods/2c66c255-9454-496c-baf3-5c331d5a2f16/volumes" Feb 02 09:27:33 crc kubenswrapper[4764]: E0202 09:27:33.849038 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"94f0f40f3100f7ed2dd2d67e2a6da28fe26a61d8d19d2128470d3558ac0876cf\": container with ID starting with 94f0f40f3100f7ed2dd2d67e2a6da28fe26a61d8d19d2128470d3558ac0876cf not found: ID does not exist" containerID="94f0f40f3100f7ed2dd2d67e2a6da28fe26a61d8d19d2128470d3558ac0876cf" Feb 02 09:27:33 crc kubenswrapper[4764]: I0202 09:27:33.849071 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"94f0f40f3100f7ed2dd2d67e2a6da28fe26a61d8d19d2128470d3558ac0876cf"} err="failed to get container status \"94f0f40f3100f7ed2dd2d67e2a6da28fe26a61d8d19d2128470d3558ac0876cf\": rpc error: code = NotFound desc = could not find container \"94f0f40f3100f7ed2dd2d67e2a6da28fe26a61d8d19d2128470d3558ac0876cf\": container with ID starting with 94f0f40f3100f7ed2dd2d67e2a6da28fe26a61d8d19d2128470d3558ac0876cf not found: ID does not exist" Feb 02 09:27:33 crc kubenswrapper[4764]: I0202 09:27:33.879846 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Feb 02 09:27:33 crc kubenswrapper[4764]: I0202 09:27:33.880121 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="7ae215e9-dc8e-4274-873d-1641cd67c6c1" containerName="nova-api-log" containerID="cri-o://2e8a49f04e50df3a7f183dd6086b65d007fe49e9d3efa236ddd67bf433ae7409" gracePeriod=30 Feb 02 09:27:33 crc kubenswrapper[4764]: I0202 09:27:33.880774 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="7ae215e9-dc8e-4274-873d-1641cd67c6c1" containerName="nova-api-api" containerID="cri-o://fa1c95d637a3c5dfe63e9ee6d8afba0bf9bc07979fd8f0b0d00c7b5cbf36b0d6" gracePeriod=30 Feb 02 09:27:33 crc kubenswrapper[4764]: I0202 09:27:33.927659 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Feb 02 09:27:33 crc kubenswrapper[4764]: I0202 09:27:33.929700 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="df690cc6-0a88-44db-8ee5-2e967e51fa59" containerName="nova-metadata-log" containerID="cri-o://0a9a3f3ea0066a3c068850c336c289ba30f34f30340cbb09b8b3c548baca6b36" gracePeriod=30 Feb 02 09:27:33 crc kubenswrapper[4764]: I0202 09:27:33.930201 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="df690cc6-0a88-44db-8ee5-2e967e51fa59" containerName="nova-metadata-metadata" containerID="cri-o://7f8ca67a7e4d45d98682b4584bb5c5eb834a5c546cc932ac81924f956a9e8f0e" gracePeriod=30 Feb 02 09:27:33 crc kubenswrapper[4764]: I0202 09:27:33.951415 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Feb 02 09:27:33 crc kubenswrapper[4764]: I0202 09:27:33.951462 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Feb 02 09:27:33 crc kubenswrapper[4764]: I0202 09:27:33.979167 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Feb 02 09:27:34 crc kubenswrapper[4764]: I0202 09:27:34.056902 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-9zzqm" Feb 02 09:27:34 crc kubenswrapper[4764]: I0202 09:27:34.084305 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j7hlw\" (UniqueName: \"kubernetes.io/projected/91e8d389-437f-4eed-b6b1-2b64802f9bc9-kube-api-access-j7hlw\") pod \"91e8d389-437f-4eed-b6b1-2b64802f9bc9\" (UID: \"91e8d389-437f-4eed-b6b1-2b64802f9bc9\") " Feb 02 09:27:34 crc kubenswrapper[4764]: I0202 09:27:34.084376 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91e8d389-437f-4eed-b6b1-2b64802f9bc9-config-data\") pod \"91e8d389-437f-4eed-b6b1-2b64802f9bc9\" (UID: \"91e8d389-437f-4eed-b6b1-2b64802f9bc9\") " Feb 02 09:27:34 crc kubenswrapper[4764]: I0202 09:27:34.084448 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91e8d389-437f-4eed-b6b1-2b64802f9bc9-combined-ca-bundle\") pod \"91e8d389-437f-4eed-b6b1-2b64802f9bc9\" (UID: \"91e8d389-437f-4eed-b6b1-2b64802f9bc9\") " Feb 02 09:27:34 crc kubenswrapper[4764]: I0202 09:27:34.084652 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/91e8d389-437f-4eed-b6b1-2b64802f9bc9-scripts\") pod \"91e8d389-437f-4eed-b6b1-2b64802f9bc9\" (UID: \"91e8d389-437f-4eed-b6b1-2b64802f9bc9\") " Feb 02 09:27:34 crc kubenswrapper[4764]: I0202 09:27:34.091166 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/91e8d389-437f-4eed-b6b1-2b64802f9bc9-kube-api-access-j7hlw" (OuterVolumeSpecName: "kube-api-access-j7hlw") pod "91e8d389-437f-4eed-b6b1-2b64802f9bc9" (UID: "91e8d389-437f-4eed-b6b1-2b64802f9bc9"). InnerVolumeSpecName "kube-api-access-j7hlw". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:27:34 crc kubenswrapper[4764]: I0202 09:27:34.097044 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/91e8d389-437f-4eed-b6b1-2b64802f9bc9-scripts" (OuterVolumeSpecName: "scripts") pod "91e8d389-437f-4eed-b6b1-2b64802f9bc9" (UID: "91e8d389-437f-4eed-b6b1-2b64802f9bc9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:27:34 crc kubenswrapper[4764]: I0202 09:27:34.118590 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/91e8d389-437f-4eed-b6b1-2b64802f9bc9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "91e8d389-437f-4eed-b6b1-2b64802f9bc9" (UID: "91e8d389-437f-4eed-b6b1-2b64802f9bc9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:27:34 crc kubenswrapper[4764]: I0202 09:27:34.148140 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/91e8d389-437f-4eed-b6b1-2b64802f9bc9-config-data" (OuterVolumeSpecName: "config-data") pod "91e8d389-437f-4eed-b6b1-2b64802f9bc9" (UID: "91e8d389-437f-4eed-b6b1-2b64802f9bc9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:27:34 crc kubenswrapper[4764]: I0202 09:27:34.186851 4764 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/91e8d389-437f-4eed-b6b1-2b64802f9bc9-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 09:27:34 crc kubenswrapper[4764]: I0202 09:27:34.186878 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j7hlw\" (UniqueName: \"kubernetes.io/projected/91e8d389-437f-4eed-b6b1-2b64802f9bc9-kube-api-access-j7hlw\") on node \"crc\" DevicePath \"\"" Feb 02 09:27:34 crc kubenswrapper[4764]: I0202 09:27:34.186890 4764 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91e8d389-437f-4eed-b6b1-2b64802f9bc9-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 09:27:34 crc kubenswrapper[4764]: I0202 09:27:34.186898 4764 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91e8d389-437f-4eed-b6b1-2b64802f9bc9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 09:27:34 crc kubenswrapper[4764]: I0202 09:27:34.675787 4764 generic.go:334] "Generic (PLEG): container finished" podID="3cad43ae-e284-446f-9dbc-ba224b11aac4" containerID="52d9f8634f508399e6f41017e512c93e0cc403c3b39ea0c39e366455b8ee787c" exitCode=0 Feb 02 09:27:34 crc kubenswrapper[4764]: I0202 09:27:34.675841 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3cad43ae-e284-446f-9dbc-ba224b11aac4","Type":"ContainerDied","Data":"52d9f8634f508399e6f41017e512c93e0cc403c3b39ea0c39e366455b8ee787c"} Feb 02 09:27:34 crc kubenswrapper[4764]: I0202 09:27:34.685316 4764 generic.go:334] "Generic (PLEG): container finished" podID="df690cc6-0a88-44db-8ee5-2e967e51fa59" containerID="7f8ca67a7e4d45d98682b4584bb5c5eb834a5c546cc932ac81924f956a9e8f0e" exitCode=0 Feb 02 09:27:34 crc kubenswrapper[4764]: I0202 09:27:34.685348 4764 generic.go:334] "Generic (PLEG): container finished" podID="df690cc6-0a88-44db-8ee5-2e967e51fa59" containerID="0a9a3f3ea0066a3c068850c336c289ba30f34f30340cbb09b8b3c548baca6b36" exitCode=143 Feb 02 09:27:34 crc kubenswrapper[4764]: I0202 09:27:34.685388 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"df690cc6-0a88-44db-8ee5-2e967e51fa59","Type":"ContainerDied","Data":"7f8ca67a7e4d45d98682b4584bb5c5eb834a5c546cc932ac81924f956a9e8f0e"} Feb 02 09:27:34 crc kubenswrapper[4764]: I0202 09:27:34.685414 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"df690cc6-0a88-44db-8ee5-2e967e51fa59","Type":"ContainerDied","Data":"0a9a3f3ea0066a3c068850c336c289ba30f34f30340cbb09b8b3c548baca6b36"} Feb 02 09:27:34 crc kubenswrapper[4764]: I0202 09:27:34.689921 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-9zzqm" event={"ID":"91e8d389-437f-4eed-b6b1-2b64802f9bc9","Type":"ContainerDied","Data":"ca23690127dab495bbd267b79e78f32157ab0f7861187b6a7d6e4e5d1fef3363"} Feb 02 09:27:34 crc kubenswrapper[4764]: I0202 09:27:34.689957 4764 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ca23690127dab495bbd267b79e78f32157ab0f7861187b6a7d6e4e5d1fef3363" Feb 02 09:27:34 crc kubenswrapper[4764]: I0202 09:27:34.690007 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-9zzqm" Feb 02 09:27:34 crc kubenswrapper[4764]: I0202 09:27:34.704688 4764 generic.go:334] "Generic (PLEG): container finished" podID="7ae215e9-dc8e-4274-873d-1641cd67c6c1" containerID="2e8a49f04e50df3a7f183dd6086b65d007fe49e9d3efa236ddd67bf433ae7409" exitCode=143 Feb 02 09:27:34 crc kubenswrapper[4764]: I0202 09:27:34.704865 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="ab56b2c6-5132-488a-bdde-b4bf3cd38b6d" containerName="nova-scheduler-scheduler" containerID="cri-o://b0f942aaab42d58433c36cefea216b2c0d65fcc55f664d23674f0b4d1b63671c" gracePeriod=30 Feb 02 09:27:34 crc kubenswrapper[4764]: I0202 09:27:34.706555 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7ae215e9-dc8e-4274-873d-1641cd67c6c1","Type":"ContainerDied","Data":"2e8a49f04e50df3a7f183dd6086b65d007fe49e9d3efa236ddd67bf433ae7409"} Feb 02 09:27:34 crc kubenswrapper[4764]: I0202 09:27:34.838525 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Feb 02 09:27:34 crc kubenswrapper[4764]: E0202 09:27:34.839141 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c66c255-9454-496c-baf3-5c331d5a2f16" containerName="dnsmasq-dns" Feb 02 09:27:34 crc kubenswrapper[4764]: I0202 09:27:34.839159 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c66c255-9454-496c-baf3-5c331d5a2f16" containerName="dnsmasq-dns" Feb 02 09:27:34 crc kubenswrapper[4764]: E0202 09:27:34.839174 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c66c255-9454-496c-baf3-5c331d5a2f16" containerName="init" Feb 02 09:27:34 crc kubenswrapper[4764]: I0202 09:27:34.839181 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c66c255-9454-496c-baf3-5c331d5a2f16" containerName="init" Feb 02 09:27:34 crc kubenswrapper[4764]: E0202 09:27:34.839193 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6408cd2c-fa48-4c79-a6dd-558008245a0f" containerName="nova-manage" Feb 02 09:27:34 crc kubenswrapper[4764]: I0202 09:27:34.839200 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="6408cd2c-fa48-4c79-a6dd-558008245a0f" containerName="nova-manage" Feb 02 09:27:34 crc kubenswrapper[4764]: E0202 09:27:34.839218 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91e8d389-437f-4eed-b6b1-2b64802f9bc9" containerName="nova-cell1-conductor-db-sync" Feb 02 09:27:34 crc kubenswrapper[4764]: I0202 09:27:34.839224 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="91e8d389-437f-4eed-b6b1-2b64802f9bc9" containerName="nova-cell1-conductor-db-sync" Feb 02 09:27:34 crc kubenswrapper[4764]: I0202 09:27:34.839386 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c66c255-9454-496c-baf3-5c331d5a2f16" containerName="dnsmasq-dns" Feb 02 09:27:34 crc kubenswrapper[4764]: I0202 09:27:34.839415 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="6408cd2c-fa48-4c79-a6dd-558008245a0f" containerName="nova-manage" Feb 02 09:27:34 crc kubenswrapper[4764]: I0202 09:27:34.839430 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="91e8d389-437f-4eed-b6b1-2b64802f9bc9" containerName="nova-cell1-conductor-db-sync" Feb 02 09:27:34 crc kubenswrapper[4764]: I0202 09:27:34.840025 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Feb 02 09:27:34 crc kubenswrapper[4764]: I0202 09:27:34.842315 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Feb 02 09:27:34 crc kubenswrapper[4764]: I0202 09:27:34.850281 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 02 09:27:34 crc kubenswrapper[4764]: I0202 09:27:34.852272 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Feb 02 09:27:34 crc kubenswrapper[4764]: I0202 09:27:34.916776 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3cad43ae-e284-446f-9dbc-ba224b11aac4-config-data\") pod \"3cad43ae-e284-446f-9dbc-ba224b11aac4\" (UID: \"3cad43ae-e284-446f-9dbc-ba224b11aac4\") " Feb 02 09:27:34 crc kubenswrapper[4764]: I0202 09:27:34.916863 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3cad43ae-e284-446f-9dbc-ba224b11aac4-run-httpd\") pod \"3cad43ae-e284-446f-9dbc-ba224b11aac4\" (UID: \"3cad43ae-e284-446f-9dbc-ba224b11aac4\") " Feb 02 09:27:34 crc kubenswrapper[4764]: I0202 09:27:34.916951 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3cad43ae-e284-446f-9dbc-ba224b11aac4-sg-core-conf-yaml\") pod \"3cad43ae-e284-446f-9dbc-ba224b11aac4\" (UID: \"3cad43ae-e284-446f-9dbc-ba224b11aac4\") " Feb 02 09:27:34 crc kubenswrapper[4764]: I0202 09:27:34.916991 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3cad43ae-e284-446f-9dbc-ba224b11aac4-scripts\") pod \"3cad43ae-e284-446f-9dbc-ba224b11aac4\" (UID: \"3cad43ae-e284-446f-9dbc-ba224b11aac4\") " Feb 02 09:27:34 crc kubenswrapper[4764]: I0202 09:27:34.917092 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3cad43ae-e284-446f-9dbc-ba224b11aac4-log-httpd\") pod \"3cad43ae-e284-446f-9dbc-ba224b11aac4\" (UID: \"3cad43ae-e284-446f-9dbc-ba224b11aac4\") " Feb 02 09:27:34 crc kubenswrapper[4764]: I0202 09:27:34.917128 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cad43ae-e284-446f-9dbc-ba224b11aac4-combined-ca-bundle\") pod \"3cad43ae-e284-446f-9dbc-ba224b11aac4\" (UID: \"3cad43ae-e284-446f-9dbc-ba224b11aac4\") " Feb 02 09:27:34 crc kubenswrapper[4764]: I0202 09:27:34.917191 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6v4xn\" (UniqueName: \"kubernetes.io/projected/3cad43ae-e284-446f-9dbc-ba224b11aac4-kube-api-access-6v4xn\") pod \"3cad43ae-e284-446f-9dbc-ba224b11aac4\" (UID: \"3cad43ae-e284-446f-9dbc-ba224b11aac4\") " Feb 02 09:27:34 crc kubenswrapper[4764]: I0202 09:27:34.917415 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b640955b-f755-48d2-9f0a-751b24424f35-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"b640955b-f755-48d2-9f0a-751b24424f35\") " pod="openstack/nova-cell1-conductor-0" Feb 02 09:27:34 crc kubenswrapper[4764]: I0202 09:27:34.917527 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b640955b-f755-48d2-9f0a-751b24424f35-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"b640955b-f755-48d2-9f0a-751b24424f35\") " pod="openstack/nova-cell1-conductor-0" Feb 02 09:27:34 crc kubenswrapper[4764]: I0202 09:27:34.917608 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qqtvn\" (UniqueName: \"kubernetes.io/projected/b640955b-f755-48d2-9f0a-751b24424f35-kube-api-access-qqtvn\") pod \"nova-cell1-conductor-0\" (UID: \"b640955b-f755-48d2-9f0a-751b24424f35\") " pod="openstack/nova-cell1-conductor-0" Feb 02 09:27:34 crc kubenswrapper[4764]: I0202 09:27:34.927451 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3cad43ae-e284-446f-9dbc-ba224b11aac4-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "3cad43ae-e284-446f-9dbc-ba224b11aac4" (UID: "3cad43ae-e284-446f-9dbc-ba224b11aac4"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 09:27:34 crc kubenswrapper[4764]: I0202 09:27:34.927625 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3cad43ae-e284-446f-9dbc-ba224b11aac4-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "3cad43ae-e284-446f-9dbc-ba224b11aac4" (UID: "3cad43ae-e284-446f-9dbc-ba224b11aac4"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 09:27:34 crc kubenswrapper[4764]: I0202 09:27:34.934109 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3cad43ae-e284-446f-9dbc-ba224b11aac4-scripts" (OuterVolumeSpecName: "scripts") pod "3cad43ae-e284-446f-9dbc-ba224b11aac4" (UID: "3cad43ae-e284-446f-9dbc-ba224b11aac4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:27:34 crc kubenswrapper[4764]: I0202 09:27:34.936110 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cad43ae-e284-446f-9dbc-ba224b11aac4-kube-api-access-6v4xn" (OuterVolumeSpecName: "kube-api-access-6v4xn") pod "3cad43ae-e284-446f-9dbc-ba224b11aac4" (UID: "3cad43ae-e284-446f-9dbc-ba224b11aac4"). InnerVolumeSpecName "kube-api-access-6v4xn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:27:34 crc kubenswrapper[4764]: I0202 09:27:34.997844 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3cad43ae-e284-446f-9dbc-ba224b11aac4-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "3cad43ae-e284-446f-9dbc-ba224b11aac4" (UID: "3cad43ae-e284-446f-9dbc-ba224b11aac4"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:27:35 crc kubenswrapper[4764]: I0202 09:27:35.011019 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 02 09:27:35 crc kubenswrapper[4764]: I0202 09:27:35.021188 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b640955b-f755-48d2-9f0a-751b24424f35-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"b640955b-f755-48d2-9f0a-751b24424f35\") " pod="openstack/nova-cell1-conductor-0" Feb 02 09:27:35 crc kubenswrapper[4764]: I0202 09:27:35.021250 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qqtvn\" (UniqueName: \"kubernetes.io/projected/b640955b-f755-48d2-9f0a-751b24424f35-kube-api-access-qqtvn\") pod \"nova-cell1-conductor-0\" (UID: \"b640955b-f755-48d2-9f0a-751b24424f35\") " pod="openstack/nova-cell1-conductor-0" Feb 02 09:27:35 crc kubenswrapper[4764]: I0202 09:27:35.021329 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b640955b-f755-48d2-9f0a-751b24424f35-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"b640955b-f755-48d2-9f0a-751b24424f35\") " pod="openstack/nova-cell1-conductor-0" Feb 02 09:27:35 crc kubenswrapper[4764]: I0202 09:27:35.021419 4764 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3cad43ae-e284-446f-9dbc-ba224b11aac4-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 02 09:27:35 crc kubenswrapper[4764]: I0202 09:27:35.021430 4764 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3cad43ae-e284-446f-9dbc-ba224b11aac4-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 02 09:27:35 crc kubenswrapper[4764]: I0202 09:27:35.021440 4764 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3cad43ae-e284-446f-9dbc-ba224b11aac4-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 09:27:35 crc kubenswrapper[4764]: I0202 09:27:35.021448 4764 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3cad43ae-e284-446f-9dbc-ba224b11aac4-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 02 09:27:35 crc kubenswrapper[4764]: I0202 09:27:35.021457 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6v4xn\" (UniqueName: \"kubernetes.io/projected/3cad43ae-e284-446f-9dbc-ba224b11aac4-kube-api-access-6v4xn\") on node \"crc\" DevicePath \"\"" Feb 02 09:27:35 crc kubenswrapper[4764]: I0202 09:27:35.024903 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b640955b-f755-48d2-9f0a-751b24424f35-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"b640955b-f755-48d2-9f0a-751b24424f35\") " pod="openstack/nova-cell1-conductor-0" Feb 02 09:27:35 crc kubenswrapper[4764]: I0202 09:27:35.035859 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b640955b-f755-48d2-9f0a-751b24424f35-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"b640955b-f755-48d2-9f0a-751b24424f35\") " pod="openstack/nova-cell1-conductor-0" Feb 02 09:27:35 crc kubenswrapper[4764]: I0202 09:27:35.047541 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qqtvn\" (UniqueName: \"kubernetes.io/projected/b640955b-f755-48d2-9f0a-751b24424f35-kube-api-access-qqtvn\") pod \"nova-cell1-conductor-0\" (UID: \"b640955b-f755-48d2-9f0a-751b24424f35\") " pod="openstack/nova-cell1-conductor-0" Feb 02 09:27:35 crc kubenswrapper[4764]: I0202 09:27:35.115255 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3cad43ae-e284-446f-9dbc-ba224b11aac4-config-data" (OuterVolumeSpecName: "config-data") pod "3cad43ae-e284-446f-9dbc-ba224b11aac4" (UID: "3cad43ae-e284-446f-9dbc-ba224b11aac4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:27:35 crc kubenswrapper[4764]: I0202 09:27:35.115574 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3cad43ae-e284-446f-9dbc-ba224b11aac4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3cad43ae-e284-446f-9dbc-ba224b11aac4" (UID: "3cad43ae-e284-446f-9dbc-ba224b11aac4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:27:35 crc kubenswrapper[4764]: I0202 09:27:35.122686 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/df690cc6-0a88-44db-8ee5-2e967e51fa59-nova-metadata-tls-certs\") pod \"df690cc6-0a88-44db-8ee5-2e967e51fa59\" (UID: \"df690cc6-0a88-44db-8ee5-2e967e51fa59\") " Feb 02 09:27:35 crc kubenswrapper[4764]: I0202 09:27:35.122886 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/df690cc6-0a88-44db-8ee5-2e967e51fa59-config-data\") pod \"df690cc6-0a88-44db-8ee5-2e967e51fa59\" (UID: \"df690cc6-0a88-44db-8ee5-2e967e51fa59\") " Feb 02 09:27:35 crc kubenswrapper[4764]: I0202 09:27:35.123104 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rhj77\" (UniqueName: \"kubernetes.io/projected/df690cc6-0a88-44db-8ee5-2e967e51fa59-kube-api-access-rhj77\") pod \"df690cc6-0a88-44db-8ee5-2e967e51fa59\" (UID: \"df690cc6-0a88-44db-8ee5-2e967e51fa59\") " Feb 02 09:27:35 crc kubenswrapper[4764]: I0202 09:27:35.123313 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/df690cc6-0a88-44db-8ee5-2e967e51fa59-logs\") pod \"df690cc6-0a88-44db-8ee5-2e967e51fa59\" (UID: \"df690cc6-0a88-44db-8ee5-2e967e51fa59\") " Feb 02 09:27:35 crc kubenswrapper[4764]: I0202 09:27:35.123423 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df690cc6-0a88-44db-8ee5-2e967e51fa59-combined-ca-bundle\") pod \"df690cc6-0a88-44db-8ee5-2e967e51fa59\" (UID: \"df690cc6-0a88-44db-8ee5-2e967e51fa59\") " Feb 02 09:27:35 crc kubenswrapper[4764]: I0202 09:27:35.123672 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/df690cc6-0a88-44db-8ee5-2e967e51fa59-logs" (OuterVolumeSpecName: "logs") pod "df690cc6-0a88-44db-8ee5-2e967e51fa59" (UID: "df690cc6-0a88-44db-8ee5-2e967e51fa59"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 09:27:35 crc kubenswrapper[4764]: I0202 09:27:35.123918 4764 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/df690cc6-0a88-44db-8ee5-2e967e51fa59-logs\") on node \"crc\" DevicePath \"\"" Feb 02 09:27:35 crc kubenswrapper[4764]: I0202 09:27:35.124016 4764 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cad43ae-e284-446f-9dbc-ba224b11aac4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 09:27:35 crc kubenswrapper[4764]: I0202 09:27:35.124104 4764 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3cad43ae-e284-446f-9dbc-ba224b11aac4-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 09:27:35 crc kubenswrapper[4764]: I0202 09:27:35.139822 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/df690cc6-0a88-44db-8ee5-2e967e51fa59-kube-api-access-rhj77" (OuterVolumeSpecName: "kube-api-access-rhj77") pod "df690cc6-0a88-44db-8ee5-2e967e51fa59" (UID: "df690cc6-0a88-44db-8ee5-2e967e51fa59"). InnerVolumeSpecName "kube-api-access-rhj77". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:27:35 crc kubenswrapper[4764]: I0202 09:27:35.158430 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Feb 02 09:27:35 crc kubenswrapper[4764]: I0202 09:27:35.172222 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df690cc6-0a88-44db-8ee5-2e967e51fa59-config-data" (OuterVolumeSpecName: "config-data") pod "df690cc6-0a88-44db-8ee5-2e967e51fa59" (UID: "df690cc6-0a88-44db-8ee5-2e967e51fa59"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:27:35 crc kubenswrapper[4764]: I0202 09:27:35.173791 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df690cc6-0a88-44db-8ee5-2e967e51fa59-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "df690cc6-0a88-44db-8ee5-2e967e51fa59" (UID: "df690cc6-0a88-44db-8ee5-2e967e51fa59"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:27:35 crc kubenswrapper[4764]: I0202 09:27:35.173996 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df690cc6-0a88-44db-8ee5-2e967e51fa59-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "df690cc6-0a88-44db-8ee5-2e967e51fa59" (UID: "df690cc6-0a88-44db-8ee5-2e967e51fa59"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:27:35 crc kubenswrapper[4764]: I0202 09:27:35.225768 4764 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/df690cc6-0a88-44db-8ee5-2e967e51fa59-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 02 09:27:35 crc kubenswrapper[4764]: I0202 09:27:35.225798 4764 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/df690cc6-0a88-44db-8ee5-2e967e51fa59-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 09:27:35 crc kubenswrapper[4764]: I0202 09:27:35.225807 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rhj77\" (UniqueName: \"kubernetes.io/projected/df690cc6-0a88-44db-8ee5-2e967e51fa59-kube-api-access-rhj77\") on node \"crc\" DevicePath \"\"" Feb 02 09:27:35 crc kubenswrapper[4764]: I0202 09:27:35.225816 4764 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df690cc6-0a88-44db-8ee5-2e967e51fa59-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 09:27:35 crc kubenswrapper[4764]: I0202 09:27:35.576392 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Feb 02 09:27:35 crc kubenswrapper[4764]: I0202 09:27:35.715173 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"b640955b-f755-48d2-9f0a-751b24424f35","Type":"ContainerStarted","Data":"adc48558303cac7d36f76b63f92999694187ed009afe67fc6f0bfff3e9472e08"} Feb 02 09:27:35 crc kubenswrapper[4764]: I0202 09:27:35.719429 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3cad43ae-e284-446f-9dbc-ba224b11aac4","Type":"ContainerDied","Data":"39fc7654d7dd07e41524b35a271817faa467c6b3dcc8ab16623b678f54955d5a"} Feb 02 09:27:35 crc kubenswrapper[4764]: I0202 09:27:35.719510 4764 scope.go:117] "RemoveContainer" containerID="9ba9869e46c55b3868c63cd7c0e91be1c9eec3f6b90f2c1a868d1065a651680a" Feb 02 09:27:35 crc kubenswrapper[4764]: I0202 09:27:35.719674 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 02 09:27:35 crc kubenswrapper[4764]: I0202 09:27:35.723683 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"df690cc6-0a88-44db-8ee5-2e967e51fa59","Type":"ContainerDied","Data":"99cfde25138a8798419ae8d9b93eec6d84a6be236156fd1c75437ac9be189a4f"} Feb 02 09:27:35 crc kubenswrapper[4764]: I0202 09:27:35.723799 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 02 09:27:35 crc kubenswrapper[4764]: I0202 09:27:35.817337 4764 scope.go:117] "RemoveContainer" containerID="ec37269bba1eeedbf80feeb128bd400885ce3747aa038d326f942e91b66f9cc1" Feb 02 09:27:35 crc kubenswrapper[4764]: I0202 09:27:35.822808 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Feb 02 09:27:35 crc kubenswrapper[4764]: I0202 09:27:35.842329 4764 scope.go:117] "RemoveContainer" containerID="52d9f8634f508399e6f41017e512c93e0cc403c3b39ea0c39e366455b8ee787c" Feb 02 09:27:35 crc kubenswrapper[4764]: I0202 09:27:35.846756 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Feb 02 09:27:35 crc kubenswrapper[4764]: I0202 09:27:35.856005 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Feb 02 09:27:35 crc kubenswrapper[4764]: E0202 09:27:35.856451 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3cad43ae-e284-446f-9dbc-ba224b11aac4" containerName="sg-core" Feb 02 09:27:35 crc kubenswrapper[4764]: I0202 09:27:35.856470 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="3cad43ae-e284-446f-9dbc-ba224b11aac4" containerName="sg-core" Feb 02 09:27:35 crc kubenswrapper[4764]: E0202 09:27:35.856485 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3cad43ae-e284-446f-9dbc-ba224b11aac4" containerName="ceilometer-central-agent" Feb 02 09:27:35 crc kubenswrapper[4764]: I0202 09:27:35.856492 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="3cad43ae-e284-446f-9dbc-ba224b11aac4" containerName="ceilometer-central-agent" Feb 02 09:27:35 crc kubenswrapper[4764]: E0202 09:27:35.856499 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3cad43ae-e284-446f-9dbc-ba224b11aac4" containerName="proxy-httpd" Feb 02 09:27:35 crc kubenswrapper[4764]: I0202 09:27:35.856506 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="3cad43ae-e284-446f-9dbc-ba224b11aac4" containerName="proxy-httpd" Feb 02 09:27:35 crc kubenswrapper[4764]: E0202 09:27:35.856529 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3cad43ae-e284-446f-9dbc-ba224b11aac4" containerName="ceilometer-notification-agent" Feb 02 09:27:35 crc kubenswrapper[4764]: I0202 09:27:35.856534 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="3cad43ae-e284-446f-9dbc-ba224b11aac4" containerName="ceilometer-notification-agent" Feb 02 09:27:35 crc kubenswrapper[4764]: E0202 09:27:35.856540 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df690cc6-0a88-44db-8ee5-2e967e51fa59" containerName="nova-metadata-log" Feb 02 09:27:35 crc kubenswrapper[4764]: I0202 09:27:35.856546 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="df690cc6-0a88-44db-8ee5-2e967e51fa59" containerName="nova-metadata-log" Feb 02 09:27:35 crc kubenswrapper[4764]: E0202 09:27:35.856557 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df690cc6-0a88-44db-8ee5-2e967e51fa59" containerName="nova-metadata-metadata" Feb 02 09:27:35 crc kubenswrapper[4764]: I0202 09:27:35.856563 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="df690cc6-0a88-44db-8ee5-2e967e51fa59" containerName="nova-metadata-metadata" Feb 02 09:27:35 crc kubenswrapper[4764]: I0202 09:27:35.856727 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="df690cc6-0a88-44db-8ee5-2e967e51fa59" containerName="nova-metadata-log" Feb 02 09:27:35 crc kubenswrapper[4764]: I0202 09:27:35.856740 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="3cad43ae-e284-446f-9dbc-ba224b11aac4" containerName="proxy-httpd" Feb 02 09:27:35 crc kubenswrapper[4764]: I0202 09:27:35.856758 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="3cad43ae-e284-446f-9dbc-ba224b11aac4" containerName="ceilometer-central-agent" Feb 02 09:27:35 crc kubenswrapper[4764]: I0202 09:27:35.856773 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="3cad43ae-e284-446f-9dbc-ba224b11aac4" containerName="sg-core" Feb 02 09:27:35 crc kubenswrapper[4764]: I0202 09:27:35.856787 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="df690cc6-0a88-44db-8ee5-2e967e51fa59" containerName="nova-metadata-metadata" Feb 02 09:27:35 crc kubenswrapper[4764]: I0202 09:27:35.856798 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="3cad43ae-e284-446f-9dbc-ba224b11aac4" containerName="ceilometer-notification-agent" Feb 02 09:27:35 crc kubenswrapper[4764]: I0202 09:27:35.857768 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 02 09:27:35 crc kubenswrapper[4764]: I0202 09:27:35.861740 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Feb 02 09:27:35 crc kubenswrapper[4764]: I0202 09:27:35.863832 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Feb 02 09:27:35 crc kubenswrapper[4764]: I0202 09:27:35.865901 4764 scope.go:117] "RemoveContainer" containerID="f0e10976314a43245db7f1bfe6ae5503a29b455557f59bd621d463a0219b2bb4" Feb 02 09:27:35 crc kubenswrapper[4764]: I0202 09:27:35.866595 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 02 09:27:35 crc kubenswrapper[4764]: I0202 09:27:35.890480 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 02 09:27:35 crc kubenswrapper[4764]: I0202 09:27:35.900134 4764 scope.go:117] "RemoveContainer" containerID="7f8ca67a7e4d45d98682b4584bb5c5eb834a5c546cc932ac81924f956a9e8f0e" Feb 02 09:27:35 crc kubenswrapper[4764]: I0202 09:27:35.910773 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 02 09:27:35 crc kubenswrapper[4764]: I0202 09:27:35.920004 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 02 09:27:35 crc kubenswrapper[4764]: I0202 09:27:35.922129 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 02 09:27:35 crc kubenswrapper[4764]: I0202 09:27:35.925298 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 02 09:27:35 crc kubenswrapper[4764]: I0202 09:27:35.925472 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Feb 02 09:27:35 crc kubenswrapper[4764]: I0202 09:27:35.926156 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 02 09:27:35 crc kubenswrapper[4764]: I0202 09:27:35.934310 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 02 09:27:35 crc kubenswrapper[4764]: I0202 09:27:35.943437 4764 scope.go:117] "RemoveContainer" containerID="0a9a3f3ea0066a3c068850c336c289ba30f34f30340cbb09b8b3c548baca6b36" Feb 02 09:27:35 crc kubenswrapper[4764]: I0202 09:27:35.944947 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5bacef6a-a70e-400b-963f-d3346b916dee-config-data\") pod \"nova-metadata-0\" (UID: \"5bacef6a-a70e-400b-963f-d3346b916dee\") " pod="openstack/nova-metadata-0" Feb 02 09:27:35 crc kubenswrapper[4764]: I0202 09:27:35.944989 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5bacef6a-a70e-400b-963f-d3346b916dee-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"5bacef6a-a70e-400b-963f-d3346b916dee\") " pod="openstack/nova-metadata-0" Feb 02 09:27:35 crc kubenswrapper[4764]: I0202 09:27:35.945021 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j84pj\" (UniqueName: \"kubernetes.io/projected/5bacef6a-a70e-400b-963f-d3346b916dee-kube-api-access-j84pj\") pod \"nova-metadata-0\" (UID: \"5bacef6a-a70e-400b-963f-d3346b916dee\") " pod="openstack/nova-metadata-0" Feb 02 09:27:35 crc kubenswrapper[4764]: I0202 09:27:35.945095 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/5bacef6a-a70e-400b-963f-d3346b916dee-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"5bacef6a-a70e-400b-963f-d3346b916dee\") " pod="openstack/nova-metadata-0" Feb 02 09:27:35 crc kubenswrapper[4764]: I0202 09:27:35.945151 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5bacef6a-a70e-400b-963f-d3346b916dee-logs\") pod \"nova-metadata-0\" (UID: \"5bacef6a-a70e-400b-963f-d3346b916dee\") " pod="openstack/nova-metadata-0" Feb 02 09:27:36 crc kubenswrapper[4764]: I0202 09:27:36.047075 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f6b65eed-dab4-4af5-ba48-c24a50e41e8f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f6b65eed-dab4-4af5-ba48-c24a50e41e8f\") " pod="openstack/ceilometer-0" Feb 02 09:27:36 crc kubenswrapper[4764]: I0202 09:27:36.047190 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j84pj\" (UniqueName: \"kubernetes.io/projected/5bacef6a-a70e-400b-963f-d3346b916dee-kube-api-access-j84pj\") pod \"nova-metadata-0\" (UID: \"5bacef6a-a70e-400b-963f-d3346b916dee\") " pod="openstack/nova-metadata-0" Feb 02 09:27:36 crc kubenswrapper[4764]: I0202 09:27:36.047272 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6b65eed-dab4-4af5-ba48-c24a50e41e8f-config-data\") pod \"ceilometer-0\" (UID: \"f6b65eed-dab4-4af5-ba48-c24a50e41e8f\") " pod="openstack/ceilometer-0" Feb 02 09:27:36 crc kubenswrapper[4764]: I0202 09:27:36.047303 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/5bacef6a-a70e-400b-963f-d3346b916dee-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"5bacef6a-a70e-400b-963f-d3346b916dee\") " pod="openstack/nova-metadata-0" Feb 02 09:27:36 crc kubenswrapper[4764]: I0202 09:27:36.047355 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f6b65eed-dab4-4af5-ba48-c24a50e41e8f-log-httpd\") pod \"ceilometer-0\" (UID: \"f6b65eed-dab4-4af5-ba48-c24a50e41e8f\") " pod="openstack/ceilometer-0" Feb 02 09:27:36 crc kubenswrapper[4764]: I0202 09:27:36.047376 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dgxbk\" (UniqueName: \"kubernetes.io/projected/f6b65eed-dab4-4af5-ba48-c24a50e41e8f-kube-api-access-dgxbk\") pod \"ceilometer-0\" (UID: \"f6b65eed-dab4-4af5-ba48-c24a50e41e8f\") " pod="openstack/ceilometer-0" Feb 02 09:27:36 crc kubenswrapper[4764]: I0202 09:27:36.047392 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5bacef6a-a70e-400b-963f-d3346b916dee-logs\") pod \"nova-metadata-0\" (UID: \"5bacef6a-a70e-400b-963f-d3346b916dee\") " pod="openstack/nova-metadata-0" Feb 02 09:27:36 crc kubenswrapper[4764]: I0202 09:27:36.047444 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f6b65eed-dab4-4af5-ba48-c24a50e41e8f-run-httpd\") pod \"ceilometer-0\" (UID: \"f6b65eed-dab4-4af5-ba48-c24a50e41e8f\") " pod="openstack/ceilometer-0" Feb 02 09:27:36 crc kubenswrapper[4764]: I0202 09:27:36.047463 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6b65eed-dab4-4af5-ba48-c24a50e41e8f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f6b65eed-dab4-4af5-ba48-c24a50e41e8f\") " pod="openstack/ceilometer-0" Feb 02 09:27:36 crc kubenswrapper[4764]: I0202 09:27:36.047522 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f6b65eed-dab4-4af5-ba48-c24a50e41e8f-scripts\") pod \"ceilometer-0\" (UID: \"f6b65eed-dab4-4af5-ba48-c24a50e41e8f\") " pod="openstack/ceilometer-0" Feb 02 09:27:36 crc kubenswrapper[4764]: I0202 09:27:36.047549 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/f6b65eed-dab4-4af5-ba48-c24a50e41e8f-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"f6b65eed-dab4-4af5-ba48-c24a50e41e8f\") " pod="openstack/ceilometer-0" Feb 02 09:27:36 crc kubenswrapper[4764]: I0202 09:27:36.047602 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5bacef6a-a70e-400b-963f-d3346b916dee-config-data\") pod \"nova-metadata-0\" (UID: \"5bacef6a-a70e-400b-963f-d3346b916dee\") " pod="openstack/nova-metadata-0" Feb 02 09:27:36 crc kubenswrapper[4764]: I0202 09:27:36.047621 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5bacef6a-a70e-400b-963f-d3346b916dee-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"5bacef6a-a70e-400b-963f-d3346b916dee\") " pod="openstack/nova-metadata-0" Feb 02 09:27:36 crc kubenswrapper[4764]: I0202 09:27:36.048040 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5bacef6a-a70e-400b-963f-d3346b916dee-logs\") pod \"nova-metadata-0\" (UID: \"5bacef6a-a70e-400b-963f-d3346b916dee\") " pod="openstack/nova-metadata-0" Feb 02 09:27:36 crc kubenswrapper[4764]: I0202 09:27:36.054472 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5bacef6a-a70e-400b-963f-d3346b916dee-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"5bacef6a-a70e-400b-963f-d3346b916dee\") " pod="openstack/nova-metadata-0" Feb 02 09:27:36 crc kubenswrapper[4764]: I0202 09:27:36.064047 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5bacef6a-a70e-400b-963f-d3346b916dee-config-data\") pod \"nova-metadata-0\" (UID: \"5bacef6a-a70e-400b-963f-d3346b916dee\") " pod="openstack/nova-metadata-0" Feb 02 09:27:36 crc kubenswrapper[4764]: I0202 09:27:36.064457 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/5bacef6a-a70e-400b-963f-d3346b916dee-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"5bacef6a-a70e-400b-963f-d3346b916dee\") " pod="openstack/nova-metadata-0" Feb 02 09:27:36 crc kubenswrapper[4764]: I0202 09:27:36.074380 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j84pj\" (UniqueName: \"kubernetes.io/projected/5bacef6a-a70e-400b-963f-d3346b916dee-kube-api-access-j84pj\") pod \"nova-metadata-0\" (UID: \"5bacef6a-a70e-400b-963f-d3346b916dee\") " pod="openstack/nova-metadata-0" Feb 02 09:27:36 crc kubenswrapper[4764]: I0202 09:27:36.148809 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/f6b65eed-dab4-4af5-ba48-c24a50e41e8f-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"f6b65eed-dab4-4af5-ba48-c24a50e41e8f\") " pod="openstack/ceilometer-0" Feb 02 09:27:36 crc kubenswrapper[4764]: I0202 09:27:36.148881 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f6b65eed-dab4-4af5-ba48-c24a50e41e8f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f6b65eed-dab4-4af5-ba48-c24a50e41e8f\") " pod="openstack/ceilometer-0" Feb 02 09:27:36 crc kubenswrapper[4764]: I0202 09:27:36.148957 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6b65eed-dab4-4af5-ba48-c24a50e41e8f-config-data\") pod \"ceilometer-0\" (UID: \"f6b65eed-dab4-4af5-ba48-c24a50e41e8f\") " pod="openstack/ceilometer-0" Feb 02 09:27:36 crc kubenswrapper[4764]: I0202 09:27:36.149002 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f6b65eed-dab4-4af5-ba48-c24a50e41e8f-log-httpd\") pod \"ceilometer-0\" (UID: \"f6b65eed-dab4-4af5-ba48-c24a50e41e8f\") " pod="openstack/ceilometer-0" Feb 02 09:27:36 crc kubenswrapper[4764]: I0202 09:27:36.149023 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dgxbk\" (UniqueName: \"kubernetes.io/projected/f6b65eed-dab4-4af5-ba48-c24a50e41e8f-kube-api-access-dgxbk\") pod \"ceilometer-0\" (UID: \"f6b65eed-dab4-4af5-ba48-c24a50e41e8f\") " pod="openstack/ceilometer-0" Feb 02 09:27:36 crc kubenswrapper[4764]: I0202 09:27:36.149049 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f6b65eed-dab4-4af5-ba48-c24a50e41e8f-run-httpd\") pod \"ceilometer-0\" (UID: \"f6b65eed-dab4-4af5-ba48-c24a50e41e8f\") " pod="openstack/ceilometer-0" Feb 02 09:27:36 crc kubenswrapper[4764]: I0202 09:27:36.149069 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6b65eed-dab4-4af5-ba48-c24a50e41e8f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f6b65eed-dab4-4af5-ba48-c24a50e41e8f\") " pod="openstack/ceilometer-0" Feb 02 09:27:36 crc kubenswrapper[4764]: I0202 09:27:36.149101 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f6b65eed-dab4-4af5-ba48-c24a50e41e8f-scripts\") pod \"ceilometer-0\" (UID: \"f6b65eed-dab4-4af5-ba48-c24a50e41e8f\") " pod="openstack/ceilometer-0" Feb 02 09:27:36 crc kubenswrapper[4764]: I0202 09:27:36.149789 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f6b65eed-dab4-4af5-ba48-c24a50e41e8f-log-httpd\") pod \"ceilometer-0\" (UID: \"f6b65eed-dab4-4af5-ba48-c24a50e41e8f\") " pod="openstack/ceilometer-0" Feb 02 09:27:36 crc kubenswrapper[4764]: I0202 09:27:36.149845 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f6b65eed-dab4-4af5-ba48-c24a50e41e8f-run-httpd\") pod \"ceilometer-0\" (UID: \"f6b65eed-dab4-4af5-ba48-c24a50e41e8f\") " pod="openstack/ceilometer-0" Feb 02 09:27:36 crc kubenswrapper[4764]: I0202 09:27:36.152260 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/f6b65eed-dab4-4af5-ba48-c24a50e41e8f-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"f6b65eed-dab4-4af5-ba48-c24a50e41e8f\") " pod="openstack/ceilometer-0" Feb 02 09:27:36 crc kubenswrapper[4764]: I0202 09:27:36.152482 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f6b65eed-dab4-4af5-ba48-c24a50e41e8f-scripts\") pod \"ceilometer-0\" (UID: \"f6b65eed-dab4-4af5-ba48-c24a50e41e8f\") " pod="openstack/ceilometer-0" Feb 02 09:27:36 crc kubenswrapper[4764]: I0202 09:27:36.154138 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6b65eed-dab4-4af5-ba48-c24a50e41e8f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f6b65eed-dab4-4af5-ba48-c24a50e41e8f\") " pod="openstack/ceilometer-0" Feb 02 09:27:36 crc kubenswrapper[4764]: I0202 09:27:36.154790 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f6b65eed-dab4-4af5-ba48-c24a50e41e8f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f6b65eed-dab4-4af5-ba48-c24a50e41e8f\") " pod="openstack/ceilometer-0" Feb 02 09:27:36 crc kubenswrapper[4764]: I0202 09:27:36.160867 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6b65eed-dab4-4af5-ba48-c24a50e41e8f-config-data\") pod \"ceilometer-0\" (UID: \"f6b65eed-dab4-4af5-ba48-c24a50e41e8f\") " pod="openstack/ceilometer-0" Feb 02 09:27:36 crc kubenswrapper[4764]: I0202 09:27:36.168763 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dgxbk\" (UniqueName: \"kubernetes.io/projected/f6b65eed-dab4-4af5-ba48-c24a50e41e8f-kube-api-access-dgxbk\") pod \"ceilometer-0\" (UID: \"f6b65eed-dab4-4af5-ba48-c24a50e41e8f\") " pod="openstack/ceilometer-0" Feb 02 09:27:36 crc kubenswrapper[4764]: I0202 09:27:36.179725 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 02 09:27:36 crc kubenswrapper[4764]: I0202 09:27:36.255227 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 02 09:27:36 crc kubenswrapper[4764]: I0202 09:27:36.504348 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 02 09:27:36 crc kubenswrapper[4764]: I0202 09:27:36.732615 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"b640955b-f755-48d2-9f0a-751b24424f35","Type":"ContainerStarted","Data":"1fb85ec28e554137f37ac47f6298fb72cf580b76df565fc0a6bd67d511aab201"} Feb 02 09:27:36 crc kubenswrapper[4764]: I0202 09:27:36.733634 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Feb 02 09:27:36 crc kubenswrapper[4764]: I0202 09:27:36.735269 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5bacef6a-a70e-400b-963f-d3346b916dee","Type":"ContainerStarted","Data":"622ff140637b2ce5b8480b66fad9ed0fef43f20de465dd7bff21c30bdb127da9"} Feb 02 09:27:36 crc kubenswrapper[4764]: I0202 09:27:36.757157 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.757139873 podStartE2EDuration="2.757139873s" podCreationTimestamp="2026-02-02 09:27:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:27:36.754727777 +0000 UTC m=+1219.688451865" watchObservedRunningTime="2026-02-02 09:27:36.757139873 +0000 UTC m=+1219.690863961" Feb 02 09:27:36 crc kubenswrapper[4764]: E0202 09:27:36.811592 4764 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="b0f942aaab42d58433c36cefea216b2c0d65fcc55f664d23674f0b4d1b63671c" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Feb 02 09:27:36 crc kubenswrapper[4764]: I0202 09:27:36.816249 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 02 09:27:36 crc kubenswrapper[4764]: E0202 09:27:36.824334 4764 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="b0f942aaab42d58433c36cefea216b2c0d65fcc55f664d23674f0b4d1b63671c" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Feb 02 09:27:36 crc kubenswrapper[4764]: W0202 09:27:36.824564 4764 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf6b65eed_dab4_4af5_ba48_c24a50e41e8f.slice/crio-11d2d903d7fe8be3d6c76d6ef8441093a875dcda0bbcc009736c64d582635e7d WatchSource:0}: Error finding container 11d2d903d7fe8be3d6c76d6ef8441093a875dcda0bbcc009736c64d582635e7d: Status 404 returned error can't find the container with id 11d2d903d7fe8be3d6c76d6ef8441093a875dcda0bbcc009736c64d582635e7d Feb 02 09:27:36 crc kubenswrapper[4764]: E0202 09:27:36.830117 4764 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="b0f942aaab42d58433c36cefea216b2c0d65fcc55f664d23674f0b4d1b63671c" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Feb 02 09:27:36 crc kubenswrapper[4764]: E0202 09:27:36.830179 4764 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="ab56b2c6-5132-488a-bdde-b4bf3cd38b6d" containerName="nova-scheduler-scheduler" Feb 02 09:27:37 crc kubenswrapper[4764]: I0202 09:27:37.746490 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5bacef6a-a70e-400b-963f-d3346b916dee","Type":"ContainerStarted","Data":"ea61bee5849658de7d6ee0a009ed7eb2d4a69344ed932c37e7cff559942691b0"} Feb 02 09:27:37 crc kubenswrapper[4764]: I0202 09:27:37.746984 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5bacef6a-a70e-400b-963f-d3346b916dee","Type":"ContainerStarted","Data":"d0a31e50c1e0faeed1483cc4d170f915250dbb3c36dd7c5c83c8f8974270a037"} Feb 02 09:27:37 crc kubenswrapper[4764]: I0202 09:27:37.751550 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f6b65eed-dab4-4af5-ba48-c24a50e41e8f","Type":"ContainerStarted","Data":"929ea7921f7027ef1b4f008d85641195d62dd7a0dbbcccbc94161ff9d6efa998"} Feb 02 09:27:37 crc kubenswrapper[4764]: I0202 09:27:37.751584 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f6b65eed-dab4-4af5-ba48-c24a50e41e8f","Type":"ContainerStarted","Data":"11d2d903d7fe8be3d6c76d6ef8441093a875dcda0bbcc009736c64d582635e7d"} Feb 02 09:27:37 crc kubenswrapper[4764]: I0202 09:27:37.767630 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.76761407 podStartE2EDuration="2.76761407s" podCreationTimestamp="2026-02-02 09:27:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:27:37.764278229 +0000 UTC m=+1220.698002317" watchObservedRunningTime="2026-02-02 09:27:37.76761407 +0000 UTC m=+1220.701338158" Feb 02 09:27:37 crc kubenswrapper[4764]: I0202 09:27:37.840620 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cad43ae-e284-446f-9dbc-ba224b11aac4" path="/var/lib/kubelet/pods/3cad43ae-e284-446f-9dbc-ba224b11aac4/volumes" Feb 02 09:27:37 crc kubenswrapper[4764]: I0202 09:27:37.841363 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="df690cc6-0a88-44db-8ee5-2e967e51fa59" path="/var/lib/kubelet/pods/df690cc6-0a88-44db-8ee5-2e967e51fa59/volumes" Feb 02 09:27:38 crc kubenswrapper[4764]: I0202 09:27:38.535616 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 02 09:27:38 crc kubenswrapper[4764]: I0202 09:27:38.603713 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4df6s\" (UniqueName: \"kubernetes.io/projected/ab56b2c6-5132-488a-bdde-b4bf3cd38b6d-kube-api-access-4df6s\") pod \"ab56b2c6-5132-488a-bdde-b4bf3cd38b6d\" (UID: \"ab56b2c6-5132-488a-bdde-b4bf3cd38b6d\") " Feb 02 09:27:38 crc kubenswrapper[4764]: I0202 09:27:38.603771 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab56b2c6-5132-488a-bdde-b4bf3cd38b6d-config-data\") pod \"ab56b2c6-5132-488a-bdde-b4bf3cd38b6d\" (UID: \"ab56b2c6-5132-488a-bdde-b4bf3cd38b6d\") " Feb 02 09:27:38 crc kubenswrapper[4764]: I0202 09:27:38.603853 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab56b2c6-5132-488a-bdde-b4bf3cd38b6d-combined-ca-bundle\") pod \"ab56b2c6-5132-488a-bdde-b4bf3cd38b6d\" (UID: \"ab56b2c6-5132-488a-bdde-b4bf3cd38b6d\") " Feb 02 09:27:38 crc kubenswrapper[4764]: I0202 09:27:38.607633 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ab56b2c6-5132-488a-bdde-b4bf3cd38b6d-kube-api-access-4df6s" (OuterVolumeSpecName: "kube-api-access-4df6s") pod "ab56b2c6-5132-488a-bdde-b4bf3cd38b6d" (UID: "ab56b2c6-5132-488a-bdde-b4bf3cd38b6d"). InnerVolumeSpecName "kube-api-access-4df6s". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:27:38 crc kubenswrapper[4764]: I0202 09:27:38.637071 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab56b2c6-5132-488a-bdde-b4bf3cd38b6d-config-data" (OuterVolumeSpecName: "config-data") pod "ab56b2c6-5132-488a-bdde-b4bf3cd38b6d" (UID: "ab56b2c6-5132-488a-bdde-b4bf3cd38b6d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:27:38 crc kubenswrapper[4764]: I0202 09:27:38.656972 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab56b2c6-5132-488a-bdde-b4bf3cd38b6d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ab56b2c6-5132-488a-bdde-b4bf3cd38b6d" (UID: "ab56b2c6-5132-488a-bdde-b4bf3cd38b6d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:27:38 crc kubenswrapper[4764]: I0202 09:27:38.708055 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4df6s\" (UniqueName: \"kubernetes.io/projected/ab56b2c6-5132-488a-bdde-b4bf3cd38b6d-kube-api-access-4df6s\") on node \"crc\" DevicePath \"\"" Feb 02 09:27:38 crc kubenswrapper[4764]: I0202 09:27:38.708088 4764 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab56b2c6-5132-488a-bdde-b4bf3cd38b6d-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 09:27:38 crc kubenswrapper[4764]: I0202 09:27:38.708097 4764 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab56b2c6-5132-488a-bdde-b4bf3cd38b6d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 09:27:38 crc kubenswrapper[4764]: I0202 09:27:38.766008 4764 generic.go:334] "Generic (PLEG): container finished" podID="7ae215e9-dc8e-4274-873d-1641cd67c6c1" containerID="fa1c95d637a3c5dfe63e9ee6d8afba0bf9bc07979fd8f0b0d00c7b5cbf36b0d6" exitCode=0 Feb 02 09:27:38 crc kubenswrapper[4764]: I0202 09:27:38.766076 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7ae215e9-dc8e-4274-873d-1641cd67c6c1","Type":"ContainerDied","Data":"fa1c95d637a3c5dfe63e9ee6d8afba0bf9bc07979fd8f0b0d00c7b5cbf36b0d6"} Feb 02 09:27:38 crc kubenswrapper[4764]: I0202 09:27:38.770153 4764 generic.go:334] "Generic (PLEG): container finished" podID="ab56b2c6-5132-488a-bdde-b4bf3cd38b6d" containerID="b0f942aaab42d58433c36cefea216b2c0d65fcc55f664d23674f0b4d1b63671c" exitCode=0 Feb 02 09:27:38 crc kubenswrapper[4764]: I0202 09:27:38.770214 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 02 09:27:38 crc kubenswrapper[4764]: I0202 09:27:38.770221 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"ab56b2c6-5132-488a-bdde-b4bf3cd38b6d","Type":"ContainerDied","Data":"b0f942aaab42d58433c36cefea216b2c0d65fcc55f664d23674f0b4d1b63671c"} Feb 02 09:27:38 crc kubenswrapper[4764]: I0202 09:27:38.770250 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"ab56b2c6-5132-488a-bdde-b4bf3cd38b6d","Type":"ContainerDied","Data":"65e121a8e165a425ccad9bf86df6ca33926b1fa84b544811507623c15b912aed"} Feb 02 09:27:38 crc kubenswrapper[4764]: I0202 09:27:38.770268 4764 scope.go:117] "RemoveContainer" containerID="b0f942aaab42d58433c36cefea216b2c0d65fcc55f664d23674f0b4d1b63671c" Feb 02 09:27:38 crc kubenswrapper[4764]: I0202 09:27:38.779949 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f6b65eed-dab4-4af5-ba48-c24a50e41e8f","Type":"ContainerStarted","Data":"a8a5259f33cae1431754c29e917f947668da51a1a4f49cc0e612e0f944314d7a"} Feb 02 09:27:38 crc kubenswrapper[4764]: I0202 09:27:38.780227 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f6b65eed-dab4-4af5-ba48-c24a50e41e8f","Type":"ContainerStarted","Data":"cc615530b6652e1802c31a65ab4877751776f4e6415b68194b67c94d1075e4af"} Feb 02 09:27:38 crc kubenswrapper[4764]: I0202 09:27:38.833847 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 02 09:27:38 crc kubenswrapper[4764]: I0202 09:27:38.848654 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Feb 02 09:27:38 crc kubenswrapper[4764]: I0202 09:27:38.858483 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Feb 02 09:27:38 crc kubenswrapper[4764]: I0202 09:27:38.858921 4764 scope.go:117] "RemoveContainer" containerID="b0f942aaab42d58433c36cefea216b2c0d65fcc55f664d23674f0b4d1b63671c" Feb 02 09:27:38 crc kubenswrapper[4764]: E0202 09:27:38.859330 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b0f942aaab42d58433c36cefea216b2c0d65fcc55f664d23674f0b4d1b63671c\": container with ID starting with b0f942aaab42d58433c36cefea216b2c0d65fcc55f664d23674f0b4d1b63671c not found: ID does not exist" containerID="b0f942aaab42d58433c36cefea216b2c0d65fcc55f664d23674f0b4d1b63671c" Feb 02 09:27:38 crc kubenswrapper[4764]: I0202 09:27:38.859353 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b0f942aaab42d58433c36cefea216b2c0d65fcc55f664d23674f0b4d1b63671c"} err="failed to get container status \"b0f942aaab42d58433c36cefea216b2c0d65fcc55f664d23674f0b4d1b63671c\": rpc error: code = NotFound desc = could not find container \"b0f942aaab42d58433c36cefea216b2c0d65fcc55f664d23674f0b4d1b63671c\": container with ID starting with b0f942aaab42d58433c36cefea216b2c0d65fcc55f664d23674f0b4d1b63671c not found: ID does not exist" Feb 02 09:27:38 crc kubenswrapper[4764]: I0202 09:27:38.903053 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Feb 02 09:27:38 crc kubenswrapper[4764]: E0202 09:27:38.903407 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ae215e9-dc8e-4274-873d-1641cd67c6c1" containerName="nova-api-api" Feb 02 09:27:38 crc kubenswrapper[4764]: I0202 09:27:38.903425 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ae215e9-dc8e-4274-873d-1641cd67c6c1" containerName="nova-api-api" Feb 02 09:27:38 crc kubenswrapper[4764]: E0202 09:27:38.903440 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ae215e9-dc8e-4274-873d-1641cd67c6c1" containerName="nova-api-log" Feb 02 09:27:38 crc kubenswrapper[4764]: I0202 09:27:38.903446 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ae215e9-dc8e-4274-873d-1641cd67c6c1" containerName="nova-api-log" Feb 02 09:27:38 crc kubenswrapper[4764]: E0202 09:27:38.903460 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab56b2c6-5132-488a-bdde-b4bf3cd38b6d" containerName="nova-scheduler-scheduler" Feb 02 09:27:38 crc kubenswrapper[4764]: I0202 09:27:38.903467 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab56b2c6-5132-488a-bdde-b4bf3cd38b6d" containerName="nova-scheduler-scheduler" Feb 02 09:27:38 crc kubenswrapper[4764]: I0202 09:27:38.903618 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ae215e9-dc8e-4274-873d-1641cd67c6c1" containerName="nova-api-api" Feb 02 09:27:38 crc kubenswrapper[4764]: I0202 09:27:38.903631 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ae215e9-dc8e-4274-873d-1641cd67c6c1" containerName="nova-api-log" Feb 02 09:27:38 crc kubenswrapper[4764]: I0202 09:27:38.903640 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab56b2c6-5132-488a-bdde-b4bf3cd38b6d" containerName="nova-scheduler-scheduler" Feb 02 09:27:38 crc kubenswrapper[4764]: I0202 09:27:38.904362 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 02 09:27:38 crc kubenswrapper[4764]: I0202 09:27:38.908415 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Feb 02 09:27:38 crc kubenswrapper[4764]: I0202 09:27:38.915539 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-trvls\" (UniqueName: \"kubernetes.io/projected/7ae215e9-dc8e-4274-873d-1641cd67c6c1-kube-api-access-trvls\") pod \"7ae215e9-dc8e-4274-873d-1641cd67c6c1\" (UID: \"7ae215e9-dc8e-4274-873d-1641cd67c6c1\") " Feb 02 09:27:38 crc kubenswrapper[4764]: I0202 09:27:38.915660 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7ae215e9-dc8e-4274-873d-1641cd67c6c1-logs\") pod \"7ae215e9-dc8e-4274-873d-1641cd67c6c1\" (UID: \"7ae215e9-dc8e-4274-873d-1641cd67c6c1\") " Feb 02 09:27:38 crc kubenswrapper[4764]: I0202 09:27:38.915710 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ae215e9-dc8e-4274-873d-1641cd67c6c1-combined-ca-bundle\") pod \"7ae215e9-dc8e-4274-873d-1641cd67c6c1\" (UID: \"7ae215e9-dc8e-4274-873d-1641cd67c6c1\") " Feb 02 09:27:38 crc kubenswrapper[4764]: I0202 09:27:38.915759 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ae215e9-dc8e-4274-873d-1641cd67c6c1-config-data\") pod \"7ae215e9-dc8e-4274-873d-1641cd67c6c1\" (UID: \"7ae215e9-dc8e-4274-873d-1641cd67c6c1\") " Feb 02 09:27:38 crc kubenswrapper[4764]: I0202 09:27:38.916257 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7ae215e9-dc8e-4274-873d-1641cd67c6c1-logs" (OuterVolumeSpecName: "logs") pod "7ae215e9-dc8e-4274-873d-1641cd67c6c1" (UID: "7ae215e9-dc8e-4274-873d-1641cd67c6c1"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 09:27:38 crc kubenswrapper[4764]: I0202 09:27:38.916768 4764 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7ae215e9-dc8e-4274-873d-1641cd67c6c1-logs\") on node \"crc\" DevicePath \"\"" Feb 02 09:27:38 crc kubenswrapper[4764]: I0202 09:27:38.929410 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Feb 02 09:27:38 crc kubenswrapper[4764]: I0202 09:27:38.943314 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ae215e9-dc8e-4274-873d-1641cd67c6c1-kube-api-access-trvls" (OuterVolumeSpecName: "kube-api-access-trvls") pod "7ae215e9-dc8e-4274-873d-1641cd67c6c1" (UID: "7ae215e9-dc8e-4274-873d-1641cd67c6c1"). InnerVolumeSpecName "kube-api-access-trvls". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:27:38 crc kubenswrapper[4764]: I0202 09:27:38.974692 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ae215e9-dc8e-4274-873d-1641cd67c6c1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7ae215e9-dc8e-4274-873d-1641cd67c6c1" (UID: "7ae215e9-dc8e-4274-873d-1641cd67c6c1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:27:39 crc kubenswrapper[4764]: I0202 09:27:39.011464 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ae215e9-dc8e-4274-873d-1641cd67c6c1-config-data" (OuterVolumeSpecName: "config-data") pod "7ae215e9-dc8e-4274-873d-1641cd67c6c1" (UID: "7ae215e9-dc8e-4274-873d-1641cd67c6c1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:27:39 crc kubenswrapper[4764]: I0202 09:27:39.020614 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e10e787-90d5-425b-8d3a-f9b1591cbc9d-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"9e10e787-90d5-425b-8d3a-f9b1591cbc9d\") " pod="openstack/nova-scheduler-0" Feb 02 09:27:39 crc kubenswrapper[4764]: I0202 09:27:39.020703 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9e10e787-90d5-425b-8d3a-f9b1591cbc9d-config-data\") pod \"nova-scheduler-0\" (UID: \"9e10e787-90d5-425b-8d3a-f9b1591cbc9d\") " pod="openstack/nova-scheduler-0" Feb 02 09:27:39 crc kubenswrapper[4764]: I0202 09:27:39.020916 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-57nkj\" (UniqueName: \"kubernetes.io/projected/9e10e787-90d5-425b-8d3a-f9b1591cbc9d-kube-api-access-57nkj\") pod \"nova-scheduler-0\" (UID: \"9e10e787-90d5-425b-8d3a-f9b1591cbc9d\") " pod="openstack/nova-scheduler-0" Feb 02 09:27:39 crc kubenswrapper[4764]: I0202 09:27:39.021060 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-trvls\" (UniqueName: \"kubernetes.io/projected/7ae215e9-dc8e-4274-873d-1641cd67c6c1-kube-api-access-trvls\") on node \"crc\" DevicePath \"\"" Feb 02 09:27:39 crc kubenswrapper[4764]: I0202 09:27:39.021076 4764 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ae215e9-dc8e-4274-873d-1641cd67c6c1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 09:27:39 crc kubenswrapper[4764]: I0202 09:27:39.021085 4764 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ae215e9-dc8e-4274-873d-1641cd67c6c1-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 09:27:39 crc kubenswrapper[4764]: I0202 09:27:39.122991 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-57nkj\" (UniqueName: \"kubernetes.io/projected/9e10e787-90d5-425b-8d3a-f9b1591cbc9d-kube-api-access-57nkj\") pod \"nova-scheduler-0\" (UID: \"9e10e787-90d5-425b-8d3a-f9b1591cbc9d\") " pod="openstack/nova-scheduler-0" Feb 02 09:27:39 crc kubenswrapper[4764]: I0202 09:27:39.123088 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e10e787-90d5-425b-8d3a-f9b1591cbc9d-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"9e10e787-90d5-425b-8d3a-f9b1591cbc9d\") " pod="openstack/nova-scheduler-0" Feb 02 09:27:39 crc kubenswrapper[4764]: I0202 09:27:39.123132 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9e10e787-90d5-425b-8d3a-f9b1591cbc9d-config-data\") pod \"nova-scheduler-0\" (UID: \"9e10e787-90d5-425b-8d3a-f9b1591cbc9d\") " pod="openstack/nova-scheduler-0" Feb 02 09:27:39 crc kubenswrapper[4764]: I0202 09:27:39.126248 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9e10e787-90d5-425b-8d3a-f9b1591cbc9d-config-data\") pod \"nova-scheduler-0\" (UID: \"9e10e787-90d5-425b-8d3a-f9b1591cbc9d\") " pod="openstack/nova-scheduler-0" Feb 02 09:27:39 crc kubenswrapper[4764]: I0202 09:27:39.126723 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e10e787-90d5-425b-8d3a-f9b1591cbc9d-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"9e10e787-90d5-425b-8d3a-f9b1591cbc9d\") " pod="openstack/nova-scheduler-0" Feb 02 09:27:39 crc kubenswrapper[4764]: I0202 09:27:39.138247 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-57nkj\" (UniqueName: \"kubernetes.io/projected/9e10e787-90d5-425b-8d3a-f9b1591cbc9d-kube-api-access-57nkj\") pod \"nova-scheduler-0\" (UID: \"9e10e787-90d5-425b-8d3a-f9b1591cbc9d\") " pod="openstack/nova-scheduler-0" Feb 02 09:27:39 crc kubenswrapper[4764]: I0202 09:27:39.325148 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 02 09:27:39 crc kubenswrapper[4764]: I0202 09:27:39.816783 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 02 09:27:39 crc kubenswrapper[4764]: I0202 09:27:39.816768 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7ae215e9-dc8e-4274-873d-1641cd67c6c1","Type":"ContainerDied","Data":"cd1add0181318b42ce8be7f09a6f35238bed8d7ff7688f486b36f99f37598193"} Feb 02 09:27:39 crc kubenswrapper[4764]: I0202 09:27:39.817003 4764 scope.go:117] "RemoveContainer" containerID="fa1c95d637a3c5dfe63e9ee6d8afba0bf9bc07979fd8f0b0d00c7b5cbf36b0d6" Feb 02 09:27:39 crc kubenswrapper[4764]: I0202 09:27:39.854188 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ab56b2c6-5132-488a-bdde-b4bf3cd38b6d" path="/var/lib/kubelet/pods/ab56b2c6-5132-488a-bdde-b4bf3cd38b6d/volumes" Feb 02 09:27:39 crc kubenswrapper[4764]: I0202 09:27:39.878346 4764 scope.go:117] "RemoveContainer" containerID="2e8a49f04e50df3a7f183dd6086b65d007fe49e9d3efa236ddd67bf433ae7409" Feb 02 09:27:39 crc kubenswrapper[4764]: I0202 09:27:39.878597 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Feb 02 09:27:39 crc kubenswrapper[4764]: I0202 09:27:39.919566 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Feb 02 09:27:39 crc kubenswrapper[4764]: I0202 09:27:39.936696 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Feb 02 09:27:39 crc kubenswrapper[4764]: I0202 09:27:39.960819 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Feb 02 09:27:39 crc kubenswrapper[4764]: I0202 09:27:39.962821 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 02 09:27:39 crc kubenswrapper[4764]: I0202 09:27:39.966676 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Feb 02 09:27:39 crc kubenswrapper[4764]: I0202 09:27:39.969022 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 02 09:27:40 crc kubenswrapper[4764]: I0202 09:27:40.073642 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e77cec7-7174-46ce-ba34-6facb5eac1c0-config-data\") pod \"nova-api-0\" (UID: \"1e77cec7-7174-46ce-ba34-6facb5eac1c0\") " pod="openstack/nova-api-0" Feb 02 09:27:40 crc kubenswrapper[4764]: I0202 09:27:40.073689 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e77cec7-7174-46ce-ba34-6facb5eac1c0-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"1e77cec7-7174-46ce-ba34-6facb5eac1c0\") " pod="openstack/nova-api-0" Feb 02 09:27:40 crc kubenswrapper[4764]: I0202 09:27:40.073738 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1e77cec7-7174-46ce-ba34-6facb5eac1c0-logs\") pod \"nova-api-0\" (UID: \"1e77cec7-7174-46ce-ba34-6facb5eac1c0\") " pod="openstack/nova-api-0" Feb 02 09:27:40 crc kubenswrapper[4764]: I0202 09:27:40.073858 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-78xx2\" (UniqueName: \"kubernetes.io/projected/1e77cec7-7174-46ce-ba34-6facb5eac1c0-kube-api-access-78xx2\") pod \"nova-api-0\" (UID: \"1e77cec7-7174-46ce-ba34-6facb5eac1c0\") " pod="openstack/nova-api-0" Feb 02 09:27:40 crc kubenswrapper[4764]: I0202 09:27:40.175428 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-78xx2\" (UniqueName: \"kubernetes.io/projected/1e77cec7-7174-46ce-ba34-6facb5eac1c0-kube-api-access-78xx2\") pod \"nova-api-0\" (UID: \"1e77cec7-7174-46ce-ba34-6facb5eac1c0\") " pod="openstack/nova-api-0" Feb 02 09:27:40 crc kubenswrapper[4764]: I0202 09:27:40.175915 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e77cec7-7174-46ce-ba34-6facb5eac1c0-config-data\") pod \"nova-api-0\" (UID: \"1e77cec7-7174-46ce-ba34-6facb5eac1c0\") " pod="openstack/nova-api-0" Feb 02 09:27:40 crc kubenswrapper[4764]: I0202 09:27:40.175969 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e77cec7-7174-46ce-ba34-6facb5eac1c0-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"1e77cec7-7174-46ce-ba34-6facb5eac1c0\") " pod="openstack/nova-api-0" Feb 02 09:27:40 crc kubenswrapper[4764]: I0202 09:27:40.176007 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1e77cec7-7174-46ce-ba34-6facb5eac1c0-logs\") pod \"nova-api-0\" (UID: \"1e77cec7-7174-46ce-ba34-6facb5eac1c0\") " pod="openstack/nova-api-0" Feb 02 09:27:40 crc kubenswrapper[4764]: I0202 09:27:40.176993 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1e77cec7-7174-46ce-ba34-6facb5eac1c0-logs\") pod \"nova-api-0\" (UID: \"1e77cec7-7174-46ce-ba34-6facb5eac1c0\") " pod="openstack/nova-api-0" Feb 02 09:27:40 crc kubenswrapper[4764]: I0202 09:27:40.183169 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e77cec7-7174-46ce-ba34-6facb5eac1c0-config-data\") pod \"nova-api-0\" (UID: \"1e77cec7-7174-46ce-ba34-6facb5eac1c0\") " pod="openstack/nova-api-0" Feb 02 09:27:40 crc kubenswrapper[4764]: I0202 09:27:40.183484 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e77cec7-7174-46ce-ba34-6facb5eac1c0-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"1e77cec7-7174-46ce-ba34-6facb5eac1c0\") " pod="openstack/nova-api-0" Feb 02 09:27:40 crc kubenswrapper[4764]: I0202 09:27:40.200387 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-78xx2\" (UniqueName: \"kubernetes.io/projected/1e77cec7-7174-46ce-ba34-6facb5eac1c0-kube-api-access-78xx2\") pod \"nova-api-0\" (UID: \"1e77cec7-7174-46ce-ba34-6facb5eac1c0\") " pod="openstack/nova-api-0" Feb 02 09:27:40 crc kubenswrapper[4764]: I0202 09:27:40.211526 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Feb 02 09:27:40 crc kubenswrapper[4764]: I0202 09:27:40.381841 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 02 09:27:40 crc kubenswrapper[4764]: I0202 09:27:40.849557 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"9e10e787-90d5-425b-8d3a-f9b1591cbc9d","Type":"ContainerStarted","Data":"0588074dea7f8b755739293bb467e526406144beeeb0f156de50d3697687609d"} Feb 02 09:27:40 crc kubenswrapper[4764]: I0202 09:27:40.850042 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"9e10e787-90d5-425b-8d3a-f9b1591cbc9d","Type":"ContainerStarted","Data":"a76c6c516fb754eed5eb905d7ab39a5c12ceb9b4166ebe04ae7cf1243a481e3b"} Feb 02 09:27:40 crc kubenswrapper[4764]: I0202 09:27:40.867627 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.867613029 podStartE2EDuration="2.867613029s" podCreationTimestamp="2026-02-02 09:27:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:27:40.864012421 +0000 UTC m=+1223.797736509" watchObservedRunningTime="2026-02-02 09:27:40.867613029 +0000 UTC m=+1223.801337117" Feb 02 09:27:40 crc kubenswrapper[4764]: I0202 09:27:40.907036 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 02 09:27:41 crc kubenswrapper[4764]: I0202 09:27:41.052054 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Feb 02 09:27:41 crc kubenswrapper[4764]: I0202 09:27:41.180023 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Feb 02 09:27:41 crc kubenswrapper[4764]: I0202 09:27:41.186031 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Feb 02 09:27:41 crc kubenswrapper[4764]: I0202 09:27:41.835766 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7ae215e9-dc8e-4274-873d-1641cd67c6c1" path="/var/lib/kubelet/pods/7ae215e9-dc8e-4274-873d-1641cd67c6c1/volumes" Feb 02 09:27:41 crc kubenswrapper[4764]: I0202 09:27:41.866901 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"1e77cec7-7174-46ce-ba34-6facb5eac1c0","Type":"ContainerStarted","Data":"78e56184f54e35c41bfd8b755dc12aff162742a4324e6e1530acf1915f448b32"} Feb 02 09:27:41 crc kubenswrapper[4764]: I0202 09:27:41.866969 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"1e77cec7-7174-46ce-ba34-6facb5eac1c0","Type":"ContainerStarted","Data":"e9c7c575d06ff5990cadd37986c09dbdc59f0d5a9a8b9a60535974f83fd45ead"} Feb 02 09:27:41 crc kubenswrapper[4764]: I0202 09:27:41.866988 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"1e77cec7-7174-46ce-ba34-6facb5eac1c0","Type":"ContainerStarted","Data":"4f7d1d1eb33324193a509c0b70d4c195c170987b5e32f212f779b8c41c86d0fd"} Feb 02 09:27:41 crc kubenswrapper[4764]: I0202 09:27:41.871889 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f6b65eed-dab4-4af5-ba48-c24a50e41e8f","Type":"ContainerStarted","Data":"636542d714cd8b7f6c52eea15ef07a6fb8036880f4581d65ed9ac2288b62b5fc"} Feb 02 09:27:41 crc kubenswrapper[4764]: I0202 09:27:41.872676 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 02 09:27:41 crc kubenswrapper[4764]: I0202 09:27:41.886983 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.8869713790000002 podStartE2EDuration="2.886971379s" podCreationTimestamp="2026-02-02 09:27:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:27:41.885377195 +0000 UTC m=+1224.819101283" watchObservedRunningTime="2026-02-02 09:27:41.886971379 +0000 UTC m=+1224.820695467" Feb 02 09:27:41 crc kubenswrapper[4764]: I0202 09:27:41.908916 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.123780377 podStartE2EDuration="6.908896756s" podCreationTimestamp="2026-02-02 09:27:35 +0000 UTC" firstStartedPulling="2026-02-02 09:27:36.836133606 +0000 UTC m=+1219.769857694" lastFinishedPulling="2026-02-02 09:27:40.621249985 +0000 UTC m=+1223.554974073" observedRunningTime="2026-02-02 09:27:41.904588569 +0000 UTC m=+1224.838312657" watchObservedRunningTime="2026-02-02 09:27:41.908896756 +0000 UTC m=+1224.842620844" Feb 02 09:27:44 crc kubenswrapper[4764]: I0202 09:27:44.326058 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Feb 02 09:27:46 crc kubenswrapper[4764]: I0202 09:27:46.180797 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Feb 02 09:27:46 crc kubenswrapper[4764]: I0202 09:27:46.180861 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Feb 02 09:27:47 crc kubenswrapper[4764]: I0202 09:27:47.193097 4764 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="5bacef6a-a70e-400b-963f-d3346b916dee" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.181:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 02 09:27:47 crc kubenswrapper[4764]: I0202 09:27:47.193097 4764 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="5bacef6a-a70e-400b-963f-d3346b916dee" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.181:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 02 09:27:49 crc kubenswrapper[4764]: I0202 09:27:49.327873 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Feb 02 09:27:49 crc kubenswrapper[4764]: I0202 09:27:49.373511 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Feb 02 09:27:50 crc kubenswrapper[4764]: I0202 09:27:50.006497 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Feb 02 09:27:50 crc kubenswrapper[4764]: I0202 09:27:50.382341 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 02 09:27:50 crc kubenswrapper[4764]: I0202 09:27:50.382845 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 02 09:27:51 crc kubenswrapper[4764]: I0202 09:27:51.465324 4764 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="1e77cec7-7174-46ce-ba34-6facb5eac1c0" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.184:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 02 09:27:51 crc kubenswrapper[4764]: I0202 09:27:51.465346 4764 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="1e77cec7-7174-46ce-ba34-6facb5eac1c0" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.184:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 02 09:27:56 crc kubenswrapper[4764]: I0202 09:27:56.191131 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Feb 02 09:27:56 crc kubenswrapper[4764]: I0202 09:27:56.191858 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Feb 02 09:27:56 crc kubenswrapper[4764]: I0202 09:27:56.200481 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Feb 02 09:27:56 crc kubenswrapper[4764]: I0202 09:27:56.200757 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Feb 02 09:27:56 crc kubenswrapper[4764]: I0202 09:27:56.911000 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Feb 02 09:27:56 crc kubenswrapper[4764]: I0202 09:27:56.990497 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51b710f2-c87f-4d7f-8eda-a3f091f39f1d-combined-ca-bundle\") pod \"51b710f2-c87f-4d7f-8eda-a3f091f39f1d\" (UID: \"51b710f2-c87f-4d7f-8eda-a3f091f39f1d\") " Feb 02 09:27:56 crc kubenswrapper[4764]: I0202 09:27:56.990587 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51b710f2-c87f-4d7f-8eda-a3f091f39f1d-config-data\") pod \"51b710f2-c87f-4d7f-8eda-a3f091f39f1d\" (UID: \"51b710f2-c87f-4d7f-8eda-a3f091f39f1d\") " Feb 02 09:27:56 crc kubenswrapper[4764]: I0202 09:27:56.990634 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8999j\" (UniqueName: \"kubernetes.io/projected/51b710f2-c87f-4d7f-8eda-a3f091f39f1d-kube-api-access-8999j\") pod \"51b710f2-c87f-4d7f-8eda-a3f091f39f1d\" (UID: \"51b710f2-c87f-4d7f-8eda-a3f091f39f1d\") " Feb 02 09:27:56 crc kubenswrapper[4764]: I0202 09:27:56.999631 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/51b710f2-c87f-4d7f-8eda-a3f091f39f1d-kube-api-access-8999j" (OuterVolumeSpecName: "kube-api-access-8999j") pod "51b710f2-c87f-4d7f-8eda-a3f091f39f1d" (UID: "51b710f2-c87f-4d7f-8eda-a3f091f39f1d"). InnerVolumeSpecName "kube-api-access-8999j". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:27:57 crc kubenswrapper[4764]: I0202 09:27:57.016218 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51b710f2-c87f-4d7f-8eda-a3f091f39f1d-config-data" (OuterVolumeSpecName: "config-data") pod "51b710f2-c87f-4d7f-8eda-a3f091f39f1d" (UID: "51b710f2-c87f-4d7f-8eda-a3f091f39f1d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:27:57 crc kubenswrapper[4764]: I0202 09:27:57.031564 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51b710f2-c87f-4d7f-8eda-a3f091f39f1d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "51b710f2-c87f-4d7f-8eda-a3f091f39f1d" (UID: "51b710f2-c87f-4d7f-8eda-a3f091f39f1d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:27:57 crc kubenswrapper[4764]: I0202 09:27:57.045912 4764 generic.go:334] "Generic (PLEG): container finished" podID="51b710f2-c87f-4d7f-8eda-a3f091f39f1d" containerID="7ec1f78b5f30dc980d60bb86d95602550ea60909e3f0b5adad9b61f7efadbcb0" exitCode=137 Feb 02 09:27:57 crc kubenswrapper[4764]: I0202 09:27:57.047240 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Feb 02 09:27:57 crc kubenswrapper[4764]: I0202 09:27:57.048011 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"51b710f2-c87f-4d7f-8eda-a3f091f39f1d","Type":"ContainerDied","Data":"7ec1f78b5f30dc980d60bb86d95602550ea60909e3f0b5adad9b61f7efadbcb0"} Feb 02 09:27:57 crc kubenswrapper[4764]: I0202 09:27:57.048058 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"51b710f2-c87f-4d7f-8eda-a3f091f39f1d","Type":"ContainerDied","Data":"c84b66c7939231229aeb51b3e319c065cd437adcd46c7f28c5b2260ca2693737"} Feb 02 09:27:57 crc kubenswrapper[4764]: I0202 09:27:57.048077 4764 scope.go:117] "RemoveContainer" containerID="7ec1f78b5f30dc980d60bb86d95602550ea60909e3f0b5adad9b61f7efadbcb0" Feb 02 09:27:57 crc kubenswrapper[4764]: I0202 09:27:57.077770 4764 scope.go:117] "RemoveContainer" containerID="7ec1f78b5f30dc980d60bb86d95602550ea60909e3f0b5adad9b61f7efadbcb0" Feb 02 09:27:57 crc kubenswrapper[4764]: E0202 09:27:57.078226 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7ec1f78b5f30dc980d60bb86d95602550ea60909e3f0b5adad9b61f7efadbcb0\": container with ID starting with 7ec1f78b5f30dc980d60bb86d95602550ea60909e3f0b5adad9b61f7efadbcb0 not found: ID does not exist" containerID="7ec1f78b5f30dc980d60bb86d95602550ea60909e3f0b5adad9b61f7efadbcb0" Feb 02 09:27:57 crc kubenswrapper[4764]: I0202 09:27:57.078270 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7ec1f78b5f30dc980d60bb86d95602550ea60909e3f0b5adad9b61f7efadbcb0"} err="failed to get container status \"7ec1f78b5f30dc980d60bb86d95602550ea60909e3f0b5adad9b61f7efadbcb0\": rpc error: code = NotFound desc = could not find container \"7ec1f78b5f30dc980d60bb86d95602550ea60909e3f0b5adad9b61f7efadbcb0\": container with ID starting with 7ec1f78b5f30dc980d60bb86d95602550ea60909e3f0b5adad9b61f7efadbcb0 not found: ID does not exist" Feb 02 09:27:57 crc kubenswrapper[4764]: I0202 09:27:57.092513 4764 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51b710f2-c87f-4d7f-8eda-a3f091f39f1d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 09:27:57 crc kubenswrapper[4764]: I0202 09:27:57.092545 4764 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51b710f2-c87f-4d7f-8eda-a3f091f39f1d-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 09:27:57 crc kubenswrapper[4764]: I0202 09:27:57.092554 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8999j\" (UniqueName: \"kubernetes.io/projected/51b710f2-c87f-4d7f-8eda-a3f091f39f1d-kube-api-access-8999j\") on node \"crc\" DevicePath \"\"" Feb 02 09:27:57 crc kubenswrapper[4764]: I0202 09:27:57.097817 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 02 09:27:57 crc kubenswrapper[4764]: I0202 09:27:57.111442 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 02 09:27:57 crc kubenswrapper[4764]: I0202 09:27:57.120353 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 02 09:27:57 crc kubenswrapper[4764]: E0202 09:27:57.120846 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51b710f2-c87f-4d7f-8eda-a3f091f39f1d" containerName="nova-cell1-novncproxy-novncproxy" Feb 02 09:27:57 crc kubenswrapper[4764]: I0202 09:27:57.120862 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="51b710f2-c87f-4d7f-8eda-a3f091f39f1d" containerName="nova-cell1-novncproxy-novncproxy" Feb 02 09:27:57 crc kubenswrapper[4764]: I0202 09:27:57.121104 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="51b710f2-c87f-4d7f-8eda-a3f091f39f1d" containerName="nova-cell1-novncproxy-novncproxy" Feb 02 09:27:57 crc kubenswrapper[4764]: I0202 09:27:57.121852 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Feb 02 09:27:57 crc kubenswrapper[4764]: I0202 09:27:57.125893 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Feb 02 09:27:57 crc kubenswrapper[4764]: I0202 09:27:57.126124 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Feb 02 09:27:57 crc kubenswrapper[4764]: I0202 09:27:57.126171 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Feb 02 09:27:57 crc kubenswrapper[4764]: I0202 09:27:57.126870 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 02 09:27:57 crc kubenswrapper[4764]: I0202 09:27:57.194492 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41b7a57b-b6a0-4af7-9304-e4c89c8e3812-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"41b7a57b-b6a0-4af7-9304-e4c89c8e3812\") " pod="openstack/nova-cell1-novncproxy-0" Feb 02 09:27:57 crc kubenswrapper[4764]: I0202 09:27:57.195675 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/41b7a57b-b6a0-4af7-9304-e4c89c8e3812-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"41b7a57b-b6a0-4af7-9304-e4c89c8e3812\") " pod="openstack/nova-cell1-novncproxy-0" Feb 02 09:27:57 crc kubenswrapper[4764]: I0202 09:27:57.195803 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41b7a57b-b6a0-4af7-9304-e4c89c8e3812-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"41b7a57b-b6a0-4af7-9304-e4c89c8e3812\") " pod="openstack/nova-cell1-novncproxy-0" Feb 02 09:27:57 crc kubenswrapper[4764]: I0202 09:27:57.195913 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/41b7a57b-b6a0-4af7-9304-e4c89c8e3812-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"41b7a57b-b6a0-4af7-9304-e4c89c8e3812\") " pod="openstack/nova-cell1-novncproxy-0" Feb 02 09:27:57 crc kubenswrapper[4764]: I0202 09:27:57.196066 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p8x8w\" (UniqueName: \"kubernetes.io/projected/41b7a57b-b6a0-4af7-9304-e4c89c8e3812-kube-api-access-p8x8w\") pod \"nova-cell1-novncproxy-0\" (UID: \"41b7a57b-b6a0-4af7-9304-e4c89c8e3812\") " pod="openstack/nova-cell1-novncproxy-0" Feb 02 09:27:57 crc kubenswrapper[4764]: I0202 09:27:57.297951 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41b7a57b-b6a0-4af7-9304-e4c89c8e3812-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"41b7a57b-b6a0-4af7-9304-e4c89c8e3812\") " pod="openstack/nova-cell1-novncproxy-0" Feb 02 09:27:57 crc kubenswrapper[4764]: I0202 09:27:57.298032 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/41b7a57b-b6a0-4af7-9304-e4c89c8e3812-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"41b7a57b-b6a0-4af7-9304-e4c89c8e3812\") " pod="openstack/nova-cell1-novncproxy-0" Feb 02 09:27:57 crc kubenswrapper[4764]: I0202 09:27:57.298103 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p8x8w\" (UniqueName: \"kubernetes.io/projected/41b7a57b-b6a0-4af7-9304-e4c89c8e3812-kube-api-access-p8x8w\") pod \"nova-cell1-novncproxy-0\" (UID: \"41b7a57b-b6a0-4af7-9304-e4c89c8e3812\") " pod="openstack/nova-cell1-novncproxy-0" Feb 02 09:27:57 crc kubenswrapper[4764]: I0202 09:27:57.298186 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41b7a57b-b6a0-4af7-9304-e4c89c8e3812-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"41b7a57b-b6a0-4af7-9304-e4c89c8e3812\") " pod="openstack/nova-cell1-novncproxy-0" Feb 02 09:27:57 crc kubenswrapper[4764]: I0202 09:27:57.298224 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/41b7a57b-b6a0-4af7-9304-e4c89c8e3812-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"41b7a57b-b6a0-4af7-9304-e4c89c8e3812\") " pod="openstack/nova-cell1-novncproxy-0" Feb 02 09:27:57 crc kubenswrapper[4764]: I0202 09:27:57.303167 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/41b7a57b-b6a0-4af7-9304-e4c89c8e3812-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"41b7a57b-b6a0-4af7-9304-e4c89c8e3812\") " pod="openstack/nova-cell1-novncproxy-0" Feb 02 09:27:57 crc kubenswrapper[4764]: I0202 09:27:57.305849 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41b7a57b-b6a0-4af7-9304-e4c89c8e3812-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"41b7a57b-b6a0-4af7-9304-e4c89c8e3812\") " pod="openstack/nova-cell1-novncproxy-0" Feb 02 09:27:57 crc kubenswrapper[4764]: I0202 09:27:57.306444 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41b7a57b-b6a0-4af7-9304-e4c89c8e3812-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"41b7a57b-b6a0-4af7-9304-e4c89c8e3812\") " pod="openstack/nova-cell1-novncproxy-0" Feb 02 09:27:57 crc kubenswrapper[4764]: I0202 09:27:57.307866 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/41b7a57b-b6a0-4af7-9304-e4c89c8e3812-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"41b7a57b-b6a0-4af7-9304-e4c89c8e3812\") " pod="openstack/nova-cell1-novncproxy-0" Feb 02 09:27:57 crc kubenswrapper[4764]: I0202 09:27:57.313816 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p8x8w\" (UniqueName: \"kubernetes.io/projected/41b7a57b-b6a0-4af7-9304-e4c89c8e3812-kube-api-access-p8x8w\") pod \"nova-cell1-novncproxy-0\" (UID: \"41b7a57b-b6a0-4af7-9304-e4c89c8e3812\") " pod="openstack/nova-cell1-novncproxy-0" Feb 02 09:27:57 crc kubenswrapper[4764]: I0202 09:27:57.437183 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Feb 02 09:27:57 crc kubenswrapper[4764]: I0202 09:27:57.837186 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="51b710f2-c87f-4d7f-8eda-a3f091f39f1d" path="/var/lib/kubelet/pods/51b710f2-c87f-4d7f-8eda-a3f091f39f1d/volumes" Feb 02 09:27:57 crc kubenswrapper[4764]: I0202 09:27:57.944503 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 02 09:27:58 crc kubenswrapper[4764]: I0202 09:27:58.078170 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"41b7a57b-b6a0-4af7-9304-e4c89c8e3812","Type":"ContainerStarted","Data":"862b30627335656a8d63d6a8186f86abce7f9989d68a6109c9b60216417980a9"} Feb 02 09:27:59 crc kubenswrapper[4764]: I0202 09:27:59.089627 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"41b7a57b-b6a0-4af7-9304-e4c89c8e3812","Type":"ContainerStarted","Data":"c2686fd9358d3bc7a9ed19ac37877c8a20e838de9a03b2fd7ead1651a30afbe2"} Feb 02 09:27:59 crc kubenswrapper[4764]: I0202 09:27:59.123911 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.123890684 podStartE2EDuration="2.123890684s" podCreationTimestamp="2026-02-02 09:27:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:27:59.118333066 +0000 UTC m=+1242.052057174" watchObservedRunningTime="2026-02-02 09:27:59.123890684 +0000 UTC m=+1242.057614782" Feb 02 09:28:00 crc kubenswrapper[4764]: I0202 09:28:00.387243 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Feb 02 09:28:00 crc kubenswrapper[4764]: I0202 09:28:00.388624 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Feb 02 09:28:00 crc kubenswrapper[4764]: I0202 09:28:00.392424 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Feb 02 09:28:00 crc kubenswrapper[4764]: I0202 09:28:00.393613 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Feb 02 09:28:01 crc kubenswrapper[4764]: I0202 09:28:01.109981 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Feb 02 09:28:01 crc kubenswrapper[4764]: I0202 09:28:01.113882 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Feb 02 09:28:01 crc kubenswrapper[4764]: I0202 09:28:01.335700 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-75956b9989-5ddn6"] Feb 02 09:28:01 crc kubenswrapper[4764]: I0202 09:28:01.342737 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75956b9989-5ddn6" Feb 02 09:28:01 crc kubenswrapper[4764]: I0202 09:28:01.355162 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-75956b9989-5ddn6"] Feb 02 09:28:01 crc kubenswrapper[4764]: I0202 09:28:01.385000 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rhd8r\" (UniqueName: \"kubernetes.io/projected/ff8a4b58-1d3f-4f7e-a0ae-8d541024983b-kube-api-access-rhd8r\") pod \"dnsmasq-dns-75956b9989-5ddn6\" (UID: \"ff8a4b58-1d3f-4f7e-a0ae-8d541024983b\") " pod="openstack/dnsmasq-dns-75956b9989-5ddn6" Feb 02 09:28:01 crc kubenswrapper[4764]: I0202 09:28:01.385113 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ff8a4b58-1d3f-4f7e-a0ae-8d541024983b-config\") pod \"dnsmasq-dns-75956b9989-5ddn6\" (UID: \"ff8a4b58-1d3f-4f7e-a0ae-8d541024983b\") " pod="openstack/dnsmasq-dns-75956b9989-5ddn6" Feb 02 09:28:01 crc kubenswrapper[4764]: I0202 09:28:01.385146 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ff8a4b58-1d3f-4f7e-a0ae-8d541024983b-ovsdbserver-sb\") pod \"dnsmasq-dns-75956b9989-5ddn6\" (UID: \"ff8a4b58-1d3f-4f7e-a0ae-8d541024983b\") " pod="openstack/dnsmasq-dns-75956b9989-5ddn6" Feb 02 09:28:01 crc kubenswrapper[4764]: I0202 09:28:01.385181 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ff8a4b58-1d3f-4f7e-a0ae-8d541024983b-ovsdbserver-nb\") pod \"dnsmasq-dns-75956b9989-5ddn6\" (UID: \"ff8a4b58-1d3f-4f7e-a0ae-8d541024983b\") " pod="openstack/dnsmasq-dns-75956b9989-5ddn6" Feb 02 09:28:01 crc kubenswrapper[4764]: I0202 09:28:01.385229 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ff8a4b58-1d3f-4f7e-a0ae-8d541024983b-dns-svc\") pod \"dnsmasq-dns-75956b9989-5ddn6\" (UID: \"ff8a4b58-1d3f-4f7e-a0ae-8d541024983b\") " pod="openstack/dnsmasq-dns-75956b9989-5ddn6" Feb 02 09:28:01 crc kubenswrapper[4764]: I0202 09:28:01.487002 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ff8a4b58-1d3f-4f7e-a0ae-8d541024983b-ovsdbserver-nb\") pod \"dnsmasq-dns-75956b9989-5ddn6\" (UID: \"ff8a4b58-1d3f-4f7e-a0ae-8d541024983b\") " pod="openstack/dnsmasq-dns-75956b9989-5ddn6" Feb 02 09:28:01 crc kubenswrapper[4764]: I0202 09:28:01.487072 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ff8a4b58-1d3f-4f7e-a0ae-8d541024983b-dns-svc\") pod \"dnsmasq-dns-75956b9989-5ddn6\" (UID: \"ff8a4b58-1d3f-4f7e-a0ae-8d541024983b\") " pod="openstack/dnsmasq-dns-75956b9989-5ddn6" Feb 02 09:28:01 crc kubenswrapper[4764]: I0202 09:28:01.487108 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rhd8r\" (UniqueName: \"kubernetes.io/projected/ff8a4b58-1d3f-4f7e-a0ae-8d541024983b-kube-api-access-rhd8r\") pod \"dnsmasq-dns-75956b9989-5ddn6\" (UID: \"ff8a4b58-1d3f-4f7e-a0ae-8d541024983b\") " pod="openstack/dnsmasq-dns-75956b9989-5ddn6" Feb 02 09:28:01 crc kubenswrapper[4764]: I0202 09:28:01.487176 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ff8a4b58-1d3f-4f7e-a0ae-8d541024983b-config\") pod \"dnsmasq-dns-75956b9989-5ddn6\" (UID: \"ff8a4b58-1d3f-4f7e-a0ae-8d541024983b\") " pod="openstack/dnsmasq-dns-75956b9989-5ddn6" Feb 02 09:28:01 crc kubenswrapper[4764]: I0202 09:28:01.487201 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ff8a4b58-1d3f-4f7e-a0ae-8d541024983b-ovsdbserver-sb\") pod \"dnsmasq-dns-75956b9989-5ddn6\" (UID: \"ff8a4b58-1d3f-4f7e-a0ae-8d541024983b\") " pod="openstack/dnsmasq-dns-75956b9989-5ddn6" Feb 02 09:28:01 crc kubenswrapper[4764]: I0202 09:28:01.487974 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ff8a4b58-1d3f-4f7e-a0ae-8d541024983b-ovsdbserver-sb\") pod \"dnsmasq-dns-75956b9989-5ddn6\" (UID: \"ff8a4b58-1d3f-4f7e-a0ae-8d541024983b\") " pod="openstack/dnsmasq-dns-75956b9989-5ddn6" Feb 02 09:28:01 crc kubenswrapper[4764]: I0202 09:28:01.488448 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ff8a4b58-1d3f-4f7e-a0ae-8d541024983b-ovsdbserver-nb\") pod \"dnsmasq-dns-75956b9989-5ddn6\" (UID: \"ff8a4b58-1d3f-4f7e-a0ae-8d541024983b\") " pod="openstack/dnsmasq-dns-75956b9989-5ddn6" Feb 02 09:28:01 crc kubenswrapper[4764]: I0202 09:28:01.488923 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ff8a4b58-1d3f-4f7e-a0ae-8d541024983b-dns-svc\") pod \"dnsmasq-dns-75956b9989-5ddn6\" (UID: \"ff8a4b58-1d3f-4f7e-a0ae-8d541024983b\") " pod="openstack/dnsmasq-dns-75956b9989-5ddn6" Feb 02 09:28:01 crc kubenswrapper[4764]: I0202 09:28:01.489699 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ff8a4b58-1d3f-4f7e-a0ae-8d541024983b-config\") pod \"dnsmasq-dns-75956b9989-5ddn6\" (UID: \"ff8a4b58-1d3f-4f7e-a0ae-8d541024983b\") " pod="openstack/dnsmasq-dns-75956b9989-5ddn6" Feb 02 09:28:01 crc kubenswrapper[4764]: I0202 09:28:01.509665 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rhd8r\" (UniqueName: \"kubernetes.io/projected/ff8a4b58-1d3f-4f7e-a0ae-8d541024983b-kube-api-access-rhd8r\") pod \"dnsmasq-dns-75956b9989-5ddn6\" (UID: \"ff8a4b58-1d3f-4f7e-a0ae-8d541024983b\") " pod="openstack/dnsmasq-dns-75956b9989-5ddn6" Feb 02 09:28:01 crc kubenswrapper[4764]: I0202 09:28:01.667902 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75956b9989-5ddn6" Feb 02 09:28:02 crc kubenswrapper[4764]: I0202 09:28:02.206961 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-75956b9989-5ddn6"] Feb 02 09:28:02 crc kubenswrapper[4764]: I0202 09:28:02.438102 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Feb 02 09:28:03 crc kubenswrapper[4764]: I0202 09:28:03.125878 4764 generic.go:334] "Generic (PLEG): container finished" podID="ff8a4b58-1d3f-4f7e-a0ae-8d541024983b" containerID="228b569a80f432fd2a8b2ae069ff1ce409d3d7c8f71ec9287bcb0e687e02d457" exitCode=0 Feb 02 09:28:03 crc kubenswrapper[4764]: I0202 09:28:03.125985 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75956b9989-5ddn6" event={"ID":"ff8a4b58-1d3f-4f7e-a0ae-8d541024983b","Type":"ContainerDied","Data":"228b569a80f432fd2a8b2ae069ff1ce409d3d7c8f71ec9287bcb0e687e02d457"} Feb 02 09:28:03 crc kubenswrapper[4764]: I0202 09:28:03.126473 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75956b9989-5ddn6" event={"ID":"ff8a4b58-1d3f-4f7e-a0ae-8d541024983b","Type":"ContainerStarted","Data":"577ed8e752e8a2300366b2364f3476e9d999aa3f4e3da7456bef1b67f7fe5a5b"} Feb 02 09:28:03 crc kubenswrapper[4764]: I0202 09:28:03.446740 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 02 09:28:03 crc kubenswrapper[4764]: I0202 09:28:03.447396 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f6b65eed-dab4-4af5-ba48-c24a50e41e8f" containerName="proxy-httpd" containerID="cri-o://636542d714cd8b7f6c52eea15ef07a6fb8036880f4581d65ed9ac2288b62b5fc" gracePeriod=30 Feb 02 09:28:03 crc kubenswrapper[4764]: I0202 09:28:03.447415 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f6b65eed-dab4-4af5-ba48-c24a50e41e8f" containerName="sg-core" containerID="cri-o://a8a5259f33cae1431754c29e917f947668da51a1a4f49cc0e612e0f944314d7a" gracePeriod=30 Feb 02 09:28:03 crc kubenswrapper[4764]: I0202 09:28:03.447537 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f6b65eed-dab4-4af5-ba48-c24a50e41e8f" containerName="ceilometer-notification-agent" containerID="cri-o://cc615530b6652e1802c31a65ab4877751776f4e6415b68194b67c94d1075e4af" gracePeriod=30 Feb 02 09:28:03 crc kubenswrapper[4764]: I0202 09:28:03.447562 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f6b65eed-dab4-4af5-ba48-c24a50e41e8f" containerName="ceilometer-central-agent" containerID="cri-o://929ea7921f7027ef1b4f008d85641195d62dd7a0dbbcccbc94161ff9d6efa998" gracePeriod=30 Feb 02 09:28:03 crc kubenswrapper[4764]: I0202 09:28:03.462146 4764 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="f6b65eed-dab4-4af5-ba48-c24a50e41e8f" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.182:3000/\": read tcp 10.217.0.2:59902->10.217.0.182:3000: read: connection reset by peer" Feb 02 09:28:03 crc kubenswrapper[4764]: I0202 09:28:03.848313 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Feb 02 09:28:04 crc kubenswrapper[4764]: I0202 09:28:04.137293 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75956b9989-5ddn6" event={"ID":"ff8a4b58-1d3f-4f7e-a0ae-8d541024983b","Type":"ContainerStarted","Data":"ba78ef785372350834e28f350c85c6581e766f4c306e316726c548005b1ee000"} Feb 02 09:28:04 crc kubenswrapper[4764]: I0202 09:28:04.137386 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-75956b9989-5ddn6" Feb 02 09:28:04 crc kubenswrapper[4764]: I0202 09:28:04.141759 4764 generic.go:334] "Generic (PLEG): container finished" podID="f6b65eed-dab4-4af5-ba48-c24a50e41e8f" containerID="636542d714cd8b7f6c52eea15ef07a6fb8036880f4581d65ed9ac2288b62b5fc" exitCode=0 Feb 02 09:28:04 crc kubenswrapper[4764]: I0202 09:28:04.141792 4764 generic.go:334] "Generic (PLEG): container finished" podID="f6b65eed-dab4-4af5-ba48-c24a50e41e8f" containerID="a8a5259f33cae1431754c29e917f947668da51a1a4f49cc0e612e0f944314d7a" exitCode=2 Feb 02 09:28:04 crc kubenswrapper[4764]: I0202 09:28:04.141804 4764 generic.go:334] "Generic (PLEG): container finished" podID="f6b65eed-dab4-4af5-ba48-c24a50e41e8f" containerID="929ea7921f7027ef1b4f008d85641195d62dd7a0dbbcccbc94161ff9d6efa998" exitCode=0 Feb 02 09:28:04 crc kubenswrapper[4764]: I0202 09:28:04.141832 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f6b65eed-dab4-4af5-ba48-c24a50e41e8f","Type":"ContainerDied","Data":"636542d714cd8b7f6c52eea15ef07a6fb8036880f4581d65ed9ac2288b62b5fc"} Feb 02 09:28:04 crc kubenswrapper[4764]: I0202 09:28:04.141881 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f6b65eed-dab4-4af5-ba48-c24a50e41e8f","Type":"ContainerDied","Data":"a8a5259f33cae1431754c29e917f947668da51a1a4f49cc0e612e0f944314d7a"} Feb 02 09:28:04 crc kubenswrapper[4764]: I0202 09:28:04.141893 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f6b65eed-dab4-4af5-ba48-c24a50e41e8f","Type":"ContainerDied","Data":"929ea7921f7027ef1b4f008d85641195d62dd7a0dbbcccbc94161ff9d6efa998"} Feb 02 09:28:04 crc kubenswrapper[4764]: I0202 09:28:04.142085 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="1e77cec7-7174-46ce-ba34-6facb5eac1c0" containerName="nova-api-api" containerID="cri-o://78e56184f54e35c41bfd8b755dc12aff162742a4324e6e1530acf1915f448b32" gracePeriod=30 Feb 02 09:28:04 crc kubenswrapper[4764]: I0202 09:28:04.146749 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="1e77cec7-7174-46ce-ba34-6facb5eac1c0" containerName="nova-api-log" containerID="cri-o://e9c7c575d06ff5990cadd37986c09dbdc59f0d5a9a8b9a60535974f83fd45ead" gracePeriod=30 Feb 02 09:28:04 crc kubenswrapper[4764]: I0202 09:28:04.165547 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-75956b9989-5ddn6" podStartSLOduration=3.16553132 podStartE2EDuration="3.16553132s" podCreationTimestamp="2026-02-02 09:28:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:28:04.161670477 +0000 UTC m=+1247.095394565" watchObservedRunningTime="2026-02-02 09:28:04.16553132 +0000 UTC m=+1247.099255408" Feb 02 09:28:04 crc kubenswrapper[4764]: I0202 09:28:04.899333 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 02 09:28:04 crc kubenswrapper[4764]: I0202 09:28:04.975607 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f6b65eed-dab4-4af5-ba48-c24a50e41e8f-scripts\") pod \"f6b65eed-dab4-4af5-ba48-c24a50e41e8f\" (UID: \"f6b65eed-dab4-4af5-ba48-c24a50e41e8f\") " Feb 02 09:28:04 crc kubenswrapper[4764]: I0202 09:28:04.975653 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/f6b65eed-dab4-4af5-ba48-c24a50e41e8f-ceilometer-tls-certs\") pod \"f6b65eed-dab4-4af5-ba48-c24a50e41e8f\" (UID: \"f6b65eed-dab4-4af5-ba48-c24a50e41e8f\") " Feb 02 09:28:04 crc kubenswrapper[4764]: I0202 09:28:04.975677 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f6b65eed-dab4-4af5-ba48-c24a50e41e8f-run-httpd\") pod \"f6b65eed-dab4-4af5-ba48-c24a50e41e8f\" (UID: \"f6b65eed-dab4-4af5-ba48-c24a50e41e8f\") " Feb 02 09:28:04 crc kubenswrapper[4764]: I0202 09:28:04.975728 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f6b65eed-dab4-4af5-ba48-c24a50e41e8f-log-httpd\") pod \"f6b65eed-dab4-4af5-ba48-c24a50e41e8f\" (UID: \"f6b65eed-dab4-4af5-ba48-c24a50e41e8f\") " Feb 02 09:28:04 crc kubenswrapper[4764]: I0202 09:28:04.975801 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f6b65eed-dab4-4af5-ba48-c24a50e41e8f-sg-core-conf-yaml\") pod \"f6b65eed-dab4-4af5-ba48-c24a50e41e8f\" (UID: \"f6b65eed-dab4-4af5-ba48-c24a50e41e8f\") " Feb 02 09:28:04 crc kubenswrapper[4764]: I0202 09:28:04.975828 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dgxbk\" (UniqueName: \"kubernetes.io/projected/f6b65eed-dab4-4af5-ba48-c24a50e41e8f-kube-api-access-dgxbk\") pod \"f6b65eed-dab4-4af5-ba48-c24a50e41e8f\" (UID: \"f6b65eed-dab4-4af5-ba48-c24a50e41e8f\") " Feb 02 09:28:04 crc kubenswrapper[4764]: I0202 09:28:04.975852 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6b65eed-dab4-4af5-ba48-c24a50e41e8f-config-data\") pod \"f6b65eed-dab4-4af5-ba48-c24a50e41e8f\" (UID: \"f6b65eed-dab4-4af5-ba48-c24a50e41e8f\") " Feb 02 09:28:04 crc kubenswrapper[4764]: I0202 09:28:04.976046 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6b65eed-dab4-4af5-ba48-c24a50e41e8f-combined-ca-bundle\") pod \"f6b65eed-dab4-4af5-ba48-c24a50e41e8f\" (UID: \"f6b65eed-dab4-4af5-ba48-c24a50e41e8f\") " Feb 02 09:28:04 crc kubenswrapper[4764]: I0202 09:28:04.976267 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f6b65eed-dab4-4af5-ba48-c24a50e41e8f-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "f6b65eed-dab4-4af5-ba48-c24a50e41e8f" (UID: "f6b65eed-dab4-4af5-ba48-c24a50e41e8f"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 09:28:04 crc kubenswrapper[4764]: I0202 09:28:04.976599 4764 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f6b65eed-dab4-4af5-ba48-c24a50e41e8f-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 02 09:28:04 crc kubenswrapper[4764]: I0202 09:28:04.977027 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f6b65eed-dab4-4af5-ba48-c24a50e41e8f-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "f6b65eed-dab4-4af5-ba48-c24a50e41e8f" (UID: "f6b65eed-dab4-4af5-ba48-c24a50e41e8f"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 09:28:04 crc kubenswrapper[4764]: I0202 09:28:04.983224 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f6b65eed-dab4-4af5-ba48-c24a50e41e8f-kube-api-access-dgxbk" (OuterVolumeSpecName: "kube-api-access-dgxbk") pod "f6b65eed-dab4-4af5-ba48-c24a50e41e8f" (UID: "f6b65eed-dab4-4af5-ba48-c24a50e41e8f"). InnerVolumeSpecName "kube-api-access-dgxbk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:28:04 crc kubenswrapper[4764]: I0202 09:28:04.988831 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6b65eed-dab4-4af5-ba48-c24a50e41e8f-scripts" (OuterVolumeSpecName: "scripts") pod "f6b65eed-dab4-4af5-ba48-c24a50e41e8f" (UID: "f6b65eed-dab4-4af5-ba48-c24a50e41e8f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:28:05 crc kubenswrapper[4764]: I0202 09:28:05.032047 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6b65eed-dab4-4af5-ba48-c24a50e41e8f-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "f6b65eed-dab4-4af5-ba48-c24a50e41e8f" (UID: "f6b65eed-dab4-4af5-ba48-c24a50e41e8f"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:28:05 crc kubenswrapper[4764]: I0202 09:28:05.072833 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6b65eed-dab4-4af5-ba48-c24a50e41e8f-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "f6b65eed-dab4-4af5-ba48-c24a50e41e8f" (UID: "f6b65eed-dab4-4af5-ba48-c24a50e41e8f"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:28:05 crc kubenswrapper[4764]: I0202 09:28:05.078108 4764 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f6b65eed-dab4-4af5-ba48-c24a50e41e8f-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 09:28:05 crc kubenswrapper[4764]: I0202 09:28:05.078136 4764 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/f6b65eed-dab4-4af5-ba48-c24a50e41e8f-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 02 09:28:05 crc kubenswrapper[4764]: I0202 09:28:05.078146 4764 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f6b65eed-dab4-4af5-ba48-c24a50e41e8f-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 02 09:28:05 crc kubenswrapper[4764]: I0202 09:28:05.078155 4764 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f6b65eed-dab4-4af5-ba48-c24a50e41e8f-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 02 09:28:05 crc kubenswrapper[4764]: I0202 09:28:05.078170 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dgxbk\" (UniqueName: \"kubernetes.io/projected/f6b65eed-dab4-4af5-ba48-c24a50e41e8f-kube-api-access-dgxbk\") on node \"crc\" DevicePath \"\"" Feb 02 09:28:05 crc kubenswrapper[4764]: I0202 09:28:05.080601 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6b65eed-dab4-4af5-ba48-c24a50e41e8f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f6b65eed-dab4-4af5-ba48-c24a50e41e8f" (UID: "f6b65eed-dab4-4af5-ba48-c24a50e41e8f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:28:05 crc kubenswrapper[4764]: I0202 09:28:05.130014 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6b65eed-dab4-4af5-ba48-c24a50e41e8f-config-data" (OuterVolumeSpecName: "config-data") pod "f6b65eed-dab4-4af5-ba48-c24a50e41e8f" (UID: "f6b65eed-dab4-4af5-ba48-c24a50e41e8f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:28:05 crc kubenswrapper[4764]: I0202 09:28:05.150457 4764 generic.go:334] "Generic (PLEG): container finished" podID="1e77cec7-7174-46ce-ba34-6facb5eac1c0" containerID="e9c7c575d06ff5990cadd37986c09dbdc59f0d5a9a8b9a60535974f83fd45ead" exitCode=143 Feb 02 09:28:05 crc kubenswrapper[4764]: I0202 09:28:05.150562 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"1e77cec7-7174-46ce-ba34-6facb5eac1c0","Type":"ContainerDied","Data":"e9c7c575d06ff5990cadd37986c09dbdc59f0d5a9a8b9a60535974f83fd45ead"} Feb 02 09:28:05 crc kubenswrapper[4764]: I0202 09:28:05.154492 4764 generic.go:334] "Generic (PLEG): container finished" podID="f6b65eed-dab4-4af5-ba48-c24a50e41e8f" containerID="cc615530b6652e1802c31a65ab4877751776f4e6415b68194b67c94d1075e4af" exitCode=0 Feb 02 09:28:05 crc kubenswrapper[4764]: I0202 09:28:05.155274 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 02 09:28:05 crc kubenswrapper[4764]: I0202 09:28:05.155423 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f6b65eed-dab4-4af5-ba48-c24a50e41e8f","Type":"ContainerDied","Data":"cc615530b6652e1802c31a65ab4877751776f4e6415b68194b67c94d1075e4af"} Feb 02 09:28:05 crc kubenswrapper[4764]: I0202 09:28:05.155473 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f6b65eed-dab4-4af5-ba48-c24a50e41e8f","Type":"ContainerDied","Data":"11d2d903d7fe8be3d6c76d6ef8441093a875dcda0bbcc009736c64d582635e7d"} Feb 02 09:28:05 crc kubenswrapper[4764]: I0202 09:28:05.155492 4764 scope.go:117] "RemoveContainer" containerID="636542d714cd8b7f6c52eea15ef07a6fb8036880f4581d65ed9ac2288b62b5fc" Feb 02 09:28:05 crc kubenswrapper[4764]: I0202 09:28:05.176513 4764 scope.go:117] "RemoveContainer" containerID="a8a5259f33cae1431754c29e917f947668da51a1a4f49cc0e612e0f944314d7a" Feb 02 09:28:05 crc kubenswrapper[4764]: I0202 09:28:05.179194 4764 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6b65eed-dab4-4af5-ba48-c24a50e41e8f-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 09:28:05 crc kubenswrapper[4764]: I0202 09:28:05.179235 4764 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6b65eed-dab4-4af5-ba48-c24a50e41e8f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 09:28:05 crc kubenswrapper[4764]: I0202 09:28:05.200795 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 02 09:28:05 crc kubenswrapper[4764]: I0202 09:28:05.206279 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 02 09:28:05 crc kubenswrapper[4764]: I0202 09:28:05.206447 4764 scope.go:117] "RemoveContainer" containerID="cc615530b6652e1802c31a65ab4877751776f4e6415b68194b67c94d1075e4af" Feb 02 09:28:05 crc kubenswrapper[4764]: I0202 09:28:05.228125 4764 scope.go:117] "RemoveContainer" containerID="929ea7921f7027ef1b4f008d85641195d62dd7a0dbbcccbc94161ff9d6efa998" Feb 02 09:28:05 crc kubenswrapper[4764]: I0202 09:28:05.233628 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 02 09:28:05 crc kubenswrapper[4764]: E0202 09:28:05.234083 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6b65eed-dab4-4af5-ba48-c24a50e41e8f" containerName="ceilometer-notification-agent" Feb 02 09:28:05 crc kubenswrapper[4764]: I0202 09:28:05.234100 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6b65eed-dab4-4af5-ba48-c24a50e41e8f" containerName="ceilometer-notification-agent" Feb 02 09:28:05 crc kubenswrapper[4764]: E0202 09:28:05.234114 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6b65eed-dab4-4af5-ba48-c24a50e41e8f" containerName="proxy-httpd" Feb 02 09:28:05 crc kubenswrapper[4764]: I0202 09:28:05.234120 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6b65eed-dab4-4af5-ba48-c24a50e41e8f" containerName="proxy-httpd" Feb 02 09:28:05 crc kubenswrapper[4764]: E0202 09:28:05.234136 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6b65eed-dab4-4af5-ba48-c24a50e41e8f" containerName="sg-core" Feb 02 09:28:05 crc kubenswrapper[4764]: I0202 09:28:05.234142 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6b65eed-dab4-4af5-ba48-c24a50e41e8f" containerName="sg-core" Feb 02 09:28:05 crc kubenswrapper[4764]: E0202 09:28:05.234154 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6b65eed-dab4-4af5-ba48-c24a50e41e8f" containerName="ceilometer-central-agent" Feb 02 09:28:05 crc kubenswrapper[4764]: I0202 09:28:05.234160 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6b65eed-dab4-4af5-ba48-c24a50e41e8f" containerName="ceilometer-central-agent" Feb 02 09:28:05 crc kubenswrapper[4764]: I0202 09:28:05.234318 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6b65eed-dab4-4af5-ba48-c24a50e41e8f" containerName="proxy-httpd" Feb 02 09:28:05 crc kubenswrapper[4764]: I0202 09:28:05.234329 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6b65eed-dab4-4af5-ba48-c24a50e41e8f" containerName="ceilometer-notification-agent" Feb 02 09:28:05 crc kubenswrapper[4764]: I0202 09:28:05.234345 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6b65eed-dab4-4af5-ba48-c24a50e41e8f" containerName="sg-core" Feb 02 09:28:05 crc kubenswrapper[4764]: I0202 09:28:05.234362 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6b65eed-dab4-4af5-ba48-c24a50e41e8f" containerName="ceilometer-central-agent" Feb 02 09:28:05 crc kubenswrapper[4764]: I0202 09:28:05.235895 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 02 09:28:05 crc kubenswrapper[4764]: I0202 09:28:05.254511 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Feb 02 09:28:05 crc kubenswrapper[4764]: I0202 09:28:05.254733 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 02 09:28:05 crc kubenswrapper[4764]: I0202 09:28:05.254922 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 02 09:28:05 crc kubenswrapper[4764]: I0202 09:28:05.286206 4764 scope.go:117] "RemoveContainer" containerID="636542d714cd8b7f6c52eea15ef07a6fb8036880f4581d65ed9ac2288b62b5fc" Feb 02 09:28:05 crc kubenswrapper[4764]: E0202 09:28:05.287326 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"636542d714cd8b7f6c52eea15ef07a6fb8036880f4581d65ed9ac2288b62b5fc\": container with ID starting with 636542d714cd8b7f6c52eea15ef07a6fb8036880f4581d65ed9ac2288b62b5fc not found: ID does not exist" containerID="636542d714cd8b7f6c52eea15ef07a6fb8036880f4581d65ed9ac2288b62b5fc" Feb 02 09:28:05 crc kubenswrapper[4764]: I0202 09:28:05.287384 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"636542d714cd8b7f6c52eea15ef07a6fb8036880f4581d65ed9ac2288b62b5fc"} err="failed to get container status \"636542d714cd8b7f6c52eea15ef07a6fb8036880f4581d65ed9ac2288b62b5fc\": rpc error: code = NotFound desc = could not find container \"636542d714cd8b7f6c52eea15ef07a6fb8036880f4581d65ed9ac2288b62b5fc\": container with ID starting with 636542d714cd8b7f6c52eea15ef07a6fb8036880f4581d65ed9ac2288b62b5fc not found: ID does not exist" Feb 02 09:28:05 crc kubenswrapper[4764]: I0202 09:28:05.287491 4764 scope.go:117] "RemoveContainer" containerID="a8a5259f33cae1431754c29e917f947668da51a1a4f49cc0e612e0f944314d7a" Feb 02 09:28:05 crc kubenswrapper[4764]: I0202 09:28:05.287544 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b-scripts\") pod \"ceilometer-0\" (UID: \"1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b\") " pod="openstack/ceilometer-0" Feb 02 09:28:05 crc kubenswrapper[4764]: I0202 09:28:05.287591 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b-config-data\") pod \"ceilometer-0\" (UID: \"1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b\") " pod="openstack/ceilometer-0" Feb 02 09:28:05 crc kubenswrapper[4764]: I0202 09:28:05.287635 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b\") " pod="openstack/ceilometer-0" Feb 02 09:28:05 crc kubenswrapper[4764]: I0202 09:28:05.287751 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b\") " pod="openstack/ceilometer-0" Feb 02 09:28:05 crc kubenswrapper[4764]: I0202 09:28:05.287840 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b-run-httpd\") pod \"ceilometer-0\" (UID: \"1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b\") " pod="openstack/ceilometer-0" Feb 02 09:28:05 crc kubenswrapper[4764]: I0202 09:28:05.287884 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b\") " pod="openstack/ceilometer-0" Feb 02 09:28:05 crc kubenswrapper[4764]: I0202 09:28:05.288225 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b-log-httpd\") pod \"ceilometer-0\" (UID: \"1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b\") " pod="openstack/ceilometer-0" Feb 02 09:28:05 crc kubenswrapper[4764]: I0202 09:28:05.288281 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nnrwj\" (UniqueName: \"kubernetes.io/projected/1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b-kube-api-access-nnrwj\") pod \"ceilometer-0\" (UID: \"1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b\") " pod="openstack/ceilometer-0" Feb 02 09:28:05 crc kubenswrapper[4764]: E0202 09:28:05.289345 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a8a5259f33cae1431754c29e917f947668da51a1a4f49cc0e612e0f944314d7a\": container with ID starting with a8a5259f33cae1431754c29e917f947668da51a1a4f49cc0e612e0f944314d7a not found: ID does not exist" containerID="a8a5259f33cae1431754c29e917f947668da51a1a4f49cc0e612e0f944314d7a" Feb 02 09:28:05 crc kubenswrapper[4764]: I0202 09:28:05.289381 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a8a5259f33cae1431754c29e917f947668da51a1a4f49cc0e612e0f944314d7a"} err="failed to get container status \"a8a5259f33cae1431754c29e917f947668da51a1a4f49cc0e612e0f944314d7a\": rpc error: code = NotFound desc = could not find container \"a8a5259f33cae1431754c29e917f947668da51a1a4f49cc0e612e0f944314d7a\": container with ID starting with a8a5259f33cae1431754c29e917f947668da51a1a4f49cc0e612e0f944314d7a not found: ID does not exist" Feb 02 09:28:05 crc kubenswrapper[4764]: I0202 09:28:05.289404 4764 scope.go:117] "RemoveContainer" containerID="cc615530b6652e1802c31a65ab4877751776f4e6415b68194b67c94d1075e4af" Feb 02 09:28:05 crc kubenswrapper[4764]: E0202 09:28:05.290222 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cc615530b6652e1802c31a65ab4877751776f4e6415b68194b67c94d1075e4af\": container with ID starting with cc615530b6652e1802c31a65ab4877751776f4e6415b68194b67c94d1075e4af not found: ID does not exist" containerID="cc615530b6652e1802c31a65ab4877751776f4e6415b68194b67c94d1075e4af" Feb 02 09:28:05 crc kubenswrapper[4764]: I0202 09:28:05.290269 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cc615530b6652e1802c31a65ab4877751776f4e6415b68194b67c94d1075e4af"} err="failed to get container status \"cc615530b6652e1802c31a65ab4877751776f4e6415b68194b67c94d1075e4af\": rpc error: code = NotFound desc = could not find container \"cc615530b6652e1802c31a65ab4877751776f4e6415b68194b67c94d1075e4af\": container with ID starting with cc615530b6652e1802c31a65ab4877751776f4e6415b68194b67c94d1075e4af not found: ID does not exist" Feb 02 09:28:05 crc kubenswrapper[4764]: I0202 09:28:05.290297 4764 scope.go:117] "RemoveContainer" containerID="929ea7921f7027ef1b4f008d85641195d62dd7a0dbbcccbc94161ff9d6efa998" Feb 02 09:28:05 crc kubenswrapper[4764]: E0202 09:28:05.290797 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"929ea7921f7027ef1b4f008d85641195d62dd7a0dbbcccbc94161ff9d6efa998\": container with ID starting with 929ea7921f7027ef1b4f008d85641195d62dd7a0dbbcccbc94161ff9d6efa998 not found: ID does not exist" containerID="929ea7921f7027ef1b4f008d85641195d62dd7a0dbbcccbc94161ff9d6efa998" Feb 02 09:28:05 crc kubenswrapper[4764]: I0202 09:28:05.290842 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"929ea7921f7027ef1b4f008d85641195d62dd7a0dbbcccbc94161ff9d6efa998"} err="failed to get container status \"929ea7921f7027ef1b4f008d85641195d62dd7a0dbbcccbc94161ff9d6efa998\": rpc error: code = NotFound desc = could not find container \"929ea7921f7027ef1b4f008d85641195d62dd7a0dbbcccbc94161ff9d6efa998\": container with ID starting with 929ea7921f7027ef1b4f008d85641195d62dd7a0dbbcccbc94161ff9d6efa998 not found: ID does not exist" Feb 02 09:28:05 crc kubenswrapper[4764]: I0202 09:28:05.303987 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 02 09:28:05 crc kubenswrapper[4764]: I0202 09:28:05.389679 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b-scripts\") pod \"ceilometer-0\" (UID: \"1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b\") " pod="openstack/ceilometer-0" Feb 02 09:28:05 crc kubenswrapper[4764]: I0202 09:28:05.389724 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b-config-data\") pod \"ceilometer-0\" (UID: \"1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b\") " pod="openstack/ceilometer-0" Feb 02 09:28:05 crc kubenswrapper[4764]: I0202 09:28:05.390284 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b\") " pod="openstack/ceilometer-0" Feb 02 09:28:05 crc kubenswrapper[4764]: I0202 09:28:05.390324 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b\") " pod="openstack/ceilometer-0" Feb 02 09:28:05 crc kubenswrapper[4764]: I0202 09:28:05.390351 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b-run-httpd\") pod \"ceilometer-0\" (UID: \"1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b\") " pod="openstack/ceilometer-0" Feb 02 09:28:05 crc kubenswrapper[4764]: I0202 09:28:05.390368 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b\") " pod="openstack/ceilometer-0" Feb 02 09:28:05 crc kubenswrapper[4764]: I0202 09:28:05.390392 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b-log-httpd\") pod \"ceilometer-0\" (UID: \"1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b\") " pod="openstack/ceilometer-0" Feb 02 09:28:05 crc kubenswrapper[4764]: I0202 09:28:05.390425 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nnrwj\" (UniqueName: \"kubernetes.io/projected/1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b-kube-api-access-nnrwj\") pod \"ceilometer-0\" (UID: \"1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b\") " pod="openstack/ceilometer-0" Feb 02 09:28:05 crc kubenswrapper[4764]: I0202 09:28:05.390844 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b-run-httpd\") pod \"ceilometer-0\" (UID: \"1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b\") " pod="openstack/ceilometer-0" Feb 02 09:28:05 crc kubenswrapper[4764]: I0202 09:28:05.391197 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b-log-httpd\") pod \"ceilometer-0\" (UID: \"1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b\") " pod="openstack/ceilometer-0" Feb 02 09:28:05 crc kubenswrapper[4764]: I0202 09:28:05.394223 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b-scripts\") pod \"ceilometer-0\" (UID: \"1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b\") " pod="openstack/ceilometer-0" Feb 02 09:28:05 crc kubenswrapper[4764]: I0202 09:28:05.394408 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b-config-data\") pod \"ceilometer-0\" (UID: \"1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b\") " pod="openstack/ceilometer-0" Feb 02 09:28:05 crc kubenswrapper[4764]: I0202 09:28:05.394465 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b\") " pod="openstack/ceilometer-0" Feb 02 09:28:05 crc kubenswrapper[4764]: I0202 09:28:05.394639 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b\") " pod="openstack/ceilometer-0" Feb 02 09:28:05 crc kubenswrapper[4764]: I0202 09:28:05.395713 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b\") " pod="openstack/ceilometer-0" Feb 02 09:28:05 crc kubenswrapper[4764]: I0202 09:28:05.412876 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nnrwj\" (UniqueName: \"kubernetes.io/projected/1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b-kube-api-access-nnrwj\") pod \"ceilometer-0\" (UID: \"1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b\") " pod="openstack/ceilometer-0" Feb 02 09:28:05 crc kubenswrapper[4764]: I0202 09:28:05.609787 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 02 09:28:05 crc kubenswrapper[4764]: I0202 09:28:05.843008 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f6b65eed-dab4-4af5-ba48-c24a50e41e8f" path="/var/lib/kubelet/pods/f6b65eed-dab4-4af5-ba48-c24a50e41e8f/volumes" Feb 02 09:28:05 crc kubenswrapper[4764]: I0202 09:28:05.865137 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 02 09:28:06 crc kubenswrapper[4764]: I0202 09:28:06.087483 4764 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 02 09:28:06 crc kubenswrapper[4764]: I0202 09:28:06.093598 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 02 09:28:06 crc kubenswrapper[4764]: I0202 09:28:06.163563 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b","Type":"ContainerStarted","Data":"6393d8044c93b4c6e3a68a44bdb8d25929b4028b7437bd38490d9f7d8d13885c"} Feb 02 09:28:07 crc kubenswrapper[4764]: I0202 09:28:07.176316 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b","Type":"ContainerStarted","Data":"37d77489aeed723c2937bac1982f551a1511c02cf3c8c7f6ec85edd4de3181f1"} Feb 02 09:28:07 crc kubenswrapper[4764]: I0202 09:28:07.437788 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Feb 02 09:28:07 crc kubenswrapper[4764]: I0202 09:28:07.456490 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Feb 02 09:28:07 crc kubenswrapper[4764]: I0202 09:28:07.653270 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 02 09:28:07 crc kubenswrapper[4764]: I0202 09:28:07.732808 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-78xx2\" (UniqueName: \"kubernetes.io/projected/1e77cec7-7174-46ce-ba34-6facb5eac1c0-kube-api-access-78xx2\") pod \"1e77cec7-7174-46ce-ba34-6facb5eac1c0\" (UID: \"1e77cec7-7174-46ce-ba34-6facb5eac1c0\") " Feb 02 09:28:07 crc kubenswrapper[4764]: I0202 09:28:07.733054 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e77cec7-7174-46ce-ba34-6facb5eac1c0-combined-ca-bundle\") pod \"1e77cec7-7174-46ce-ba34-6facb5eac1c0\" (UID: \"1e77cec7-7174-46ce-ba34-6facb5eac1c0\") " Feb 02 09:28:07 crc kubenswrapper[4764]: I0202 09:28:07.733522 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1e77cec7-7174-46ce-ba34-6facb5eac1c0-logs" (OuterVolumeSpecName: "logs") pod "1e77cec7-7174-46ce-ba34-6facb5eac1c0" (UID: "1e77cec7-7174-46ce-ba34-6facb5eac1c0"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 09:28:07 crc kubenswrapper[4764]: I0202 09:28:07.733666 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1e77cec7-7174-46ce-ba34-6facb5eac1c0-logs\") pod \"1e77cec7-7174-46ce-ba34-6facb5eac1c0\" (UID: \"1e77cec7-7174-46ce-ba34-6facb5eac1c0\") " Feb 02 09:28:07 crc kubenswrapper[4764]: I0202 09:28:07.733776 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e77cec7-7174-46ce-ba34-6facb5eac1c0-config-data\") pod \"1e77cec7-7174-46ce-ba34-6facb5eac1c0\" (UID: \"1e77cec7-7174-46ce-ba34-6facb5eac1c0\") " Feb 02 09:28:07 crc kubenswrapper[4764]: I0202 09:28:07.734176 4764 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1e77cec7-7174-46ce-ba34-6facb5eac1c0-logs\") on node \"crc\" DevicePath \"\"" Feb 02 09:28:07 crc kubenswrapper[4764]: I0202 09:28:07.740701 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1e77cec7-7174-46ce-ba34-6facb5eac1c0-kube-api-access-78xx2" (OuterVolumeSpecName: "kube-api-access-78xx2") pod "1e77cec7-7174-46ce-ba34-6facb5eac1c0" (UID: "1e77cec7-7174-46ce-ba34-6facb5eac1c0"). InnerVolumeSpecName "kube-api-access-78xx2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:28:07 crc kubenswrapper[4764]: I0202 09:28:07.774053 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e77cec7-7174-46ce-ba34-6facb5eac1c0-config-data" (OuterVolumeSpecName: "config-data") pod "1e77cec7-7174-46ce-ba34-6facb5eac1c0" (UID: "1e77cec7-7174-46ce-ba34-6facb5eac1c0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:28:07 crc kubenswrapper[4764]: I0202 09:28:07.794012 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e77cec7-7174-46ce-ba34-6facb5eac1c0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1e77cec7-7174-46ce-ba34-6facb5eac1c0" (UID: "1e77cec7-7174-46ce-ba34-6facb5eac1c0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:28:07 crc kubenswrapper[4764]: I0202 09:28:07.835925 4764 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e77cec7-7174-46ce-ba34-6facb5eac1c0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 09:28:07 crc kubenswrapper[4764]: I0202 09:28:07.835966 4764 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e77cec7-7174-46ce-ba34-6facb5eac1c0-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 09:28:07 crc kubenswrapper[4764]: I0202 09:28:07.835976 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-78xx2\" (UniqueName: \"kubernetes.io/projected/1e77cec7-7174-46ce-ba34-6facb5eac1c0-kube-api-access-78xx2\") on node \"crc\" DevicePath \"\"" Feb 02 09:28:07 crc kubenswrapper[4764]: E0202 09:28:07.931413 4764 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1e77cec7_7174_46ce_ba34_6facb5eac1c0.slice/crio-4f7d1d1eb33324193a509c0b70d4c195c170987b5e32f212f779b8c41c86d0fd\": RecentStats: unable to find data in memory cache]" Feb 02 09:28:08 crc kubenswrapper[4764]: I0202 09:28:08.191143 4764 generic.go:334] "Generic (PLEG): container finished" podID="1e77cec7-7174-46ce-ba34-6facb5eac1c0" containerID="78e56184f54e35c41bfd8b755dc12aff162742a4324e6e1530acf1915f448b32" exitCode=0 Feb 02 09:28:08 crc kubenswrapper[4764]: I0202 09:28:08.191202 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"1e77cec7-7174-46ce-ba34-6facb5eac1c0","Type":"ContainerDied","Data":"78e56184f54e35c41bfd8b755dc12aff162742a4324e6e1530acf1915f448b32"} Feb 02 09:28:08 crc kubenswrapper[4764]: I0202 09:28:08.191444 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"1e77cec7-7174-46ce-ba34-6facb5eac1c0","Type":"ContainerDied","Data":"4f7d1d1eb33324193a509c0b70d4c195c170987b5e32f212f779b8c41c86d0fd"} Feb 02 09:28:08 crc kubenswrapper[4764]: I0202 09:28:08.191277 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 02 09:28:08 crc kubenswrapper[4764]: I0202 09:28:08.191468 4764 scope.go:117] "RemoveContainer" containerID="78e56184f54e35c41bfd8b755dc12aff162742a4324e6e1530acf1915f448b32" Feb 02 09:28:08 crc kubenswrapper[4764]: I0202 09:28:08.194304 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b","Type":"ContainerStarted","Data":"7f419b8f8922d19daee559647678b9398eeab3b0ea628415a1e0020220e120e5"} Feb 02 09:28:08 crc kubenswrapper[4764]: I0202 09:28:08.194331 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b","Type":"ContainerStarted","Data":"505c759195911603bb96aa1c52c9f6906f245c02c4c7dbbce9a3c65acf37f246"} Feb 02 09:28:08 crc kubenswrapper[4764]: I0202 09:28:08.222252 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Feb 02 09:28:08 crc kubenswrapper[4764]: I0202 09:28:08.232689 4764 scope.go:117] "RemoveContainer" containerID="e9c7c575d06ff5990cadd37986c09dbdc59f0d5a9a8b9a60535974f83fd45ead" Feb 02 09:28:08 crc kubenswrapper[4764]: I0202 09:28:08.235994 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Feb 02 09:28:08 crc kubenswrapper[4764]: I0202 09:28:08.243841 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Feb 02 09:28:08 crc kubenswrapper[4764]: I0202 09:28:08.280071 4764 scope.go:117] "RemoveContainer" containerID="78e56184f54e35c41bfd8b755dc12aff162742a4324e6e1530acf1915f448b32" Feb 02 09:28:08 crc kubenswrapper[4764]: E0202 09:28:08.282537 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"78e56184f54e35c41bfd8b755dc12aff162742a4324e6e1530acf1915f448b32\": container with ID starting with 78e56184f54e35c41bfd8b755dc12aff162742a4324e6e1530acf1915f448b32 not found: ID does not exist" containerID="78e56184f54e35c41bfd8b755dc12aff162742a4324e6e1530acf1915f448b32" Feb 02 09:28:08 crc kubenswrapper[4764]: I0202 09:28:08.282567 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"78e56184f54e35c41bfd8b755dc12aff162742a4324e6e1530acf1915f448b32"} err="failed to get container status \"78e56184f54e35c41bfd8b755dc12aff162742a4324e6e1530acf1915f448b32\": rpc error: code = NotFound desc = could not find container \"78e56184f54e35c41bfd8b755dc12aff162742a4324e6e1530acf1915f448b32\": container with ID starting with 78e56184f54e35c41bfd8b755dc12aff162742a4324e6e1530acf1915f448b32 not found: ID does not exist" Feb 02 09:28:08 crc kubenswrapper[4764]: I0202 09:28:08.282588 4764 scope.go:117] "RemoveContainer" containerID="e9c7c575d06ff5990cadd37986c09dbdc59f0d5a9a8b9a60535974f83fd45ead" Feb 02 09:28:08 crc kubenswrapper[4764]: E0202 09:28:08.288091 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e9c7c575d06ff5990cadd37986c09dbdc59f0d5a9a8b9a60535974f83fd45ead\": container with ID starting with e9c7c575d06ff5990cadd37986c09dbdc59f0d5a9a8b9a60535974f83fd45ead not found: ID does not exist" containerID="e9c7c575d06ff5990cadd37986c09dbdc59f0d5a9a8b9a60535974f83fd45ead" Feb 02 09:28:08 crc kubenswrapper[4764]: I0202 09:28:08.288121 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e9c7c575d06ff5990cadd37986c09dbdc59f0d5a9a8b9a60535974f83fd45ead"} err="failed to get container status \"e9c7c575d06ff5990cadd37986c09dbdc59f0d5a9a8b9a60535974f83fd45ead\": rpc error: code = NotFound desc = could not find container \"e9c7c575d06ff5990cadd37986c09dbdc59f0d5a9a8b9a60535974f83fd45ead\": container with ID starting with e9c7c575d06ff5990cadd37986c09dbdc59f0d5a9a8b9a60535974f83fd45ead not found: ID does not exist" Feb 02 09:28:08 crc kubenswrapper[4764]: I0202 09:28:08.290375 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Feb 02 09:28:08 crc kubenswrapper[4764]: E0202 09:28:08.290790 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e77cec7-7174-46ce-ba34-6facb5eac1c0" containerName="nova-api-api" Feb 02 09:28:08 crc kubenswrapper[4764]: I0202 09:28:08.290807 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e77cec7-7174-46ce-ba34-6facb5eac1c0" containerName="nova-api-api" Feb 02 09:28:08 crc kubenswrapper[4764]: E0202 09:28:08.290830 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e77cec7-7174-46ce-ba34-6facb5eac1c0" containerName="nova-api-log" Feb 02 09:28:08 crc kubenswrapper[4764]: I0202 09:28:08.290839 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e77cec7-7174-46ce-ba34-6facb5eac1c0" containerName="nova-api-log" Feb 02 09:28:08 crc kubenswrapper[4764]: I0202 09:28:08.291002 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e77cec7-7174-46ce-ba34-6facb5eac1c0" containerName="nova-api-log" Feb 02 09:28:08 crc kubenswrapper[4764]: I0202 09:28:08.291019 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e77cec7-7174-46ce-ba34-6facb5eac1c0" containerName="nova-api-api" Feb 02 09:28:08 crc kubenswrapper[4764]: I0202 09:28:08.291916 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 02 09:28:08 crc kubenswrapper[4764]: I0202 09:28:08.301595 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Feb 02 09:28:08 crc kubenswrapper[4764]: I0202 09:28:08.301917 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Feb 02 09:28:08 crc kubenswrapper[4764]: I0202 09:28:08.302176 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Feb 02 09:28:08 crc kubenswrapper[4764]: I0202 09:28:08.344112 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ff2c8e9-62a5-4e5a-93e8-8a82957bcece-config-data\") pod \"nova-api-0\" (UID: \"2ff2c8e9-62a5-4e5a-93e8-8a82957bcece\") " pod="openstack/nova-api-0" Feb 02 09:28:08 crc kubenswrapper[4764]: I0202 09:28:08.344444 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2ff2c8e9-62a5-4e5a-93e8-8a82957bcece-internal-tls-certs\") pod \"nova-api-0\" (UID: \"2ff2c8e9-62a5-4e5a-93e8-8a82957bcece\") " pod="openstack/nova-api-0" Feb 02 09:28:08 crc kubenswrapper[4764]: I0202 09:28:08.344495 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2ff2c8e9-62a5-4e5a-93e8-8a82957bcece-logs\") pod \"nova-api-0\" (UID: \"2ff2c8e9-62a5-4e5a-93e8-8a82957bcece\") " pod="openstack/nova-api-0" Feb 02 09:28:08 crc kubenswrapper[4764]: I0202 09:28:08.344520 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2ff2c8e9-62a5-4e5a-93e8-8a82957bcece-public-tls-certs\") pod \"nova-api-0\" (UID: \"2ff2c8e9-62a5-4e5a-93e8-8a82957bcece\") " pod="openstack/nova-api-0" Feb 02 09:28:08 crc kubenswrapper[4764]: I0202 09:28:08.344570 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vvq5m\" (UniqueName: \"kubernetes.io/projected/2ff2c8e9-62a5-4e5a-93e8-8a82957bcece-kube-api-access-vvq5m\") pod \"nova-api-0\" (UID: \"2ff2c8e9-62a5-4e5a-93e8-8a82957bcece\") " pod="openstack/nova-api-0" Feb 02 09:28:08 crc kubenswrapper[4764]: I0202 09:28:08.344601 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ff2c8e9-62a5-4e5a-93e8-8a82957bcece-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"2ff2c8e9-62a5-4e5a-93e8-8a82957bcece\") " pod="openstack/nova-api-0" Feb 02 09:28:08 crc kubenswrapper[4764]: I0202 09:28:08.352800 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 02 09:28:08 crc kubenswrapper[4764]: I0202 09:28:08.448147 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ff2c8e9-62a5-4e5a-93e8-8a82957bcece-config-data\") pod \"nova-api-0\" (UID: \"2ff2c8e9-62a5-4e5a-93e8-8a82957bcece\") " pod="openstack/nova-api-0" Feb 02 09:28:08 crc kubenswrapper[4764]: I0202 09:28:08.448962 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2ff2c8e9-62a5-4e5a-93e8-8a82957bcece-internal-tls-certs\") pod \"nova-api-0\" (UID: \"2ff2c8e9-62a5-4e5a-93e8-8a82957bcece\") " pod="openstack/nova-api-0" Feb 02 09:28:08 crc kubenswrapper[4764]: I0202 09:28:08.449022 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2ff2c8e9-62a5-4e5a-93e8-8a82957bcece-logs\") pod \"nova-api-0\" (UID: \"2ff2c8e9-62a5-4e5a-93e8-8a82957bcece\") " pod="openstack/nova-api-0" Feb 02 09:28:08 crc kubenswrapper[4764]: I0202 09:28:08.449043 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2ff2c8e9-62a5-4e5a-93e8-8a82957bcece-public-tls-certs\") pod \"nova-api-0\" (UID: \"2ff2c8e9-62a5-4e5a-93e8-8a82957bcece\") " pod="openstack/nova-api-0" Feb 02 09:28:08 crc kubenswrapper[4764]: I0202 09:28:08.449073 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vvq5m\" (UniqueName: \"kubernetes.io/projected/2ff2c8e9-62a5-4e5a-93e8-8a82957bcece-kube-api-access-vvq5m\") pod \"nova-api-0\" (UID: \"2ff2c8e9-62a5-4e5a-93e8-8a82957bcece\") " pod="openstack/nova-api-0" Feb 02 09:28:08 crc kubenswrapper[4764]: I0202 09:28:08.449104 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ff2c8e9-62a5-4e5a-93e8-8a82957bcece-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"2ff2c8e9-62a5-4e5a-93e8-8a82957bcece\") " pod="openstack/nova-api-0" Feb 02 09:28:08 crc kubenswrapper[4764]: I0202 09:28:08.449895 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2ff2c8e9-62a5-4e5a-93e8-8a82957bcece-logs\") pod \"nova-api-0\" (UID: \"2ff2c8e9-62a5-4e5a-93e8-8a82957bcece\") " pod="openstack/nova-api-0" Feb 02 09:28:08 crc kubenswrapper[4764]: I0202 09:28:08.455663 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ff2c8e9-62a5-4e5a-93e8-8a82957bcece-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"2ff2c8e9-62a5-4e5a-93e8-8a82957bcece\") " pod="openstack/nova-api-0" Feb 02 09:28:08 crc kubenswrapper[4764]: I0202 09:28:08.465634 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2ff2c8e9-62a5-4e5a-93e8-8a82957bcece-internal-tls-certs\") pod \"nova-api-0\" (UID: \"2ff2c8e9-62a5-4e5a-93e8-8a82957bcece\") " pod="openstack/nova-api-0" Feb 02 09:28:08 crc kubenswrapper[4764]: I0202 09:28:08.467856 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2ff2c8e9-62a5-4e5a-93e8-8a82957bcece-public-tls-certs\") pod \"nova-api-0\" (UID: \"2ff2c8e9-62a5-4e5a-93e8-8a82957bcece\") " pod="openstack/nova-api-0" Feb 02 09:28:08 crc kubenswrapper[4764]: I0202 09:28:08.475576 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ff2c8e9-62a5-4e5a-93e8-8a82957bcece-config-data\") pod \"nova-api-0\" (UID: \"2ff2c8e9-62a5-4e5a-93e8-8a82957bcece\") " pod="openstack/nova-api-0" Feb 02 09:28:08 crc kubenswrapper[4764]: I0202 09:28:08.527537 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vvq5m\" (UniqueName: \"kubernetes.io/projected/2ff2c8e9-62a5-4e5a-93e8-8a82957bcece-kube-api-access-vvq5m\") pod \"nova-api-0\" (UID: \"2ff2c8e9-62a5-4e5a-93e8-8a82957bcece\") " pod="openstack/nova-api-0" Feb 02 09:28:08 crc kubenswrapper[4764]: I0202 09:28:08.642185 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 02 09:28:08 crc kubenswrapper[4764]: I0202 09:28:08.712792 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-q58jt"] Feb 02 09:28:08 crc kubenswrapper[4764]: I0202 09:28:08.713924 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-q58jt" Feb 02 09:28:08 crc kubenswrapper[4764]: I0202 09:28:08.716013 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Feb 02 09:28:08 crc kubenswrapper[4764]: I0202 09:28:08.720998 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Feb 02 09:28:08 crc kubenswrapper[4764]: I0202 09:28:08.728519 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-q58jt"] Feb 02 09:28:08 crc kubenswrapper[4764]: I0202 09:28:08.763292 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f83a9caf-7a27-4adf-90ea-072142c914f9-config-data\") pod \"nova-cell1-cell-mapping-q58jt\" (UID: \"f83a9caf-7a27-4adf-90ea-072142c914f9\") " pod="openstack/nova-cell1-cell-mapping-q58jt" Feb 02 09:28:08 crc kubenswrapper[4764]: I0202 09:28:08.763330 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f83a9caf-7a27-4adf-90ea-072142c914f9-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-q58jt\" (UID: \"f83a9caf-7a27-4adf-90ea-072142c914f9\") " pod="openstack/nova-cell1-cell-mapping-q58jt" Feb 02 09:28:08 crc kubenswrapper[4764]: I0202 09:28:08.763378 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f83a9caf-7a27-4adf-90ea-072142c914f9-scripts\") pod \"nova-cell1-cell-mapping-q58jt\" (UID: \"f83a9caf-7a27-4adf-90ea-072142c914f9\") " pod="openstack/nova-cell1-cell-mapping-q58jt" Feb 02 09:28:08 crc kubenswrapper[4764]: I0202 09:28:08.763477 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bxfzn\" (UniqueName: \"kubernetes.io/projected/f83a9caf-7a27-4adf-90ea-072142c914f9-kube-api-access-bxfzn\") pod \"nova-cell1-cell-mapping-q58jt\" (UID: \"f83a9caf-7a27-4adf-90ea-072142c914f9\") " pod="openstack/nova-cell1-cell-mapping-q58jt" Feb 02 09:28:08 crc kubenswrapper[4764]: I0202 09:28:08.867063 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f83a9caf-7a27-4adf-90ea-072142c914f9-scripts\") pod \"nova-cell1-cell-mapping-q58jt\" (UID: \"f83a9caf-7a27-4adf-90ea-072142c914f9\") " pod="openstack/nova-cell1-cell-mapping-q58jt" Feb 02 09:28:08 crc kubenswrapper[4764]: I0202 09:28:08.868367 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bxfzn\" (UniqueName: \"kubernetes.io/projected/f83a9caf-7a27-4adf-90ea-072142c914f9-kube-api-access-bxfzn\") pod \"nova-cell1-cell-mapping-q58jt\" (UID: \"f83a9caf-7a27-4adf-90ea-072142c914f9\") " pod="openstack/nova-cell1-cell-mapping-q58jt" Feb 02 09:28:08 crc kubenswrapper[4764]: I0202 09:28:08.868430 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f83a9caf-7a27-4adf-90ea-072142c914f9-config-data\") pod \"nova-cell1-cell-mapping-q58jt\" (UID: \"f83a9caf-7a27-4adf-90ea-072142c914f9\") " pod="openstack/nova-cell1-cell-mapping-q58jt" Feb 02 09:28:08 crc kubenswrapper[4764]: I0202 09:28:08.868466 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f83a9caf-7a27-4adf-90ea-072142c914f9-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-q58jt\" (UID: \"f83a9caf-7a27-4adf-90ea-072142c914f9\") " pod="openstack/nova-cell1-cell-mapping-q58jt" Feb 02 09:28:08 crc kubenswrapper[4764]: I0202 09:28:08.871285 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f83a9caf-7a27-4adf-90ea-072142c914f9-scripts\") pod \"nova-cell1-cell-mapping-q58jt\" (UID: \"f83a9caf-7a27-4adf-90ea-072142c914f9\") " pod="openstack/nova-cell1-cell-mapping-q58jt" Feb 02 09:28:08 crc kubenswrapper[4764]: I0202 09:28:08.874482 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f83a9caf-7a27-4adf-90ea-072142c914f9-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-q58jt\" (UID: \"f83a9caf-7a27-4adf-90ea-072142c914f9\") " pod="openstack/nova-cell1-cell-mapping-q58jt" Feb 02 09:28:08 crc kubenswrapper[4764]: I0202 09:28:08.875232 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f83a9caf-7a27-4adf-90ea-072142c914f9-config-data\") pod \"nova-cell1-cell-mapping-q58jt\" (UID: \"f83a9caf-7a27-4adf-90ea-072142c914f9\") " pod="openstack/nova-cell1-cell-mapping-q58jt" Feb 02 09:28:08 crc kubenswrapper[4764]: I0202 09:28:08.885070 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bxfzn\" (UniqueName: \"kubernetes.io/projected/f83a9caf-7a27-4adf-90ea-072142c914f9-kube-api-access-bxfzn\") pod \"nova-cell1-cell-mapping-q58jt\" (UID: \"f83a9caf-7a27-4adf-90ea-072142c914f9\") " pod="openstack/nova-cell1-cell-mapping-q58jt" Feb 02 09:28:09 crc kubenswrapper[4764]: I0202 09:28:09.052166 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-q58jt" Feb 02 09:28:09 crc kubenswrapper[4764]: I0202 09:28:09.182320 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 02 09:28:09 crc kubenswrapper[4764]: W0202 09:28:09.188948 4764 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2ff2c8e9_62a5_4e5a_93e8_8a82957bcece.slice/crio-e42f3294ec437c06b9fba240dfafed74ccd91751df00eebfde730972b7fb61b6 WatchSource:0}: Error finding container e42f3294ec437c06b9fba240dfafed74ccd91751df00eebfde730972b7fb61b6: Status 404 returned error can't find the container with id e42f3294ec437c06b9fba240dfafed74ccd91751df00eebfde730972b7fb61b6 Feb 02 09:28:09 crc kubenswrapper[4764]: I0202 09:28:09.502907 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-q58jt"] Feb 02 09:28:09 crc kubenswrapper[4764]: W0202 09:28:09.512720 4764 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf83a9caf_7a27_4adf_90ea_072142c914f9.slice/crio-2741080fa42b3ef53cbe3b139cb0a2ca59a00fb52fa751b7d4b0e917afc3b462 WatchSource:0}: Error finding container 2741080fa42b3ef53cbe3b139cb0a2ca59a00fb52fa751b7d4b0e917afc3b462: Status 404 returned error can't find the container with id 2741080fa42b3ef53cbe3b139cb0a2ca59a00fb52fa751b7d4b0e917afc3b462 Feb 02 09:28:09 crc kubenswrapper[4764]: I0202 09:28:09.840097 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1e77cec7-7174-46ce-ba34-6facb5eac1c0" path="/var/lib/kubelet/pods/1e77cec7-7174-46ce-ba34-6facb5eac1c0/volumes" Feb 02 09:28:10 crc kubenswrapper[4764]: I0202 09:28:10.223741 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-q58jt" event={"ID":"f83a9caf-7a27-4adf-90ea-072142c914f9","Type":"ContainerStarted","Data":"11f32a8c2d61df5ff6d2656eb4ba0d2fe0897343eea8c23f62baa26ab81563c5"} Feb 02 09:28:10 crc kubenswrapper[4764]: I0202 09:28:10.224064 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-q58jt" event={"ID":"f83a9caf-7a27-4adf-90ea-072142c914f9","Type":"ContainerStarted","Data":"2741080fa42b3ef53cbe3b139cb0a2ca59a00fb52fa751b7d4b0e917afc3b462"} Feb 02 09:28:10 crc kubenswrapper[4764]: I0202 09:28:10.229186 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2ff2c8e9-62a5-4e5a-93e8-8a82957bcece","Type":"ContainerStarted","Data":"6e6e8f093d6566dd63a9af8b076e024cf869b1c69f6366fd544397ce51b91669"} Feb 02 09:28:10 crc kubenswrapper[4764]: I0202 09:28:10.229250 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2ff2c8e9-62a5-4e5a-93e8-8a82957bcece","Type":"ContainerStarted","Data":"281c60052638c3fe057a661ce1953d611a63021984e7458c31a49065277d5bbb"} Feb 02 09:28:10 crc kubenswrapper[4764]: I0202 09:28:10.229270 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2ff2c8e9-62a5-4e5a-93e8-8a82957bcece","Type":"ContainerStarted","Data":"e42f3294ec437c06b9fba240dfafed74ccd91751df00eebfde730972b7fb61b6"} Feb 02 09:28:10 crc kubenswrapper[4764]: I0202 09:28:10.257221 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-q58jt" podStartSLOduration=2.257202953 podStartE2EDuration="2.257202953s" podCreationTimestamp="2026-02-02 09:28:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:28:10.247617987 +0000 UTC m=+1253.181342075" watchObservedRunningTime="2026-02-02 09:28:10.257202953 +0000 UTC m=+1253.190927041" Feb 02 09:28:10 crc kubenswrapper[4764]: I0202 09:28:10.328353 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.328325268 podStartE2EDuration="2.328325268s" podCreationTimestamp="2026-02-02 09:28:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:28:10.306010713 +0000 UTC m=+1253.239734821" watchObservedRunningTime="2026-02-02 09:28:10.328325268 +0000 UTC m=+1253.262049356" Feb 02 09:28:11 crc kubenswrapper[4764]: I0202 09:28:11.241152 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b" containerName="ceilometer-central-agent" containerID="cri-o://37d77489aeed723c2937bac1982f551a1511c02cf3c8c7f6ec85edd4de3181f1" gracePeriod=30 Feb 02 09:28:11 crc kubenswrapper[4764]: I0202 09:28:11.241870 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b","Type":"ContainerStarted","Data":"35079f5ebbbb42df91e138a36f0a8863357759ec735c7b6cd16cc29f28a95881"} Feb 02 09:28:11 crc kubenswrapper[4764]: I0202 09:28:11.243068 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 02 09:28:11 crc kubenswrapper[4764]: I0202 09:28:11.243302 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b" containerName="proxy-httpd" containerID="cri-o://35079f5ebbbb42df91e138a36f0a8863357759ec735c7b6cd16cc29f28a95881" gracePeriod=30 Feb 02 09:28:11 crc kubenswrapper[4764]: I0202 09:28:11.243344 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b" containerName="sg-core" containerID="cri-o://7f419b8f8922d19daee559647678b9398eeab3b0ea628415a1e0020220e120e5" gracePeriod=30 Feb 02 09:28:11 crc kubenswrapper[4764]: I0202 09:28:11.243375 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b" containerName="ceilometer-notification-agent" containerID="cri-o://505c759195911603bb96aa1c52c9f6906f245c02c4c7dbbce9a3c65acf37f246" gracePeriod=30 Feb 02 09:28:11 crc kubenswrapper[4764]: I0202 09:28:11.276814 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.129495285 podStartE2EDuration="6.27679801s" podCreationTimestamp="2026-02-02 09:28:05 +0000 UTC" firstStartedPulling="2026-02-02 09:28:06.087285695 +0000 UTC m=+1249.021009783" lastFinishedPulling="2026-02-02 09:28:10.23458842 +0000 UTC m=+1253.168312508" observedRunningTime="2026-02-02 09:28:11.267808171 +0000 UTC m=+1254.201532269" watchObservedRunningTime="2026-02-02 09:28:11.27679801 +0000 UTC m=+1254.210522098" Feb 02 09:28:11 crc kubenswrapper[4764]: I0202 09:28:11.670794 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-75956b9989-5ddn6" Feb 02 09:28:11 crc kubenswrapper[4764]: I0202 09:28:11.727545 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-65b8795f57-z4kh6"] Feb 02 09:28:11 crc kubenswrapper[4764]: I0202 09:28:11.727768 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-65b8795f57-z4kh6" podUID="39a2a1fd-e483-46ec-807d-7a8b63f4d403" containerName="dnsmasq-dns" containerID="cri-o://a1b11461d74d94922fb1108c376a1ebcd01bd6511e487c221b592c97deda502d" gracePeriod=10 Feb 02 09:28:13 crc kubenswrapper[4764]: I0202 09:28:13.065097 4764 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-65b8795f57-z4kh6" podUID="39a2a1fd-e483-46ec-807d-7a8b63f4d403" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.176:5353: connect: connection refused" Feb 02 09:28:13 crc kubenswrapper[4764]: E0202 09:28:13.088928 4764 kubelet.go:2526] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="1.264s" Feb 02 09:28:13 crc kubenswrapper[4764]: I0202 09:28:13.156031 4764 generic.go:334] "Generic (PLEG): container finished" podID="1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b" containerID="35079f5ebbbb42df91e138a36f0a8863357759ec735c7b6cd16cc29f28a95881" exitCode=0 Feb 02 09:28:13 crc kubenswrapper[4764]: I0202 09:28:13.156066 4764 generic.go:334] "Generic (PLEG): container finished" podID="1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b" containerID="7f419b8f8922d19daee559647678b9398eeab3b0ea628415a1e0020220e120e5" exitCode=2 Feb 02 09:28:13 crc kubenswrapper[4764]: I0202 09:28:13.156086 4764 generic.go:334] "Generic (PLEG): container finished" podID="1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b" containerID="505c759195911603bb96aa1c52c9f6906f245c02c4c7dbbce9a3c65acf37f246" exitCode=0 Feb 02 09:28:13 crc kubenswrapper[4764]: I0202 09:28:13.156112 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b","Type":"ContainerDied","Data":"35079f5ebbbb42df91e138a36f0a8863357759ec735c7b6cd16cc29f28a95881"} Feb 02 09:28:13 crc kubenswrapper[4764]: I0202 09:28:13.156143 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b","Type":"ContainerDied","Data":"7f419b8f8922d19daee559647678b9398eeab3b0ea628415a1e0020220e120e5"} Feb 02 09:28:13 crc kubenswrapper[4764]: I0202 09:28:13.156156 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b","Type":"ContainerDied","Data":"505c759195911603bb96aa1c52c9f6906f245c02c4c7dbbce9a3c65acf37f246"} Feb 02 09:28:13 crc kubenswrapper[4764]: I0202 09:28:13.554771 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-65b8795f57-z4kh6" Feb 02 09:28:13 crc kubenswrapper[4764]: I0202 09:28:13.578944 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/39a2a1fd-e483-46ec-807d-7a8b63f4d403-ovsdbserver-nb\") pod \"39a2a1fd-e483-46ec-807d-7a8b63f4d403\" (UID: \"39a2a1fd-e483-46ec-807d-7a8b63f4d403\") " Feb 02 09:28:13 crc kubenswrapper[4764]: I0202 09:28:13.579109 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/39a2a1fd-e483-46ec-807d-7a8b63f4d403-dns-svc\") pod \"39a2a1fd-e483-46ec-807d-7a8b63f4d403\" (UID: \"39a2a1fd-e483-46ec-807d-7a8b63f4d403\") " Feb 02 09:28:13 crc kubenswrapper[4764]: I0202 09:28:13.579175 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/39a2a1fd-e483-46ec-807d-7a8b63f4d403-config\") pod \"39a2a1fd-e483-46ec-807d-7a8b63f4d403\" (UID: \"39a2a1fd-e483-46ec-807d-7a8b63f4d403\") " Feb 02 09:28:13 crc kubenswrapper[4764]: I0202 09:28:13.579227 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2lhpm\" (UniqueName: \"kubernetes.io/projected/39a2a1fd-e483-46ec-807d-7a8b63f4d403-kube-api-access-2lhpm\") pod \"39a2a1fd-e483-46ec-807d-7a8b63f4d403\" (UID: \"39a2a1fd-e483-46ec-807d-7a8b63f4d403\") " Feb 02 09:28:13 crc kubenswrapper[4764]: I0202 09:28:13.579250 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/39a2a1fd-e483-46ec-807d-7a8b63f4d403-ovsdbserver-sb\") pod \"39a2a1fd-e483-46ec-807d-7a8b63f4d403\" (UID: \"39a2a1fd-e483-46ec-807d-7a8b63f4d403\") " Feb 02 09:28:13 crc kubenswrapper[4764]: I0202 09:28:13.625558 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/39a2a1fd-e483-46ec-807d-7a8b63f4d403-kube-api-access-2lhpm" (OuterVolumeSpecName: "kube-api-access-2lhpm") pod "39a2a1fd-e483-46ec-807d-7a8b63f4d403" (UID: "39a2a1fd-e483-46ec-807d-7a8b63f4d403"). InnerVolumeSpecName "kube-api-access-2lhpm". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:28:13 crc kubenswrapper[4764]: I0202 09:28:13.649476 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 02 09:28:13 crc kubenswrapper[4764]: I0202 09:28:13.682748 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b-sg-core-conf-yaml\") pod \"1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b\" (UID: \"1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b\") " Feb 02 09:28:13 crc kubenswrapper[4764]: I0202 09:28:13.693881 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b-scripts\") pod \"1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b\" (UID: \"1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b\") " Feb 02 09:28:13 crc kubenswrapper[4764]: I0202 09:28:13.694385 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b-run-httpd\") pod \"1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b\" (UID: \"1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b\") " Feb 02 09:28:13 crc kubenswrapper[4764]: I0202 09:28:13.694526 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b-combined-ca-bundle\") pod \"1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b\" (UID: \"1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b\") " Feb 02 09:28:13 crc kubenswrapper[4764]: I0202 09:28:13.694654 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b-config-data\") pod \"1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b\" (UID: \"1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b\") " Feb 02 09:28:13 crc kubenswrapper[4764]: I0202 09:28:13.694767 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b-log-httpd\") pod \"1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b\" (UID: \"1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b\") " Feb 02 09:28:13 crc kubenswrapper[4764]: I0202 09:28:13.694877 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nnrwj\" (UniqueName: \"kubernetes.io/projected/1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b-kube-api-access-nnrwj\") pod \"1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b\" (UID: \"1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b\") " Feb 02 09:28:13 crc kubenswrapper[4764]: I0202 09:28:13.695021 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b-ceilometer-tls-certs\") pod \"1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b\" (UID: \"1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b\") " Feb 02 09:28:13 crc kubenswrapper[4764]: I0202 09:28:13.696173 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2lhpm\" (UniqueName: \"kubernetes.io/projected/39a2a1fd-e483-46ec-807d-7a8b63f4d403-kube-api-access-2lhpm\") on node \"crc\" DevicePath \"\"" Feb 02 09:28:13 crc kubenswrapper[4764]: I0202 09:28:13.695264 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b" (UID: "1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 09:28:13 crc kubenswrapper[4764]: I0202 09:28:13.695286 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b" (UID: "1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 09:28:13 crc kubenswrapper[4764]: I0202 09:28:13.697914 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/39a2a1fd-e483-46ec-807d-7a8b63f4d403-config" (OuterVolumeSpecName: "config") pod "39a2a1fd-e483-46ec-807d-7a8b63f4d403" (UID: "39a2a1fd-e483-46ec-807d-7a8b63f4d403"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:28:13 crc kubenswrapper[4764]: I0202 09:28:13.704091 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b-scripts" (OuterVolumeSpecName: "scripts") pod "1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b" (UID: "1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:28:13 crc kubenswrapper[4764]: I0202 09:28:13.710442 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b-kube-api-access-nnrwj" (OuterVolumeSpecName: "kube-api-access-nnrwj") pod "1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b" (UID: "1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b"). InnerVolumeSpecName "kube-api-access-nnrwj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:28:13 crc kubenswrapper[4764]: I0202 09:28:13.719323 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b" (UID: "1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:28:13 crc kubenswrapper[4764]: I0202 09:28:13.719624 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/39a2a1fd-e483-46ec-807d-7a8b63f4d403-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "39a2a1fd-e483-46ec-807d-7a8b63f4d403" (UID: "39a2a1fd-e483-46ec-807d-7a8b63f4d403"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:28:13 crc kubenswrapper[4764]: I0202 09:28:13.721403 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/39a2a1fd-e483-46ec-807d-7a8b63f4d403-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "39a2a1fd-e483-46ec-807d-7a8b63f4d403" (UID: "39a2a1fd-e483-46ec-807d-7a8b63f4d403"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:28:13 crc kubenswrapper[4764]: I0202 09:28:13.735500 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/39a2a1fd-e483-46ec-807d-7a8b63f4d403-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "39a2a1fd-e483-46ec-807d-7a8b63f4d403" (UID: "39a2a1fd-e483-46ec-807d-7a8b63f4d403"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:28:13 crc kubenswrapper[4764]: I0202 09:28:13.751693 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b" (UID: "1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:28:13 crc kubenswrapper[4764]: I0202 09:28:13.770532 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b" (UID: "1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:28:13 crc kubenswrapper[4764]: I0202 09:28:13.787412 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b-config-data" (OuterVolumeSpecName: "config-data") pod "1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b" (UID: "1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:28:13 crc kubenswrapper[4764]: I0202 09:28:13.798759 4764 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 02 09:28:13 crc kubenswrapper[4764]: I0202 09:28:13.798788 4764 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 09:28:13 crc kubenswrapper[4764]: I0202 09:28:13.798799 4764 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/39a2a1fd-e483-46ec-807d-7a8b63f4d403-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 02 09:28:13 crc kubenswrapper[4764]: I0202 09:28:13.798808 4764 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/39a2a1fd-e483-46ec-807d-7a8b63f4d403-config\") on node \"crc\" DevicePath \"\"" Feb 02 09:28:13 crc kubenswrapper[4764]: I0202 09:28:13.798817 4764 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 02 09:28:13 crc kubenswrapper[4764]: I0202 09:28:13.798825 4764 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/39a2a1fd-e483-46ec-807d-7a8b63f4d403-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 02 09:28:13 crc kubenswrapper[4764]: I0202 09:28:13.798833 4764 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 09:28:13 crc kubenswrapper[4764]: I0202 09:28:13.798841 4764 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 09:28:13 crc kubenswrapper[4764]: I0202 09:28:13.798848 4764 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 02 09:28:13 crc kubenswrapper[4764]: I0202 09:28:13.798857 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nnrwj\" (UniqueName: \"kubernetes.io/projected/1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b-kube-api-access-nnrwj\") on node \"crc\" DevicePath \"\"" Feb 02 09:28:13 crc kubenswrapper[4764]: I0202 09:28:13.798869 4764 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 02 09:28:13 crc kubenswrapper[4764]: I0202 09:28:13.798876 4764 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/39a2a1fd-e483-46ec-807d-7a8b63f4d403-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 02 09:28:14 crc kubenswrapper[4764]: I0202 09:28:14.168626 4764 generic.go:334] "Generic (PLEG): container finished" podID="39a2a1fd-e483-46ec-807d-7a8b63f4d403" containerID="a1b11461d74d94922fb1108c376a1ebcd01bd6511e487c221b592c97deda502d" exitCode=0 Feb 02 09:28:14 crc kubenswrapper[4764]: I0202 09:28:14.168823 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-65b8795f57-z4kh6" Feb 02 09:28:14 crc kubenswrapper[4764]: I0202 09:28:14.170429 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-65b8795f57-z4kh6" event={"ID":"39a2a1fd-e483-46ec-807d-7a8b63f4d403","Type":"ContainerDied","Data":"a1b11461d74d94922fb1108c376a1ebcd01bd6511e487c221b592c97deda502d"} Feb 02 09:28:14 crc kubenswrapper[4764]: I0202 09:28:14.170474 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-65b8795f57-z4kh6" event={"ID":"39a2a1fd-e483-46ec-807d-7a8b63f4d403","Type":"ContainerDied","Data":"3ecf47785943539ce4dde17602b85eabb4236aa1a8b1df77f23f50a2db37f417"} Feb 02 09:28:14 crc kubenswrapper[4764]: I0202 09:28:14.171073 4764 scope.go:117] "RemoveContainer" containerID="a1b11461d74d94922fb1108c376a1ebcd01bd6511e487c221b592c97deda502d" Feb 02 09:28:14 crc kubenswrapper[4764]: I0202 09:28:14.177448 4764 generic.go:334] "Generic (PLEG): container finished" podID="1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b" containerID="37d77489aeed723c2937bac1982f551a1511c02cf3c8c7f6ec85edd4de3181f1" exitCode=0 Feb 02 09:28:14 crc kubenswrapper[4764]: I0202 09:28:14.177512 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 02 09:28:14 crc kubenswrapper[4764]: I0202 09:28:14.177550 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b","Type":"ContainerDied","Data":"37d77489aeed723c2937bac1982f551a1511c02cf3c8c7f6ec85edd4de3181f1"} Feb 02 09:28:14 crc kubenswrapper[4764]: I0202 09:28:14.178125 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b","Type":"ContainerDied","Data":"6393d8044c93b4c6e3a68a44bdb8d25929b4028b7437bd38490d9f7d8d13885c"} Feb 02 09:28:14 crc kubenswrapper[4764]: I0202 09:28:14.216031 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 02 09:28:14 crc kubenswrapper[4764]: I0202 09:28:14.234153 4764 scope.go:117] "RemoveContainer" containerID="53dec1d4f771e36dacae86b0a9391c587b248da8664f3f08dd707ad478ced95d" Feb 02 09:28:14 crc kubenswrapper[4764]: I0202 09:28:14.249148 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 02 09:28:14 crc kubenswrapper[4764]: I0202 09:28:14.262618 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-65b8795f57-z4kh6"] Feb 02 09:28:14 crc kubenswrapper[4764]: I0202 09:28:14.294164 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-65b8795f57-z4kh6"] Feb 02 09:28:14 crc kubenswrapper[4764]: I0202 09:28:14.295762 4764 scope.go:117] "RemoveContainer" containerID="a1b11461d74d94922fb1108c376a1ebcd01bd6511e487c221b592c97deda502d" Feb 02 09:28:14 crc kubenswrapper[4764]: E0202 09:28:14.296349 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a1b11461d74d94922fb1108c376a1ebcd01bd6511e487c221b592c97deda502d\": container with ID starting with a1b11461d74d94922fb1108c376a1ebcd01bd6511e487c221b592c97deda502d not found: ID does not exist" containerID="a1b11461d74d94922fb1108c376a1ebcd01bd6511e487c221b592c97deda502d" Feb 02 09:28:14 crc kubenswrapper[4764]: I0202 09:28:14.296416 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a1b11461d74d94922fb1108c376a1ebcd01bd6511e487c221b592c97deda502d"} err="failed to get container status \"a1b11461d74d94922fb1108c376a1ebcd01bd6511e487c221b592c97deda502d\": rpc error: code = NotFound desc = could not find container \"a1b11461d74d94922fb1108c376a1ebcd01bd6511e487c221b592c97deda502d\": container with ID starting with a1b11461d74d94922fb1108c376a1ebcd01bd6511e487c221b592c97deda502d not found: ID does not exist" Feb 02 09:28:14 crc kubenswrapper[4764]: I0202 09:28:14.296442 4764 scope.go:117] "RemoveContainer" containerID="53dec1d4f771e36dacae86b0a9391c587b248da8664f3f08dd707ad478ced95d" Feb 02 09:28:14 crc kubenswrapper[4764]: E0202 09:28:14.296756 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"53dec1d4f771e36dacae86b0a9391c587b248da8664f3f08dd707ad478ced95d\": container with ID starting with 53dec1d4f771e36dacae86b0a9391c587b248da8664f3f08dd707ad478ced95d not found: ID does not exist" containerID="53dec1d4f771e36dacae86b0a9391c587b248da8664f3f08dd707ad478ced95d" Feb 02 09:28:14 crc kubenswrapper[4764]: I0202 09:28:14.297040 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"53dec1d4f771e36dacae86b0a9391c587b248da8664f3f08dd707ad478ced95d"} err="failed to get container status \"53dec1d4f771e36dacae86b0a9391c587b248da8664f3f08dd707ad478ced95d\": rpc error: code = NotFound desc = could not find container \"53dec1d4f771e36dacae86b0a9391c587b248da8664f3f08dd707ad478ced95d\": container with ID starting with 53dec1d4f771e36dacae86b0a9391c587b248da8664f3f08dd707ad478ced95d not found: ID does not exist" Feb 02 09:28:14 crc kubenswrapper[4764]: I0202 09:28:14.297068 4764 scope.go:117] "RemoveContainer" containerID="35079f5ebbbb42df91e138a36f0a8863357759ec735c7b6cd16cc29f28a95881" Feb 02 09:28:14 crc kubenswrapper[4764]: I0202 09:28:14.303460 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 02 09:28:14 crc kubenswrapper[4764]: E0202 09:28:14.303839 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39a2a1fd-e483-46ec-807d-7a8b63f4d403" containerName="dnsmasq-dns" Feb 02 09:28:14 crc kubenswrapper[4764]: I0202 09:28:14.303865 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="39a2a1fd-e483-46ec-807d-7a8b63f4d403" containerName="dnsmasq-dns" Feb 02 09:28:14 crc kubenswrapper[4764]: E0202 09:28:14.303880 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b" containerName="ceilometer-notification-agent" Feb 02 09:28:14 crc kubenswrapper[4764]: I0202 09:28:14.303887 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b" containerName="ceilometer-notification-agent" Feb 02 09:28:14 crc kubenswrapper[4764]: E0202 09:28:14.303897 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b" containerName="proxy-httpd" Feb 02 09:28:14 crc kubenswrapper[4764]: I0202 09:28:14.303903 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b" containerName="proxy-httpd" Feb 02 09:28:14 crc kubenswrapper[4764]: E0202 09:28:14.303923 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39a2a1fd-e483-46ec-807d-7a8b63f4d403" containerName="init" Feb 02 09:28:14 crc kubenswrapper[4764]: I0202 09:28:14.303929 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="39a2a1fd-e483-46ec-807d-7a8b63f4d403" containerName="init" Feb 02 09:28:14 crc kubenswrapper[4764]: E0202 09:28:14.303937 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b" containerName="ceilometer-central-agent" Feb 02 09:28:14 crc kubenswrapper[4764]: I0202 09:28:14.303944 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b" containerName="ceilometer-central-agent" Feb 02 09:28:14 crc kubenswrapper[4764]: E0202 09:28:14.303985 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b" containerName="sg-core" Feb 02 09:28:14 crc kubenswrapper[4764]: I0202 09:28:14.303992 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b" containerName="sg-core" Feb 02 09:28:14 crc kubenswrapper[4764]: I0202 09:28:14.304136 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b" containerName="proxy-httpd" Feb 02 09:28:14 crc kubenswrapper[4764]: I0202 09:28:14.304148 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b" containerName="ceilometer-central-agent" Feb 02 09:28:14 crc kubenswrapper[4764]: I0202 09:28:14.304163 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b" containerName="sg-core" Feb 02 09:28:14 crc kubenswrapper[4764]: I0202 09:28:14.304174 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="39a2a1fd-e483-46ec-807d-7a8b63f4d403" containerName="dnsmasq-dns" Feb 02 09:28:14 crc kubenswrapper[4764]: I0202 09:28:14.304187 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b" containerName="ceilometer-notification-agent" Feb 02 09:28:14 crc kubenswrapper[4764]: I0202 09:28:14.310185 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 02 09:28:14 crc kubenswrapper[4764]: I0202 09:28:14.312782 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Feb 02 09:28:14 crc kubenswrapper[4764]: I0202 09:28:14.312974 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 02 09:28:14 crc kubenswrapper[4764]: I0202 09:28:14.318033 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 02 09:28:14 crc kubenswrapper[4764]: I0202 09:28:14.321630 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 02 09:28:14 crc kubenswrapper[4764]: I0202 09:28:14.342448 4764 scope.go:117] "RemoveContainer" containerID="7f419b8f8922d19daee559647678b9398eeab3b0ea628415a1e0020220e120e5" Feb 02 09:28:14 crc kubenswrapper[4764]: I0202 09:28:14.359646 4764 scope.go:117] "RemoveContainer" containerID="505c759195911603bb96aa1c52c9f6906f245c02c4c7dbbce9a3c65acf37f246" Feb 02 09:28:14 crc kubenswrapper[4764]: I0202 09:28:14.378802 4764 scope.go:117] "RemoveContainer" containerID="37d77489aeed723c2937bac1982f551a1511c02cf3c8c7f6ec85edd4de3181f1" Feb 02 09:28:14 crc kubenswrapper[4764]: I0202 09:28:14.396680 4764 scope.go:117] "RemoveContainer" containerID="35079f5ebbbb42df91e138a36f0a8863357759ec735c7b6cd16cc29f28a95881" Feb 02 09:28:14 crc kubenswrapper[4764]: E0202 09:28:14.397214 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"35079f5ebbbb42df91e138a36f0a8863357759ec735c7b6cd16cc29f28a95881\": container with ID starting with 35079f5ebbbb42df91e138a36f0a8863357759ec735c7b6cd16cc29f28a95881 not found: ID does not exist" containerID="35079f5ebbbb42df91e138a36f0a8863357759ec735c7b6cd16cc29f28a95881" Feb 02 09:28:14 crc kubenswrapper[4764]: I0202 09:28:14.397263 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"35079f5ebbbb42df91e138a36f0a8863357759ec735c7b6cd16cc29f28a95881"} err="failed to get container status \"35079f5ebbbb42df91e138a36f0a8863357759ec735c7b6cd16cc29f28a95881\": rpc error: code = NotFound desc = could not find container \"35079f5ebbbb42df91e138a36f0a8863357759ec735c7b6cd16cc29f28a95881\": container with ID starting with 35079f5ebbbb42df91e138a36f0a8863357759ec735c7b6cd16cc29f28a95881 not found: ID does not exist" Feb 02 09:28:14 crc kubenswrapper[4764]: I0202 09:28:14.397284 4764 scope.go:117] "RemoveContainer" containerID="7f419b8f8922d19daee559647678b9398eeab3b0ea628415a1e0020220e120e5" Feb 02 09:28:14 crc kubenswrapper[4764]: E0202 09:28:14.397648 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7f419b8f8922d19daee559647678b9398eeab3b0ea628415a1e0020220e120e5\": container with ID starting with 7f419b8f8922d19daee559647678b9398eeab3b0ea628415a1e0020220e120e5 not found: ID does not exist" containerID="7f419b8f8922d19daee559647678b9398eeab3b0ea628415a1e0020220e120e5" Feb 02 09:28:14 crc kubenswrapper[4764]: I0202 09:28:14.397671 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7f419b8f8922d19daee559647678b9398eeab3b0ea628415a1e0020220e120e5"} err="failed to get container status \"7f419b8f8922d19daee559647678b9398eeab3b0ea628415a1e0020220e120e5\": rpc error: code = NotFound desc = could not find container \"7f419b8f8922d19daee559647678b9398eeab3b0ea628415a1e0020220e120e5\": container with ID starting with 7f419b8f8922d19daee559647678b9398eeab3b0ea628415a1e0020220e120e5 not found: ID does not exist" Feb 02 09:28:14 crc kubenswrapper[4764]: I0202 09:28:14.397700 4764 scope.go:117] "RemoveContainer" containerID="505c759195911603bb96aa1c52c9f6906f245c02c4c7dbbce9a3c65acf37f246" Feb 02 09:28:14 crc kubenswrapper[4764]: E0202 09:28:14.398065 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"505c759195911603bb96aa1c52c9f6906f245c02c4c7dbbce9a3c65acf37f246\": container with ID starting with 505c759195911603bb96aa1c52c9f6906f245c02c4c7dbbce9a3c65acf37f246 not found: ID does not exist" containerID="505c759195911603bb96aa1c52c9f6906f245c02c4c7dbbce9a3c65acf37f246" Feb 02 09:28:14 crc kubenswrapper[4764]: I0202 09:28:14.398107 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"505c759195911603bb96aa1c52c9f6906f245c02c4c7dbbce9a3c65acf37f246"} err="failed to get container status \"505c759195911603bb96aa1c52c9f6906f245c02c4c7dbbce9a3c65acf37f246\": rpc error: code = NotFound desc = could not find container \"505c759195911603bb96aa1c52c9f6906f245c02c4c7dbbce9a3c65acf37f246\": container with ID starting with 505c759195911603bb96aa1c52c9f6906f245c02c4c7dbbce9a3c65acf37f246 not found: ID does not exist" Feb 02 09:28:14 crc kubenswrapper[4764]: I0202 09:28:14.398122 4764 scope.go:117] "RemoveContainer" containerID="37d77489aeed723c2937bac1982f551a1511c02cf3c8c7f6ec85edd4de3181f1" Feb 02 09:28:14 crc kubenswrapper[4764]: E0202 09:28:14.398365 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"37d77489aeed723c2937bac1982f551a1511c02cf3c8c7f6ec85edd4de3181f1\": container with ID starting with 37d77489aeed723c2937bac1982f551a1511c02cf3c8c7f6ec85edd4de3181f1 not found: ID does not exist" containerID="37d77489aeed723c2937bac1982f551a1511c02cf3c8c7f6ec85edd4de3181f1" Feb 02 09:28:14 crc kubenswrapper[4764]: I0202 09:28:14.398385 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"37d77489aeed723c2937bac1982f551a1511c02cf3c8c7f6ec85edd4de3181f1"} err="failed to get container status \"37d77489aeed723c2937bac1982f551a1511c02cf3c8c7f6ec85edd4de3181f1\": rpc error: code = NotFound desc = could not find container \"37d77489aeed723c2937bac1982f551a1511c02cf3c8c7f6ec85edd4de3181f1\": container with ID starting with 37d77489aeed723c2937bac1982f551a1511c02cf3c8c7f6ec85edd4de3181f1 not found: ID does not exist" Feb 02 09:28:14 crc kubenswrapper[4764]: I0202 09:28:14.432590 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab13eb4e-3576-47f7-bda0-dbac43dd1200-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ab13eb4e-3576-47f7-bda0-dbac43dd1200\") " pod="openstack/ceilometer-0" Feb 02 09:28:14 crc kubenswrapper[4764]: I0202 09:28:14.432660 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ab13eb4e-3576-47f7-bda0-dbac43dd1200-run-httpd\") pod \"ceilometer-0\" (UID: \"ab13eb4e-3576-47f7-bda0-dbac43dd1200\") " pod="openstack/ceilometer-0" Feb 02 09:28:14 crc kubenswrapper[4764]: I0202 09:28:14.432684 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ab13eb4e-3576-47f7-bda0-dbac43dd1200-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ab13eb4e-3576-47f7-bda0-dbac43dd1200\") " pod="openstack/ceilometer-0" Feb 02 09:28:14 crc kubenswrapper[4764]: I0202 09:28:14.432844 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ab13eb4e-3576-47f7-bda0-dbac43dd1200-log-httpd\") pod \"ceilometer-0\" (UID: \"ab13eb4e-3576-47f7-bda0-dbac43dd1200\") " pod="openstack/ceilometer-0" Feb 02 09:28:14 crc kubenswrapper[4764]: I0202 09:28:14.433000 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wx9sw\" (UniqueName: \"kubernetes.io/projected/ab13eb4e-3576-47f7-bda0-dbac43dd1200-kube-api-access-wx9sw\") pod \"ceilometer-0\" (UID: \"ab13eb4e-3576-47f7-bda0-dbac43dd1200\") " pod="openstack/ceilometer-0" Feb 02 09:28:14 crc kubenswrapper[4764]: I0202 09:28:14.433029 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab13eb4e-3576-47f7-bda0-dbac43dd1200-config-data\") pod \"ceilometer-0\" (UID: \"ab13eb4e-3576-47f7-bda0-dbac43dd1200\") " pod="openstack/ceilometer-0" Feb 02 09:28:14 crc kubenswrapper[4764]: I0202 09:28:14.433166 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab13eb4e-3576-47f7-bda0-dbac43dd1200-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"ab13eb4e-3576-47f7-bda0-dbac43dd1200\") " pod="openstack/ceilometer-0" Feb 02 09:28:14 crc kubenswrapper[4764]: I0202 09:28:14.433229 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ab13eb4e-3576-47f7-bda0-dbac43dd1200-scripts\") pod \"ceilometer-0\" (UID: \"ab13eb4e-3576-47f7-bda0-dbac43dd1200\") " pod="openstack/ceilometer-0" Feb 02 09:28:14 crc kubenswrapper[4764]: I0202 09:28:14.534610 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ab13eb4e-3576-47f7-bda0-dbac43dd1200-log-httpd\") pod \"ceilometer-0\" (UID: \"ab13eb4e-3576-47f7-bda0-dbac43dd1200\") " pod="openstack/ceilometer-0" Feb 02 09:28:14 crc kubenswrapper[4764]: I0202 09:28:14.534668 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wx9sw\" (UniqueName: \"kubernetes.io/projected/ab13eb4e-3576-47f7-bda0-dbac43dd1200-kube-api-access-wx9sw\") pod \"ceilometer-0\" (UID: \"ab13eb4e-3576-47f7-bda0-dbac43dd1200\") " pod="openstack/ceilometer-0" Feb 02 09:28:14 crc kubenswrapper[4764]: I0202 09:28:14.534687 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab13eb4e-3576-47f7-bda0-dbac43dd1200-config-data\") pod \"ceilometer-0\" (UID: \"ab13eb4e-3576-47f7-bda0-dbac43dd1200\") " pod="openstack/ceilometer-0" Feb 02 09:28:14 crc kubenswrapper[4764]: I0202 09:28:14.534734 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab13eb4e-3576-47f7-bda0-dbac43dd1200-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"ab13eb4e-3576-47f7-bda0-dbac43dd1200\") " pod="openstack/ceilometer-0" Feb 02 09:28:14 crc kubenswrapper[4764]: I0202 09:28:14.534796 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ab13eb4e-3576-47f7-bda0-dbac43dd1200-scripts\") pod \"ceilometer-0\" (UID: \"ab13eb4e-3576-47f7-bda0-dbac43dd1200\") " pod="openstack/ceilometer-0" Feb 02 09:28:14 crc kubenswrapper[4764]: I0202 09:28:14.534825 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab13eb4e-3576-47f7-bda0-dbac43dd1200-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ab13eb4e-3576-47f7-bda0-dbac43dd1200\") " pod="openstack/ceilometer-0" Feb 02 09:28:14 crc kubenswrapper[4764]: I0202 09:28:14.534887 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ab13eb4e-3576-47f7-bda0-dbac43dd1200-run-httpd\") pod \"ceilometer-0\" (UID: \"ab13eb4e-3576-47f7-bda0-dbac43dd1200\") " pod="openstack/ceilometer-0" Feb 02 09:28:14 crc kubenswrapper[4764]: I0202 09:28:14.534905 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ab13eb4e-3576-47f7-bda0-dbac43dd1200-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ab13eb4e-3576-47f7-bda0-dbac43dd1200\") " pod="openstack/ceilometer-0" Feb 02 09:28:14 crc kubenswrapper[4764]: I0202 09:28:14.537284 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ab13eb4e-3576-47f7-bda0-dbac43dd1200-log-httpd\") pod \"ceilometer-0\" (UID: \"ab13eb4e-3576-47f7-bda0-dbac43dd1200\") " pod="openstack/ceilometer-0" Feb 02 09:28:14 crc kubenswrapper[4764]: I0202 09:28:14.537708 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ab13eb4e-3576-47f7-bda0-dbac43dd1200-run-httpd\") pod \"ceilometer-0\" (UID: \"ab13eb4e-3576-47f7-bda0-dbac43dd1200\") " pod="openstack/ceilometer-0" Feb 02 09:28:14 crc kubenswrapper[4764]: I0202 09:28:14.539482 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab13eb4e-3576-47f7-bda0-dbac43dd1200-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"ab13eb4e-3576-47f7-bda0-dbac43dd1200\") " pod="openstack/ceilometer-0" Feb 02 09:28:14 crc kubenswrapper[4764]: I0202 09:28:14.540331 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ab13eb4e-3576-47f7-bda0-dbac43dd1200-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ab13eb4e-3576-47f7-bda0-dbac43dd1200\") " pod="openstack/ceilometer-0" Feb 02 09:28:14 crc kubenswrapper[4764]: I0202 09:28:14.541356 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ab13eb4e-3576-47f7-bda0-dbac43dd1200-scripts\") pod \"ceilometer-0\" (UID: \"ab13eb4e-3576-47f7-bda0-dbac43dd1200\") " pod="openstack/ceilometer-0" Feb 02 09:28:14 crc kubenswrapper[4764]: I0202 09:28:14.541816 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab13eb4e-3576-47f7-bda0-dbac43dd1200-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ab13eb4e-3576-47f7-bda0-dbac43dd1200\") " pod="openstack/ceilometer-0" Feb 02 09:28:14 crc kubenswrapper[4764]: I0202 09:28:14.555133 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab13eb4e-3576-47f7-bda0-dbac43dd1200-config-data\") pod \"ceilometer-0\" (UID: \"ab13eb4e-3576-47f7-bda0-dbac43dd1200\") " pod="openstack/ceilometer-0" Feb 02 09:28:14 crc kubenswrapper[4764]: I0202 09:28:14.559049 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wx9sw\" (UniqueName: \"kubernetes.io/projected/ab13eb4e-3576-47f7-bda0-dbac43dd1200-kube-api-access-wx9sw\") pod \"ceilometer-0\" (UID: \"ab13eb4e-3576-47f7-bda0-dbac43dd1200\") " pod="openstack/ceilometer-0" Feb 02 09:28:14 crc kubenswrapper[4764]: I0202 09:28:14.638620 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 02 09:28:15 crc kubenswrapper[4764]: I0202 09:28:15.077223 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 02 09:28:15 crc kubenswrapper[4764]: I0202 09:28:15.191528 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ab13eb4e-3576-47f7-bda0-dbac43dd1200","Type":"ContainerStarted","Data":"26d9f36a2d01e51dbe8e4b17a4ea22ca650308bfb4f9f1aaad22cd30ffe2f479"} Feb 02 09:28:15 crc kubenswrapper[4764]: I0202 09:28:15.194112 4764 generic.go:334] "Generic (PLEG): container finished" podID="f83a9caf-7a27-4adf-90ea-072142c914f9" containerID="11f32a8c2d61df5ff6d2656eb4ba0d2fe0897343eea8c23f62baa26ab81563c5" exitCode=0 Feb 02 09:28:15 crc kubenswrapper[4764]: I0202 09:28:15.194152 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-q58jt" event={"ID":"f83a9caf-7a27-4adf-90ea-072142c914f9","Type":"ContainerDied","Data":"11f32a8c2d61df5ff6d2656eb4ba0d2fe0897343eea8c23f62baa26ab81563c5"} Feb 02 09:28:15 crc kubenswrapper[4764]: I0202 09:28:15.838325 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b" path="/var/lib/kubelet/pods/1bc3a40e-ccaa-4a6f-91a1-7f3853b3467b/volumes" Feb 02 09:28:15 crc kubenswrapper[4764]: I0202 09:28:15.839184 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="39a2a1fd-e483-46ec-807d-7a8b63f4d403" path="/var/lib/kubelet/pods/39a2a1fd-e483-46ec-807d-7a8b63f4d403/volumes" Feb 02 09:28:16 crc kubenswrapper[4764]: I0202 09:28:16.209193 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ab13eb4e-3576-47f7-bda0-dbac43dd1200","Type":"ContainerStarted","Data":"1f234f6678d369cfb7c8faf829aa7ccf331f7c163b39d33b6f9c87ecb5534564"} Feb 02 09:28:16 crc kubenswrapper[4764]: I0202 09:28:16.600291 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-q58jt" Feb 02 09:28:16 crc kubenswrapper[4764]: I0202 09:28:16.683875 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f83a9caf-7a27-4adf-90ea-072142c914f9-config-data\") pod \"f83a9caf-7a27-4adf-90ea-072142c914f9\" (UID: \"f83a9caf-7a27-4adf-90ea-072142c914f9\") " Feb 02 09:28:16 crc kubenswrapper[4764]: I0202 09:28:16.684017 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f83a9caf-7a27-4adf-90ea-072142c914f9-combined-ca-bundle\") pod \"f83a9caf-7a27-4adf-90ea-072142c914f9\" (UID: \"f83a9caf-7a27-4adf-90ea-072142c914f9\") " Feb 02 09:28:16 crc kubenswrapper[4764]: I0202 09:28:16.684088 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f83a9caf-7a27-4adf-90ea-072142c914f9-scripts\") pod \"f83a9caf-7a27-4adf-90ea-072142c914f9\" (UID: \"f83a9caf-7a27-4adf-90ea-072142c914f9\") " Feb 02 09:28:16 crc kubenswrapper[4764]: I0202 09:28:16.684136 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bxfzn\" (UniqueName: \"kubernetes.io/projected/f83a9caf-7a27-4adf-90ea-072142c914f9-kube-api-access-bxfzn\") pod \"f83a9caf-7a27-4adf-90ea-072142c914f9\" (UID: \"f83a9caf-7a27-4adf-90ea-072142c914f9\") " Feb 02 09:28:16 crc kubenswrapper[4764]: I0202 09:28:16.690049 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f83a9caf-7a27-4adf-90ea-072142c914f9-scripts" (OuterVolumeSpecName: "scripts") pod "f83a9caf-7a27-4adf-90ea-072142c914f9" (UID: "f83a9caf-7a27-4adf-90ea-072142c914f9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:28:16 crc kubenswrapper[4764]: I0202 09:28:16.690159 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f83a9caf-7a27-4adf-90ea-072142c914f9-kube-api-access-bxfzn" (OuterVolumeSpecName: "kube-api-access-bxfzn") pod "f83a9caf-7a27-4adf-90ea-072142c914f9" (UID: "f83a9caf-7a27-4adf-90ea-072142c914f9"). InnerVolumeSpecName "kube-api-access-bxfzn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:28:16 crc kubenswrapper[4764]: I0202 09:28:16.713443 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f83a9caf-7a27-4adf-90ea-072142c914f9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f83a9caf-7a27-4adf-90ea-072142c914f9" (UID: "f83a9caf-7a27-4adf-90ea-072142c914f9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:28:16 crc kubenswrapper[4764]: I0202 09:28:16.730179 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f83a9caf-7a27-4adf-90ea-072142c914f9-config-data" (OuterVolumeSpecName: "config-data") pod "f83a9caf-7a27-4adf-90ea-072142c914f9" (UID: "f83a9caf-7a27-4adf-90ea-072142c914f9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:28:16 crc kubenswrapper[4764]: I0202 09:28:16.785834 4764 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f83a9caf-7a27-4adf-90ea-072142c914f9-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 09:28:16 crc kubenswrapper[4764]: I0202 09:28:16.785869 4764 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f83a9caf-7a27-4adf-90ea-072142c914f9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 09:28:16 crc kubenswrapper[4764]: I0202 09:28:16.785879 4764 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f83a9caf-7a27-4adf-90ea-072142c914f9-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 09:28:16 crc kubenswrapper[4764]: I0202 09:28:16.785890 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bxfzn\" (UniqueName: \"kubernetes.io/projected/f83a9caf-7a27-4adf-90ea-072142c914f9-kube-api-access-bxfzn\") on node \"crc\" DevicePath \"\"" Feb 02 09:28:17 crc kubenswrapper[4764]: I0202 09:28:17.220899 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ab13eb4e-3576-47f7-bda0-dbac43dd1200","Type":"ContainerStarted","Data":"1429d52ae9bdcecbc9724b64fe86de6efb703fdea6c88978d63974d8de4cb204"} Feb 02 09:28:17 crc kubenswrapper[4764]: I0202 09:28:17.224054 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-q58jt" event={"ID":"f83a9caf-7a27-4adf-90ea-072142c914f9","Type":"ContainerDied","Data":"2741080fa42b3ef53cbe3b139cb0a2ca59a00fb52fa751b7d4b0e917afc3b462"} Feb 02 09:28:17 crc kubenswrapper[4764]: I0202 09:28:17.224120 4764 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2741080fa42b3ef53cbe3b139cb0a2ca59a00fb52fa751b7d4b0e917afc3b462" Feb 02 09:28:17 crc kubenswrapper[4764]: I0202 09:28:17.224122 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-q58jt" Feb 02 09:28:17 crc kubenswrapper[4764]: I0202 09:28:17.420166 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Feb 02 09:28:17 crc kubenswrapper[4764]: I0202 09:28:17.420591 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="2ff2c8e9-62a5-4e5a-93e8-8a82957bcece" containerName="nova-api-log" containerID="cri-o://281c60052638c3fe057a661ce1953d611a63021984e7458c31a49065277d5bbb" gracePeriod=30 Feb 02 09:28:17 crc kubenswrapper[4764]: I0202 09:28:17.420709 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="2ff2c8e9-62a5-4e5a-93e8-8a82957bcece" containerName="nova-api-api" containerID="cri-o://6e6e8f093d6566dd63a9af8b076e024cf869b1c69f6366fd544397ce51b91669" gracePeriod=30 Feb 02 09:28:17 crc kubenswrapper[4764]: I0202 09:28:17.437045 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Feb 02 09:28:17 crc kubenswrapper[4764]: I0202 09:28:17.437434 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="9e10e787-90d5-425b-8d3a-f9b1591cbc9d" containerName="nova-scheduler-scheduler" containerID="cri-o://0588074dea7f8b755739293bb467e526406144beeeb0f156de50d3697687609d" gracePeriod=30 Feb 02 09:28:17 crc kubenswrapper[4764]: I0202 09:28:17.452325 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Feb 02 09:28:17 crc kubenswrapper[4764]: I0202 09:28:17.452645 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="5bacef6a-a70e-400b-963f-d3346b916dee" containerName="nova-metadata-metadata" containerID="cri-o://ea61bee5849658de7d6ee0a009ed7eb2d4a69344ed932c37e7cff559942691b0" gracePeriod=30 Feb 02 09:28:17 crc kubenswrapper[4764]: I0202 09:28:17.452976 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="5bacef6a-a70e-400b-963f-d3346b916dee" containerName="nova-metadata-log" containerID="cri-o://d0a31e50c1e0faeed1483cc4d170f915250dbb3c36dd7c5c83c8f8974270a037" gracePeriod=30 Feb 02 09:28:18 crc kubenswrapper[4764]: I0202 09:28:18.079230 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 02 09:28:18 crc kubenswrapper[4764]: I0202 09:28:18.209498 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2ff2c8e9-62a5-4e5a-93e8-8a82957bcece-internal-tls-certs\") pod \"2ff2c8e9-62a5-4e5a-93e8-8a82957bcece\" (UID: \"2ff2c8e9-62a5-4e5a-93e8-8a82957bcece\") " Feb 02 09:28:18 crc kubenswrapper[4764]: I0202 09:28:18.209617 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ff2c8e9-62a5-4e5a-93e8-8a82957bcece-config-data\") pod \"2ff2c8e9-62a5-4e5a-93e8-8a82957bcece\" (UID: \"2ff2c8e9-62a5-4e5a-93e8-8a82957bcece\") " Feb 02 09:28:18 crc kubenswrapper[4764]: I0202 09:28:18.209635 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ff2c8e9-62a5-4e5a-93e8-8a82957bcece-combined-ca-bundle\") pod \"2ff2c8e9-62a5-4e5a-93e8-8a82957bcece\" (UID: \"2ff2c8e9-62a5-4e5a-93e8-8a82957bcece\") " Feb 02 09:28:18 crc kubenswrapper[4764]: I0202 09:28:18.209713 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2ff2c8e9-62a5-4e5a-93e8-8a82957bcece-public-tls-certs\") pod \"2ff2c8e9-62a5-4e5a-93e8-8a82957bcece\" (UID: \"2ff2c8e9-62a5-4e5a-93e8-8a82957bcece\") " Feb 02 09:28:18 crc kubenswrapper[4764]: I0202 09:28:18.209738 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vvq5m\" (UniqueName: \"kubernetes.io/projected/2ff2c8e9-62a5-4e5a-93e8-8a82957bcece-kube-api-access-vvq5m\") pod \"2ff2c8e9-62a5-4e5a-93e8-8a82957bcece\" (UID: \"2ff2c8e9-62a5-4e5a-93e8-8a82957bcece\") " Feb 02 09:28:18 crc kubenswrapper[4764]: I0202 09:28:18.209760 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2ff2c8e9-62a5-4e5a-93e8-8a82957bcece-logs\") pod \"2ff2c8e9-62a5-4e5a-93e8-8a82957bcece\" (UID: \"2ff2c8e9-62a5-4e5a-93e8-8a82957bcece\") " Feb 02 09:28:18 crc kubenswrapper[4764]: I0202 09:28:18.210849 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2ff2c8e9-62a5-4e5a-93e8-8a82957bcece-logs" (OuterVolumeSpecName: "logs") pod "2ff2c8e9-62a5-4e5a-93e8-8a82957bcece" (UID: "2ff2c8e9-62a5-4e5a-93e8-8a82957bcece"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 09:28:18 crc kubenswrapper[4764]: I0202 09:28:18.219413 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2ff2c8e9-62a5-4e5a-93e8-8a82957bcece-kube-api-access-vvq5m" (OuterVolumeSpecName: "kube-api-access-vvq5m") pod "2ff2c8e9-62a5-4e5a-93e8-8a82957bcece" (UID: "2ff2c8e9-62a5-4e5a-93e8-8a82957bcece"). InnerVolumeSpecName "kube-api-access-vvq5m". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:28:18 crc kubenswrapper[4764]: I0202 09:28:18.244434 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ab13eb4e-3576-47f7-bda0-dbac43dd1200","Type":"ContainerStarted","Data":"9b1db71ed2689d7dbec046dab3a016f1284b7c3571476ce84e62b27a58ae9a6a"} Feb 02 09:28:18 crc kubenswrapper[4764]: I0202 09:28:18.247647 4764 generic.go:334] "Generic (PLEG): container finished" podID="5bacef6a-a70e-400b-963f-d3346b916dee" containerID="d0a31e50c1e0faeed1483cc4d170f915250dbb3c36dd7c5c83c8f8974270a037" exitCode=143 Feb 02 09:28:18 crc kubenswrapper[4764]: I0202 09:28:18.247688 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5bacef6a-a70e-400b-963f-d3346b916dee","Type":"ContainerDied","Data":"d0a31e50c1e0faeed1483cc4d170f915250dbb3c36dd7c5c83c8f8974270a037"} Feb 02 09:28:18 crc kubenswrapper[4764]: I0202 09:28:18.258673 4764 generic.go:334] "Generic (PLEG): container finished" podID="2ff2c8e9-62a5-4e5a-93e8-8a82957bcece" containerID="6e6e8f093d6566dd63a9af8b076e024cf869b1c69f6366fd544397ce51b91669" exitCode=0 Feb 02 09:28:18 crc kubenswrapper[4764]: I0202 09:28:18.258693 4764 generic.go:334] "Generic (PLEG): container finished" podID="2ff2c8e9-62a5-4e5a-93e8-8a82957bcece" containerID="281c60052638c3fe057a661ce1953d611a63021984e7458c31a49065277d5bbb" exitCode=143 Feb 02 09:28:18 crc kubenswrapper[4764]: I0202 09:28:18.258708 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2ff2c8e9-62a5-4e5a-93e8-8a82957bcece","Type":"ContainerDied","Data":"6e6e8f093d6566dd63a9af8b076e024cf869b1c69f6366fd544397ce51b91669"} Feb 02 09:28:18 crc kubenswrapper[4764]: I0202 09:28:18.258725 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2ff2c8e9-62a5-4e5a-93e8-8a82957bcece","Type":"ContainerDied","Data":"281c60052638c3fe057a661ce1953d611a63021984e7458c31a49065277d5bbb"} Feb 02 09:28:18 crc kubenswrapper[4764]: I0202 09:28:18.258734 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2ff2c8e9-62a5-4e5a-93e8-8a82957bcece","Type":"ContainerDied","Data":"e42f3294ec437c06b9fba240dfafed74ccd91751df00eebfde730972b7fb61b6"} Feb 02 09:28:18 crc kubenswrapper[4764]: I0202 09:28:18.258749 4764 scope.go:117] "RemoveContainer" containerID="6e6e8f093d6566dd63a9af8b076e024cf869b1c69f6366fd544397ce51b91669" Feb 02 09:28:18 crc kubenswrapper[4764]: I0202 09:28:18.258861 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 02 09:28:18 crc kubenswrapper[4764]: I0202 09:28:18.260027 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ff2c8e9-62a5-4e5a-93e8-8a82957bcece-config-data" (OuterVolumeSpecName: "config-data") pod "2ff2c8e9-62a5-4e5a-93e8-8a82957bcece" (UID: "2ff2c8e9-62a5-4e5a-93e8-8a82957bcece"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:28:18 crc kubenswrapper[4764]: I0202 09:28:18.264861 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ff2c8e9-62a5-4e5a-93e8-8a82957bcece-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2ff2c8e9-62a5-4e5a-93e8-8a82957bcece" (UID: "2ff2c8e9-62a5-4e5a-93e8-8a82957bcece"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:28:18 crc kubenswrapper[4764]: I0202 09:28:18.266056 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ff2c8e9-62a5-4e5a-93e8-8a82957bcece-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "2ff2c8e9-62a5-4e5a-93e8-8a82957bcece" (UID: "2ff2c8e9-62a5-4e5a-93e8-8a82957bcece"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:28:18 crc kubenswrapper[4764]: I0202 09:28:18.286096 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ff2c8e9-62a5-4e5a-93e8-8a82957bcece-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "2ff2c8e9-62a5-4e5a-93e8-8a82957bcece" (UID: "2ff2c8e9-62a5-4e5a-93e8-8a82957bcece"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:28:18 crc kubenswrapper[4764]: I0202 09:28:18.302147 4764 scope.go:117] "RemoveContainer" containerID="281c60052638c3fe057a661ce1953d611a63021984e7458c31a49065277d5bbb" Feb 02 09:28:18 crc kubenswrapper[4764]: I0202 09:28:18.311810 4764 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2ff2c8e9-62a5-4e5a-93e8-8a82957bcece-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 02 09:28:18 crc kubenswrapper[4764]: I0202 09:28:18.311843 4764 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ff2c8e9-62a5-4e5a-93e8-8a82957bcece-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 09:28:18 crc kubenswrapper[4764]: I0202 09:28:18.311864 4764 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ff2c8e9-62a5-4e5a-93e8-8a82957bcece-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 09:28:18 crc kubenswrapper[4764]: I0202 09:28:18.311872 4764 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2ff2c8e9-62a5-4e5a-93e8-8a82957bcece-public-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 02 09:28:18 crc kubenswrapper[4764]: I0202 09:28:18.311881 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vvq5m\" (UniqueName: \"kubernetes.io/projected/2ff2c8e9-62a5-4e5a-93e8-8a82957bcece-kube-api-access-vvq5m\") on node \"crc\" DevicePath \"\"" Feb 02 09:28:18 crc kubenswrapper[4764]: I0202 09:28:18.311891 4764 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2ff2c8e9-62a5-4e5a-93e8-8a82957bcece-logs\") on node \"crc\" DevicePath \"\"" Feb 02 09:28:18 crc kubenswrapper[4764]: I0202 09:28:18.343685 4764 scope.go:117] "RemoveContainer" containerID="6e6e8f093d6566dd63a9af8b076e024cf869b1c69f6366fd544397ce51b91669" Feb 02 09:28:18 crc kubenswrapper[4764]: E0202 09:28:18.345344 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6e6e8f093d6566dd63a9af8b076e024cf869b1c69f6366fd544397ce51b91669\": container with ID starting with 6e6e8f093d6566dd63a9af8b076e024cf869b1c69f6366fd544397ce51b91669 not found: ID does not exist" containerID="6e6e8f093d6566dd63a9af8b076e024cf869b1c69f6366fd544397ce51b91669" Feb 02 09:28:18 crc kubenswrapper[4764]: I0202 09:28:18.345377 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6e6e8f093d6566dd63a9af8b076e024cf869b1c69f6366fd544397ce51b91669"} err="failed to get container status \"6e6e8f093d6566dd63a9af8b076e024cf869b1c69f6366fd544397ce51b91669\": rpc error: code = NotFound desc = could not find container \"6e6e8f093d6566dd63a9af8b076e024cf869b1c69f6366fd544397ce51b91669\": container with ID starting with 6e6e8f093d6566dd63a9af8b076e024cf869b1c69f6366fd544397ce51b91669 not found: ID does not exist" Feb 02 09:28:18 crc kubenswrapper[4764]: I0202 09:28:18.345398 4764 scope.go:117] "RemoveContainer" containerID="281c60052638c3fe057a661ce1953d611a63021984e7458c31a49065277d5bbb" Feb 02 09:28:18 crc kubenswrapper[4764]: E0202 09:28:18.349263 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"281c60052638c3fe057a661ce1953d611a63021984e7458c31a49065277d5bbb\": container with ID starting with 281c60052638c3fe057a661ce1953d611a63021984e7458c31a49065277d5bbb not found: ID does not exist" containerID="281c60052638c3fe057a661ce1953d611a63021984e7458c31a49065277d5bbb" Feb 02 09:28:18 crc kubenswrapper[4764]: I0202 09:28:18.349317 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"281c60052638c3fe057a661ce1953d611a63021984e7458c31a49065277d5bbb"} err="failed to get container status \"281c60052638c3fe057a661ce1953d611a63021984e7458c31a49065277d5bbb\": rpc error: code = NotFound desc = could not find container \"281c60052638c3fe057a661ce1953d611a63021984e7458c31a49065277d5bbb\": container with ID starting with 281c60052638c3fe057a661ce1953d611a63021984e7458c31a49065277d5bbb not found: ID does not exist" Feb 02 09:28:18 crc kubenswrapper[4764]: I0202 09:28:18.349349 4764 scope.go:117] "RemoveContainer" containerID="6e6e8f093d6566dd63a9af8b076e024cf869b1c69f6366fd544397ce51b91669" Feb 02 09:28:18 crc kubenswrapper[4764]: I0202 09:28:18.353140 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6e6e8f093d6566dd63a9af8b076e024cf869b1c69f6366fd544397ce51b91669"} err="failed to get container status \"6e6e8f093d6566dd63a9af8b076e024cf869b1c69f6366fd544397ce51b91669\": rpc error: code = NotFound desc = could not find container \"6e6e8f093d6566dd63a9af8b076e024cf869b1c69f6366fd544397ce51b91669\": container with ID starting with 6e6e8f093d6566dd63a9af8b076e024cf869b1c69f6366fd544397ce51b91669 not found: ID does not exist" Feb 02 09:28:18 crc kubenswrapper[4764]: I0202 09:28:18.353166 4764 scope.go:117] "RemoveContainer" containerID="281c60052638c3fe057a661ce1953d611a63021984e7458c31a49065277d5bbb" Feb 02 09:28:18 crc kubenswrapper[4764]: I0202 09:28:18.358062 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"281c60052638c3fe057a661ce1953d611a63021984e7458c31a49065277d5bbb"} err="failed to get container status \"281c60052638c3fe057a661ce1953d611a63021984e7458c31a49065277d5bbb\": rpc error: code = NotFound desc = could not find container \"281c60052638c3fe057a661ce1953d611a63021984e7458c31a49065277d5bbb\": container with ID starting with 281c60052638c3fe057a661ce1953d611a63021984e7458c31a49065277d5bbb not found: ID does not exist" Feb 02 09:28:18 crc kubenswrapper[4764]: I0202 09:28:18.594962 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Feb 02 09:28:18 crc kubenswrapper[4764]: I0202 09:28:18.618011 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Feb 02 09:28:18 crc kubenswrapper[4764]: I0202 09:28:18.631331 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Feb 02 09:28:18 crc kubenswrapper[4764]: E0202 09:28:18.631675 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ff2c8e9-62a5-4e5a-93e8-8a82957bcece" containerName="nova-api-api" Feb 02 09:28:18 crc kubenswrapper[4764]: I0202 09:28:18.631687 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ff2c8e9-62a5-4e5a-93e8-8a82957bcece" containerName="nova-api-api" Feb 02 09:28:18 crc kubenswrapper[4764]: E0202 09:28:18.631694 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ff2c8e9-62a5-4e5a-93e8-8a82957bcece" containerName="nova-api-log" Feb 02 09:28:18 crc kubenswrapper[4764]: I0202 09:28:18.631700 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ff2c8e9-62a5-4e5a-93e8-8a82957bcece" containerName="nova-api-log" Feb 02 09:28:18 crc kubenswrapper[4764]: E0202 09:28:18.631729 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f83a9caf-7a27-4adf-90ea-072142c914f9" containerName="nova-manage" Feb 02 09:28:18 crc kubenswrapper[4764]: I0202 09:28:18.631735 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="f83a9caf-7a27-4adf-90ea-072142c914f9" containerName="nova-manage" Feb 02 09:28:18 crc kubenswrapper[4764]: I0202 09:28:18.631881 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ff2c8e9-62a5-4e5a-93e8-8a82957bcece" containerName="nova-api-api" Feb 02 09:28:18 crc kubenswrapper[4764]: I0202 09:28:18.631892 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ff2c8e9-62a5-4e5a-93e8-8a82957bcece" containerName="nova-api-log" Feb 02 09:28:18 crc kubenswrapper[4764]: I0202 09:28:18.631907 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="f83a9caf-7a27-4adf-90ea-072142c914f9" containerName="nova-manage" Feb 02 09:28:18 crc kubenswrapper[4764]: I0202 09:28:18.632709 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 02 09:28:18 crc kubenswrapper[4764]: I0202 09:28:18.638577 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Feb 02 09:28:18 crc kubenswrapper[4764]: I0202 09:28:18.638711 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Feb 02 09:28:18 crc kubenswrapper[4764]: I0202 09:28:18.639134 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Feb 02 09:28:18 crc kubenswrapper[4764]: I0202 09:28:18.653508 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 02 09:28:18 crc kubenswrapper[4764]: I0202 09:28:18.717058 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db89c9c9-cad1-4e6b-aaf0-708e5b8b32b4-config-data\") pod \"nova-api-0\" (UID: \"db89c9c9-cad1-4e6b-aaf0-708e5b8b32b4\") " pod="openstack/nova-api-0" Feb 02 09:28:18 crc kubenswrapper[4764]: I0202 09:28:18.717147 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db89c9c9-cad1-4e6b-aaf0-708e5b8b32b4-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"db89c9c9-cad1-4e6b-aaf0-708e5b8b32b4\") " pod="openstack/nova-api-0" Feb 02 09:28:18 crc kubenswrapper[4764]: I0202 09:28:18.717177 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/db89c9c9-cad1-4e6b-aaf0-708e5b8b32b4-internal-tls-certs\") pod \"nova-api-0\" (UID: \"db89c9c9-cad1-4e6b-aaf0-708e5b8b32b4\") " pod="openstack/nova-api-0" Feb 02 09:28:18 crc kubenswrapper[4764]: I0202 09:28:18.717314 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/db89c9c9-cad1-4e6b-aaf0-708e5b8b32b4-public-tls-certs\") pod \"nova-api-0\" (UID: \"db89c9c9-cad1-4e6b-aaf0-708e5b8b32b4\") " pod="openstack/nova-api-0" Feb 02 09:28:18 crc kubenswrapper[4764]: I0202 09:28:18.717458 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5zpkx\" (UniqueName: \"kubernetes.io/projected/db89c9c9-cad1-4e6b-aaf0-708e5b8b32b4-kube-api-access-5zpkx\") pod \"nova-api-0\" (UID: \"db89c9c9-cad1-4e6b-aaf0-708e5b8b32b4\") " pod="openstack/nova-api-0" Feb 02 09:28:18 crc kubenswrapper[4764]: I0202 09:28:18.717739 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/db89c9c9-cad1-4e6b-aaf0-708e5b8b32b4-logs\") pod \"nova-api-0\" (UID: \"db89c9c9-cad1-4e6b-aaf0-708e5b8b32b4\") " pod="openstack/nova-api-0" Feb 02 09:28:18 crc kubenswrapper[4764]: I0202 09:28:18.820319 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/db89c9c9-cad1-4e6b-aaf0-708e5b8b32b4-logs\") pod \"nova-api-0\" (UID: \"db89c9c9-cad1-4e6b-aaf0-708e5b8b32b4\") " pod="openstack/nova-api-0" Feb 02 09:28:18 crc kubenswrapper[4764]: I0202 09:28:18.820428 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db89c9c9-cad1-4e6b-aaf0-708e5b8b32b4-config-data\") pod \"nova-api-0\" (UID: \"db89c9c9-cad1-4e6b-aaf0-708e5b8b32b4\") " pod="openstack/nova-api-0" Feb 02 09:28:18 crc kubenswrapper[4764]: I0202 09:28:18.820495 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db89c9c9-cad1-4e6b-aaf0-708e5b8b32b4-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"db89c9c9-cad1-4e6b-aaf0-708e5b8b32b4\") " pod="openstack/nova-api-0" Feb 02 09:28:18 crc kubenswrapper[4764]: I0202 09:28:18.820528 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/db89c9c9-cad1-4e6b-aaf0-708e5b8b32b4-internal-tls-certs\") pod \"nova-api-0\" (UID: \"db89c9c9-cad1-4e6b-aaf0-708e5b8b32b4\") " pod="openstack/nova-api-0" Feb 02 09:28:18 crc kubenswrapper[4764]: I0202 09:28:18.820551 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/db89c9c9-cad1-4e6b-aaf0-708e5b8b32b4-public-tls-certs\") pod \"nova-api-0\" (UID: \"db89c9c9-cad1-4e6b-aaf0-708e5b8b32b4\") " pod="openstack/nova-api-0" Feb 02 09:28:18 crc kubenswrapper[4764]: I0202 09:28:18.820597 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5zpkx\" (UniqueName: \"kubernetes.io/projected/db89c9c9-cad1-4e6b-aaf0-708e5b8b32b4-kube-api-access-5zpkx\") pod \"nova-api-0\" (UID: \"db89c9c9-cad1-4e6b-aaf0-708e5b8b32b4\") " pod="openstack/nova-api-0" Feb 02 09:28:18 crc kubenswrapper[4764]: I0202 09:28:18.823280 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/db89c9c9-cad1-4e6b-aaf0-708e5b8b32b4-logs\") pod \"nova-api-0\" (UID: \"db89c9c9-cad1-4e6b-aaf0-708e5b8b32b4\") " pod="openstack/nova-api-0" Feb 02 09:28:18 crc kubenswrapper[4764]: I0202 09:28:18.826619 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/db89c9c9-cad1-4e6b-aaf0-708e5b8b32b4-internal-tls-certs\") pod \"nova-api-0\" (UID: \"db89c9c9-cad1-4e6b-aaf0-708e5b8b32b4\") " pod="openstack/nova-api-0" Feb 02 09:28:18 crc kubenswrapper[4764]: I0202 09:28:18.833534 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/db89c9c9-cad1-4e6b-aaf0-708e5b8b32b4-public-tls-certs\") pod \"nova-api-0\" (UID: \"db89c9c9-cad1-4e6b-aaf0-708e5b8b32b4\") " pod="openstack/nova-api-0" Feb 02 09:28:18 crc kubenswrapper[4764]: I0202 09:28:18.835880 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db89c9c9-cad1-4e6b-aaf0-708e5b8b32b4-config-data\") pod \"nova-api-0\" (UID: \"db89c9c9-cad1-4e6b-aaf0-708e5b8b32b4\") " pod="openstack/nova-api-0" Feb 02 09:28:18 crc kubenswrapper[4764]: I0202 09:28:18.836265 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db89c9c9-cad1-4e6b-aaf0-708e5b8b32b4-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"db89c9c9-cad1-4e6b-aaf0-708e5b8b32b4\") " pod="openstack/nova-api-0" Feb 02 09:28:18 crc kubenswrapper[4764]: I0202 09:28:18.839685 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5zpkx\" (UniqueName: \"kubernetes.io/projected/db89c9c9-cad1-4e6b-aaf0-708e5b8b32b4-kube-api-access-5zpkx\") pod \"nova-api-0\" (UID: \"db89c9c9-cad1-4e6b-aaf0-708e5b8b32b4\") " pod="openstack/nova-api-0" Feb 02 09:28:18 crc kubenswrapper[4764]: I0202 09:28:18.975768 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 02 09:28:19 crc kubenswrapper[4764]: I0202 09:28:19.027421 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9e10e787-90d5-425b-8d3a-f9b1591cbc9d-config-data\") pod \"9e10e787-90d5-425b-8d3a-f9b1591cbc9d\" (UID: \"9e10e787-90d5-425b-8d3a-f9b1591cbc9d\") " Feb 02 09:28:19 crc kubenswrapper[4764]: I0202 09:28:19.027734 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e10e787-90d5-425b-8d3a-f9b1591cbc9d-combined-ca-bundle\") pod \"9e10e787-90d5-425b-8d3a-f9b1591cbc9d\" (UID: \"9e10e787-90d5-425b-8d3a-f9b1591cbc9d\") " Feb 02 09:28:19 crc kubenswrapper[4764]: I0202 09:28:19.027945 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-57nkj\" (UniqueName: \"kubernetes.io/projected/9e10e787-90d5-425b-8d3a-f9b1591cbc9d-kube-api-access-57nkj\") pod \"9e10e787-90d5-425b-8d3a-f9b1591cbc9d\" (UID: \"9e10e787-90d5-425b-8d3a-f9b1591cbc9d\") " Feb 02 09:28:19 crc kubenswrapper[4764]: I0202 09:28:19.043634 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e10e787-90d5-425b-8d3a-f9b1591cbc9d-kube-api-access-57nkj" (OuterVolumeSpecName: "kube-api-access-57nkj") pod "9e10e787-90d5-425b-8d3a-f9b1591cbc9d" (UID: "9e10e787-90d5-425b-8d3a-f9b1591cbc9d"). InnerVolumeSpecName "kube-api-access-57nkj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:28:19 crc kubenswrapper[4764]: I0202 09:28:19.047864 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 02 09:28:19 crc kubenswrapper[4764]: I0202 09:28:19.059419 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9e10e787-90d5-425b-8d3a-f9b1591cbc9d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9e10e787-90d5-425b-8d3a-f9b1591cbc9d" (UID: "9e10e787-90d5-425b-8d3a-f9b1591cbc9d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:28:19 crc kubenswrapper[4764]: I0202 09:28:19.064560 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9e10e787-90d5-425b-8d3a-f9b1591cbc9d-config-data" (OuterVolumeSpecName: "config-data") pod "9e10e787-90d5-425b-8d3a-f9b1591cbc9d" (UID: "9e10e787-90d5-425b-8d3a-f9b1591cbc9d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:28:19 crc kubenswrapper[4764]: I0202 09:28:19.130826 4764 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e10e787-90d5-425b-8d3a-f9b1591cbc9d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 09:28:19 crc kubenswrapper[4764]: I0202 09:28:19.130848 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-57nkj\" (UniqueName: \"kubernetes.io/projected/9e10e787-90d5-425b-8d3a-f9b1591cbc9d-kube-api-access-57nkj\") on node \"crc\" DevicePath \"\"" Feb 02 09:28:19 crc kubenswrapper[4764]: I0202 09:28:19.130858 4764 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9e10e787-90d5-425b-8d3a-f9b1591cbc9d-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 09:28:19 crc kubenswrapper[4764]: I0202 09:28:19.276389 4764 generic.go:334] "Generic (PLEG): container finished" podID="9e10e787-90d5-425b-8d3a-f9b1591cbc9d" containerID="0588074dea7f8b755739293bb467e526406144beeeb0f156de50d3697687609d" exitCode=0 Feb 02 09:28:19 crc kubenswrapper[4764]: I0202 09:28:19.276697 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"9e10e787-90d5-425b-8d3a-f9b1591cbc9d","Type":"ContainerDied","Data":"0588074dea7f8b755739293bb467e526406144beeeb0f156de50d3697687609d"} Feb 02 09:28:19 crc kubenswrapper[4764]: I0202 09:28:19.276728 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"9e10e787-90d5-425b-8d3a-f9b1591cbc9d","Type":"ContainerDied","Data":"a76c6c516fb754eed5eb905d7ab39a5c12ceb9b4166ebe04ae7cf1243a481e3b"} Feb 02 09:28:19 crc kubenswrapper[4764]: I0202 09:28:19.276744 4764 scope.go:117] "RemoveContainer" containerID="0588074dea7f8b755739293bb467e526406144beeeb0f156de50d3697687609d" Feb 02 09:28:19 crc kubenswrapper[4764]: I0202 09:28:19.276942 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 02 09:28:19 crc kubenswrapper[4764]: I0202 09:28:19.341131 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Feb 02 09:28:19 crc kubenswrapper[4764]: I0202 09:28:19.349084 4764 scope.go:117] "RemoveContainer" containerID="0588074dea7f8b755739293bb467e526406144beeeb0f156de50d3697687609d" Feb 02 09:28:19 crc kubenswrapper[4764]: E0202 09:28:19.349565 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0588074dea7f8b755739293bb467e526406144beeeb0f156de50d3697687609d\": container with ID starting with 0588074dea7f8b755739293bb467e526406144beeeb0f156de50d3697687609d not found: ID does not exist" containerID="0588074dea7f8b755739293bb467e526406144beeeb0f156de50d3697687609d" Feb 02 09:28:19 crc kubenswrapper[4764]: I0202 09:28:19.349613 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0588074dea7f8b755739293bb467e526406144beeeb0f156de50d3697687609d"} err="failed to get container status \"0588074dea7f8b755739293bb467e526406144beeeb0f156de50d3697687609d\": rpc error: code = NotFound desc = could not find container \"0588074dea7f8b755739293bb467e526406144beeeb0f156de50d3697687609d\": container with ID starting with 0588074dea7f8b755739293bb467e526406144beeeb0f156de50d3697687609d not found: ID does not exist" Feb 02 09:28:19 crc kubenswrapper[4764]: I0202 09:28:19.356212 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Feb 02 09:28:19 crc kubenswrapper[4764]: I0202 09:28:19.363860 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Feb 02 09:28:19 crc kubenswrapper[4764]: E0202 09:28:19.364230 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e10e787-90d5-425b-8d3a-f9b1591cbc9d" containerName="nova-scheduler-scheduler" Feb 02 09:28:19 crc kubenswrapper[4764]: I0202 09:28:19.364251 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e10e787-90d5-425b-8d3a-f9b1591cbc9d" containerName="nova-scheduler-scheduler" Feb 02 09:28:19 crc kubenswrapper[4764]: I0202 09:28:19.364439 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e10e787-90d5-425b-8d3a-f9b1591cbc9d" containerName="nova-scheduler-scheduler" Feb 02 09:28:19 crc kubenswrapper[4764]: I0202 09:28:19.365036 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 02 09:28:19 crc kubenswrapper[4764]: I0202 09:28:19.366482 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Feb 02 09:28:19 crc kubenswrapper[4764]: I0202 09:28:19.378580 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Feb 02 09:28:19 crc kubenswrapper[4764]: I0202 09:28:19.448990 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd11a817-6357-4405-983e-181451ce4daf-config-data\") pod \"nova-scheduler-0\" (UID: \"dd11a817-6357-4405-983e-181451ce4daf\") " pod="openstack/nova-scheduler-0" Feb 02 09:28:19 crc kubenswrapper[4764]: I0202 09:28:19.449099 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd11a817-6357-4405-983e-181451ce4daf-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"dd11a817-6357-4405-983e-181451ce4daf\") " pod="openstack/nova-scheduler-0" Feb 02 09:28:19 crc kubenswrapper[4764]: I0202 09:28:19.449187 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4hh6b\" (UniqueName: \"kubernetes.io/projected/dd11a817-6357-4405-983e-181451ce4daf-kube-api-access-4hh6b\") pod \"nova-scheduler-0\" (UID: \"dd11a817-6357-4405-983e-181451ce4daf\") " pod="openstack/nova-scheduler-0" Feb 02 09:28:19 crc kubenswrapper[4764]: I0202 09:28:19.476714 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 02 09:28:19 crc kubenswrapper[4764]: I0202 09:28:19.550236 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd11a817-6357-4405-983e-181451ce4daf-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"dd11a817-6357-4405-983e-181451ce4daf\") " pod="openstack/nova-scheduler-0" Feb 02 09:28:19 crc kubenswrapper[4764]: I0202 09:28:19.550319 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4hh6b\" (UniqueName: \"kubernetes.io/projected/dd11a817-6357-4405-983e-181451ce4daf-kube-api-access-4hh6b\") pod \"nova-scheduler-0\" (UID: \"dd11a817-6357-4405-983e-181451ce4daf\") " pod="openstack/nova-scheduler-0" Feb 02 09:28:19 crc kubenswrapper[4764]: I0202 09:28:19.550373 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd11a817-6357-4405-983e-181451ce4daf-config-data\") pod \"nova-scheduler-0\" (UID: \"dd11a817-6357-4405-983e-181451ce4daf\") " pod="openstack/nova-scheduler-0" Feb 02 09:28:19 crc kubenswrapper[4764]: I0202 09:28:19.556660 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd11a817-6357-4405-983e-181451ce4daf-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"dd11a817-6357-4405-983e-181451ce4daf\") " pod="openstack/nova-scheduler-0" Feb 02 09:28:19 crc kubenswrapper[4764]: I0202 09:28:19.564491 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd11a817-6357-4405-983e-181451ce4daf-config-data\") pod \"nova-scheduler-0\" (UID: \"dd11a817-6357-4405-983e-181451ce4daf\") " pod="openstack/nova-scheduler-0" Feb 02 09:28:19 crc kubenswrapper[4764]: I0202 09:28:19.566494 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4hh6b\" (UniqueName: \"kubernetes.io/projected/dd11a817-6357-4405-983e-181451ce4daf-kube-api-access-4hh6b\") pod \"nova-scheduler-0\" (UID: \"dd11a817-6357-4405-983e-181451ce4daf\") " pod="openstack/nova-scheduler-0" Feb 02 09:28:19 crc kubenswrapper[4764]: I0202 09:28:19.680380 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 02 09:28:19 crc kubenswrapper[4764]: I0202 09:28:19.838559 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2ff2c8e9-62a5-4e5a-93e8-8a82957bcece" path="/var/lib/kubelet/pods/2ff2c8e9-62a5-4e5a-93e8-8a82957bcece/volumes" Feb 02 09:28:19 crc kubenswrapper[4764]: I0202 09:28:19.839918 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9e10e787-90d5-425b-8d3a-f9b1591cbc9d" path="/var/lib/kubelet/pods/9e10e787-90d5-425b-8d3a-f9b1591cbc9d/volumes" Feb 02 09:28:20 crc kubenswrapper[4764]: I0202 09:28:20.140508 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Feb 02 09:28:20 crc kubenswrapper[4764]: I0202 09:28:20.286475 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"dd11a817-6357-4405-983e-181451ce4daf","Type":"ContainerStarted","Data":"e7303c019c09e65846cd5b0818f7e2c7f83bab252a8dce954a9d923aa6734681"} Feb 02 09:28:20 crc kubenswrapper[4764]: I0202 09:28:20.286524 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"dd11a817-6357-4405-983e-181451ce4daf","Type":"ContainerStarted","Data":"8c6750769ee84aff600210a6c9a5715c9b1ea3c0b28025664963a19564358c71"} Feb 02 09:28:20 crc kubenswrapper[4764]: I0202 09:28:20.292517 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ab13eb4e-3576-47f7-bda0-dbac43dd1200","Type":"ContainerStarted","Data":"024682a0e94744dfba3481e7a6baee331b2323d6d1006332b15cda00582b0e74"} Feb 02 09:28:20 crc kubenswrapper[4764]: I0202 09:28:20.293185 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 02 09:28:20 crc kubenswrapper[4764]: I0202 09:28:20.294899 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"db89c9c9-cad1-4e6b-aaf0-708e5b8b32b4","Type":"ContainerStarted","Data":"364f7ca7f3e0b07d6d74eef23611102bc2c5f6c04ef7b10ec6135d53af43c541"} Feb 02 09:28:20 crc kubenswrapper[4764]: I0202 09:28:20.294949 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"db89c9c9-cad1-4e6b-aaf0-708e5b8b32b4","Type":"ContainerStarted","Data":"88d761afd6d275c73087182a3d1b278a81ba94ab058cbea01715e9249f9b0010"} Feb 02 09:28:20 crc kubenswrapper[4764]: I0202 09:28:20.294961 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"db89c9c9-cad1-4e6b-aaf0-708e5b8b32b4","Type":"ContainerStarted","Data":"b7444f7c39fa0e7ea4074509cae8c2c2132b0a972ebec961fba5ff7dd02805a3"} Feb 02 09:28:20 crc kubenswrapper[4764]: I0202 09:28:20.305259 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=1.305228603 podStartE2EDuration="1.305228603s" podCreationTimestamp="2026-02-02 09:28:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:28:20.298075213 +0000 UTC m=+1263.231799301" watchObservedRunningTime="2026-02-02 09:28:20.305228603 +0000 UTC m=+1263.238952691" Feb 02 09:28:20 crc kubenswrapper[4764]: I0202 09:28:20.328007 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.302429199 podStartE2EDuration="6.32799176s" podCreationTimestamp="2026-02-02 09:28:14 +0000 UTC" firstStartedPulling="2026-02-02 09:28:15.077760228 +0000 UTC m=+1258.011484356" lastFinishedPulling="2026-02-02 09:28:19.103322829 +0000 UTC m=+1262.037046917" observedRunningTime="2026-02-02 09:28:20.325074452 +0000 UTC m=+1263.258798540" watchObservedRunningTime="2026-02-02 09:28:20.32799176 +0000 UTC m=+1263.261715848" Feb 02 09:28:20 crc kubenswrapper[4764]: I0202 09:28:20.348197 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.348180678 podStartE2EDuration="2.348180678s" podCreationTimestamp="2026-02-02 09:28:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:28:20.344158231 +0000 UTC m=+1263.277882319" watchObservedRunningTime="2026-02-02 09:28:20.348180678 +0000 UTC m=+1263.281904756" Feb 02 09:28:21 crc kubenswrapper[4764]: I0202 09:28:21.040069 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 02 09:28:21 crc kubenswrapper[4764]: I0202 09:28:21.180482 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5bacef6a-a70e-400b-963f-d3346b916dee-config-data\") pod \"5bacef6a-a70e-400b-963f-d3346b916dee\" (UID: \"5bacef6a-a70e-400b-963f-d3346b916dee\") " Feb 02 09:28:21 crc kubenswrapper[4764]: I0202 09:28:21.180848 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j84pj\" (UniqueName: \"kubernetes.io/projected/5bacef6a-a70e-400b-963f-d3346b916dee-kube-api-access-j84pj\") pod \"5bacef6a-a70e-400b-963f-d3346b916dee\" (UID: \"5bacef6a-a70e-400b-963f-d3346b916dee\") " Feb 02 09:28:21 crc kubenswrapper[4764]: I0202 09:28:21.180963 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5bacef6a-a70e-400b-963f-d3346b916dee-logs\") pod \"5bacef6a-a70e-400b-963f-d3346b916dee\" (UID: \"5bacef6a-a70e-400b-963f-d3346b916dee\") " Feb 02 09:28:21 crc kubenswrapper[4764]: I0202 09:28:21.180995 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5bacef6a-a70e-400b-963f-d3346b916dee-combined-ca-bundle\") pod \"5bacef6a-a70e-400b-963f-d3346b916dee\" (UID: \"5bacef6a-a70e-400b-963f-d3346b916dee\") " Feb 02 09:28:21 crc kubenswrapper[4764]: I0202 09:28:21.181033 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/5bacef6a-a70e-400b-963f-d3346b916dee-nova-metadata-tls-certs\") pod \"5bacef6a-a70e-400b-963f-d3346b916dee\" (UID: \"5bacef6a-a70e-400b-963f-d3346b916dee\") " Feb 02 09:28:21 crc kubenswrapper[4764]: I0202 09:28:21.182580 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5bacef6a-a70e-400b-963f-d3346b916dee-logs" (OuterVolumeSpecName: "logs") pod "5bacef6a-a70e-400b-963f-d3346b916dee" (UID: "5bacef6a-a70e-400b-963f-d3346b916dee"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 09:28:21 crc kubenswrapper[4764]: I0202 09:28:21.207376 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5bacef6a-a70e-400b-963f-d3346b916dee-kube-api-access-j84pj" (OuterVolumeSpecName: "kube-api-access-j84pj") pod "5bacef6a-a70e-400b-963f-d3346b916dee" (UID: "5bacef6a-a70e-400b-963f-d3346b916dee"). InnerVolumeSpecName "kube-api-access-j84pj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:28:21 crc kubenswrapper[4764]: I0202 09:28:21.212417 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5bacef6a-a70e-400b-963f-d3346b916dee-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5bacef6a-a70e-400b-963f-d3346b916dee" (UID: "5bacef6a-a70e-400b-963f-d3346b916dee"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:28:21 crc kubenswrapper[4764]: I0202 09:28:21.222414 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5bacef6a-a70e-400b-963f-d3346b916dee-config-data" (OuterVolumeSpecName: "config-data") pod "5bacef6a-a70e-400b-963f-d3346b916dee" (UID: "5bacef6a-a70e-400b-963f-d3346b916dee"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:28:21 crc kubenswrapper[4764]: I0202 09:28:21.237136 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5bacef6a-a70e-400b-963f-d3346b916dee-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "5bacef6a-a70e-400b-963f-d3346b916dee" (UID: "5bacef6a-a70e-400b-963f-d3346b916dee"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:28:21 crc kubenswrapper[4764]: I0202 09:28:21.283301 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j84pj\" (UniqueName: \"kubernetes.io/projected/5bacef6a-a70e-400b-963f-d3346b916dee-kube-api-access-j84pj\") on node \"crc\" DevicePath \"\"" Feb 02 09:28:21 crc kubenswrapper[4764]: I0202 09:28:21.283330 4764 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5bacef6a-a70e-400b-963f-d3346b916dee-logs\") on node \"crc\" DevicePath \"\"" Feb 02 09:28:21 crc kubenswrapper[4764]: I0202 09:28:21.283340 4764 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5bacef6a-a70e-400b-963f-d3346b916dee-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 09:28:21 crc kubenswrapper[4764]: I0202 09:28:21.283348 4764 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/5bacef6a-a70e-400b-963f-d3346b916dee-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 02 09:28:21 crc kubenswrapper[4764]: I0202 09:28:21.283356 4764 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5bacef6a-a70e-400b-963f-d3346b916dee-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 09:28:21 crc kubenswrapper[4764]: I0202 09:28:21.308184 4764 generic.go:334] "Generic (PLEG): container finished" podID="5bacef6a-a70e-400b-963f-d3346b916dee" containerID="ea61bee5849658de7d6ee0a009ed7eb2d4a69344ed932c37e7cff559942691b0" exitCode=0 Feb 02 09:28:21 crc kubenswrapper[4764]: I0202 09:28:21.308984 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 02 09:28:21 crc kubenswrapper[4764]: I0202 09:28:21.309171 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5bacef6a-a70e-400b-963f-d3346b916dee","Type":"ContainerDied","Data":"ea61bee5849658de7d6ee0a009ed7eb2d4a69344ed932c37e7cff559942691b0"} Feb 02 09:28:21 crc kubenswrapper[4764]: I0202 09:28:21.309219 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5bacef6a-a70e-400b-963f-d3346b916dee","Type":"ContainerDied","Data":"622ff140637b2ce5b8480b66fad9ed0fef43f20de465dd7bff21c30bdb127da9"} Feb 02 09:28:21 crc kubenswrapper[4764]: I0202 09:28:21.309239 4764 scope.go:117] "RemoveContainer" containerID="ea61bee5849658de7d6ee0a009ed7eb2d4a69344ed932c37e7cff559942691b0" Feb 02 09:28:21 crc kubenswrapper[4764]: I0202 09:28:21.328196 4764 scope.go:117] "RemoveContainer" containerID="d0a31e50c1e0faeed1483cc4d170f915250dbb3c36dd7c5c83c8f8974270a037" Feb 02 09:28:21 crc kubenswrapper[4764]: I0202 09:28:21.353144 4764 scope.go:117] "RemoveContainer" containerID="ea61bee5849658de7d6ee0a009ed7eb2d4a69344ed932c37e7cff559942691b0" Feb 02 09:28:21 crc kubenswrapper[4764]: E0202 09:28:21.353704 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ea61bee5849658de7d6ee0a009ed7eb2d4a69344ed932c37e7cff559942691b0\": container with ID starting with ea61bee5849658de7d6ee0a009ed7eb2d4a69344ed932c37e7cff559942691b0 not found: ID does not exist" containerID="ea61bee5849658de7d6ee0a009ed7eb2d4a69344ed932c37e7cff559942691b0" Feb 02 09:28:21 crc kubenswrapper[4764]: I0202 09:28:21.353772 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ea61bee5849658de7d6ee0a009ed7eb2d4a69344ed932c37e7cff559942691b0"} err="failed to get container status \"ea61bee5849658de7d6ee0a009ed7eb2d4a69344ed932c37e7cff559942691b0\": rpc error: code = NotFound desc = could not find container \"ea61bee5849658de7d6ee0a009ed7eb2d4a69344ed932c37e7cff559942691b0\": container with ID starting with ea61bee5849658de7d6ee0a009ed7eb2d4a69344ed932c37e7cff559942691b0 not found: ID does not exist" Feb 02 09:28:21 crc kubenswrapper[4764]: I0202 09:28:21.353814 4764 scope.go:117] "RemoveContainer" containerID="d0a31e50c1e0faeed1483cc4d170f915250dbb3c36dd7c5c83c8f8974270a037" Feb 02 09:28:21 crc kubenswrapper[4764]: E0202 09:28:21.354278 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d0a31e50c1e0faeed1483cc4d170f915250dbb3c36dd7c5c83c8f8974270a037\": container with ID starting with d0a31e50c1e0faeed1483cc4d170f915250dbb3c36dd7c5c83c8f8974270a037 not found: ID does not exist" containerID="d0a31e50c1e0faeed1483cc4d170f915250dbb3c36dd7c5c83c8f8974270a037" Feb 02 09:28:21 crc kubenswrapper[4764]: I0202 09:28:21.354338 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d0a31e50c1e0faeed1483cc4d170f915250dbb3c36dd7c5c83c8f8974270a037"} err="failed to get container status \"d0a31e50c1e0faeed1483cc4d170f915250dbb3c36dd7c5c83c8f8974270a037\": rpc error: code = NotFound desc = could not find container \"d0a31e50c1e0faeed1483cc4d170f915250dbb3c36dd7c5c83c8f8974270a037\": container with ID starting with d0a31e50c1e0faeed1483cc4d170f915250dbb3c36dd7c5c83c8f8974270a037 not found: ID does not exist" Feb 02 09:28:21 crc kubenswrapper[4764]: I0202 09:28:21.373848 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Feb 02 09:28:21 crc kubenswrapper[4764]: I0202 09:28:21.410013 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Feb 02 09:28:21 crc kubenswrapper[4764]: I0202 09:28:21.417450 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Feb 02 09:28:21 crc kubenswrapper[4764]: E0202 09:28:21.417852 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5bacef6a-a70e-400b-963f-d3346b916dee" containerName="nova-metadata-log" Feb 02 09:28:21 crc kubenswrapper[4764]: I0202 09:28:21.417873 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="5bacef6a-a70e-400b-963f-d3346b916dee" containerName="nova-metadata-log" Feb 02 09:28:21 crc kubenswrapper[4764]: E0202 09:28:21.417883 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5bacef6a-a70e-400b-963f-d3346b916dee" containerName="nova-metadata-metadata" Feb 02 09:28:21 crc kubenswrapper[4764]: I0202 09:28:21.417891 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="5bacef6a-a70e-400b-963f-d3346b916dee" containerName="nova-metadata-metadata" Feb 02 09:28:21 crc kubenswrapper[4764]: I0202 09:28:21.418128 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="5bacef6a-a70e-400b-963f-d3346b916dee" containerName="nova-metadata-log" Feb 02 09:28:21 crc kubenswrapper[4764]: I0202 09:28:21.418154 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="5bacef6a-a70e-400b-963f-d3346b916dee" containerName="nova-metadata-metadata" Feb 02 09:28:21 crc kubenswrapper[4764]: I0202 09:28:21.419199 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 02 09:28:21 crc kubenswrapper[4764]: I0202 09:28:21.423698 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Feb 02 09:28:21 crc kubenswrapper[4764]: I0202 09:28:21.425338 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Feb 02 09:28:21 crc kubenswrapper[4764]: I0202 09:28:21.427283 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 02 09:28:21 crc kubenswrapper[4764]: I0202 09:28:21.486569 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/61eec87e-6dfc-41b8-a670-0914cec8b593-logs\") pod \"nova-metadata-0\" (UID: \"61eec87e-6dfc-41b8-a670-0914cec8b593\") " pod="openstack/nova-metadata-0" Feb 02 09:28:21 crc kubenswrapper[4764]: I0202 09:28:21.486609 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61eec87e-6dfc-41b8-a670-0914cec8b593-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"61eec87e-6dfc-41b8-a670-0914cec8b593\") " pod="openstack/nova-metadata-0" Feb 02 09:28:21 crc kubenswrapper[4764]: I0202 09:28:21.486847 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/61eec87e-6dfc-41b8-a670-0914cec8b593-config-data\") pod \"nova-metadata-0\" (UID: \"61eec87e-6dfc-41b8-a670-0914cec8b593\") " pod="openstack/nova-metadata-0" Feb 02 09:28:21 crc kubenswrapper[4764]: I0202 09:28:21.487133 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/61eec87e-6dfc-41b8-a670-0914cec8b593-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"61eec87e-6dfc-41b8-a670-0914cec8b593\") " pod="openstack/nova-metadata-0" Feb 02 09:28:21 crc kubenswrapper[4764]: I0202 09:28:21.487230 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s77nz\" (UniqueName: \"kubernetes.io/projected/61eec87e-6dfc-41b8-a670-0914cec8b593-kube-api-access-s77nz\") pod \"nova-metadata-0\" (UID: \"61eec87e-6dfc-41b8-a670-0914cec8b593\") " pod="openstack/nova-metadata-0" Feb 02 09:28:21 crc kubenswrapper[4764]: I0202 09:28:21.588396 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/61eec87e-6dfc-41b8-a670-0914cec8b593-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"61eec87e-6dfc-41b8-a670-0914cec8b593\") " pod="openstack/nova-metadata-0" Feb 02 09:28:21 crc kubenswrapper[4764]: I0202 09:28:21.588481 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s77nz\" (UniqueName: \"kubernetes.io/projected/61eec87e-6dfc-41b8-a670-0914cec8b593-kube-api-access-s77nz\") pod \"nova-metadata-0\" (UID: \"61eec87e-6dfc-41b8-a670-0914cec8b593\") " pod="openstack/nova-metadata-0" Feb 02 09:28:21 crc kubenswrapper[4764]: I0202 09:28:21.588543 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/61eec87e-6dfc-41b8-a670-0914cec8b593-logs\") pod \"nova-metadata-0\" (UID: \"61eec87e-6dfc-41b8-a670-0914cec8b593\") " pod="openstack/nova-metadata-0" Feb 02 09:28:21 crc kubenswrapper[4764]: I0202 09:28:21.588578 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61eec87e-6dfc-41b8-a670-0914cec8b593-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"61eec87e-6dfc-41b8-a670-0914cec8b593\") " pod="openstack/nova-metadata-0" Feb 02 09:28:21 crc kubenswrapper[4764]: I0202 09:28:21.588676 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/61eec87e-6dfc-41b8-a670-0914cec8b593-config-data\") pod \"nova-metadata-0\" (UID: \"61eec87e-6dfc-41b8-a670-0914cec8b593\") " pod="openstack/nova-metadata-0" Feb 02 09:28:21 crc kubenswrapper[4764]: I0202 09:28:21.589097 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/61eec87e-6dfc-41b8-a670-0914cec8b593-logs\") pod \"nova-metadata-0\" (UID: \"61eec87e-6dfc-41b8-a670-0914cec8b593\") " pod="openstack/nova-metadata-0" Feb 02 09:28:21 crc kubenswrapper[4764]: I0202 09:28:21.591234 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/61eec87e-6dfc-41b8-a670-0914cec8b593-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"61eec87e-6dfc-41b8-a670-0914cec8b593\") " pod="openstack/nova-metadata-0" Feb 02 09:28:21 crc kubenswrapper[4764]: I0202 09:28:21.592329 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61eec87e-6dfc-41b8-a670-0914cec8b593-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"61eec87e-6dfc-41b8-a670-0914cec8b593\") " pod="openstack/nova-metadata-0" Feb 02 09:28:21 crc kubenswrapper[4764]: I0202 09:28:21.596673 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/61eec87e-6dfc-41b8-a670-0914cec8b593-config-data\") pod \"nova-metadata-0\" (UID: \"61eec87e-6dfc-41b8-a670-0914cec8b593\") " pod="openstack/nova-metadata-0" Feb 02 09:28:21 crc kubenswrapper[4764]: I0202 09:28:21.608990 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s77nz\" (UniqueName: \"kubernetes.io/projected/61eec87e-6dfc-41b8-a670-0914cec8b593-kube-api-access-s77nz\") pod \"nova-metadata-0\" (UID: \"61eec87e-6dfc-41b8-a670-0914cec8b593\") " pod="openstack/nova-metadata-0" Feb 02 09:28:21 crc kubenswrapper[4764]: I0202 09:28:21.772542 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 02 09:28:21 crc kubenswrapper[4764]: I0202 09:28:21.836192 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5bacef6a-a70e-400b-963f-d3346b916dee" path="/var/lib/kubelet/pods/5bacef6a-a70e-400b-963f-d3346b916dee/volumes" Feb 02 09:28:22 crc kubenswrapper[4764]: W0202 09:28:22.244532 4764 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod61eec87e_6dfc_41b8_a670_0914cec8b593.slice/crio-e5a8b7fcc08a7e7c12b3f65e3de1c63479de630ac1e1ea7d0fa61addc2b48945 WatchSource:0}: Error finding container e5a8b7fcc08a7e7c12b3f65e3de1c63479de630ac1e1ea7d0fa61addc2b48945: Status 404 returned error can't find the container with id e5a8b7fcc08a7e7c12b3f65e3de1c63479de630ac1e1ea7d0fa61addc2b48945 Feb 02 09:28:22 crc kubenswrapper[4764]: I0202 09:28:22.249734 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 02 09:28:22 crc kubenswrapper[4764]: I0202 09:28:22.320556 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"61eec87e-6dfc-41b8-a670-0914cec8b593","Type":"ContainerStarted","Data":"e5a8b7fcc08a7e7c12b3f65e3de1c63479de630ac1e1ea7d0fa61addc2b48945"} Feb 02 09:28:23 crc kubenswrapper[4764]: I0202 09:28:23.340196 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"61eec87e-6dfc-41b8-a670-0914cec8b593","Type":"ContainerStarted","Data":"260fe9a7168afb4f75618d3b61dd3e4a6fcc890eeb18ad9c9a54b93140f6278b"} Feb 02 09:28:23 crc kubenswrapper[4764]: I0202 09:28:23.342218 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"61eec87e-6dfc-41b8-a670-0914cec8b593","Type":"ContainerStarted","Data":"1c9a26cc3bad76786cbc83090552c99e816fdcc04307dc46b270419ed70211b2"} Feb 02 09:28:23 crc kubenswrapper[4764]: I0202 09:28:23.373373 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.373349045 podStartE2EDuration="2.373349045s" podCreationTimestamp="2026-02-02 09:28:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:28:23.36151787 +0000 UTC m=+1266.295241968" watchObservedRunningTime="2026-02-02 09:28:23.373349045 +0000 UTC m=+1266.307073143" Feb 02 09:28:24 crc kubenswrapper[4764]: I0202 09:28:24.681709 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Feb 02 09:28:26 crc kubenswrapper[4764]: I0202 09:28:26.773331 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Feb 02 09:28:26 crc kubenswrapper[4764]: I0202 09:28:26.773405 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Feb 02 09:28:29 crc kubenswrapper[4764]: I0202 09:28:29.049185 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 02 09:28:29 crc kubenswrapper[4764]: I0202 09:28:29.049589 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 02 09:28:29 crc kubenswrapper[4764]: I0202 09:28:29.681255 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Feb 02 09:28:29 crc kubenswrapper[4764]: I0202 09:28:29.739316 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Feb 02 09:28:30 crc kubenswrapper[4764]: I0202 09:28:30.063304 4764 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="db89c9c9-cad1-4e6b-aaf0-708e5b8b32b4" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.191:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 02 09:28:30 crc kubenswrapper[4764]: I0202 09:28:30.063499 4764 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="db89c9c9-cad1-4e6b-aaf0-708e5b8b32b4" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.191:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 02 09:28:30 crc kubenswrapper[4764]: I0202 09:28:30.493001 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Feb 02 09:28:31 crc kubenswrapper[4764]: I0202 09:28:31.773623 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Feb 02 09:28:31 crc kubenswrapper[4764]: I0202 09:28:31.774077 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Feb 02 09:28:32 crc kubenswrapper[4764]: I0202 09:28:32.786159 4764 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="61eec87e-6dfc-41b8-a670-0914cec8b593" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.193:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 02 09:28:32 crc kubenswrapper[4764]: I0202 09:28:32.786253 4764 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="61eec87e-6dfc-41b8-a670-0914cec8b593" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.193:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 02 09:28:39 crc kubenswrapper[4764]: I0202 09:28:39.059603 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Feb 02 09:28:39 crc kubenswrapper[4764]: I0202 09:28:39.060586 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Feb 02 09:28:39 crc kubenswrapper[4764]: I0202 09:28:39.062060 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Feb 02 09:28:39 crc kubenswrapper[4764]: I0202 09:28:39.062264 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Feb 02 09:28:39 crc kubenswrapper[4764]: I0202 09:28:39.073831 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Feb 02 09:28:39 crc kubenswrapper[4764]: I0202 09:28:39.075530 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Feb 02 09:28:41 crc kubenswrapper[4764]: I0202 09:28:41.782321 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Feb 02 09:28:41 crc kubenswrapper[4764]: I0202 09:28:41.782783 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Feb 02 09:28:41 crc kubenswrapper[4764]: I0202 09:28:41.818740 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Feb 02 09:28:41 crc kubenswrapper[4764]: I0202 09:28:41.823629 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Feb 02 09:28:43 crc kubenswrapper[4764]: I0202 09:28:43.523591 4764 patch_prober.go:28] interesting pod/machine-config-daemon-4ndm4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 09:28:43 crc kubenswrapper[4764]: I0202 09:28:43.524072 4764 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 09:28:44 crc kubenswrapper[4764]: I0202 09:28:44.651636 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Feb 02 09:28:52 crc kubenswrapper[4764]: I0202 09:28:52.917741 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 02 09:28:53 crc kubenswrapper[4764]: I0202 09:28:53.858661 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 02 09:28:57 crc kubenswrapper[4764]: I0202 09:28:57.747178 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="53d5d300-11ae-408d-8bb8-347defb93991" containerName="rabbitmq" containerID="cri-o://352eac9c43f64896394cf8c78ba1f8a36be490958f2022459a08cd79e7b607ea" gracePeriod=604796 Feb 02 09:28:58 crc kubenswrapper[4764]: I0202 09:28:58.444028 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="03d9f090-4a93-4226-8fce-6fd0a425d929" containerName="rabbitmq" containerID="cri-o://b0556eef8b622b39f414f040258f94501b1ff75d65d0e71032e333b6b15e8b4f" gracePeriod=604796 Feb 02 09:29:04 crc kubenswrapper[4764]: I0202 09:29:04.031640 4764 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="53d5d300-11ae-408d-8bb8-347defb93991" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.97:5671: connect: connection refused" Feb 02 09:29:04 crc kubenswrapper[4764]: I0202 09:29:04.315135 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Feb 02 09:29:04 crc kubenswrapper[4764]: I0202 09:29:04.435872 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/53d5d300-11ae-408d-8bb8-347defb93991-erlang-cookie-secret\") pod \"53d5d300-11ae-408d-8bb8-347defb93991\" (UID: \"53d5d300-11ae-408d-8bb8-347defb93991\") " Feb 02 09:29:04 crc kubenswrapper[4764]: I0202 09:29:04.435946 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"53d5d300-11ae-408d-8bb8-347defb93991\" (UID: \"53d5d300-11ae-408d-8bb8-347defb93991\") " Feb 02 09:29:04 crc kubenswrapper[4764]: I0202 09:29:04.435984 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/53d5d300-11ae-408d-8bb8-347defb93991-server-conf\") pod \"53d5d300-11ae-408d-8bb8-347defb93991\" (UID: \"53d5d300-11ae-408d-8bb8-347defb93991\") " Feb 02 09:29:04 crc kubenswrapper[4764]: I0202 09:29:04.436034 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vs58s\" (UniqueName: \"kubernetes.io/projected/53d5d300-11ae-408d-8bb8-347defb93991-kube-api-access-vs58s\") pod \"53d5d300-11ae-408d-8bb8-347defb93991\" (UID: \"53d5d300-11ae-408d-8bb8-347defb93991\") " Feb 02 09:29:04 crc kubenswrapper[4764]: I0202 09:29:04.436053 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/53d5d300-11ae-408d-8bb8-347defb93991-config-data\") pod \"53d5d300-11ae-408d-8bb8-347defb93991\" (UID: \"53d5d300-11ae-408d-8bb8-347defb93991\") " Feb 02 09:29:04 crc kubenswrapper[4764]: I0202 09:29:04.436070 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/53d5d300-11ae-408d-8bb8-347defb93991-rabbitmq-tls\") pod \"53d5d300-11ae-408d-8bb8-347defb93991\" (UID: \"53d5d300-11ae-408d-8bb8-347defb93991\") " Feb 02 09:29:04 crc kubenswrapper[4764]: I0202 09:29:04.436134 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/53d5d300-11ae-408d-8bb8-347defb93991-rabbitmq-plugins\") pod \"53d5d300-11ae-408d-8bb8-347defb93991\" (UID: \"53d5d300-11ae-408d-8bb8-347defb93991\") " Feb 02 09:29:04 crc kubenswrapper[4764]: I0202 09:29:04.436157 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/53d5d300-11ae-408d-8bb8-347defb93991-plugins-conf\") pod \"53d5d300-11ae-408d-8bb8-347defb93991\" (UID: \"53d5d300-11ae-408d-8bb8-347defb93991\") " Feb 02 09:29:04 crc kubenswrapper[4764]: I0202 09:29:04.436203 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/53d5d300-11ae-408d-8bb8-347defb93991-rabbitmq-confd\") pod \"53d5d300-11ae-408d-8bb8-347defb93991\" (UID: \"53d5d300-11ae-408d-8bb8-347defb93991\") " Feb 02 09:29:04 crc kubenswrapper[4764]: I0202 09:29:04.436246 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/53d5d300-11ae-408d-8bb8-347defb93991-pod-info\") pod \"53d5d300-11ae-408d-8bb8-347defb93991\" (UID: \"53d5d300-11ae-408d-8bb8-347defb93991\") " Feb 02 09:29:04 crc kubenswrapper[4764]: I0202 09:29:04.436329 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/53d5d300-11ae-408d-8bb8-347defb93991-rabbitmq-erlang-cookie\") pod \"53d5d300-11ae-408d-8bb8-347defb93991\" (UID: \"53d5d300-11ae-408d-8bb8-347defb93991\") " Feb 02 09:29:04 crc kubenswrapper[4764]: I0202 09:29:04.437054 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/53d5d300-11ae-408d-8bb8-347defb93991-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "53d5d300-11ae-408d-8bb8-347defb93991" (UID: "53d5d300-11ae-408d-8bb8-347defb93991"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 09:29:04 crc kubenswrapper[4764]: I0202 09:29:04.437261 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/53d5d300-11ae-408d-8bb8-347defb93991-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "53d5d300-11ae-408d-8bb8-347defb93991" (UID: "53d5d300-11ae-408d-8bb8-347defb93991"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 09:29:04 crc kubenswrapper[4764]: I0202 09:29:04.437715 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/53d5d300-11ae-408d-8bb8-347defb93991-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "53d5d300-11ae-408d-8bb8-347defb93991" (UID: "53d5d300-11ae-408d-8bb8-347defb93991"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:29:04 crc kubenswrapper[4764]: I0202 09:29:04.449719 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "persistence") pod "53d5d300-11ae-408d-8bb8-347defb93991" (UID: "53d5d300-11ae-408d-8bb8-347defb93991"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 02 09:29:04 crc kubenswrapper[4764]: I0202 09:29:04.451435 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53d5d300-11ae-408d-8bb8-347defb93991-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "53d5d300-11ae-408d-8bb8-347defb93991" (UID: "53d5d300-11ae-408d-8bb8-347defb93991"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:29:04 crc kubenswrapper[4764]: I0202 09:29:04.479001 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/53d5d300-11ae-408d-8bb8-347defb93991-kube-api-access-vs58s" (OuterVolumeSpecName: "kube-api-access-vs58s") pod "53d5d300-11ae-408d-8bb8-347defb93991" (UID: "53d5d300-11ae-408d-8bb8-347defb93991"). InnerVolumeSpecName "kube-api-access-vs58s". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:29:04 crc kubenswrapper[4764]: I0202 09:29:04.479130 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/53d5d300-11ae-408d-8bb8-347defb93991-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "53d5d300-11ae-408d-8bb8-347defb93991" (UID: "53d5d300-11ae-408d-8bb8-347defb93991"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:29:04 crc kubenswrapper[4764]: I0202 09:29:04.479155 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/53d5d300-11ae-408d-8bb8-347defb93991-pod-info" (OuterVolumeSpecName: "pod-info") pod "53d5d300-11ae-408d-8bb8-347defb93991" (UID: "53d5d300-11ae-408d-8bb8-347defb93991"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Feb 02 09:29:04 crc kubenswrapper[4764]: I0202 09:29:04.506274 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/53d5d300-11ae-408d-8bb8-347defb93991-config-data" (OuterVolumeSpecName: "config-data") pod "53d5d300-11ae-408d-8bb8-347defb93991" (UID: "53d5d300-11ae-408d-8bb8-347defb93991"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:29:04 crc kubenswrapper[4764]: I0202 09:29:04.525436 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/53d5d300-11ae-408d-8bb8-347defb93991-server-conf" (OuterVolumeSpecName: "server-conf") pod "53d5d300-11ae-408d-8bb8-347defb93991" (UID: "53d5d300-11ae-408d-8bb8-347defb93991"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:29:04 crc kubenswrapper[4764]: I0202 09:29:04.537415 4764 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="03d9f090-4a93-4226-8fce-6fd0a425d929" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.98:5671: connect: connection refused" Feb 02 09:29:04 crc kubenswrapper[4764]: I0202 09:29:04.537973 4764 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/53d5d300-11ae-408d-8bb8-347defb93991-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Feb 02 09:29:04 crc kubenswrapper[4764]: I0202 09:29:04.538017 4764 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Feb 02 09:29:04 crc kubenswrapper[4764]: I0202 09:29:04.538028 4764 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/53d5d300-11ae-408d-8bb8-347defb93991-server-conf\") on node \"crc\" DevicePath \"\"" Feb 02 09:29:04 crc kubenswrapper[4764]: I0202 09:29:04.538038 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vs58s\" (UniqueName: \"kubernetes.io/projected/53d5d300-11ae-408d-8bb8-347defb93991-kube-api-access-vs58s\") on node \"crc\" DevicePath \"\"" Feb 02 09:29:04 crc kubenswrapper[4764]: I0202 09:29:04.538049 4764 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/53d5d300-11ae-408d-8bb8-347defb93991-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 09:29:04 crc kubenswrapper[4764]: I0202 09:29:04.538058 4764 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/53d5d300-11ae-408d-8bb8-347defb93991-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Feb 02 09:29:04 crc kubenswrapper[4764]: I0202 09:29:04.538067 4764 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/53d5d300-11ae-408d-8bb8-347defb93991-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Feb 02 09:29:04 crc kubenswrapper[4764]: I0202 09:29:04.538075 4764 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/53d5d300-11ae-408d-8bb8-347defb93991-plugins-conf\") on node \"crc\" DevicePath \"\"" Feb 02 09:29:04 crc kubenswrapper[4764]: I0202 09:29:04.538084 4764 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/53d5d300-11ae-408d-8bb8-347defb93991-pod-info\") on node \"crc\" DevicePath \"\"" Feb 02 09:29:04 crc kubenswrapper[4764]: I0202 09:29:04.538093 4764 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/53d5d300-11ae-408d-8bb8-347defb93991-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Feb 02 09:29:04 crc kubenswrapper[4764]: I0202 09:29:04.557235 4764 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Feb 02 09:29:04 crc kubenswrapper[4764]: I0202 09:29:04.565129 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/53d5d300-11ae-408d-8bb8-347defb93991-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "53d5d300-11ae-408d-8bb8-347defb93991" (UID: "53d5d300-11ae-408d-8bb8-347defb93991"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:29:04 crc kubenswrapper[4764]: I0202 09:29:04.639976 4764 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Feb 02 09:29:04 crc kubenswrapper[4764]: I0202 09:29:04.640006 4764 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/53d5d300-11ae-408d-8bb8-347defb93991-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Feb 02 09:29:04 crc kubenswrapper[4764]: I0202 09:29:04.853838 4764 generic.go:334] "Generic (PLEG): container finished" podID="03d9f090-4a93-4226-8fce-6fd0a425d929" containerID="b0556eef8b622b39f414f040258f94501b1ff75d65d0e71032e333b6b15e8b4f" exitCode=0 Feb 02 09:29:04 crc kubenswrapper[4764]: I0202 09:29:04.853899 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"03d9f090-4a93-4226-8fce-6fd0a425d929","Type":"ContainerDied","Data":"b0556eef8b622b39f414f040258f94501b1ff75d65d0e71032e333b6b15e8b4f"} Feb 02 09:29:04 crc kubenswrapper[4764]: I0202 09:29:04.864914 4764 generic.go:334] "Generic (PLEG): container finished" podID="53d5d300-11ae-408d-8bb8-347defb93991" containerID="352eac9c43f64896394cf8c78ba1f8a36be490958f2022459a08cd79e7b607ea" exitCode=0 Feb 02 09:29:04 crc kubenswrapper[4764]: I0202 09:29:04.865169 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"53d5d300-11ae-408d-8bb8-347defb93991","Type":"ContainerDied","Data":"352eac9c43f64896394cf8c78ba1f8a36be490958f2022459a08cd79e7b607ea"} Feb 02 09:29:04 crc kubenswrapper[4764]: I0202 09:29:04.865253 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"53d5d300-11ae-408d-8bb8-347defb93991","Type":"ContainerDied","Data":"5111e0090c4901a2d85fc188815016dd3aa41b8c15a1b826e3b6edbb5e86e2cc"} Feb 02 09:29:04 crc kubenswrapper[4764]: I0202 09:29:04.865329 4764 scope.go:117] "RemoveContainer" containerID="352eac9c43f64896394cf8c78ba1f8a36be490958f2022459a08cd79e7b607ea" Feb 02 09:29:04 crc kubenswrapper[4764]: I0202 09:29:04.865718 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Feb 02 09:29:04 crc kubenswrapper[4764]: I0202 09:29:04.906433 4764 scope.go:117] "RemoveContainer" containerID="f5641821f6b2d966369e040f6849f892c7244535560aa2d911c4599545382008" Feb 02 09:29:04 crc kubenswrapper[4764]: I0202 09:29:04.980276 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 02 09:29:04 crc kubenswrapper[4764]: I0202 09:29:04.987261 4764 scope.go:117] "RemoveContainer" containerID="352eac9c43f64896394cf8c78ba1f8a36be490958f2022459a08cd79e7b607ea" Feb 02 09:29:04 crc kubenswrapper[4764]: I0202 09:29:04.995189 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 02 09:29:04 crc kubenswrapper[4764]: E0202 09:29:04.996552 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"352eac9c43f64896394cf8c78ba1f8a36be490958f2022459a08cd79e7b607ea\": container with ID starting with 352eac9c43f64896394cf8c78ba1f8a36be490958f2022459a08cd79e7b607ea not found: ID does not exist" containerID="352eac9c43f64896394cf8c78ba1f8a36be490958f2022459a08cd79e7b607ea" Feb 02 09:29:04 crc kubenswrapper[4764]: I0202 09:29:04.996587 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"352eac9c43f64896394cf8c78ba1f8a36be490958f2022459a08cd79e7b607ea"} err="failed to get container status \"352eac9c43f64896394cf8c78ba1f8a36be490958f2022459a08cd79e7b607ea\": rpc error: code = NotFound desc = could not find container \"352eac9c43f64896394cf8c78ba1f8a36be490958f2022459a08cd79e7b607ea\": container with ID starting with 352eac9c43f64896394cf8c78ba1f8a36be490958f2022459a08cd79e7b607ea not found: ID does not exist" Feb 02 09:29:04 crc kubenswrapper[4764]: I0202 09:29:04.996614 4764 scope.go:117] "RemoveContainer" containerID="f5641821f6b2d966369e040f6849f892c7244535560aa2d911c4599545382008" Feb 02 09:29:05 crc kubenswrapper[4764]: I0202 09:29:05.004353 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Feb 02 09:29:05 crc kubenswrapper[4764]: E0202 09:29:05.004773 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53d5d300-11ae-408d-8bb8-347defb93991" containerName="rabbitmq" Feb 02 09:29:05 crc kubenswrapper[4764]: I0202 09:29:05.004786 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="53d5d300-11ae-408d-8bb8-347defb93991" containerName="rabbitmq" Feb 02 09:29:05 crc kubenswrapper[4764]: E0202 09:29:05.004814 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53d5d300-11ae-408d-8bb8-347defb93991" containerName="setup-container" Feb 02 09:29:05 crc kubenswrapper[4764]: I0202 09:29:05.004820 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="53d5d300-11ae-408d-8bb8-347defb93991" containerName="setup-container" Feb 02 09:29:05 crc kubenswrapper[4764]: I0202 09:29:05.005013 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="53d5d300-11ae-408d-8bb8-347defb93991" containerName="rabbitmq" Feb 02 09:29:05 crc kubenswrapper[4764]: I0202 09:29:05.006330 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Feb 02 09:29:05 crc kubenswrapper[4764]: E0202 09:29:05.006637 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f5641821f6b2d966369e040f6849f892c7244535560aa2d911c4599545382008\": container with ID starting with f5641821f6b2d966369e040f6849f892c7244535560aa2d911c4599545382008 not found: ID does not exist" containerID="f5641821f6b2d966369e040f6849f892c7244535560aa2d911c4599545382008" Feb 02 09:29:05 crc kubenswrapper[4764]: I0202 09:29:05.006785 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f5641821f6b2d966369e040f6849f892c7244535560aa2d911c4599545382008"} err="failed to get container status \"f5641821f6b2d966369e040f6849f892c7244535560aa2d911c4599545382008\": rpc error: code = NotFound desc = could not find container \"f5641821f6b2d966369e040f6849f892c7244535560aa2d911c4599545382008\": container with ID starting with f5641821f6b2d966369e040f6849f892c7244535560aa2d911c4599545382008 not found: ID does not exist" Feb 02 09:29:05 crc kubenswrapper[4764]: I0202 09:29:05.011562 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 02 09:29:05 crc kubenswrapper[4764]: I0202 09:29:05.014765 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Feb 02 09:29:05 crc kubenswrapper[4764]: I0202 09:29:05.014982 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-9q4h7" Feb 02 09:29:05 crc kubenswrapper[4764]: I0202 09:29:05.015124 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Feb 02 09:29:05 crc kubenswrapper[4764]: I0202 09:29:05.015225 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Feb 02 09:29:05 crc kubenswrapper[4764]: I0202 09:29:05.015476 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Feb 02 09:29:05 crc kubenswrapper[4764]: I0202 09:29:05.015644 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Feb 02 09:29:05 crc kubenswrapper[4764]: I0202 09:29:05.015797 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Feb 02 09:29:05 crc kubenswrapper[4764]: I0202 09:29:05.084631 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Feb 02 09:29:05 crc kubenswrapper[4764]: I0202 09:29:05.169895 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ec28045f-e1d4-4b22-bb7d-3c6d2a1f1a38-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"ec28045f-e1d4-4b22-bb7d-3c6d2a1f1a38\") " pod="openstack/rabbitmq-server-0" Feb 02 09:29:05 crc kubenswrapper[4764]: I0202 09:29:05.170027 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ec28045f-e1d4-4b22-bb7d-3c6d2a1f1a38-config-data\") pod \"rabbitmq-server-0\" (UID: \"ec28045f-e1d4-4b22-bb7d-3c6d2a1f1a38\") " pod="openstack/rabbitmq-server-0" Feb 02 09:29:05 crc kubenswrapper[4764]: I0202 09:29:05.170050 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xzvdh\" (UniqueName: \"kubernetes.io/projected/ec28045f-e1d4-4b22-bb7d-3c6d2a1f1a38-kube-api-access-xzvdh\") pod \"rabbitmq-server-0\" (UID: \"ec28045f-e1d4-4b22-bb7d-3c6d2a1f1a38\") " pod="openstack/rabbitmq-server-0" Feb 02 09:29:05 crc kubenswrapper[4764]: I0202 09:29:05.170106 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ec28045f-e1d4-4b22-bb7d-3c6d2a1f1a38-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"ec28045f-e1d4-4b22-bb7d-3c6d2a1f1a38\") " pod="openstack/rabbitmq-server-0" Feb 02 09:29:05 crc kubenswrapper[4764]: I0202 09:29:05.170152 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ec28045f-e1d4-4b22-bb7d-3c6d2a1f1a38-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"ec28045f-e1d4-4b22-bb7d-3c6d2a1f1a38\") " pod="openstack/rabbitmq-server-0" Feb 02 09:29:05 crc kubenswrapper[4764]: I0202 09:29:05.170183 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ec28045f-e1d4-4b22-bb7d-3c6d2a1f1a38-server-conf\") pod \"rabbitmq-server-0\" (UID: \"ec28045f-e1d4-4b22-bb7d-3c6d2a1f1a38\") " pod="openstack/rabbitmq-server-0" Feb 02 09:29:05 crc kubenswrapper[4764]: I0202 09:29:05.170226 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-server-0\" (UID: \"ec28045f-e1d4-4b22-bb7d-3c6d2a1f1a38\") " pod="openstack/rabbitmq-server-0" Feb 02 09:29:05 crc kubenswrapper[4764]: I0202 09:29:05.170309 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ec28045f-e1d4-4b22-bb7d-3c6d2a1f1a38-pod-info\") pod \"rabbitmq-server-0\" (UID: \"ec28045f-e1d4-4b22-bb7d-3c6d2a1f1a38\") " pod="openstack/rabbitmq-server-0" Feb 02 09:29:05 crc kubenswrapper[4764]: I0202 09:29:05.170360 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ec28045f-e1d4-4b22-bb7d-3c6d2a1f1a38-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"ec28045f-e1d4-4b22-bb7d-3c6d2a1f1a38\") " pod="openstack/rabbitmq-server-0" Feb 02 09:29:05 crc kubenswrapper[4764]: I0202 09:29:05.170377 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ec28045f-e1d4-4b22-bb7d-3c6d2a1f1a38-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"ec28045f-e1d4-4b22-bb7d-3c6d2a1f1a38\") " pod="openstack/rabbitmq-server-0" Feb 02 09:29:05 crc kubenswrapper[4764]: I0202 09:29:05.170411 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ec28045f-e1d4-4b22-bb7d-3c6d2a1f1a38-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"ec28045f-e1d4-4b22-bb7d-3c6d2a1f1a38\") " pod="openstack/rabbitmq-server-0" Feb 02 09:29:05 crc kubenswrapper[4764]: I0202 09:29:05.276536 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/03d9f090-4a93-4226-8fce-6fd0a425d929-rabbitmq-plugins\") pod \"03d9f090-4a93-4226-8fce-6fd0a425d929\" (UID: \"03d9f090-4a93-4226-8fce-6fd0a425d929\") " Feb 02 09:29:05 crc kubenswrapper[4764]: I0202 09:29:05.276600 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-27wk5\" (UniqueName: \"kubernetes.io/projected/03d9f090-4a93-4226-8fce-6fd0a425d929-kube-api-access-27wk5\") pod \"03d9f090-4a93-4226-8fce-6fd0a425d929\" (UID: \"03d9f090-4a93-4226-8fce-6fd0a425d929\") " Feb 02 09:29:05 crc kubenswrapper[4764]: I0202 09:29:05.276625 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/03d9f090-4a93-4226-8fce-6fd0a425d929-rabbitmq-confd\") pod \"03d9f090-4a93-4226-8fce-6fd0a425d929\" (UID: \"03d9f090-4a93-4226-8fce-6fd0a425d929\") " Feb 02 09:29:05 crc kubenswrapper[4764]: I0202 09:29:05.276672 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"03d9f090-4a93-4226-8fce-6fd0a425d929\" (UID: \"03d9f090-4a93-4226-8fce-6fd0a425d929\") " Feb 02 09:29:05 crc kubenswrapper[4764]: I0202 09:29:05.276691 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/03d9f090-4a93-4226-8fce-6fd0a425d929-erlang-cookie-secret\") pod \"03d9f090-4a93-4226-8fce-6fd0a425d929\" (UID: \"03d9f090-4a93-4226-8fce-6fd0a425d929\") " Feb 02 09:29:05 crc kubenswrapper[4764]: I0202 09:29:05.276729 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/03d9f090-4a93-4226-8fce-6fd0a425d929-plugins-conf\") pod \"03d9f090-4a93-4226-8fce-6fd0a425d929\" (UID: \"03d9f090-4a93-4226-8fce-6fd0a425d929\") " Feb 02 09:29:05 crc kubenswrapper[4764]: I0202 09:29:05.276743 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/03d9f090-4a93-4226-8fce-6fd0a425d929-config-data\") pod \"03d9f090-4a93-4226-8fce-6fd0a425d929\" (UID: \"03d9f090-4a93-4226-8fce-6fd0a425d929\") " Feb 02 09:29:05 crc kubenswrapper[4764]: I0202 09:29:05.276764 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/03d9f090-4a93-4226-8fce-6fd0a425d929-rabbitmq-tls\") pod \"03d9f090-4a93-4226-8fce-6fd0a425d929\" (UID: \"03d9f090-4a93-4226-8fce-6fd0a425d929\") " Feb 02 09:29:05 crc kubenswrapper[4764]: I0202 09:29:05.276893 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/03d9f090-4a93-4226-8fce-6fd0a425d929-rabbitmq-erlang-cookie\") pod \"03d9f090-4a93-4226-8fce-6fd0a425d929\" (UID: \"03d9f090-4a93-4226-8fce-6fd0a425d929\") " Feb 02 09:29:05 crc kubenswrapper[4764]: I0202 09:29:05.276956 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/03d9f090-4a93-4226-8fce-6fd0a425d929-pod-info\") pod \"03d9f090-4a93-4226-8fce-6fd0a425d929\" (UID: \"03d9f090-4a93-4226-8fce-6fd0a425d929\") " Feb 02 09:29:05 crc kubenswrapper[4764]: I0202 09:29:05.276981 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/03d9f090-4a93-4226-8fce-6fd0a425d929-server-conf\") pod \"03d9f090-4a93-4226-8fce-6fd0a425d929\" (UID: \"03d9f090-4a93-4226-8fce-6fd0a425d929\") " Feb 02 09:29:05 crc kubenswrapper[4764]: I0202 09:29:05.277231 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ec28045f-e1d4-4b22-bb7d-3c6d2a1f1a38-pod-info\") pod \"rabbitmq-server-0\" (UID: \"ec28045f-e1d4-4b22-bb7d-3c6d2a1f1a38\") " pod="openstack/rabbitmq-server-0" Feb 02 09:29:05 crc kubenswrapper[4764]: I0202 09:29:05.277274 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ec28045f-e1d4-4b22-bb7d-3c6d2a1f1a38-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"ec28045f-e1d4-4b22-bb7d-3c6d2a1f1a38\") " pod="openstack/rabbitmq-server-0" Feb 02 09:29:05 crc kubenswrapper[4764]: I0202 09:29:05.277291 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ec28045f-e1d4-4b22-bb7d-3c6d2a1f1a38-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"ec28045f-e1d4-4b22-bb7d-3c6d2a1f1a38\") " pod="openstack/rabbitmq-server-0" Feb 02 09:29:05 crc kubenswrapper[4764]: I0202 09:29:05.277313 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ec28045f-e1d4-4b22-bb7d-3c6d2a1f1a38-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"ec28045f-e1d4-4b22-bb7d-3c6d2a1f1a38\") " pod="openstack/rabbitmq-server-0" Feb 02 09:29:05 crc kubenswrapper[4764]: I0202 09:29:05.277341 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ec28045f-e1d4-4b22-bb7d-3c6d2a1f1a38-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"ec28045f-e1d4-4b22-bb7d-3c6d2a1f1a38\") " pod="openstack/rabbitmq-server-0" Feb 02 09:29:05 crc kubenswrapper[4764]: I0202 09:29:05.277408 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ec28045f-e1d4-4b22-bb7d-3c6d2a1f1a38-config-data\") pod \"rabbitmq-server-0\" (UID: \"ec28045f-e1d4-4b22-bb7d-3c6d2a1f1a38\") " pod="openstack/rabbitmq-server-0" Feb 02 09:29:05 crc kubenswrapper[4764]: I0202 09:29:05.277425 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xzvdh\" (UniqueName: \"kubernetes.io/projected/ec28045f-e1d4-4b22-bb7d-3c6d2a1f1a38-kube-api-access-xzvdh\") pod \"rabbitmq-server-0\" (UID: \"ec28045f-e1d4-4b22-bb7d-3c6d2a1f1a38\") " pod="openstack/rabbitmq-server-0" Feb 02 09:29:05 crc kubenswrapper[4764]: I0202 09:29:05.277457 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ec28045f-e1d4-4b22-bb7d-3c6d2a1f1a38-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"ec28045f-e1d4-4b22-bb7d-3c6d2a1f1a38\") " pod="openstack/rabbitmq-server-0" Feb 02 09:29:05 crc kubenswrapper[4764]: I0202 09:29:05.277477 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ec28045f-e1d4-4b22-bb7d-3c6d2a1f1a38-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"ec28045f-e1d4-4b22-bb7d-3c6d2a1f1a38\") " pod="openstack/rabbitmq-server-0" Feb 02 09:29:05 crc kubenswrapper[4764]: I0202 09:29:05.277509 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ec28045f-e1d4-4b22-bb7d-3c6d2a1f1a38-server-conf\") pod \"rabbitmq-server-0\" (UID: \"ec28045f-e1d4-4b22-bb7d-3c6d2a1f1a38\") " pod="openstack/rabbitmq-server-0" Feb 02 09:29:05 crc kubenswrapper[4764]: I0202 09:29:05.277527 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-server-0\" (UID: \"ec28045f-e1d4-4b22-bb7d-3c6d2a1f1a38\") " pod="openstack/rabbitmq-server-0" Feb 02 09:29:05 crc kubenswrapper[4764]: I0202 09:29:05.277887 4764 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-server-0\" (UID: \"ec28045f-e1d4-4b22-bb7d-3c6d2a1f1a38\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/rabbitmq-server-0" Feb 02 09:29:05 crc kubenswrapper[4764]: I0202 09:29:05.288459 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ec28045f-e1d4-4b22-bb7d-3c6d2a1f1a38-config-data\") pod \"rabbitmq-server-0\" (UID: \"ec28045f-e1d4-4b22-bb7d-3c6d2a1f1a38\") " pod="openstack/rabbitmq-server-0" Feb 02 09:29:05 crc kubenswrapper[4764]: I0202 09:29:05.291627 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ec28045f-e1d4-4b22-bb7d-3c6d2a1f1a38-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"ec28045f-e1d4-4b22-bb7d-3c6d2a1f1a38\") " pod="openstack/rabbitmq-server-0" Feb 02 09:29:05 crc kubenswrapper[4764]: I0202 09:29:05.294018 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/03d9f090-4a93-4226-8fce-6fd0a425d929-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "03d9f090-4a93-4226-8fce-6fd0a425d929" (UID: "03d9f090-4a93-4226-8fce-6fd0a425d929"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:29:05 crc kubenswrapper[4764]: I0202 09:29:05.295423 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/03d9f090-4a93-4226-8fce-6fd0a425d929-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "03d9f090-4a93-4226-8fce-6fd0a425d929" (UID: "03d9f090-4a93-4226-8fce-6fd0a425d929"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 09:29:05 crc kubenswrapper[4764]: I0202 09:29:05.295747 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/03d9f090-4a93-4226-8fce-6fd0a425d929-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "03d9f090-4a93-4226-8fce-6fd0a425d929" (UID: "03d9f090-4a93-4226-8fce-6fd0a425d929"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:29:05 crc kubenswrapper[4764]: I0202 09:29:05.296118 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/03d9f090-4a93-4226-8fce-6fd0a425d929-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "03d9f090-4a93-4226-8fce-6fd0a425d929" (UID: "03d9f090-4a93-4226-8fce-6fd0a425d929"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 09:29:05 crc kubenswrapper[4764]: I0202 09:29:05.308654 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ec28045f-e1d4-4b22-bb7d-3c6d2a1f1a38-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"ec28045f-e1d4-4b22-bb7d-3c6d2a1f1a38\") " pod="openstack/rabbitmq-server-0" Feb 02 09:29:05 crc kubenswrapper[4764]: I0202 09:29:05.316110 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/03d9f090-4a93-4226-8fce-6fd0a425d929-kube-api-access-27wk5" (OuterVolumeSpecName: "kube-api-access-27wk5") pod "03d9f090-4a93-4226-8fce-6fd0a425d929" (UID: "03d9f090-4a93-4226-8fce-6fd0a425d929"). InnerVolumeSpecName "kube-api-access-27wk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:29:05 crc kubenswrapper[4764]: I0202 09:29:05.317250 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ec28045f-e1d4-4b22-bb7d-3c6d2a1f1a38-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"ec28045f-e1d4-4b22-bb7d-3c6d2a1f1a38\") " pod="openstack/rabbitmq-server-0" Feb 02 09:29:05 crc kubenswrapper[4764]: I0202 09:29:05.317817 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ec28045f-e1d4-4b22-bb7d-3c6d2a1f1a38-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"ec28045f-e1d4-4b22-bb7d-3c6d2a1f1a38\") " pod="openstack/rabbitmq-server-0" Feb 02 09:29:05 crc kubenswrapper[4764]: I0202 09:29:05.320211 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ec28045f-e1d4-4b22-bb7d-3c6d2a1f1a38-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"ec28045f-e1d4-4b22-bb7d-3c6d2a1f1a38\") " pod="openstack/rabbitmq-server-0" Feb 02 09:29:05 crc kubenswrapper[4764]: I0202 09:29:05.321268 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ec28045f-e1d4-4b22-bb7d-3c6d2a1f1a38-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"ec28045f-e1d4-4b22-bb7d-3c6d2a1f1a38\") " pod="openstack/rabbitmq-server-0" Feb 02 09:29:05 crc kubenswrapper[4764]: I0202 09:29:05.330605 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ec28045f-e1d4-4b22-bb7d-3c6d2a1f1a38-server-conf\") pod \"rabbitmq-server-0\" (UID: \"ec28045f-e1d4-4b22-bb7d-3c6d2a1f1a38\") " pod="openstack/rabbitmq-server-0" Feb 02 09:29:05 crc kubenswrapper[4764]: I0202 09:29:05.335731 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/03d9f090-4a93-4226-8fce-6fd0a425d929-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "03d9f090-4a93-4226-8fce-6fd0a425d929" (UID: "03d9f090-4a93-4226-8fce-6fd0a425d929"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:29:05 crc kubenswrapper[4764]: I0202 09:29:05.337695 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ec28045f-e1d4-4b22-bb7d-3c6d2a1f1a38-pod-info\") pod \"rabbitmq-server-0\" (UID: \"ec28045f-e1d4-4b22-bb7d-3c6d2a1f1a38\") " pod="openstack/rabbitmq-server-0" Feb 02 09:29:05 crc kubenswrapper[4764]: I0202 09:29:05.342671 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "persistence") pod "03d9f090-4a93-4226-8fce-6fd0a425d929" (UID: "03d9f090-4a93-4226-8fce-6fd0a425d929"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 02 09:29:05 crc kubenswrapper[4764]: I0202 09:29:05.343059 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/03d9f090-4a93-4226-8fce-6fd0a425d929-pod-info" (OuterVolumeSpecName: "pod-info") pod "03d9f090-4a93-4226-8fce-6fd0a425d929" (UID: "03d9f090-4a93-4226-8fce-6fd0a425d929"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Feb 02 09:29:05 crc kubenswrapper[4764]: I0202 09:29:05.353653 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xzvdh\" (UniqueName: \"kubernetes.io/projected/ec28045f-e1d4-4b22-bb7d-3c6d2a1f1a38-kube-api-access-xzvdh\") pod \"rabbitmq-server-0\" (UID: \"ec28045f-e1d4-4b22-bb7d-3c6d2a1f1a38\") " pod="openstack/rabbitmq-server-0" Feb 02 09:29:05 crc kubenswrapper[4764]: I0202 09:29:05.379620 4764 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/03d9f090-4a93-4226-8fce-6fd0a425d929-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Feb 02 09:29:05 crc kubenswrapper[4764]: I0202 09:29:05.379778 4764 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/03d9f090-4a93-4226-8fce-6fd0a425d929-pod-info\") on node \"crc\" DevicePath \"\"" Feb 02 09:29:05 crc kubenswrapper[4764]: I0202 09:29:05.379838 4764 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/03d9f090-4a93-4226-8fce-6fd0a425d929-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Feb 02 09:29:05 crc kubenswrapper[4764]: I0202 09:29:05.379895 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-27wk5\" (UniqueName: \"kubernetes.io/projected/03d9f090-4a93-4226-8fce-6fd0a425d929-kube-api-access-27wk5\") on node \"crc\" DevicePath \"\"" Feb 02 09:29:05 crc kubenswrapper[4764]: I0202 09:29:05.379983 4764 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Feb 02 09:29:05 crc kubenswrapper[4764]: I0202 09:29:05.380142 4764 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/03d9f090-4a93-4226-8fce-6fd0a425d929-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Feb 02 09:29:05 crc kubenswrapper[4764]: I0202 09:29:05.380278 4764 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/03d9f090-4a93-4226-8fce-6fd0a425d929-plugins-conf\") on node \"crc\" DevicePath \"\"" Feb 02 09:29:05 crc kubenswrapper[4764]: I0202 09:29:05.380407 4764 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/03d9f090-4a93-4226-8fce-6fd0a425d929-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Feb 02 09:29:05 crc kubenswrapper[4764]: I0202 09:29:05.395324 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-server-0\" (UID: \"ec28045f-e1d4-4b22-bb7d-3c6d2a1f1a38\") " pod="openstack/rabbitmq-server-0" Feb 02 09:29:05 crc kubenswrapper[4764]: I0202 09:29:05.399126 4764 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Feb 02 09:29:05 crc kubenswrapper[4764]: I0202 09:29:05.411516 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/03d9f090-4a93-4226-8fce-6fd0a425d929-config-data" (OuterVolumeSpecName: "config-data") pod "03d9f090-4a93-4226-8fce-6fd0a425d929" (UID: "03d9f090-4a93-4226-8fce-6fd0a425d929"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:29:05 crc kubenswrapper[4764]: I0202 09:29:05.424677 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/03d9f090-4a93-4226-8fce-6fd0a425d929-server-conf" (OuterVolumeSpecName: "server-conf") pod "03d9f090-4a93-4226-8fce-6fd0a425d929" (UID: "03d9f090-4a93-4226-8fce-6fd0a425d929"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:29:05 crc kubenswrapper[4764]: I0202 09:29:05.482199 4764 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/03d9f090-4a93-4226-8fce-6fd0a425d929-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 09:29:05 crc kubenswrapper[4764]: I0202 09:29:05.482422 4764 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/03d9f090-4a93-4226-8fce-6fd0a425d929-server-conf\") on node \"crc\" DevicePath \"\"" Feb 02 09:29:05 crc kubenswrapper[4764]: I0202 09:29:05.482483 4764 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Feb 02 09:29:05 crc kubenswrapper[4764]: I0202 09:29:05.506791 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/03d9f090-4a93-4226-8fce-6fd0a425d929-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "03d9f090-4a93-4226-8fce-6fd0a425d929" (UID: "03d9f090-4a93-4226-8fce-6fd0a425d929"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:29:05 crc kubenswrapper[4764]: I0202 09:29:05.583799 4764 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/03d9f090-4a93-4226-8fce-6fd0a425d929-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Feb 02 09:29:05 crc kubenswrapper[4764]: I0202 09:29:05.681654 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Feb 02 09:29:05 crc kubenswrapper[4764]: I0202 09:29:05.850167 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="53d5d300-11ae-408d-8bb8-347defb93991" path="/var/lib/kubelet/pods/53d5d300-11ae-408d-8bb8-347defb93991/volumes" Feb 02 09:29:05 crc kubenswrapper[4764]: I0202 09:29:05.878616 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"03d9f090-4a93-4226-8fce-6fd0a425d929","Type":"ContainerDied","Data":"d923a6203fb8b53db25d87f66f30d098c92cff9a522b96722479fb01f94e1e95"} Feb 02 09:29:05 crc kubenswrapper[4764]: I0202 09:29:05.878660 4764 scope.go:117] "RemoveContainer" containerID="b0556eef8b622b39f414f040258f94501b1ff75d65d0e71032e333b6b15e8b4f" Feb 02 09:29:05 crc kubenswrapper[4764]: I0202 09:29:05.878745 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Feb 02 09:29:05 crc kubenswrapper[4764]: I0202 09:29:05.919905 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 02 09:29:05 crc kubenswrapper[4764]: I0202 09:29:05.982323 4764 scope.go:117] "RemoveContainer" containerID="d08998db15d7c2f061c9aa78b783e5e53996e688590d41367d09011c4224893d" Feb 02 09:29:06 crc kubenswrapper[4764]: I0202 09:29:06.025037 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 02 09:29:06 crc kubenswrapper[4764]: I0202 09:29:06.034000 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 02 09:29:06 crc kubenswrapper[4764]: E0202 09:29:06.034712 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03d9f090-4a93-4226-8fce-6fd0a425d929" containerName="setup-container" Feb 02 09:29:06 crc kubenswrapper[4764]: I0202 09:29:06.034730 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="03d9f090-4a93-4226-8fce-6fd0a425d929" containerName="setup-container" Feb 02 09:29:06 crc kubenswrapper[4764]: E0202 09:29:06.034758 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03d9f090-4a93-4226-8fce-6fd0a425d929" containerName="rabbitmq" Feb 02 09:29:06 crc kubenswrapper[4764]: I0202 09:29:06.034764 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="03d9f090-4a93-4226-8fce-6fd0a425d929" containerName="rabbitmq" Feb 02 09:29:06 crc kubenswrapper[4764]: I0202 09:29:06.034957 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="03d9f090-4a93-4226-8fce-6fd0a425d929" containerName="rabbitmq" Feb 02 09:29:06 crc kubenswrapper[4764]: I0202 09:29:06.036177 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Feb 02 09:29:06 crc kubenswrapper[4764]: I0202 09:29:06.039677 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Feb 02 09:29:06 crc kubenswrapper[4764]: I0202 09:29:06.040985 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-4mcsh" Feb 02 09:29:06 crc kubenswrapper[4764]: I0202 09:29:06.042239 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Feb 02 09:29:06 crc kubenswrapper[4764]: I0202 09:29:06.050275 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Feb 02 09:29:06 crc kubenswrapper[4764]: I0202 09:29:06.050475 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Feb 02 09:29:06 crc kubenswrapper[4764]: I0202 09:29:06.050617 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Feb 02 09:29:06 crc kubenswrapper[4764]: I0202 09:29:06.050718 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Feb 02 09:29:06 crc kubenswrapper[4764]: I0202 09:29:06.050720 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 02 09:29:06 crc kubenswrapper[4764]: I0202 09:29:06.194018 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c60b198c-58ad-4363-afea-257f51f64e51-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"c60b198c-58ad-4363-afea-257f51f64e51\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 09:29:06 crc kubenswrapper[4764]: I0202 09:29:06.194141 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c60b198c-58ad-4363-afea-257f51f64e51-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"c60b198c-58ad-4363-afea-257f51f64e51\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 09:29:06 crc kubenswrapper[4764]: I0202 09:29:06.194226 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"c60b198c-58ad-4363-afea-257f51f64e51\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 09:29:06 crc kubenswrapper[4764]: I0202 09:29:06.194265 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c60b198c-58ad-4363-afea-257f51f64e51-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"c60b198c-58ad-4363-afea-257f51f64e51\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 09:29:06 crc kubenswrapper[4764]: I0202 09:29:06.194315 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c60b198c-58ad-4363-afea-257f51f64e51-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"c60b198c-58ad-4363-afea-257f51f64e51\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 09:29:06 crc kubenswrapper[4764]: I0202 09:29:06.194346 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c60b198c-58ad-4363-afea-257f51f64e51-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"c60b198c-58ad-4363-afea-257f51f64e51\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 09:29:06 crc kubenswrapper[4764]: I0202 09:29:06.194386 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c60b198c-58ad-4363-afea-257f51f64e51-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"c60b198c-58ad-4363-afea-257f51f64e51\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 09:29:06 crc kubenswrapper[4764]: I0202 09:29:06.194410 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h2vdf\" (UniqueName: \"kubernetes.io/projected/c60b198c-58ad-4363-afea-257f51f64e51-kube-api-access-h2vdf\") pod \"rabbitmq-cell1-server-0\" (UID: \"c60b198c-58ad-4363-afea-257f51f64e51\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 09:29:06 crc kubenswrapper[4764]: I0202 09:29:06.194439 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c60b198c-58ad-4363-afea-257f51f64e51-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"c60b198c-58ad-4363-afea-257f51f64e51\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 09:29:06 crc kubenswrapper[4764]: I0202 09:29:06.194455 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c60b198c-58ad-4363-afea-257f51f64e51-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"c60b198c-58ad-4363-afea-257f51f64e51\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 09:29:06 crc kubenswrapper[4764]: I0202 09:29:06.194477 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c60b198c-58ad-4363-afea-257f51f64e51-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"c60b198c-58ad-4363-afea-257f51f64e51\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 09:29:06 crc kubenswrapper[4764]: I0202 09:29:06.296275 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c60b198c-58ad-4363-afea-257f51f64e51-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"c60b198c-58ad-4363-afea-257f51f64e51\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 09:29:06 crc kubenswrapper[4764]: I0202 09:29:06.296567 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c60b198c-58ad-4363-afea-257f51f64e51-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"c60b198c-58ad-4363-afea-257f51f64e51\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 09:29:06 crc kubenswrapper[4764]: I0202 09:29:06.296608 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"c60b198c-58ad-4363-afea-257f51f64e51\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 09:29:06 crc kubenswrapper[4764]: I0202 09:29:06.296639 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c60b198c-58ad-4363-afea-257f51f64e51-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"c60b198c-58ad-4363-afea-257f51f64e51\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 09:29:06 crc kubenswrapper[4764]: I0202 09:29:06.296673 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c60b198c-58ad-4363-afea-257f51f64e51-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"c60b198c-58ad-4363-afea-257f51f64e51\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 09:29:06 crc kubenswrapper[4764]: I0202 09:29:06.296701 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c60b198c-58ad-4363-afea-257f51f64e51-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"c60b198c-58ad-4363-afea-257f51f64e51\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 09:29:06 crc kubenswrapper[4764]: I0202 09:29:06.296730 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c60b198c-58ad-4363-afea-257f51f64e51-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"c60b198c-58ad-4363-afea-257f51f64e51\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 09:29:06 crc kubenswrapper[4764]: I0202 09:29:06.296755 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h2vdf\" (UniqueName: \"kubernetes.io/projected/c60b198c-58ad-4363-afea-257f51f64e51-kube-api-access-h2vdf\") pod \"rabbitmq-cell1-server-0\" (UID: \"c60b198c-58ad-4363-afea-257f51f64e51\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 09:29:06 crc kubenswrapper[4764]: I0202 09:29:06.296781 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c60b198c-58ad-4363-afea-257f51f64e51-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"c60b198c-58ad-4363-afea-257f51f64e51\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 09:29:06 crc kubenswrapper[4764]: I0202 09:29:06.296802 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c60b198c-58ad-4363-afea-257f51f64e51-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"c60b198c-58ad-4363-afea-257f51f64e51\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 09:29:06 crc kubenswrapper[4764]: I0202 09:29:06.296829 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c60b198c-58ad-4363-afea-257f51f64e51-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"c60b198c-58ad-4363-afea-257f51f64e51\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 09:29:06 crc kubenswrapper[4764]: I0202 09:29:06.297798 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c60b198c-58ad-4363-afea-257f51f64e51-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"c60b198c-58ad-4363-afea-257f51f64e51\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 09:29:06 crc kubenswrapper[4764]: I0202 09:29:06.297819 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c60b198c-58ad-4363-afea-257f51f64e51-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"c60b198c-58ad-4363-afea-257f51f64e51\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 09:29:06 crc kubenswrapper[4764]: I0202 09:29:06.298185 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c60b198c-58ad-4363-afea-257f51f64e51-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"c60b198c-58ad-4363-afea-257f51f64e51\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 09:29:06 crc kubenswrapper[4764]: I0202 09:29:06.298988 4764 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"c60b198c-58ad-4363-afea-257f51f64e51\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/rabbitmq-cell1-server-0" Feb 02 09:29:06 crc kubenswrapper[4764]: I0202 09:29:06.299171 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c60b198c-58ad-4363-afea-257f51f64e51-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"c60b198c-58ad-4363-afea-257f51f64e51\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 09:29:06 crc kubenswrapper[4764]: I0202 09:29:06.299355 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c60b198c-58ad-4363-afea-257f51f64e51-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"c60b198c-58ad-4363-afea-257f51f64e51\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 09:29:06 crc kubenswrapper[4764]: I0202 09:29:06.316984 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c60b198c-58ad-4363-afea-257f51f64e51-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"c60b198c-58ad-4363-afea-257f51f64e51\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 09:29:06 crc kubenswrapper[4764]: I0202 09:29:06.316990 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c60b198c-58ad-4363-afea-257f51f64e51-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"c60b198c-58ad-4363-afea-257f51f64e51\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 09:29:06 crc kubenswrapper[4764]: I0202 09:29:06.317359 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c60b198c-58ad-4363-afea-257f51f64e51-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"c60b198c-58ad-4363-afea-257f51f64e51\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 09:29:06 crc kubenswrapper[4764]: I0202 09:29:06.318403 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c60b198c-58ad-4363-afea-257f51f64e51-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"c60b198c-58ad-4363-afea-257f51f64e51\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 09:29:06 crc kubenswrapper[4764]: I0202 09:29:06.324810 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h2vdf\" (UniqueName: \"kubernetes.io/projected/c60b198c-58ad-4363-afea-257f51f64e51-kube-api-access-h2vdf\") pod \"rabbitmq-cell1-server-0\" (UID: \"c60b198c-58ad-4363-afea-257f51f64e51\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 09:29:06 crc kubenswrapper[4764]: I0202 09:29:06.357179 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"c60b198c-58ad-4363-afea-257f51f64e51\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 09:29:06 crc kubenswrapper[4764]: I0202 09:29:06.384526 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Feb 02 09:29:06 crc kubenswrapper[4764]: I0202 09:29:06.405819 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 02 09:29:06 crc kubenswrapper[4764]: I0202 09:29:06.859309 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 02 09:29:06 crc kubenswrapper[4764]: I0202 09:29:06.924551 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"c60b198c-58ad-4363-afea-257f51f64e51","Type":"ContainerStarted","Data":"b0c402201c40e99e3700e22eea2358655fe7dd45a88ed01ba99ac4459b9fe3ea"} Feb 02 09:29:06 crc kubenswrapper[4764]: I0202 09:29:06.931534 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"ec28045f-e1d4-4b22-bb7d-3c6d2a1f1a38","Type":"ContainerStarted","Data":"4a4011d47ffda0c8d5fa08661ca20ab8900ff7c9adc5fcb0bf156f356c305722"} Feb 02 09:29:07 crc kubenswrapper[4764]: I0202 09:29:07.105049 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-b4d48ddf9-kbdwn"] Feb 02 09:29:07 crc kubenswrapper[4764]: I0202 09:29:07.106374 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b4d48ddf9-kbdwn" Feb 02 09:29:07 crc kubenswrapper[4764]: I0202 09:29:07.108670 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Feb 02 09:29:07 crc kubenswrapper[4764]: I0202 09:29:07.173370 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b4d48ddf9-kbdwn"] Feb 02 09:29:07 crc kubenswrapper[4764]: I0202 09:29:07.217613 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a71471c2-6a47-403f-b511-04a512577dcf-ovsdbserver-nb\") pod \"dnsmasq-dns-b4d48ddf9-kbdwn\" (UID: \"a71471c2-6a47-403f-b511-04a512577dcf\") " pod="openstack/dnsmasq-dns-b4d48ddf9-kbdwn" Feb 02 09:29:07 crc kubenswrapper[4764]: I0202 09:29:07.217671 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a71471c2-6a47-403f-b511-04a512577dcf-ovsdbserver-sb\") pod \"dnsmasq-dns-b4d48ddf9-kbdwn\" (UID: \"a71471c2-6a47-403f-b511-04a512577dcf\") " pod="openstack/dnsmasq-dns-b4d48ddf9-kbdwn" Feb 02 09:29:07 crc kubenswrapper[4764]: I0202 09:29:07.217698 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a71471c2-6a47-403f-b511-04a512577dcf-dns-svc\") pod \"dnsmasq-dns-b4d48ddf9-kbdwn\" (UID: \"a71471c2-6a47-403f-b511-04a512577dcf\") " pod="openstack/dnsmasq-dns-b4d48ddf9-kbdwn" Feb 02 09:29:07 crc kubenswrapper[4764]: I0202 09:29:07.217720 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6gpc2\" (UniqueName: \"kubernetes.io/projected/a71471c2-6a47-403f-b511-04a512577dcf-kube-api-access-6gpc2\") pod \"dnsmasq-dns-b4d48ddf9-kbdwn\" (UID: \"a71471c2-6a47-403f-b511-04a512577dcf\") " pod="openstack/dnsmasq-dns-b4d48ddf9-kbdwn" Feb 02 09:29:07 crc kubenswrapper[4764]: I0202 09:29:07.217742 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a71471c2-6a47-403f-b511-04a512577dcf-config\") pod \"dnsmasq-dns-b4d48ddf9-kbdwn\" (UID: \"a71471c2-6a47-403f-b511-04a512577dcf\") " pod="openstack/dnsmasq-dns-b4d48ddf9-kbdwn" Feb 02 09:29:07 crc kubenswrapper[4764]: I0202 09:29:07.217780 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/a71471c2-6a47-403f-b511-04a512577dcf-openstack-edpm-ipam\") pod \"dnsmasq-dns-b4d48ddf9-kbdwn\" (UID: \"a71471c2-6a47-403f-b511-04a512577dcf\") " pod="openstack/dnsmasq-dns-b4d48ddf9-kbdwn" Feb 02 09:29:07 crc kubenswrapper[4764]: I0202 09:29:07.319654 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a71471c2-6a47-403f-b511-04a512577dcf-ovsdbserver-nb\") pod \"dnsmasq-dns-b4d48ddf9-kbdwn\" (UID: \"a71471c2-6a47-403f-b511-04a512577dcf\") " pod="openstack/dnsmasq-dns-b4d48ddf9-kbdwn" Feb 02 09:29:07 crc kubenswrapper[4764]: I0202 09:29:07.319711 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a71471c2-6a47-403f-b511-04a512577dcf-ovsdbserver-sb\") pod \"dnsmasq-dns-b4d48ddf9-kbdwn\" (UID: \"a71471c2-6a47-403f-b511-04a512577dcf\") " pod="openstack/dnsmasq-dns-b4d48ddf9-kbdwn" Feb 02 09:29:07 crc kubenswrapper[4764]: I0202 09:29:07.319739 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a71471c2-6a47-403f-b511-04a512577dcf-dns-svc\") pod \"dnsmasq-dns-b4d48ddf9-kbdwn\" (UID: \"a71471c2-6a47-403f-b511-04a512577dcf\") " pod="openstack/dnsmasq-dns-b4d48ddf9-kbdwn" Feb 02 09:29:07 crc kubenswrapper[4764]: I0202 09:29:07.319758 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6gpc2\" (UniqueName: \"kubernetes.io/projected/a71471c2-6a47-403f-b511-04a512577dcf-kube-api-access-6gpc2\") pod \"dnsmasq-dns-b4d48ddf9-kbdwn\" (UID: \"a71471c2-6a47-403f-b511-04a512577dcf\") " pod="openstack/dnsmasq-dns-b4d48ddf9-kbdwn" Feb 02 09:29:07 crc kubenswrapper[4764]: I0202 09:29:07.319780 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a71471c2-6a47-403f-b511-04a512577dcf-config\") pod \"dnsmasq-dns-b4d48ddf9-kbdwn\" (UID: \"a71471c2-6a47-403f-b511-04a512577dcf\") " pod="openstack/dnsmasq-dns-b4d48ddf9-kbdwn" Feb 02 09:29:07 crc kubenswrapper[4764]: I0202 09:29:07.319816 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/a71471c2-6a47-403f-b511-04a512577dcf-openstack-edpm-ipam\") pod \"dnsmasq-dns-b4d48ddf9-kbdwn\" (UID: \"a71471c2-6a47-403f-b511-04a512577dcf\") " pod="openstack/dnsmasq-dns-b4d48ddf9-kbdwn" Feb 02 09:29:07 crc kubenswrapper[4764]: I0202 09:29:07.320565 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/a71471c2-6a47-403f-b511-04a512577dcf-openstack-edpm-ipam\") pod \"dnsmasq-dns-b4d48ddf9-kbdwn\" (UID: \"a71471c2-6a47-403f-b511-04a512577dcf\") " pod="openstack/dnsmasq-dns-b4d48ddf9-kbdwn" Feb 02 09:29:07 crc kubenswrapper[4764]: I0202 09:29:07.321097 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a71471c2-6a47-403f-b511-04a512577dcf-ovsdbserver-nb\") pod \"dnsmasq-dns-b4d48ddf9-kbdwn\" (UID: \"a71471c2-6a47-403f-b511-04a512577dcf\") " pod="openstack/dnsmasq-dns-b4d48ddf9-kbdwn" Feb 02 09:29:07 crc kubenswrapper[4764]: I0202 09:29:07.321593 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a71471c2-6a47-403f-b511-04a512577dcf-ovsdbserver-sb\") pod \"dnsmasq-dns-b4d48ddf9-kbdwn\" (UID: \"a71471c2-6a47-403f-b511-04a512577dcf\") " pod="openstack/dnsmasq-dns-b4d48ddf9-kbdwn" Feb 02 09:29:07 crc kubenswrapper[4764]: I0202 09:29:07.322308 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a71471c2-6a47-403f-b511-04a512577dcf-dns-svc\") pod \"dnsmasq-dns-b4d48ddf9-kbdwn\" (UID: \"a71471c2-6a47-403f-b511-04a512577dcf\") " pod="openstack/dnsmasq-dns-b4d48ddf9-kbdwn" Feb 02 09:29:07 crc kubenswrapper[4764]: I0202 09:29:07.322894 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a71471c2-6a47-403f-b511-04a512577dcf-config\") pod \"dnsmasq-dns-b4d48ddf9-kbdwn\" (UID: \"a71471c2-6a47-403f-b511-04a512577dcf\") " pod="openstack/dnsmasq-dns-b4d48ddf9-kbdwn" Feb 02 09:29:07 crc kubenswrapper[4764]: I0202 09:29:07.339861 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6gpc2\" (UniqueName: \"kubernetes.io/projected/a71471c2-6a47-403f-b511-04a512577dcf-kube-api-access-6gpc2\") pod \"dnsmasq-dns-b4d48ddf9-kbdwn\" (UID: \"a71471c2-6a47-403f-b511-04a512577dcf\") " pod="openstack/dnsmasq-dns-b4d48ddf9-kbdwn" Feb 02 09:29:07 crc kubenswrapper[4764]: I0202 09:29:07.603313 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b4d48ddf9-kbdwn" Feb 02 09:29:07 crc kubenswrapper[4764]: I0202 09:29:07.865979 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="03d9f090-4a93-4226-8fce-6fd0a425d929" path="/var/lib/kubelet/pods/03d9f090-4a93-4226-8fce-6fd0a425d929/volumes" Feb 02 09:29:07 crc kubenswrapper[4764]: I0202 09:29:07.943274 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"ec28045f-e1d4-4b22-bb7d-3c6d2a1f1a38","Type":"ContainerStarted","Data":"99db28642a3c01b0c74f9ee5f2768815ab4b293028c9e0891a0d582b53104ab3"} Feb 02 09:29:08 crc kubenswrapper[4764]: I0202 09:29:08.104276 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b4d48ddf9-kbdwn"] Feb 02 09:29:08 crc kubenswrapper[4764]: I0202 09:29:08.954011 4764 generic.go:334] "Generic (PLEG): container finished" podID="a71471c2-6a47-403f-b511-04a512577dcf" containerID="8fffdb31bc7dce696ff8f7e94c471934ee94135606ee981115620d1db6cd2dc5" exitCode=0 Feb 02 09:29:08 crc kubenswrapper[4764]: I0202 09:29:08.954057 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b4d48ddf9-kbdwn" event={"ID":"a71471c2-6a47-403f-b511-04a512577dcf","Type":"ContainerDied","Data":"8fffdb31bc7dce696ff8f7e94c471934ee94135606ee981115620d1db6cd2dc5"} Feb 02 09:29:08 crc kubenswrapper[4764]: I0202 09:29:08.954572 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b4d48ddf9-kbdwn" event={"ID":"a71471c2-6a47-403f-b511-04a512577dcf","Type":"ContainerStarted","Data":"2757a1ce774d687add422263ad13e527ea2501d77368a5104a32323fc7a52f41"} Feb 02 09:29:08 crc kubenswrapper[4764]: I0202 09:29:08.959062 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"c60b198c-58ad-4363-afea-257f51f64e51","Type":"ContainerStarted","Data":"48f8feab56900bd296e165775c719351ff9243c9a64a42f3116e3c49c612c42f"} Feb 02 09:29:09 crc kubenswrapper[4764]: I0202 09:29:09.968879 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b4d48ddf9-kbdwn" event={"ID":"a71471c2-6a47-403f-b511-04a512577dcf","Type":"ContainerStarted","Data":"af4ab96a20f13c81bcc82ad0e0602fbdf553d56a33c0bd91c84deca64f54c681"} Feb 02 09:29:10 crc kubenswrapper[4764]: I0202 09:29:10.000649 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-b4d48ddf9-kbdwn" podStartSLOduration=3.000604283 podStartE2EDuration="3.000604283s" podCreationTimestamp="2026-02-02 09:29:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:29:09.994534981 +0000 UTC m=+1312.928259079" watchObservedRunningTime="2026-02-02 09:29:10.000604283 +0000 UTC m=+1312.934328371" Feb 02 09:29:10 crc kubenswrapper[4764]: I0202 09:29:10.979843 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-b4d48ddf9-kbdwn" Feb 02 09:29:13 crc kubenswrapper[4764]: I0202 09:29:13.523031 4764 patch_prober.go:28] interesting pod/machine-config-daemon-4ndm4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 09:29:13 crc kubenswrapper[4764]: I0202 09:29:13.523547 4764 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 09:29:17 crc kubenswrapper[4764]: I0202 09:29:17.606570 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-b4d48ddf9-kbdwn" Feb 02 09:29:17 crc kubenswrapper[4764]: I0202 09:29:17.703128 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-75956b9989-5ddn6"] Feb 02 09:29:17 crc kubenswrapper[4764]: I0202 09:29:17.712757 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-75956b9989-5ddn6" podUID="ff8a4b58-1d3f-4f7e-a0ae-8d541024983b" containerName="dnsmasq-dns" containerID="cri-o://ba78ef785372350834e28f350c85c6581e766f4c306e316726c548005b1ee000" gracePeriod=10 Feb 02 09:29:17 crc kubenswrapper[4764]: I0202 09:29:17.967259 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-77755c997f-6996z"] Feb 02 09:29:17 crc kubenswrapper[4764]: I0202 09:29:17.968846 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77755c997f-6996z" Feb 02 09:29:17 crc kubenswrapper[4764]: I0202 09:29:17.977301 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-77755c997f-6996z"] Feb 02 09:29:18 crc kubenswrapper[4764]: I0202 09:29:18.010677 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/33b1de07-e5a5-472a-acfb-102dbfe83e0e-dns-svc\") pod \"dnsmasq-dns-77755c997f-6996z\" (UID: \"33b1de07-e5a5-472a-acfb-102dbfe83e0e\") " pod="openstack/dnsmasq-dns-77755c997f-6996z" Feb 02 09:29:18 crc kubenswrapper[4764]: I0202 09:29:18.011062 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/33b1de07-e5a5-472a-acfb-102dbfe83e0e-ovsdbserver-nb\") pod \"dnsmasq-dns-77755c997f-6996z\" (UID: \"33b1de07-e5a5-472a-acfb-102dbfe83e0e\") " pod="openstack/dnsmasq-dns-77755c997f-6996z" Feb 02 09:29:18 crc kubenswrapper[4764]: I0202 09:29:18.011096 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/33b1de07-e5a5-472a-acfb-102dbfe83e0e-openstack-edpm-ipam\") pod \"dnsmasq-dns-77755c997f-6996z\" (UID: \"33b1de07-e5a5-472a-acfb-102dbfe83e0e\") " pod="openstack/dnsmasq-dns-77755c997f-6996z" Feb 02 09:29:18 crc kubenswrapper[4764]: I0202 09:29:18.011154 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/33b1de07-e5a5-472a-acfb-102dbfe83e0e-ovsdbserver-sb\") pod \"dnsmasq-dns-77755c997f-6996z\" (UID: \"33b1de07-e5a5-472a-acfb-102dbfe83e0e\") " pod="openstack/dnsmasq-dns-77755c997f-6996z" Feb 02 09:29:18 crc kubenswrapper[4764]: I0202 09:29:18.011174 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-496bg\" (UniqueName: \"kubernetes.io/projected/33b1de07-e5a5-472a-acfb-102dbfe83e0e-kube-api-access-496bg\") pod \"dnsmasq-dns-77755c997f-6996z\" (UID: \"33b1de07-e5a5-472a-acfb-102dbfe83e0e\") " pod="openstack/dnsmasq-dns-77755c997f-6996z" Feb 02 09:29:18 crc kubenswrapper[4764]: I0202 09:29:18.011200 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/33b1de07-e5a5-472a-acfb-102dbfe83e0e-config\") pod \"dnsmasq-dns-77755c997f-6996z\" (UID: \"33b1de07-e5a5-472a-acfb-102dbfe83e0e\") " pod="openstack/dnsmasq-dns-77755c997f-6996z" Feb 02 09:29:18 crc kubenswrapper[4764]: I0202 09:29:18.055374 4764 generic.go:334] "Generic (PLEG): container finished" podID="ff8a4b58-1d3f-4f7e-a0ae-8d541024983b" containerID="ba78ef785372350834e28f350c85c6581e766f4c306e316726c548005b1ee000" exitCode=0 Feb 02 09:29:18 crc kubenswrapper[4764]: I0202 09:29:18.055424 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75956b9989-5ddn6" event={"ID":"ff8a4b58-1d3f-4f7e-a0ae-8d541024983b","Type":"ContainerDied","Data":"ba78ef785372350834e28f350c85c6581e766f4c306e316726c548005b1ee000"} Feb 02 09:29:18 crc kubenswrapper[4764]: I0202 09:29:18.113367 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/33b1de07-e5a5-472a-acfb-102dbfe83e0e-ovsdbserver-sb\") pod \"dnsmasq-dns-77755c997f-6996z\" (UID: \"33b1de07-e5a5-472a-acfb-102dbfe83e0e\") " pod="openstack/dnsmasq-dns-77755c997f-6996z" Feb 02 09:29:18 crc kubenswrapper[4764]: I0202 09:29:18.113439 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-496bg\" (UniqueName: \"kubernetes.io/projected/33b1de07-e5a5-472a-acfb-102dbfe83e0e-kube-api-access-496bg\") pod \"dnsmasq-dns-77755c997f-6996z\" (UID: \"33b1de07-e5a5-472a-acfb-102dbfe83e0e\") " pod="openstack/dnsmasq-dns-77755c997f-6996z" Feb 02 09:29:18 crc kubenswrapper[4764]: I0202 09:29:18.114395 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/33b1de07-e5a5-472a-acfb-102dbfe83e0e-ovsdbserver-sb\") pod \"dnsmasq-dns-77755c997f-6996z\" (UID: \"33b1de07-e5a5-472a-acfb-102dbfe83e0e\") " pod="openstack/dnsmasq-dns-77755c997f-6996z" Feb 02 09:29:18 crc kubenswrapper[4764]: I0202 09:29:18.114455 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/33b1de07-e5a5-472a-acfb-102dbfe83e0e-config\") pod \"dnsmasq-dns-77755c997f-6996z\" (UID: \"33b1de07-e5a5-472a-acfb-102dbfe83e0e\") " pod="openstack/dnsmasq-dns-77755c997f-6996z" Feb 02 09:29:18 crc kubenswrapper[4764]: I0202 09:29:18.115241 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/33b1de07-e5a5-472a-acfb-102dbfe83e0e-config\") pod \"dnsmasq-dns-77755c997f-6996z\" (UID: \"33b1de07-e5a5-472a-acfb-102dbfe83e0e\") " pod="openstack/dnsmasq-dns-77755c997f-6996z" Feb 02 09:29:18 crc kubenswrapper[4764]: I0202 09:29:18.115288 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/33b1de07-e5a5-472a-acfb-102dbfe83e0e-dns-svc\") pod \"dnsmasq-dns-77755c997f-6996z\" (UID: \"33b1de07-e5a5-472a-acfb-102dbfe83e0e\") " pod="openstack/dnsmasq-dns-77755c997f-6996z" Feb 02 09:29:18 crc kubenswrapper[4764]: I0202 09:29:18.115334 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/33b1de07-e5a5-472a-acfb-102dbfe83e0e-ovsdbserver-nb\") pod \"dnsmasq-dns-77755c997f-6996z\" (UID: \"33b1de07-e5a5-472a-acfb-102dbfe83e0e\") " pod="openstack/dnsmasq-dns-77755c997f-6996z" Feb 02 09:29:18 crc kubenswrapper[4764]: I0202 09:29:18.115377 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/33b1de07-e5a5-472a-acfb-102dbfe83e0e-openstack-edpm-ipam\") pod \"dnsmasq-dns-77755c997f-6996z\" (UID: \"33b1de07-e5a5-472a-acfb-102dbfe83e0e\") " pod="openstack/dnsmasq-dns-77755c997f-6996z" Feb 02 09:29:18 crc kubenswrapper[4764]: I0202 09:29:18.116254 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/33b1de07-e5a5-472a-acfb-102dbfe83e0e-dns-svc\") pod \"dnsmasq-dns-77755c997f-6996z\" (UID: \"33b1de07-e5a5-472a-acfb-102dbfe83e0e\") " pod="openstack/dnsmasq-dns-77755c997f-6996z" Feb 02 09:29:18 crc kubenswrapper[4764]: I0202 09:29:18.117126 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/33b1de07-e5a5-472a-acfb-102dbfe83e0e-ovsdbserver-nb\") pod \"dnsmasq-dns-77755c997f-6996z\" (UID: \"33b1de07-e5a5-472a-acfb-102dbfe83e0e\") " pod="openstack/dnsmasq-dns-77755c997f-6996z" Feb 02 09:29:18 crc kubenswrapper[4764]: I0202 09:29:18.117627 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/33b1de07-e5a5-472a-acfb-102dbfe83e0e-openstack-edpm-ipam\") pod \"dnsmasq-dns-77755c997f-6996z\" (UID: \"33b1de07-e5a5-472a-acfb-102dbfe83e0e\") " pod="openstack/dnsmasq-dns-77755c997f-6996z" Feb 02 09:29:18 crc kubenswrapper[4764]: I0202 09:29:18.135987 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-496bg\" (UniqueName: \"kubernetes.io/projected/33b1de07-e5a5-472a-acfb-102dbfe83e0e-kube-api-access-496bg\") pod \"dnsmasq-dns-77755c997f-6996z\" (UID: \"33b1de07-e5a5-472a-acfb-102dbfe83e0e\") " pod="openstack/dnsmasq-dns-77755c997f-6996z" Feb 02 09:29:18 crc kubenswrapper[4764]: I0202 09:29:18.266313 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75956b9989-5ddn6" Feb 02 09:29:18 crc kubenswrapper[4764]: I0202 09:29:18.299341 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77755c997f-6996z" Feb 02 09:29:18 crc kubenswrapper[4764]: I0202 09:29:18.431614 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rhd8r\" (UniqueName: \"kubernetes.io/projected/ff8a4b58-1d3f-4f7e-a0ae-8d541024983b-kube-api-access-rhd8r\") pod \"ff8a4b58-1d3f-4f7e-a0ae-8d541024983b\" (UID: \"ff8a4b58-1d3f-4f7e-a0ae-8d541024983b\") " Feb 02 09:29:18 crc kubenswrapper[4764]: I0202 09:29:18.431816 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ff8a4b58-1d3f-4f7e-a0ae-8d541024983b-ovsdbserver-sb\") pod \"ff8a4b58-1d3f-4f7e-a0ae-8d541024983b\" (UID: \"ff8a4b58-1d3f-4f7e-a0ae-8d541024983b\") " Feb 02 09:29:18 crc kubenswrapper[4764]: I0202 09:29:18.431850 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ff8a4b58-1d3f-4f7e-a0ae-8d541024983b-dns-svc\") pod \"ff8a4b58-1d3f-4f7e-a0ae-8d541024983b\" (UID: \"ff8a4b58-1d3f-4f7e-a0ae-8d541024983b\") " Feb 02 09:29:18 crc kubenswrapper[4764]: I0202 09:29:18.431872 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ff8a4b58-1d3f-4f7e-a0ae-8d541024983b-config\") pod \"ff8a4b58-1d3f-4f7e-a0ae-8d541024983b\" (UID: \"ff8a4b58-1d3f-4f7e-a0ae-8d541024983b\") " Feb 02 09:29:18 crc kubenswrapper[4764]: I0202 09:29:18.431945 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ff8a4b58-1d3f-4f7e-a0ae-8d541024983b-ovsdbserver-nb\") pod \"ff8a4b58-1d3f-4f7e-a0ae-8d541024983b\" (UID: \"ff8a4b58-1d3f-4f7e-a0ae-8d541024983b\") " Feb 02 09:29:18 crc kubenswrapper[4764]: I0202 09:29:18.444225 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff8a4b58-1d3f-4f7e-a0ae-8d541024983b-kube-api-access-rhd8r" (OuterVolumeSpecName: "kube-api-access-rhd8r") pod "ff8a4b58-1d3f-4f7e-a0ae-8d541024983b" (UID: "ff8a4b58-1d3f-4f7e-a0ae-8d541024983b"). InnerVolumeSpecName "kube-api-access-rhd8r". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:29:19 crc kubenswrapper[4764]: I0202 09:29:18.487632 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ff8a4b58-1d3f-4f7e-a0ae-8d541024983b-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ff8a4b58-1d3f-4f7e-a0ae-8d541024983b" (UID: "ff8a4b58-1d3f-4f7e-a0ae-8d541024983b"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:29:19 crc kubenswrapper[4764]: I0202 09:29:18.499039 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ff8a4b58-1d3f-4f7e-a0ae-8d541024983b-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ff8a4b58-1d3f-4f7e-a0ae-8d541024983b" (UID: "ff8a4b58-1d3f-4f7e-a0ae-8d541024983b"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:29:19 crc kubenswrapper[4764]: I0202 09:29:18.512752 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ff8a4b58-1d3f-4f7e-a0ae-8d541024983b-config" (OuterVolumeSpecName: "config") pod "ff8a4b58-1d3f-4f7e-a0ae-8d541024983b" (UID: "ff8a4b58-1d3f-4f7e-a0ae-8d541024983b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:29:19 crc kubenswrapper[4764]: I0202 09:29:18.515498 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ff8a4b58-1d3f-4f7e-a0ae-8d541024983b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ff8a4b58-1d3f-4f7e-a0ae-8d541024983b" (UID: "ff8a4b58-1d3f-4f7e-a0ae-8d541024983b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:29:19 crc kubenswrapper[4764]: I0202 09:29:18.537573 4764 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ff8a4b58-1d3f-4f7e-a0ae-8d541024983b-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 02 09:29:19 crc kubenswrapper[4764]: I0202 09:29:18.537596 4764 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ff8a4b58-1d3f-4f7e-a0ae-8d541024983b-config\") on node \"crc\" DevicePath \"\"" Feb 02 09:29:19 crc kubenswrapper[4764]: I0202 09:29:18.537605 4764 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ff8a4b58-1d3f-4f7e-a0ae-8d541024983b-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 02 09:29:19 crc kubenswrapper[4764]: I0202 09:29:18.537615 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rhd8r\" (UniqueName: \"kubernetes.io/projected/ff8a4b58-1d3f-4f7e-a0ae-8d541024983b-kube-api-access-rhd8r\") on node \"crc\" DevicePath \"\"" Feb 02 09:29:19 crc kubenswrapper[4764]: I0202 09:29:18.537624 4764 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ff8a4b58-1d3f-4f7e-a0ae-8d541024983b-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 02 09:29:19 crc kubenswrapper[4764]: I0202 09:29:19.064796 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75956b9989-5ddn6" event={"ID":"ff8a4b58-1d3f-4f7e-a0ae-8d541024983b","Type":"ContainerDied","Data":"577ed8e752e8a2300366b2364f3476e9d999aa3f4e3da7456bef1b67f7fe5a5b"} Feb 02 09:29:19 crc kubenswrapper[4764]: I0202 09:29:19.064841 4764 scope.go:117] "RemoveContainer" containerID="ba78ef785372350834e28f350c85c6581e766f4c306e316726c548005b1ee000" Feb 02 09:29:19 crc kubenswrapper[4764]: I0202 09:29:19.064863 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75956b9989-5ddn6" Feb 02 09:29:19 crc kubenswrapper[4764]: I0202 09:29:19.089459 4764 scope.go:117] "RemoveContainer" containerID="228b569a80f432fd2a8b2ae069ff1ce409d3d7c8f71ec9287bcb0e687e02d457" Feb 02 09:29:19 crc kubenswrapper[4764]: I0202 09:29:19.117775 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-75956b9989-5ddn6"] Feb 02 09:29:19 crc kubenswrapper[4764]: I0202 09:29:19.125663 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-75956b9989-5ddn6"] Feb 02 09:29:19 crc kubenswrapper[4764]: I0202 09:29:19.457833 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-77755c997f-6996z"] Feb 02 09:29:19 crc kubenswrapper[4764]: W0202 09:29:19.467792 4764 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod33b1de07_e5a5_472a_acfb_102dbfe83e0e.slice/crio-9a5a2a5be6affd06eb000aba6a885f6c506bcd48084acc13349fd4216e72945e WatchSource:0}: Error finding container 9a5a2a5be6affd06eb000aba6a885f6c506bcd48084acc13349fd4216e72945e: Status 404 returned error can't find the container with id 9a5a2a5be6affd06eb000aba6a885f6c506bcd48084acc13349fd4216e72945e Feb 02 09:29:19 crc kubenswrapper[4764]: I0202 09:29:19.838525 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ff8a4b58-1d3f-4f7e-a0ae-8d541024983b" path="/var/lib/kubelet/pods/ff8a4b58-1d3f-4f7e-a0ae-8d541024983b/volumes" Feb 02 09:29:20 crc kubenswrapper[4764]: I0202 09:29:20.074461 4764 generic.go:334] "Generic (PLEG): container finished" podID="33b1de07-e5a5-472a-acfb-102dbfe83e0e" containerID="781e28062c683f31972368a29d5067daf908ad95169f41f525d8e9446af75809" exitCode=0 Feb 02 09:29:20 crc kubenswrapper[4764]: I0202 09:29:20.074498 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77755c997f-6996z" event={"ID":"33b1de07-e5a5-472a-acfb-102dbfe83e0e","Type":"ContainerDied","Data":"781e28062c683f31972368a29d5067daf908ad95169f41f525d8e9446af75809"} Feb 02 09:29:20 crc kubenswrapper[4764]: I0202 09:29:20.074521 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77755c997f-6996z" event={"ID":"33b1de07-e5a5-472a-acfb-102dbfe83e0e","Type":"ContainerStarted","Data":"9a5a2a5be6affd06eb000aba6a885f6c506bcd48084acc13349fd4216e72945e"} Feb 02 09:29:21 crc kubenswrapper[4764]: I0202 09:29:21.087214 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77755c997f-6996z" event={"ID":"33b1de07-e5a5-472a-acfb-102dbfe83e0e","Type":"ContainerStarted","Data":"b42ec966d617e60cef3c0bd4351d46ef96d749b7d8a53fafb653e8d7beb3a126"} Feb 02 09:29:21 crc kubenswrapper[4764]: I0202 09:29:21.088991 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-77755c997f-6996z" Feb 02 09:29:28 crc kubenswrapper[4764]: I0202 09:29:28.302154 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-77755c997f-6996z" Feb 02 09:29:28 crc kubenswrapper[4764]: I0202 09:29:28.331191 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-77755c997f-6996z" podStartSLOduration=11.331173331 podStartE2EDuration="11.331173331s" podCreationTimestamp="2026-02-02 09:29:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:29:21.111809191 +0000 UTC m=+1324.045533299" watchObservedRunningTime="2026-02-02 09:29:28.331173331 +0000 UTC m=+1331.264897419" Feb 02 09:29:28 crc kubenswrapper[4764]: I0202 09:29:28.435921 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b4d48ddf9-kbdwn"] Feb 02 09:29:28 crc kubenswrapper[4764]: I0202 09:29:28.436147 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-b4d48ddf9-kbdwn" podUID="a71471c2-6a47-403f-b511-04a512577dcf" containerName="dnsmasq-dns" containerID="cri-o://af4ab96a20f13c81bcc82ad0e0602fbdf553d56a33c0bd91c84deca64f54c681" gracePeriod=10 Feb 02 09:29:28 crc kubenswrapper[4764]: I0202 09:29:28.919226 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b4d48ddf9-kbdwn" Feb 02 09:29:29 crc kubenswrapper[4764]: I0202 09:29:29.066158 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a71471c2-6a47-403f-b511-04a512577dcf-config\") pod \"a71471c2-6a47-403f-b511-04a512577dcf\" (UID: \"a71471c2-6a47-403f-b511-04a512577dcf\") " Feb 02 09:29:29 crc kubenswrapper[4764]: I0202 09:29:29.066204 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/a71471c2-6a47-403f-b511-04a512577dcf-openstack-edpm-ipam\") pod \"a71471c2-6a47-403f-b511-04a512577dcf\" (UID: \"a71471c2-6a47-403f-b511-04a512577dcf\") " Feb 02 09:29:29 crc kubenswrapper[4764]: I0202 09:29:29.066257 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a71471c2-6a47-403f-b511-04a512577dcf-dns-svc\") pod \"a71471c2-6a47-403f-b511-04a512577dcf\" (UID: \"a71471c2-6a47-403f-b511-04a512577dcf\") " Feb 02 09:29:29 crc kubenswrapper[4764]: I0202 09:29:29.066276 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a71471c2-6a47-403f-b511-04a512577dcf-ovsdbserver-nb\") pod \"a71471c2-6a47-403f-b511-04a512577dcf\" (UID: \"a71471c2-6a47-403f-b511-04a512577dcf\") " Feb 02 09:29:29 crc kubenswrapper[4764]: I0202 09:29:29.066352 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6gpc2\" (UniqueName: \"kubernetes.io/projected/a71471c2-6a47-403f-b511-04a512577dcf-kube-api-access-6gpc2\") pod \"a71471c2-6a47-403f-b511-04a512577dcf\" (UID: \"a71471c2-6a47-403f-b511-04a512577dcf\") " Feb 02 09:29:29 crc kubenswrapper[4764]: I0202 09:29:29.066371 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a71471c2-6a47-403f-b511-04a512577dcf-ovsdbserver-sb\") pod \"a71471c2-6a47-403f-b511-04a512577dcf\" (UID: \"a71471c2-6a47-403f-b511-04a512577dcf\") " Feb 02 09:29:29 crc kubenswrapper[4764]: I0202 09:29:29.083194 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a71471c2-6a47-403f-b511-04a512577dcf-kube-api-access-6gpc2" (OuterVolumeSpecName: "kube-api-access-6gpc2") pod "a71471c2-6a47-403f-b511-04a512577dcf" (UID: "a71471c2-6a47-403f-b511-04a512577dcf"). InnerVolumeSpecName "kube-api-access-6gpc2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:29:29 crc kubenswrapper[4764]: I0202 09:29:29.152213 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a71471c2-6a47-403f-b511-04a512577dcf-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "a71471c2-6a47-403f-b511-04a512577dcf" (UID: "a71471c2-6a47-403f-b511-04a512577dcf"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:29:29 crc kubenswrapper[4764]: I0202 09:29:29.156397 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a71471c2-6a47-403f-b511-04a512577dcf-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "a71471c2-6a47-403f-b511-04a512577dcf" (UID: "a71471c2-6a47-403f-b511-04a512577dcf"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:29:29 crc kubenswrapper[4764]: I0202 09:29:29.156500 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a71471c2-6a47-403f-b511-04a512577dcf-config" (OuterVolumeSpecName: "config") pod "a71471c2-6a47-403f-b511-04a512577dcf" (UID: "a71471c2-6a47-403f-b511-04a512577dcf"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:29:29 crc kubenswrapper[4764]: I0202 09:29:29.168115 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a71471c2-6a47-403f-b511-04a512577dcf-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "a71471c2-6a47-403f-b511-04a512577dcf" (UID: "a71471c2-6a47-403f-b511-04a512577dcf"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:29:29 crc kubenswrapper[4764]: I0202 09:29:29.168374 4764 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a71471c2-6a47-403f-b511-04a512577dcf-config\") on node \"crc\" DevicePath \"\"" Feb 02 09:29:29 crc kubenswrapper[4764]: I0202 09:29:29.168396 4764 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/a71471c2-6a47-403f-b511-04a512577dcf-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 02 09:29:29 crc kubenswrapper[4764]: I0202 09:29:29.168406 4764 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a71471c2-6a47-403f-b511-04a512577dcf-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 02 09:29:29 crc kubenswrapper[4764]: I0202 09:29:29.168414 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6gpc2\" (UniqueName: \"kubernetes.io/projected/a71471c2-6a47-403f-b511-04a512577dcf-kube-api-access-6gpc2\") on node \"crc\" DevicePath \"\"" Feb 02 09:29:29 crc kubenswrapper[4764]: I0202 09:29:29.168421 4764 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a71471c2-6a47-403f-b511-04a512577dcf-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 02 09:29:29 crc kubenswrapper[4764]: I0202 09:29:29.178135 4764 generic.go:334] "Generic (PLEG): container finished" podID="a71471c2-6a47-403f-b511-04a512577dcf" containerID="af4ab96a20f13c81bcc82ad0e0602fbdf553d56a33c0bd91c84deca64f54c681" exitCode=0 Feb 02 09:29:29 crc kubenswrapper[4764]: I0202 09:29:29.178177 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b4d48ddf9-kbdwn" event={"ID":"a71471c2-6a47-403f-b511-04a512577dcf","Type":"ContainerDied","Data":"af4ab96a20f13c81bcc82ad0e0602fbdf553d56a33c0bd91c84deca64f54c681"} Feb 02 09:29:29 crc kubenswrapper[4764]: I0202 09:29:29.178205 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b4d48ddf9-kbdwn" event={"ID":"a71471c2-6a47-403f-b511-04a512577dcf","Type":"ContainerDied","Data":"2757a1ce774d687add422263ad13e527ea2501d77368a5104a32323fc7a52f41"} Feb 02 09:29:29 crc kubenswrapper[4764]: I0202 09:29:29.178223 4764 scope.go:117] "RemoveContainer" containerID="af4ab96a20f13c81bcc82ad0e0602fbdf553d56a33c0bd91c84deca64f54c681" Feb 02 09:29:29 crc kubenswrapper[4764]: I0202 09:29:29.178338 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b4d48ddf9-kbdwn" Feb 02 09:29:29 crc kubenswrapper[4764]: I0202 09:29:29.195980 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a71471c2-6a47-403f-b511-04a512577dcf-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a71471c2-6a47-403f-b511-04a512577dcf" (UID: "a71471c2-6a47-403f-b511-04a512577dcf"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:29:29 crc kubenswrapper[4764]: I0202 09:29:29.197035 4764 scope.go:117] "RemoveContainer" containerID="8fffdb31bc7dce696ff8f7e94c471934ee94135606ee981115620d1db6cd2dc5" Feb 02 09:29:29 crc kubenswrapper[4764]: I0202 09:29:29.214152 4764 scope.go:117] "RemoveContainer" containerID="af4ab96a20f13c81bcc82ad0e0602fbdf553d56a33c0bd91c84deca64f54c681" Feb 02 09:29:29 crc kubenswrapper[4764]: E0202 09:29:29.214612 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"af4ab96a20f13c81bcc82ad0e0602fbdf553d56a33c0bd91c84deca64f54c681\": container with ID starting with af4ab96a20f13c81bcc82ad0e0602fbdf553d56a33c0bd91c84deca64f54c681 not found: ID does not exist" containerID="af4ab96a20f13c81bcc82ad0e0602fbdf553d56a33c0bd91c84deca64f54c681" Feb 02 09:29:29 crc kubenswrapper[4764]: I0202 09:29:29.214651 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af4ab96a20f13c81bcc82ad0e0602fbdf553d56a33c0bd91c84deca64f54c681"} err="failed to get container status \"af4ab96a20f13c81bcc82ad0e0602fbdf553d56a33c0bd91c84deca64f54c681\": rpc error: code = NotFound desc = could not find container \"af4ab96a20f13c81bcc82ad0e0602fbdf553d56a33c0bd91c84deca64f54c681\": container with ID starting with af4ab96a20f13c81bcc82ad0e0602fbdf553d56a33c0bd91c84deca64f54c681 not found: ID does not exist" Feb 02 09:29:29 crc kubenswrapper[4764]: I0202 09:29:29.214677 4764 scope.go:117] "RemoveContainer" containerID="8fffdb31bc7dce696ff8f7e94c471934ee94135606ee981115620d1db6cd2dc5" Feb 02 09:29:29 crc kubenswrapper[4764]: E0202 09:29:29.215017 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8fffdb31bc7dce696ff8f7e94c471934ee94135606ee981115620d1db6cd2dc5\": container with ID starting with 8fffdb31bc7dce696ff8f7e94c471934ee94135606ee981115620d1db6cd2dc5 not found: ID does not exist" containerID="8fffdb31bc7dce696ff8f7e94c471934ee94135606ee981115620d1db6cd2dc5" Feb 02 09:29:29 crc kubenswrapper[4764]: I0202 09:29:29.215039 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8fffdb31bc7dce696ff8f7e94c471934ee94135606ee981115620d1db6cd2dc5"} err="failed to get container status \"8fffdb31bc7dce696ff8f7e94c471934ee94135606ee981115620d1db6cd2dc5\": rpc error: code = NotFound desc = could not find container \"8fffdb31bc7dce696ff8f7e94c471934ee94135606ee981115620d1db6cd2dc5\": container with ID starting with 8fffdb31bc7dce696ff8f7e94c471934ee94135606ee981115620d1db6cd2dc5 not found: ID does not exist" Feb 02 09:29:29 crc kubenswrapper[4764]: I0202 09:29:29.270387 4764 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a71471c2-6a47-403f-b511-04a512577dcf-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 02 09:29:29 crc kubenswrapper[4764]: I0202 09:29:29.529663 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b4d48ddf9-kbdwn"] Feb 02 09:29:29 crc kubenswrapper[4764]: I0202 09:29:29.538286 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-b4d48ddf9-kbdwn"] Feb 02 09:29:29 crc kubenswrapper[4764]: I0202 09:29:29.835572 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a71471c2-6a47-403f-b511-04a512577dcf" path="/var/lib/kubelet/pods/a71471c2-6a47-403f-b511-04a512577dcf/volumes" Feb 02 09:29:38 crc kubenswrapper[4764]: I0202 09:29:38.171302 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gv65q"] Feb 02 09:29:38 crc kubenswrapper[4764]: E0202 09:29:38.172400 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a71471c2-6a47-403f-b511-04a512577dcf" containerName="init" Feb 02 09:29:38 crc kubenswrapper[4764]: I0202 09:29:38.172421 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="a71471c2-6a47-403f-b511-04a512577dcf" containerName="init" Feb 02 09:29:38 crc kubenswrapper[4764]: E0202 09:29:38.172438 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a71471c2-6a47-403f-b511-04a512577dcf" containerName="dnsmasq-dns" Feb 02 09:29:38 crc kubenswrapper[4764]: I0202 09:29:38.172451 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="a71471c2-6a47-403f-b511-04a512577dcf" containerName="dnsmasq-dns" Feb 02 09:29:38 crc kubenswrapper[4764]: E0202 09:29:38.172478 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff8a4b58-1d3f-4f7e-a0ae-8d541024983b" containerName="init" Feb 02 09:29:38 crc kubenswrapper[4764]: I0202 09:29:38.172491 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff8a4b58-1d3f-4f7e-a0ae-8d541024983b" containerName="init" Feb 02 09:29:38 crc kubenswrapper[4764]: E0202 09:29:38.172541 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff8a4b58-1d3f-4f7e-a0ae-8d541024983b" containerName="dnsmasq-dns" Feb 02 09:29:38 crc kubenswrapper[4764]: I0202 09:29:38.172554 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff8a4b58-1d3f-4f7e-a0ae-8d541024983b" containerName="dnsmasq-dns" Feb 02 09:29:38 crc kubenswrapper[4764]: I0202 09:29:38.172856 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff8a4b58-1d3f-4f7e-a0ae-8d541024983b" containerName="dnsmasq-dns" Feb 02 09:29:38 crc kubenswrapper[4764]: I0202 09:29:38.172881 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="a71471c2-6a47-403f-b511-04a512577dcf" containerName="dnsmasq-dns" Feb 02 09:29:38 crc kubenswrapper[4764]: I0202 09:29:38.173876 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gv65q" Feb 02 09:29:38 crc kubenswrapper[4764]: I0202 09:29:38.177131 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 02 09:29:38 crc kubenswrapper[4764]: I0202 09:29:38.177557 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 02 09:29:38 crc kubenswrapper[4764]: I0202 09:29:38.177867 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 02 09:29:38 crc kubenswrapper[4764]: I0202 09:29:38.178302 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-xksq5" Feb 02 09:29:38 crc kubenswrapper[4764]: I0202 09:29:38.198051 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gv65q"] Feb 02 09:29:38 crc kubenswrapper[4764]: I0202 09:29:38.333586 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d04cc0fd-f303-47b2-a8d4-6efd0a591b77-ssh-key-openstack-edpm-ipam\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-gv65q\" (UID: \"d04cc0fd-f303-47b2-a8d4-6efd0a591b77\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gv65q" Feb 02 09:29:38 crc kubenswrapper[4764]: I0202 09:29:38.333803 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c7vzb\" (UniqueName: \"kubernetes.io/projected/d04cc0fd-f303-47b2-a8d4-6efd0a591b77-kube-api-access-c7vzb\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-gv65q\" (UID: \"d04cc0fd-f303-47b2-a8d4-6efd0a591b77\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gv65q" Feb 02 09:29:38 crc kubenswrapper[4764]: I0202 09:29:38.333849 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d04cc0fd-f303-47b2-a8d4-6efd0a591b77-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-gv65q\" (UID: \"d04cc0fd-f303-47b2-a8d4-6efd0a591b77\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gv65q" Feb 02 09:29:38 crc kubenswrapper[4764]: I0202 09:29:38.333925 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d04cc0fd-f303-47b2-a8d4-6efd0a591b77-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-gv65q\" (UID: \"d04cc0fd-f303-47b2-a8d4-6efd0a591b77\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gv65q" Feb 02 09:29:38 crc kubenswrapper[4764]: I0202 09:29:38.435405 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c7vzb\" (UniqueName: \"kubernetes.io/projected/d04cc0fd-f303-47b2-a8d4-6efd0a591b77-kube-api-access-c7vzb\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-gv65q\" (UID: \"d04cc0fd-f303-47b2-a8d4-6efd0a591b77\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gv65q" Feb 02 09:29:38 crc kubenswrapper[4764]: I0202 09:29:38.435448 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d04cc0fd-f303-47b2-a8d4-6efd0a591b77-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-gv65q\" (UID: \"d04cc0fd-f303-47b2-a8d4-6efd0a591b77\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gv65q" Feb 02 09:29:38 crc kubenswrapper[4764]: I0202 09:29:38.435468 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d04cc0fd-f303-47b2-a8d4-6efd0a591b77-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-gv65q\" (UID: \"d04cc0fd-f303-47b2-a8d4-6efd0a591b77\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gv65q" Feb 02 09:29:38 crc kubenswrapper[4764]: I0202 09:29:38.435511 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d04cc0fd-f303-47b2-a8d4-6efd0a591b77-ssh-key-openstack-edpm-ipam\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-gv65q\" (UID: \"d04cc0fd-f303-47b2-a8d4-6efd0a591b77\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gv65q" Feb 02 09:29:38 crc kubenswrapper[4764]: I0202 09:29:38.441418 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d04cc0fd-f303-47b2-a8d4-6efd0a591b77-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-gv65q\" (UID: \"d04cc0fd-f303-47b2-a8d4-6efd0a591b77\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gv65q" Feb 02 09:29:38 crc kubenswrapper[4764]: I0202 09:29:38.441481 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d04cc0fd-f303-47b2-a8d4-6efd0a591b77-ssh-key-openstack-edpm-ipam\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-gv65q\" (UID: \"d04cc0fd-f303-47b2-a8d4-6efd0a591b77\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gv65q" Feb 02 09:29:38 crc kubenswrapper[4764]: I0202 09:29:38.444048 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d04cc0fd-f303-47b2-a8d4-6efd0a591b77-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-gv65q\" (UID: \"d04cc0fd-f303-47b2-a8d4-6efd0a591b77\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gv65q" Feb 02 09:29:38 crc kubenswrapper[4764]: I0202 09:29:38.457483 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c7vzb\" (UniqueName: \"kubernetes.io/projected/d04cc0fd-f303-47b2-a8d4-6efd0a591b77-kube-api-access-c7vzb\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-gv65q\" (UID: \"d04cc0fd-f303-47b2-a8d4-6efd0a591b77\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gv65q" Feb 02 09:29:38 crc kubenswrapper[4764]: I0202 09:29:38.504207 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gv65q" Feb 02 09:29:39 crc kubenswrapper[4764]: I0202 09:29:39.114178 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gv65q"] Feb 02 09:29:39 crc kubenswrapper[4764]: I0202 09:29:39.308845 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gv65q" event={"ID":"d04cc0fd-f303-47b2-a8d4-6efd0a591b77","Type":"ContainerStarted","Data":"f30fe347ddf4634495d2a806b4d7582fe0a60feb8a6f73e3e028091306ef2c2e"} Feb 02 09:29:40 crc kubenswrapper[4764]: I0202 09:29:40.320016 4764 generic.go:334] "Generic (PLEG): container finished" podID="ec28045f-e1d4-4b22-bb7d-3c6d2a1f1a38" containerID="99db28642a3c01b0c74f9ee5f2768815ab4b293028c9e0891a0d582b53104ab3" exitCode=0 Feb 02 09:29:40 crc kubenswrapper[4764]: I0202 09:29:40.320169 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"ec28045f-e1d4-4b22-bb7d-3c6d2a1f1a38","Type":"ContainerDied","Data":"99db28642a3c01b0c74f9ee5f2768815ab4b293028c9e0891a0d582b53104ab3"} Feb 02 09:29:41 crc kubenswrapper[4764]: I0202 09:29:41.344220 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"ec28045f-e1d4-4b22-bb7d-3c6d2a1f1a38","Type":"ContainerStarted","Data":"197df6a680d886ced44059ae14abca0383aad6ddeefbc7128bf326a55c7aa312"} Feb 02 09:29:41 crc kubenswrapper[4764]: I0202 09:29:41.345283 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Feb 02 09:29:41 crc kubenswrapper[4764]: I0202 09:29:41.348002 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"c60b198c-58ad-4363-afea-257f51f64e51","Type":"ContainerDied","Data":"48f8feab56900bd296e165775c719351ff9243c9a64a42f3116e3c49c612c42f"} Feb 02 09:29:41 crc kubenswrapper[4764]: I0202 09:29:41.348163 4764 generic.go:334] "Generic (PLEG): container finished" podID="c60b198c-58ad-4363-afea-257f51f64e51" containerID="48f8feab56900bd296e165775c719351ff9243c9a64a42f3116e3c49c612c42f" exitCode=0 Feb 02 09:29:41 crc kubenswrapper[4764]: I0202 09:29:41.378248 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=37.378230782 podStartE2EDuration="37.378230782s" podCreationTimestamp="2026-02-02 09:29:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:29:41.36612146 +0000 UTC m=+1344.299845548" watchObservedRunningTime="2026-02-02 09:29:41.378230782 +0000 UTC m=+1344.311954870" Feb 02 09:29:42 crc kubenswrapper[4764]: I0202 09:29:42.359965 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"c60b198c-58ad-4363-afea-257f51f64e51","Type":"ContainerStarted","Data":"71e293c225e5d3a4758fe594adfb83604bdde097286fbc25981d7ce0f167655e"} Feb 02 09:29:42 crc kubenswrapper[4764]: I0202 09:29:42.360493 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Feb 02 09:29:42 crc kubenswrapper[4764]: I0202 09:29:42.420036 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=37.42002062 podStartE2EDuration="37.42002062s" podCreationTimestamp="2026-02-02 09:29:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:29:42.407428775 +0000 UTC m=+1345.341152863" watchObservedRunningTime="2026-02-02 09:29:42.42002062 +0000 UTC m=+1345.353744708" Feb 02 09:29:43 crc kubenswrapper[4764]: I0202 09:29:43.523187 4764 patch_prober.go:28] interesting pod/machine-config-daemon-4ndm4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 09:29:43 crc kubenswrapper[4764]: I0202 09:29:43.523488 4764 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 09:29:43 crc kubenswrapper[4764]: I0202 09:29:43.523530 4764 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" Feb 02 09:29:43 crc kubenswrapper[4764]: I0202 09:29:43.524178 4764 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7e0e1c40ccad424f2dce09dc2a68260202cf2abc147b119996f40fbb4f3a453b"} pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 02 09:29:43 crc kubenswrapper[4764]: I0202 09:29:43.524235 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" containerName="machine-config-daemon" containerID="cri-o://7e0e1c40ccad424f2dce09dc2a68260202cf2abc147b119996f40fbb4f3a453b" gracePeriod=600 Feb 02 09:29:44 crc kubenswrapper[4764]: I0202 09:29:44.380480 4764 generic.go:334] "Generic (PLEG): container finished" podID="d192f670-9f9d-4539-9641-e4bed73acdd4" containerID="7e0e1c40ccad424f2dce09dc2a68260202cf2abc147b119996f40fbb4f3a453b" exitCode=0 Feb 02 09:29:44 crc kubenswrapper[4764]: I0202 09:29:44.380527 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" event={"ID":"d192f670-9f9d-4539-9641-e4bed73acdd4","Type":"ContainerDied","Data":"7e0e1c40ccad424f2dce09dc2a68260202cf2abc147b119996f40fbb4f3a453b"} Feb 02 09:29:44 crc kubenswrapper[4764]: I0202 09:29:44.380569 4764 scope.go:117] "RemoveContainer" containerID="bbb9c6cc86c729fedc84038f3af8b700a433083aa4652e832aca9d3d773be9fd" Feb 02 09:29:50 crc kubenswrapper[4764]: I0202 09:29:50.436418 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gv65q" event={"ID":"d04cc0fd-f303-47b2-a8d4-6efd0a591b77","Type":"ContainerStarted","Data":"f130cc11505e280b34eb9b6df346654a0fad23fde9ea5194446dd5fed996e091"} Feb 02 09:29:50 crc kubenswrapper[4764]: I0202 09:29:50.439414 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" event={"ID":"d192f670-9f9d-4539-9641-e4bed73acdd4","Type":"ContainerStarted","Data":"6db43ab36406cdd4dbe61d7d8688257aba4271b8601a5ba5594675ae309c761d"} Feb 02 09:29:50 crc kubenswrapper[4764]: I0202 09:29:50.462239 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gv65q" podStartSLOduration=1.6829634119999999 podStartE2EDuration="12.462220446s" podCreationTimestamp="2026-02-02 09:29:38 +0000 UTC" firstStartedPulling="2026-02-02 09:29:39.118864861 +0000 UTC m=+1342.052588979" lastFinishedPulling="2026-02-02 09:29:49.898121935 +0000 UTC m=+1352.831846013" observedRunningTime="2026-02-02 09:29:50.451030798 +0000 UTC m=+1353.384754886" watchObservedRunningTime="2026-02-02 09:29:50.462220446 +0000 UTC m=+1353.395944534" Feb 02 09:29:55 crc kubenswrapper[4764]: I0202 09:29:55.686284 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Feb 02 09:29:56 crc kubenswrapper[4764]: I0202 09:29:56.389115 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Feb 02 09:30:00 crc kubenswrapper[4764]: I0202 09:30:00.142375 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29500410-b8znn"] Feb 02 09:30:00 crc kubenswrapper[4764]: I0202 09:30:00.143884 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29500410-b8znn" Feb 02 09:30:00 crc kubenswrapper[4764]: I0202 09:30:00.145764 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 02 09:30:00 crc kubenswrapper[4764]: I0202 09:30:00.149185 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 02 09:30:00 crc kubenswrapper[4764]: I0202 09:30:00.167066 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29500410-b8znn"] Feb 02 09:30:00 crc kubenswrapper[4764]: I0202 09:30:00.249391 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fd431fc8-7193-4ad6-b7f6-eb83b4dbe824-config-volume\") pod \"collect-profiles-29500410-b8znn\" (UID: \"fd431fc8-7193-4ad6-b7f6-eb83b4dbe824\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500410-b8znn" Feb 02 09:30:00 crc kubenswrapper[4764]: I0202 09:30:00.249491 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j8xmz\" (UniqueName: \"kubernetes.io/projected/fd431fc8-7193-4ad6-b7f6-eb83b4dbe824-kube-api-access-j8xmz\") pod \"collect-profiles-29500410-b8znn\" (UID: \"fd431fc8-7193-4ad6-b7f6-eb83b4dbe824\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500410-b8znn" Feb 02 09:30:00 crc kubenswrapper[4764]: I0202 09:30:00.249531 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/fd431fc8-7193-4ad6-b7f6-eb83b4dbe824-secret-volume\") pod \"collect-profiles-29500410-b8znn\" (UID: \"fd431fc8-7193-4ad6-b7f6-eb83b4dbe824\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500410-b8znn" Feb 02 09:30:00 crc kubenswrapper[4764]: I0202 09:30:00.351487 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j8xmz\" (UniqueName: \"kubernetes.io/projected/fd431fc8-7193-4ad6-b7f6-eb83b4dbe824-kube-api-access-j8xmz\") pod \"collect-profiles-29500410-b8znn\" (UID: \"fd431fc8-7193-4ad6-b7f6-eb83b4dbe824\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500410-b8znn" Feb 02 09:30:00 crc kubenswrapper[4764]: I0202 09:30:00.351618 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/fd431fc8-7193-4ad6-b7f6-eb83b4dbe824-secret-volume\") pod \"collect-profiles-29500410-b8znn\" (UID: \"fd431fc8-7193-4ad6-b7f6-eb83b4dbe824\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500410-b8znn" Feb 02 09:30:00 crc kubenswrapper[4764]: I0202 09:30:00.351744 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fd431fc8-7193-4ad6-b7f6-eb83b4dbe824-config-volume\") pod \"collect-profiles-29500410-b8znn\" (UID: \"fd431fc8-7193-4ad6-b7f6-eb83b4dbe824\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500410-b8znn" Feb 02 09:30:00 crc kubenswrapper[4764]: I0202 09:30:00.352904 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fd431fc8-7193-4ad6-b7f6-eb83b4dbe824-config-volume\") pod \"collect-profiles-29500410-b8znn\" (UID: \"fd431fc8-7193-4ad6-b7f6-eb83b4dbe824\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500410-b8znn" Feb 02 09:30:00 crc kubenswrapper[4764]: I0202 09:30:00.366680 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/fd431fc8-7193-4ad6-b7f6-eb83b4dbe824-secret-volume\") pod \"collect-profiles-29500410-b8znn\" (UID: \"fd431fc8-7193-4ad6-b7f6-eb83b4dbe824\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500410-b8znn" Feb 02 09:30:00 crc kubenswrapper[4764]: I0202 09:30:00.373534 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j8xmz\" (UniqueName: \"kubernetes.io/projected/fd431fc8-7193-4ad6-b7f6-eb83b4dbe824-kube-api-access-j8xmz\") pod \"collect-profiles-29500410-b8znn\" (UID: \"fd431fc8-7193-4ad6-b7f6-eb83b4dbe824\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500410-b8znn" Feb 02 09:30:00 crc kubenswrapper[4764]: I0202 09:30:00.468031 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29500410-b8znn" Feb 02 09:30:00 crc kubenswrapper[4764]: I0202 09:30:00.927076 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29500410-b8znn"] Feb 02 09:30:00 crc kubenswrapper[4764]: W0202 09:30:00.929978 4764 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfd431fc8_7193_4ad6_b7f6_eb83b4dbe824.slice/crio-74443e3a86ed26b871c478b98a5d2a1d608321b3b629861c30cec9dc662a7f10 WatchSource:0}: Error finding container 74443e3a86ed26b871c478b98a5d2a1d608321b3b629861c30cec9dc662a7f10: Status 404 returned error can't find the container with id 74443e3a86ed26b871c478b98a5d2a1d608321b3b629861c30cec9dc662a7f10 Feb 02 09:30:01 crc kubenswrapper[4764]: I0202 09:30:01.542369 4764 generic.go:334] "Generic (PLEG): container finished" podID="d04cc0fd-f303-47b2-a8d4-6efd0a591b77" containerID="f130cc11505e280b34eb9b6df346654a0fad23fde9ea5194446dd5fed996e091" exitCode=0 Feb 02 09:30:01 crc kubenswrapper[4764]: I0202 09:30:01.542579 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gv65q" event={"ID":"d04cc0fd-f303-47b2-a8d4-6efd0a591b77","Type":"ContainerDied","Data":"f130cc11505e280b34eb9b6df346654a0fad23fde9ea5194446dd5fed996e091"} Feb 02 09:30:01 crc kubenswrapper[4764]: I0202 09:30:01.545569 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29500410-b8znn" event={"ID":"fd431fc8-7193-4ad6-b7f6-eb83b4dbe824","Type":"ContainerStarted","Data":"5e600332710ba1ec80f0870d729d873b95fe41f319c1d38d3d8d14af9bb59bae"} Feb 02 09:30:01 crc kubenswrapper[4764]: I0202 09:30:01.545610 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29500410-b8znn" event={"ID":"fd431fc8-7193-4ad6-b7f6-eb83b4dbe824","Type":"ContainerStarted","Data":"74443e3a86ed26b871c478b98a5d2a1d608321b3b629861c30cec9dc662a7f10"} Feb 02 09:30:01 crc kubenswrapper[4764]: I0202 09:30:01.594067 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29500410-b8znn" podStartSLOduration=1.5940387980000001 podStartE2EDuration="1.594038798s" podCreationTimestamp="2026-02-02 09:30:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:30:01.586888534 +0000 UTC m=+1364.520612632" watchObservedRunningTime="2026-02-02 09:30:01.594038798 +0000 UTC m=+1364.527762916" Feb 02 09:30:02 crc kubenswrapper[4764]: I0202 09:30:02.561215 4764 generic.go:334] "Generic (PLEG): container finished" podID="fd431fc8-7193-4ad6-b7f6-eb83b4dbe824" containerID="5e600332710ba1ec80f0870d729d873b95fe41f319c1d38d3d8d14af9bb59bae" exitCode=0 Feb 02 09:30:02 crc kubenswrapper[4764]: I0202 09:30:02.561339 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29500410-b8znn" event={"ID":"fd431fc8-7193-4ad6-b7f6-eb83b4dbe824","Type":"ContainerDied","Data":"5e600332710ba1ec80f0870d729d873b95fe41f319c1d38d3d8d14af9bb59bae"} Feb 02 09:30:03 crc kubenswrapper[4764]: I0202 09:30:03.019297 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gv65q" Feb 02 09:30:03 crc kubenswrapper[4764]: I0202 09:30:03.118054 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d04cc0fd-f303-47b2-a8d4-6efd0a591b77-inventory\") pod \"d04cc0fd-f303-47b2-a8d4-6efd0a591b77\" (UID: \"d04cc0fd-f303-47b2-a8d4-6efd0a591b77\") " Feb 02 09:30:03 crc kubenswrapper[4764]: I0202 09:30:03.118185 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c7vzb\" (UniqueName: \"kubernetes.io/projected/d04cc0fd-f303-47b2-a8d4-6efd0a591b77-kube-api-access-c7vzb\") pod \"d04cc0fd-f303-47b2-a8d4-6efd0a591b77\" (UID: \"d04cc0fd-f303-47b2-a8d4-6efd0a591b77\") " Feb 02 09:30:03 crc kubenswrapper[4764]: I0202 09:30:03.118559 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d04cc0fd-f303-47b2-a8d4-6efd0a591b77-repo-setup-combined-ca-bundle\") pod \"d04cc0fd-f303-47b2-a8d4-6efd0a591b77\" (UID: \"d04cc0fd-f303-47b2-a8d4-6efd0a591b77\") " Feb 02 09:30:03 crc kubenswrapper[4764]: I0202 09:30:03.118628 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d04cc0fd-f303-47b2-a8d4-6efd0a591b77-ssh-key-openstack-edpm-ipam\") pod \"d04cc0fd-f303-47b2-a8d4-6efd0a591b77\" (UID: \"d04cc0fd-f303-47b2-a8d4-6efd0a591b77\") " Feb 02 09:30:03 crc kubenswrapper[4764]: I0202 09:30:03.124702 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d04cc0fd-f303-47b2-a8d4-6efd0a591b77-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "d04cc0fd-f303-47b2-a8d4-6efd0a591b77" (UID: "d04cc0fd-f303-47b2-a8d4-6efd0a591b77"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:30:03 crc kubenswrapper[4764]: I0202 09:30:03.129911 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d04cc0fd-f303-47b2-a8d4-6efd0a591b77-kube-api-access-c7vzb" (OuterVolumeSpecName: "kube-api-access-c7vzb") pod "d04cc0fd-f303-47b2-a8d4-6efd0a591b77" (UID: "d04cc0fd-f303-47b2-a8d4-6efd0a591b77"). InnerVolumeSpecName "kube-api-access-c7vzb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:30:03 crc kubenswrapper[4764]: I0202 09:30:03.152410 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d04cc0fd-f303-47b2-a8d4-6efd0a591b77-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "d04cc0fd-f303-47b2-a8d4-6efd0a591b77" (UID: "d04cc0fd-f303-47b2-a8d4-6efd0a591b77"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:30:03 crc kubenswrapper[4764]: I0202 09:30:03.153080 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d04cc0fd-f303-47b2-a8d4-6efd0a591b77-inventory" (OuterVolumeSpecName: "inventory") pod "d04cc0fd-f303-47b2-a8d4-6efd0a591b77" (UID: "d04cc0fd-f303-47b2-a8d4-6efd0a591b77"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:30:03 crc kubenswrapper[4764]: I0202 09:30:03.220106 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c7vzb\" (UniqueName: \"kubernetes.io/projected/d04cc0fd-f303-47b2-a8d4-6efd0a591b77-kube-api-access-c7vzb\") on node \"crc\" DevicePath \"\"" Feb 02 09:30:03 crc kubenswrapper[4764]: I0202 09:30:03.220137 4764 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d04cc0fd-f303-47b2-a8d4-6efd0a591b77-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 09:30:03 crc kubenswrapper[4764]: I0202 09:30:03.220149 4764 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d04cc0fd-f303-47b2-a8d4-6efd0a591b77-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 02 09:30:03 crc kubenswrapper[4764]: I0202 09:30:03.220159 4764 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d04cc0fd-f303-47b2-a8d4-6efd0a591b77-inventory\") on node \"crc\" DevicePath \"\"" Feb 02 09:30:03 crc kubenswrapper[4764]: I0202 09:30:03.577036 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gv65q" event={"ID":"d04cc0fd-f303-47b2-a8d4-6efd0a591b77","Type":"ContainerDied","Data":"f30fe347ddf4634495d2a806b4d7582fe0a60feb8a6f73e3e028091306ef2c2e"} Feb 02 09:30:03 crc kubenswrapper[4764]: I0202 09:30:03.577130 4764 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f30fe347ddf4634495d2a806b4d7582fe0a60feb8a6f73e3e028091306ef2c2e" Feb 02 09:30:03 crc kubenswrapper[4764]: I0202 09:30:03.577068 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gv65q" Feb 02 09:30:03 crc kubenswrapper[4764]: I0202 09:30:03.685499 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5qvdd"] Feb 02 09:30:03 crc kubenswrapper[4764]: E0202 09:30:03.686000 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d04cc0fd-f303-47b2-a8d4-6efd0a591b77" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Feb 02 09:30:03 crc kubenswrapper[4764]: I0202 09:30:03.686019 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="d04cc0fd-f303-47b2-a8d4-6efd0a591b77" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Feb 02 09:30:03 crc kubenswrapper[4764]: I0202 09:30:03.686293 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="d04cc0fd-f303-47b2-a8d4-6efd0a591b77" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Feb 02 09:30:03 crc kubenswrapper[4764]: I0202 09:30:03.687006 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5qvdd" Feb 02 09:30:03 crc kubenswrapper[4764]: I0202 09:30:03.721409 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 02 09:30:03 crc kubenswrapper[4764]: I0202 09:30:03.723365 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 02 09:30:03 crc kubenswrapper[4764]: I0202 09:30:03.723493 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-xksq5" Feb 02 09:30:03 crc kubenswrapper[4764]: I0202 09:30:03.731553 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b60a0abc-6093-4ab4-b8ac-5bd6220d1533-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-5qvdd\" (UID: \"b60a0abc-6093-4ab4-b8ac-5bd6220d1533\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5qvdd" Feb 02 09:30:03 crc kubenswrapper[4764]: I0202 09:30:03.731664 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b60a0abc-6093-4ab4-b8ac-5bd6220d1533-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-5qvdd\" (UID: \"b60a0abc-6093-4ab4-b8ac-5bd6220d1533\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5qvdd" Feb 02 09:30:03 crc kubenswrapper[4764]: I0202 09:30:03.731750 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mkxzn\" (UniqueName: \"kubernetes.io/projected/b60a0abc-6093-4ab4-b8ac-5bd6220d1533-kube-api-access-mkxzn\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-5qvdd\" (UID: \"b60a0abc-6093-4ab4-b8ac-5bd6220d1533\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5qvdd" Feb 02 09:30:03 crc kubenswrapper[4764]: I0202 09:30:03.731851 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b60a0abc-6093-4ab4-b8ac-5bd6220d1533-ssh-key-openstack-edpm-ipam\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-5qvdd\" (UID: \"b60a0abc-6093-4ab4-b8ac-5bd6220d1533\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5qvdd" Feb 02 09:30:03 crc kubenswrapper[4764]: I0202 09:30:03.737425 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 02 09:30:03 crc kubenswrapper[4764]: I0202 09:30:03.746890 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5qvdd"] Feb 02 09:30:03 crc kubenswrapper[4764]: I0202 09:30:03.834453 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b60a0abc-6093-4ab4-b8ac-5bd6220d1533-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-5qvdd\" (UID: \"b60a0abc-6093-4ab4-b8ac-5bd6220d1533\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5qvdd" Feb 02 09:30:03 crc kubenswrapper[4764]: I0202 09:30:03.834523 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b60a0abc-6093-4ab4-b8ac-5bd6220d1533-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-5qvdd\" (UID: \"b60a0abc-6093-4ab4-b8ac-5bd6220d1533\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5qvdd" Feb 02 09:30:03 crc kubenswrapper[4764]: I0202 09:30:03.834582 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mkxzn\" (UniqueName: \"kubernetes.io/projected/b60a0abc-6093-4ab4-b8ac-5bd6220d1533-kube-api-access-mkxzn\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-5qvdd\" (UID: \"b60a0abc-6093-4ab4-b8ac-5bd6220d1533\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5qvdd" Feb 02 09:30:03 crc kubenswrapper[4764]: I0202 09:30:03.834670 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b60a0abc-6093-4ab4-b8ac-5bd6220d1533-ssh-key-openstack-edpm-ipam\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-5qvdd\" (UID: \"b60a0abc-6093-4ab4-b8ac-5bd6220d1533\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5qvdd" Feb 02 09:30:03 crc kubenswrapper[4764]: I0202 09:30:03.849455 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b60a0abc-6093-4ab4-b8ac-5bd6220d1533-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-5qvdd\" (UID: \"b60a0abc-6093-4ab4-b8ac-5bd6220d1533\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5qvdd" Feb 02 09:30:03 crc kubenswrapper[4764]: I0202 09:30:03.852626 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mkxzn\" (UniqueName: \"kubernetes.io/projected/b60a0abc-6093-4ab4-b8ac-5bd6220d1533-kube-api-access-mkxzn\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-5qvdd\" (UID: \"b60a0abc-6093-4ab4-b8ac-5bd6220d1533\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5qvdd" Feb 02 09:30:03 crc kubenswrapper[4764]: I0202 09:30:03.852728 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b60a0abc-6093-4ab4-b8ac-5bd6220d1533-ssh-key-openstack-edpm-ipam\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-5qvdd\" (UID: \"b60a0abc-6093-4ab4-b8ac-5bd6220d1533\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5qvdd" Feb 02 09:30:03 crc kubenswrapper[4764]: I0202 09:30:03.852794 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b60a0abc-6093-4ab4-b8ac-5bd6220d1533-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-5qvdd\" (UID: \"b60a0abc-6093-4ab4-b8ac-5bd6220d1533\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5qvdd" Feb 02 09:30:03 crc kubenswrapper[4764]: I0202 09:30:03.896993 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29500410-b8znn" Feb 02 09:30:03 crc kubenswrapper[4764]: I0202 09:30:03.936851 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j8xmz\" (UniqueName: \"kubernetes.io/projected/fd431fc8-7193-4ad6-b7f6-eb83b4dbe824-kube-api-access-j8xmz\") pod \"fd431fc8-7193-4ad6-b7f6-eb83b4dbe824\" (UID: \"fd431fc8-7193-4ad6-b7f6-eb83b4dbe824\") " Feb 02 09:30:03 crc kubenswrapper[4764]: I0202 09:30:03.937023 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/fd431fc8-7193-4ad6-b7f6-eb83b4dbe824-secret-volume\") pod \"fd431fc8-7193-4ad6-b7f6-eb83b4dbe824\" (UID: \"fd431fc8-7193-4ad6-b7f6-eb83b4dbe824\") " Feb 02 09:30:03 crc kubenswrapper[4764]: I0202 09:30:03.937089 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fd431fc8-7193-4ad6-b7f6-eb83b4dbe824-config-volume\") pod \"fd431fc8-7193-4ad6-b7f6-eb83b4dbe824\" (UID: \"fd431fc8-7193-4ad6-b7f6-eb83b4dbe824\") " Feb 02 09:30:03 crc kubenswrapper[4764]: I0202 09:30:03.937997 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fd431fc8-7193-4ad6-b7f6-eb83b4dbe824-config-volume" (OuterVolumeSpecName: "config-volume") pod "fd431fc8-7193-4ad6-b7f6-eb83b4dbe824" (UID: "fd431fc8-7193-4ad6-b7f6-eb83b4dbe824"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:30:03 crc kubenswrapper[4764]: I0202 09:30:03.940169 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd431fc8-7193-4ad6-b7f6-eb83b4dbe824-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "fd431fc8-7193-4ad6-b7f6-eb83b4dbe824" (UID: "fd431fc8-7193-4ad6-b7f6-eb83b4dbe824"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:30:03 crc kubenswrapper[4764]: I0202 09:30:03.942488 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fd431fc8-7193-4ad6-b7f6-eb83b4dbe824-kube-api-access-j8xmz" (OuterVolumeSpecName: "kube-api-access-j8xmz") pod "fd431fc8-7193-4ad6-b7f6-eb83b4dbe824" (UID: "fd431fc8-7193-4ad6-b7f6-eb83b4dbe824"). InnerVolumeSpecName "kube-api-access-j8xmz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:30:04 crc kubenswrapper[4764]: I0202 09:30:04.039829 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5qvdd" Feb 02 09:30:04 crc kubenswrapper[4764]: I0202 09:30:04.040005 4764 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/fd431fc8-7193-4ad6-b7f6-eb83b4dbe824-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 02 09:30:04 crc kubenswrapper[4764]: I0202 09:30:04.040507 4764 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fd431fc8-7193-4ad6-b7f6-eb83b4dbe824-config-volume\") on node \"crc\" DevicePath \"\"" Feb 02 09:30:04 crc kubenswrapper[4764]: I0202 09:30:04.040532 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j8xmz\" (UniqueName: \"kubernetes.io/projected/fd431fc8-7193-4ad6-b7f6-eb83b4dbe824-kube-api-access-j8xmz\") on node \"crc\" DevicePath \"\"" Feb 02 09:30:04 crc kubenswrapper[4764]: I0202 09:30:04.595462 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29500410-b8znn" event={"ID":"fd431fc8-7193-4ad6-b7f6-eb83b4dbe824","Type":"ContainerDied","Data":"74443e3a86ed26b871c478b98a5d2a1d608321b3b629861c30cec9dc662a7f10"} Feb 02 09:30:04 crc kubenswrapper[4764]: I0202 09:30:04.595853 4764 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="74443e3a86ed26b871c478b98a5d2a1d608321b3b629861c30cec9dc662a7f10" Feb 02 09:30:04 crc kubenswrapper[4764]: I0202 09:30:04.595544 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29500410-b8znn" Feb 02 09:30:04 crc kubenswrapper[4764]: I0202 09:30:04.601114 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5qvdd"] Feb 02 09:30:05 crc kubenswrapper[4764]: I0202 09:30:05.603134 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5qvdd" event={"ID":"b60a0abc-6093-4ab4-b8ac-5bd6220d1533","Type":"ContainerStarted","Data":"72a3eb32dbfc5f69f955c69ab48f7b3367908aed17f2d9f5d6311123e797ccbc"} Feb 02 09:30:05 crc kubenswrapper[4764]: I0202 09:30:05.604089 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5qvdd" event={"ID":"b60a0abc-6093-4ab4-b8ac-5bd6220d1533","Type":"ContainerStarted","Data":"f048ceb463eb12026dfd2d93dc0c94c5e952ac06aebd1b65b3ddbd787bcfe0e4"} Feb 02 09:30:05 crc kubenswrapper[4764]: I0202 09:30:05.627046 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5qvdd" podStartSLOduration=2.208929237 podStartE2EDuration="2.627032225s" podCreationTimestamp="2026-02-02 09:30:03 +0000 UTC" firstStartedPulling="2026-02-02 09:30:04.612233883 +0000 UTC m=+1367.545957981" lastFinishedPulling="2026-02-02 09:30:05.030336841 +0000 UTC m=+1367.964060969" observedRunningTime="2026-02-02 09:30:05.619800228 +0000 UTC m=+1368.553524316" watchObservedRunningTime="2026-02-02 09:30:05.627032225 +0000 UTC m=+1368.560756313" Feb 02 09:30:24 crc kubenswrapper[4764]: I0202 09:30:24.296199 4764 scope.go:117] "RemoveContainer" containerID="588e10f44de4046cd270174efdbfc218e7cd319ddead99865bd4a949dab2c722" Feb 02 09:30:49 crc kubenswrapper[4764]: I0202 09:30:49.448178 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-5b5rd"] Feb 02 09:30:49 crc kubenswrapper[4764]: E0202 09:30:49.450635 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd431fc8-7193-4ad6-b7f6-eb83b4dbe824" containerName="collect-profiles" Feb 02 09:30:49 crc kubenswrapper[4764]: I0202 09:30:49.450666 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd431fc8-7193-4ad6-b7f6-eb83b4dbe824" containerName="collect-profiles" Feb 02 09:30:49 crc kubenswrapper[4764]: I0202 09:30:49.450899 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd431fc8-7193-4ad6-b7f6-eb83b4dbe824" containerName="collect-profiles" Feb 02 09:30:49 crc kubenswrapper[4764]: I0202 09:30:49.452467 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5b5rd" Feb 02 09:30:49 crc kubenswrapper[4764]: I0202 09:30:49.463528 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5b5rd"] Feb 02 09:30:49 crc kubenswrapper[4764]: I0202 09:30:49.512246 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/082f313d-80e1-4d9c-9ddf-8487c56a1554-utilities\") pod \"redhat-operators-5b5rd\" (UID: \"082f313d-80e1-4d9c-9ddf-8487c56a1554\") " pod="openshift-marketplace/redhat-operators-5b5rd" Feb 02 09:30:49 crc kubenswrapper[4764]: I0202 09:30:49.512407 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/082f313d-80e1-4d9c-9ddf-8487c56a1554-catalog-content\") pod \"redhat-operators-5b5rd\" (UID: \"082f313d-80e1-4d9c-9ddf-8487c56a1554\") " pod="openshift-marketplace/redhat-operators-5b5rd" Feb 02 09:30:49 crc kubenswrapper[4764]: I0202 09:30:49.512532 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7nxpc\" (UniqueName: \"kubernetes.io/projected/082f313d-80e1-4d9c-9ddf-8487c56a1554-kube-api-access-7nxpc\") pod \"redhat-operators-5b5rd\" (UID: \"082f313d-80e1-4d9c-9ddf-8487c56a1554\") " pod="openshift-marketplace/redhat-operators-5b5rd" Feb 02 09:30:49 crc kubenswrapper[4764]: I0202 09:30:49.615379 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/082f313d-80e1-4d9c-9ddf-8487c56a1554-catalog-content\") pod \"redhat-operators-5b5rd\" (UID: \"082f313d-80e1-4d9c-9ddf-8487c56a1554\") " pod="openshift-marketplace/redhat-operators-5b5rd" Feb 02 09:30:49 crc kubenswrapper[4764]: I0202 09:30:49.615506 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7nxpc\" (UniqueName: \"kubernetes.io/projected/082f313d-80e1-4d9c-9ddf-8487c56a1554-kube-api-access-7nxpc\") pod \"redhat-operators-5b5rd\" (UID: \"082f313d-80e1-4d9c-9ddf-8487c56a1554\") " pod="openshift-marketplace/redhat-operators-5b5rd" Feb 02 09:30:49 crc kubenswrapper[4764]: I0202 09:30:49.615667 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/082f313d-80e1-4d9c-9ddf-8487c56a1554-utilities\") pod \"redhat-operators-5b5rd\" (UID: \"082f313d-80e1-4d9c-9ddf-8487c56a1554\") " pod="openshift-marketplace/redhat-operators-5b5rd" Feb 02 09:30:49 crc kubenswrapper[4764]: I0202 09:30:49.615864 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/082f313d-80e1-4d9c-9ddf-8487c56a1554-catalog-content\") pod \"redhat-operators-5b5rd\" (UID: \"082f313d-80e1-4d9c-9ddf-8487c56a1554\") " pod="openshift-marketplace/redhat-operators-5b5rd" Feb 02 09:30:49 crc kubenswrapper[4764]: I0202 09:30:49.616176 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/082f313d-80e1-4d9c-9ddf-8487c56a1554-utilities\") pod \"redhat-operators-5b5rd\" (UID: \"082f313d-80e1-4d9c-9ddf-8487c56a1554\") " pod="openshift-marketplace/redhat-operators-5b5rd" Feb 02 09:30:49 crc kubenswrapper[4764]: I0202 09:30:49.632832 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7nxpc\" (UniqueName: \"kubernetes.io/projected/082f313d-80e1-4d9c-9ddf-8487c56a1554-kube-api-access-7nxpc\") pod \"redhat-operators-5b5rd\" (UID: \"082f313d-80e1-4d9c-9ddf-8487c56a1554\") " pod="openshift-marketplace/redhat-operators-5b5rd" Feb 02 09:30:49 crc kubenswrapper[4764]: I0202 09:30:49.784790 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5b5rd" Feb 02 09:30:50 crc kubenswrapper[4764]: I0202 09:30:50.317842 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5b5rd"] Feb 02 09:30:51 crc kubenswrapper[4764]: I0202 09:30:51.194335 4764 generic.go:334] "Generic (PLEG): container finished" podID="082f313d-80e1-4d9c-9ddf-8487c56a1554" containerID="ab10a13b1b55fd630b385b1748e7870bb5dd16732bd4d494bc9135542cb7e9f0" exitCode=0 Feb 02 09:30:51 crc kubenswrapper[4764]: I0202 09:30:51.194620 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5b5rd" event={"ID":"082f313d-80e1-4d9c-9ddf-8487c56a1554","Type":"ContainerDied","Data":"ab10a13b1b55fd630b385b1748e7870bb5dd16732bd4d494bc9135542cb7e9f0"} Feb 02 09:30:51 crc kubenswrapper[4764]: I0202 09:30:51.195129 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5b5rd" event={"ID":"082f313d-80e1-4d9c-9ddf-8487c56a1554","Type":"ContainerStarted","Data":"4b8333f7949fc3512a22e8126138ea2dcf71b7e98d58772d5db77c0f10de28ca"} Feb 02 09:30:52 crc kubenswrapper[4764]: I0202 09:30:52.206260 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5b5rd" event={"ID":"082f313d-80e1-4d9c-9ddf-8487c56a1554","Type":"ContainerStarted","Data":"d3273b22995e5d8a66aef76d5c0da20dae3f2f654a991e8f45cd02e398196cbb"} Feb 02 09:30:57 crc kubenswrapper[4764]: I0202 09:30:57.276379 4764 generic.go:334] "Generic (PLEG): container finished" podID="082f313d-80e1-4d9c-9ddf-8487c56a1554" containerID="d3273b22995e5d8a66aef76d5c0da20dae3f2f654a991e8f45cd02e398196cbb" exitCode=0 Feb 02 09:30:57 crc kubenswrapper[4764]: I0202 09:30:57.276481 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5b5rd" event={"ID":"082f313d-80e1-4d9c-9ddf-8487c56a1554","Type":"ContainerDied","Data":"d3273b22995e5d8a66aef76d5c0da20dae3f2f654a991e8f45cd02e398196cbb"} Feb 02 09:30:58 crc kubenswrapper[4764]: I0202 09:30:58.291402 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5b5rd" event={"ID":"082f313d-80e1-4d9c-9ddf-8487c56a1554","Type":"ContainerStarted","Data":"86402475dbe6c0e52f0174e79d373b5f079db7734554bf96c573c69df4656441"} Feb 02 09:30:58 crc kubenswrapper[4764]: I0202 09:30:58.326127 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-5b5rd" podStartSLOduration=2.784903 podStartE2EDuration="9.326104435s" podCreationTimestamp="2026-02-02 09:30:49 +0000 UTC" firstStartedPulling="2026-02-02 09:30:51.199527644 +0000 UTC m=+1414.133251722" lastFinishedPulling="2026-02-02 09:30:57.740729059 +0000 UTC m=+1420.674453157" observedRunningTime="2026-02-02 09:30:58.312236268 +0000 UTC m=+1421.245960356" watchObservedRunningTime="2026-02-02 09:30:58.326104435 +0000 UTC m=+1421.259828533" Feb 02 09:30:59 crc kubenswrapper[4764]: I0202 09:30:59.785614 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-5b5rd" Feb 02 09:30:59 crc kubenswrapper[4764]: I0202 09:30:59.786228 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-5b5rd" Feb 02 09:31:00 crc kubenswrapper[4764]: I0202 09:31:00.842885 4764 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-5b5rd" podUID="082f313d-80e1-4d9c-9ddf-8487c56a1554" containerName="registry-server" probeResult="failure" output=< Feb 02 09:31:00 crc kubenswrapper[4764]: timeout: failed to connect service ":50051" within 1s Feb 02 09:31:00 crc kubenswrapper[4764]: > Feb 02 09:31:10 crc kubenswrapper[4764]: I0202 09:31:10.868619 4764 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-5b5rd" podUID="082f313d-80e1-4d9c-9ddf-8487c56a1554" containerName="registry-server" probeResult="failure" output=< Feb 02 09:31:10 crc kubenswrapper[4764]: timeout: failed to connect service ":50051" within 1s Feb 02 09:31:10 crc kubenswrapper[4764]: > Feb 02 09:31:19 crc kubenswrapper[4764]: I0202 09:31:19.877827 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-5b5rd" Feb 02 09:31:19 crc kubenswrapper[4764]: I0202 09:31:19.954245 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-5b5rd" Feb 02 09:31:20 crc kubenswrapper[4764]: I0202 09:31:20.655431 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5b5rd"] Feb 02 09:31:21 crc kubenswrapper[4764]: I0202 09:31:21.644017 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-5b5rd" podUID="082f313d-80e1-4d9c-9ddf-8487c56a1554" containerName="registry-server" containerID="cri-o://86402475dbe6c0e52f0174e79d373b5f079db7734554bf96c573c69df4656441" gracePeriod=2 Feb 02 09:31:22 crc kubenswrapper[4764]: I0202 09:31:22.132731 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5b5rd" Feb 02 09:31:22 crc kubenswrapper[4764]: I0202 09:31:22.211578 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7nxpc\" (UniqueName: \"kubernetes.io/projected/082f313d-80e1-4d9c-9ddf-8487c56a1554-kube-api-access-7nxpc\") pod \"082f313d-80e1-4d9c-9ddf-8487c56a1554\" (UID: \"082f313d-80e1-4d9c-9ddf-8487c56a1554\") " Feb 02 09:31:22 crc kubenswrapper[4764]: I0202 09:31:22.211653 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/082f313d-80e1-4d9c-9ddf-8487c56a1554-catalog-content\") pod \"082f313d-80e1-4d9c-9ddf-8487c56a1554\" (UID: \"082f313d-80e1-4d9c-9ddf-8487c56a1554\") " Feb 02 09:31:22 crc kubenswrapper[4764]: I0202 09:31:22.211736 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/082f313d-80e1-4d9c-9ddf-8487c56a1554-utilities\") pod \"082f313d-80e1-4d9c-9ddf-8487c56a1554\" (UID: \"082f313d-80e1-4d9c-9ddf-8487c56a1554\") " Feb 02 09:31:22 crc kubenswrapper[4764]: I0202 09:31:22.212761 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/082f313d-80e1-4d9c-9ddf-8487c56a1554-utilities" (OuterVolumeSpecName: "utilities") pod "082f313d-80e1-4d9c-9ddf-8487c56a1554" (UID: "082f313d-80e1-4d9c-9ddf-8487c56a1554"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 09:31:22 crc kubenswrapper[4764]: I0202 09:31:22.219290 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/082f313d-80e1-4d9c-9ddf-8487c56a1554-kube-api-access-7nxpc" (OuterVolumeSpecName: "kube-api-access-7nxpc") pod "082f313d-80e1-4d9c-9ddf-8487c56a1554" (UID: "082f313d-80e1-4d9c-9ddf-8487c56a1554"). InnerVolumeSpecName "kube-api-access-7nxpc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:31:22 crc kubenswrapper[4764]: I0202 09:31:22.313665 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7nxpc\" (UniqueName: \"kubernetes.io/projected/082f313d-80e1-4d9c-9ddf-8487c56a1554-kube-api-access-7nxpc\") on node \"crc\" DevicePath \"\"" Feb 02 09:31:22 crc kubenswrapper[4764]: I0202 09:31:22.313695 4764 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/082f313d-80e1-4d9c-9ddf-8487c56a1554-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 09:31:22 crc kubenswrapper[4764]: I0202 09:31:22.353149 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/082f313d-80e1-4d9c-9ddf-8487c56a1554-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "082f313d-80e1-4d9c-9ddf-8487c56a1554" (UID: "082f313d-80e1-4d9c-9ddf-8487c56a1554"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 09:31:22 crc kubenswrapper[4764]: I0202 09:31:22.418305 4764 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/082f313d-80e1-4d9c-9ddf-8487c56a1554-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 09:31:22 crc kubenswrapper[4764]: I0202 09:31:22.657846 4764 generic.go:334] "Generic (PLEG): container finished" podID="082f313d-80e1-4d9c-9ddf-8487c56a1554" containerID="86402475dbe6c0e52f0174e79d373b5f079db7734554bf96c573c69df4656441" exitCode=0 Feb 02 09:31:22 crc kubenswrapper[4764]: I0202 09:31:22.657967 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5b5rd" Feb 02 09:31:22 crc kubenswrapper[4764]: I0202 09:31:22.657990 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5b5rd" event={"ID":"082f313d-80e1-4d9c-9ddf-8487c56a1554","Type":"ContainerDied","Data":"86402475dbe6c0e52f0174e79d373b5f079db7734554bf96c573c69df4656441"} Feb 02 09:31:22 crc kubenswrapper[4764]: I0202 09:31:22.658576 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5b5rd" event={"ID":"082f313d-80e1-4d9c-9ddf-8487c56a1554","Type":"ContainerDied","Data":"4b8333f7949fc3512a22e8126138ea2dcf71b7e98d58772d5db77c0f10de28ca"} Feb 02 09:31:22 crc kubenswrapper[4764]: I0202 09:31:22.658607 4764 scope.go:117] "RemoveContainer" containerID="86402475dbe6c0e52f0174e79d373b5f079db7734554bf96c573c69df4656441" Feb 02 09:31:22 crc kubenswrapper[4764]: I0202 09:31:22.686115 4764 scope.go:117] "RemoveContainer" containerID="d3273b22995e5d8a66aef76d5c0da20dae3f2f654a991e8f45cd02e398196cbb" Feb 02 09:31:22 crc kubenswrapper[4764]: I0202 09:31:22.730435 4764 scope.go:117] "RemoveContainer" containerID="ab10a13b1b55fd630b385b1748e7870bb5dd16732bd4d494bc9135542cb7e9f0" Feb 02 09:31:22 crc kubenswrapper[4764]: I0202 09:31:22.731693 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5b5rd"] Feb 02 09:31:22 crc kubenswrapper[4764]: I0202 09:31:22.749254 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-5b5rd"] Feb 02 09:31:22 crc kubenswrapper[4764]: I0202 09:31:22.785002 4764 scope.go:117] "RemoveContainer" containerID="86402475dbe6c0e52f0174e79d373b5f079db7734554bf96c573c69df4656441" Feb 02 09:31:22 crc kubenswrapper[4764]: E0202 09:31:22.785488 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"86402475dbe6c0e52f0174e79d373b5f079db7734554bf96c573c69df4656441\": container with ID starting with 86402475dbe6c0e52f0174e79d373b5f079db7734554bf96c573c69df4656441 not found: ID does not exist" containerID="86402475dbe6c0e52f0174e79d373b5f079db7734554bf96c573c69df4656441" Feb 02 09:31:22 crc kubenswrapper[4764]: I0202 09:31:22.785520 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"86402475dbe6c0e52f0174e79d373b5f079db7734554bf96c573c69df4656441"} err="failed to get container status \"86402475dbe6c0e52f0174e79d373b5f079db7734554bf96c573c69df4656441\": rpc error: code = NotFound desc = could not find container \"86402475dbe6c0e52f0174e79d373b5f079db7734554bf96c573c69df4656441\": container with ID starting with 86402475dbe6c0e52f0174e79d373b5f079db7734554bf96c573c69df4656441 not found: ID does not exist" Feb 02 09:31:22 crc kubenswrapper[4764]: I0202 09:31:22.785544 4764 scope.go:117] "RemoveContainer" containerID="d3273b22995e5d8a66aef76d5c0da20dae3f2f654a991e8f45cd02e398196cbb" Feb 02 09:31:22 crc kubenswrapper[4764]: E0202 09:31:22.786063 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d3273b22995e5d8a66aef76d5c0da20dae3f2f654a991e8f45cd02e398196cbb\": container with ID starting with d3273b22995e5d8a66aef76d5c0da20dae3f2f654a991e8f45cd02e398196cbb not found: ID does not exist" containerID="d3273b22995e5d8a66aef76d5c0da20dae3f2f654a991e8f45cd02e398196cbb" Feb 02 09:31:22 crc kubenswrapper[4764]: I0202 09:31:22.786095 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d3273b22995e5d8a66aef76d5c0da20dae3f2f654a991e8f45cd02e398196cbb"} err="failed to get container status \"d3273b22995e5d8a66aef76d5c0da20dae3f2f654a991e8f45cd02e398196cbb\": rpc error: code = NotFound desc = could not find container \"d3273b22995e5d8a66aef76d5c0da20dae3f2f654a991e8f45cd02e398196cbb\": container with ID starting with d3273b22995e5d8a66aef76d5c0da20dae3f2f654a991e8f45cd02e398196cbb not found: ID does not exist" Feb 02 09:31:22 crc kubenswrapper[4764]: I0202 09:31:22.786110 4764 scope.go:117] "RemoveContainer" containerID="ab10a13b1b55fd630b385b1748e7870bb5dd16732bd4d494bc9135542cb7e9f0" Feb 02 09:31:22 crc kubenswrapper[4764]: E0202 09:31:22.786381 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ab10a13b1b55fd630b385b1748e7870bb5dd16732bd4d494bc9135542cb7e9f0\": container with ID starting with ab10a13b1b55fd630b385b1748e7870bb5dd16732bd4d494bc9135542cb7e9f0 not found: ID does not exist" containerID="ab10a13b1b55fd630b385b1748e7870bb5dd16732bd4d494bc9135542cb7e9f0" Feb 02 09:31:22 crc kubenswrapper[4764]: I0202 09:31:22.786401 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ab10a13b1b55fd630b385b1748e7870bb5dd16732bd4d494bc9135542cb7e9f0"} err="failed to get container status \"ab10a13b1b55fd630b385b1748e7870bb5dd16732bd4d494bc9135542cb7e9f0\": rpc error: code = NotFound desc = could not find container \"ab10a13b1b55fd630b385b1748e7870bb5dd16732bd4d494bc9135542cb7e9f0\": container with ID starting with ab10a13b1b55fd630b385b1748e7870bb5dd16732bd4d494bc9135542cb7e9f0 not found: ID does not exist" Feb 02 09:31:23 crc kubenswrapper[4764]: I0202 09:31:23.844969 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="082f313d-80e1-4d9c-9ddf-8487c56a1554" path="/var/lib/kubelet/pods/082f313d-80e1-4d9c-9ddf-8487c56a1554/volumes" Feb 02 09:31:24 crc kubenswrapper[4764]: I0202 09:31:24.401498 4764 scope.go:117] "RemoveContainer" containerID="ceb8422566de0c1e2db3ceb17cfb3eb09bd59d5db0c9758695b7c739686570c0" Feb 02 09:31:24 crc kubenswrapper[4764]: I0202 09:31:24.439548 4764 scope.go:117] "RemoveContainer" containerID="074b62c56ecca7f047050603bbafc675a69dad1f862acb1313c25578ceb1bc02" Feb 02 09:32:13 crc kubenswrapper[4764]: I0202 09:32:13.522545 4764 patch_prober.go:28] interesting pod/machine-config-daemon-4ndm4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 09:32:13 crc kubenswrapper[4764]: I0202 09:32:13.523097 4764 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 09:32:24 crc kubenswrapper[4764]: I0202 09:32:24.607507 4764 scope.go:117] "RemoveContainer" containerID="294d44cc286817e23d8f3b1d9417fc50ada00fd33354cb6a949b1259efb86ba9" Feb 02 09:32:24 crc kubenswrapper[4764]: I0202 09:32:24.667329 4764 scope.go:117] "RemoveContainer" containerID="c301ea8b2f199dad2ba71f47cdbeb2bc337210f442107f205bf2728ba0595afd" Feb 02 09:32:24 crc kubenswrapper[4764]: I0202 09:32:24.713365 4764 scope.go:117] "RemoveContainer" containerID="46ac0362dd33241a49276203ed445721439244f4f312c3c5389ac4126d87b43f" Feb 02 09:32:24 crc kubenswrapper[4764]: I0202 09:32:24.738618 4764 scope.go:117] "RemoveContainer" containerID="c6c4d99dacaf78f359aaf41aedad12c157da6d0de970ce57f1147991e1f002d5" Feb 02 09:32:24 crc kubenswrapper[4764]: I0202 09:32:24.771355 4764 scope.go:117] "RemoveContainer" containerID="51619052573cea6bacf4e847a8df670a9036620cd4def95428f044400cefce9a" Feb 02 09:32:24 crc kubenswrapper[4764]: I0202 09:32:24.791921 4764 scope.go:117] "RemoveContainer" containerID="502205bc8cf64e06eabe340873459163683f16c8530ea72161ca0adfc975501a" Feb 02 09:32:43 crc kubenswrapper[4764]: I0202 09:32:43.522598 4764 patch_prober.go:28] interesting pod/machine-config-daemon-4ndm4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 09:32:43 crc kubenswrapper[4764]: I0202 09:32:43.523194 4764 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 09:33:02 crc kubenswrapper[4764]: I0202 09:33:02.853139 4764 generic.go:334] "Generic (PLEG): container finished" podID="b60a0abc-6093-4ab4-b8ac-5bd6220d1533" containerID="72a3eb32dbfc5f69f955c69ab48f7b3367908aed17f2d9f5d6311123e797ccbc" exitCode=0 Feb 02 09:33:02 crc kubenswrapper[4764]: I0202 09:33:02.853240 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5qvdd" event={"ID":"b60a0abc-6093-4ab4-b8ac-5bd6220d1533","Type":"ContainerDied","Data":"72a3eb32dbfc5f69f955c69ab48f7b3367908aed17f2d9f5d6311123e797ccbc"} Feb 02 09:33:04 crc kubenswrapper[4764]: I0202 09:33:04.373518 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5qvdd" Feb 02 09:33:04 crc kubenswrapper[4764]: I0202 09:33:04.480317 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b60a0abc-6093-4ab4-b8ac-5bd6220d1533-bootstrap-combined-ca-bundle\") pod \"b60a0abc-6093-4ab4-b8ac-5bd6220d1533\" (UID: \"b60a0abc-6093-4ab4-b8ac-5bd6220d1533\") " Feb 02 09:33:04 crc kubenswrapper[4764]: I0202 09:33:04.480401 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b60a0abc-6093-4ab4-b8ac-5bd6220d1533-ssh-key-openstack-edpm-ipam\") pod \"b60a0abc-6093-4ab4-b8ac-5bd6220d1533\" (UID: \"b60a0abc-6093-4ab4-b8ac-5bd6220d1533\") " Feb 02 09:33:04 crc kubenswrapper[4764]: I0202 09:33:04.480527 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mkxzn\" (UniqueName: \"kubernetes.io/projected/b60a0abc-6093-4ab4-b8ac-5bd6220d1533-kube-api-access-mkxzn\") pod \"b60a0abc-6093-4ab4-b8ac-5bd6220d1533\" (UID: \"b60a0abc-6093-4ab4-b8ac-5bd6220d1533\") " Feb 02 09:33:04 crc kubenswrapper[4764]: I0202 09:33:04.480581 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b60a0abc-6093-4ab4-b8ac-5bd6220d1533-inventory\") pod \"b60a0abc-6093-4ab4-b8ac-5bd6220d1533\" (UID: \"b60a0abc-6093-4ab4-b8ac-5bd6220d1533\") " Feb 02 09:33:04 crc kubenswrapper[4764]: I0202 09:33:04.487542 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b60a0abc-6093-4ab4-b8ac-5bd6220d1533-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "b60a0abc-6093-4ab4-b8ac-5bd6220d1533" (UID: "b60a0abc-6093-4ab4-b8ac-5bd6220d1533"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:33:04 crc kubenswrapper[4764]: I0202 09:33:04.488217 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b60a0abc-6093-4ab4-b8ac-5bd6220d1533-kube-api-access-mkxzn" (OuterVolumeSpecName: "kube-api-access-mkxzn") pod "b60a0abc-6093-4ab4-b8ac-5bd6220d1533" (UID: "b60a0abc-6093-4ab4-b8ac-5bd6220d1533"). InnerVolumeSpecName "kube-api-access-mkxzn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:33:04 crc kubenswrapper[4764]: I0202 09:33:04.512924 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b60a0abc-6093-4ab4-b8ac-5bd6220d1533-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "b60a0abc-6093-4ab4-b8ac-5bd6220d1533" (UID: "b60a0abc-6093-4ab4-b8ac-5bd6220d1533"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:33:04 crc kubenswrapper[4764]: I0202 09:33:04.521022 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b60a0abc-6093-4ab4-b8ac-5bd6220d1533-inventory" (OuterVolumeSpecName: "inventory") pod "b60a0abc-6093-4ab4-b8ac-5bd6220d1533" (UID: "b60a0abc-6093-4ab4-b8ac-5bd6220d1533"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:33:04 crc kubenswrapper[4764]: I0202 09:33:04.583270 4764 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b60a0abc-6093-4ab4-b8ac-5bd6220d1533-inventory\") on node \"crc\" DevicePath \"\"" Feb 02 09:33:04 crc kubenswrapper[4764]: I0202 09:33:04.583304 4764 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b60a0abc-6093-4ab4-b8ac-5bd6220d1533-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 09:33:04 crc kubenswrapper[4764]: I0202 09:33:04.583315 4764 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b60a0abc-6093-4ab4-b8ac-5bd6220d1533-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 02 09:33:04 crc kubenswrapper[4764]: I0202 09:33:04.583324 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mkxzn\" (UniqueName: \"kubernetes.io/projected/b60a0abc-6093-4ab4-b8ac-5bd6220d1533-kube-api-access-mkxzn\") on node \"crc\" DevicePath \"\"" Feb 02 09:33:04 crc kubenswrapper[4764]: I0202 09:33:04.878505 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5qvdd" event={"ID":"b60a0abc-6093-4ab4-b8ac-5bd6220d1533","Type":"ContainerDied","Data":"f048ceb463eb12026dfd2d93dc0c94c5e952ac06aebd1b65b3ddbd787bcfe0e4"} Feb 02 09:33:04 crc kubenswrapper[4764]: I0202 09:33:04.878553 4764 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f048ceb463eb12026dfd2d93dc0c94c5e952ac06aebd1b65b3ddbd787bcfe0e4" Feb 02 09:33:04 crc kubenswrapper[4764]: I0202 09:33:04.878611 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5qvdd" Feb 02 09:33:04 crc kubenswrapper[4764]: I0202 09:33:04.985138 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-nx4kb"] Feb 02 09:33:04 crc kubenswrapper[4764]: E0202 09:33:04.985498 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="082f313d-80e1-4d9c-9ddf-8487c56a1554" containerName="extract-content" Feb 02 09:33:04 crc kubenswrapper[4764]: I0202 09:33:04.985515 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="082f313d-80e1-4d9c-9ddf-8487c56a1554" containerName="extract-content" Feb 02 09:33:04 crc kubenswrapper[4764]: E0202 09:33:04.985528 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="082f313d-80e1-4d9c-9ddf-8487c56a1554" containerName="extract-utilities" Feb 02 09:33:04 crc kubenswrapper[4764]: I0202 09:33:04.985534 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="082f313d-80e1-4d9c-9ddf-8487c56a1554" containerName="extract-utilities" Feb 02 09:33:04 crc kubenswrapper[4764]: E0202 09:33:04.985558 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="082f313d-80e1-4d9c-9ddf-8487c56a1554" containerName="registry-server" Feb 02 09:33:04 crc kubenswrapper[4764]: I0202 09:33:04.985564 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="082f313d-80e1-4d9c-9ddf-8487c56a1554" containerName="registry-server" Feb 02 09:33:04 crc kubenswrapper[4764]: E0202 09:33:04.985575 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b60a0abc-6093-4ab4-b8ac-5bd6220d1533" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Feb 02 09:33:04 crc kubenswrapper[4764]: I0202 09:33:04.985581 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="b60a0abc-6093-4ab4-b8ac-5bd6220d1533" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Feb 02 09:33:04 crc kubenswrapper[4764]: I0202 09:33:04.985740 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="b60a0abc-6093-4ab4-b8ac-5bd6220d1533" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Feb 02 09:33:04 crc kubenswrapper[4764]: I0202 09:33:04.985759 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="082f313d-80e1-4d9c-9ddf-8487c56a1554" containerName="registry-server" Feb 02 09:33:04 crc kubenswrapper[4764]: I0202 09:33:04.986330 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-nx4kb" Feb 02 09:33:04 crc kubenswrapper[4764]: I0202 09:33:04.988273 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 02 09:33:04 crc kubenswrapper[4764]: I0202 09:33:04.988768 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 02 09:33:04 crc kubenswrapper[4764]: I0202 09:33:04.989596 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 02 09:33:04 crc kubenswrapper[4764]: I0202 09:33:04.990065 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-xksq5" Feb 02 09:33:04 crc kubenswrapper[4764]: I0202 09:33:04.993050 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-nx4kb"] Feb 02 09:33:05 crc kubenswrapper[4764]: I0202 09:33:05.093988 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hlbft\" (UniqueName: \"kubernetes.io/projected/667bdcf6-12ac-4bfc-b652-48ce7aca1fc2-kube-api-access-hlbft\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-nx4kb\" (UID: \"667bdcf6-12ac-4bfc-b652-48ce7aca1fc2\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-nx4kb" Feb 02 09:33:05 crc kubenswrapper[4764]: I0202 09:33:05.094231 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/667bdcf6-12ac-4bfc-b652-48ce7aca1fc2-ssh-key-openstack-edpm-ipam\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-nx4kb\" (UID: \"667bdcf6-12ac-4bfc-b652-48ce7aca1fc2\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-nx4kb" Feb 02 09:33:05 crc kubenswrapper[4764]: I0202 09:33:05.094303 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/667bdcf6-12ac-4bfc-b652-48ce7aca1fc2-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-nx4kb\" (UID: \"667bdcf6-12ac-4bfc-b652-48ce7aca1fc2\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-nx4kb" Feb 02 09:33:05 crc kubenswrapper[4764]: I0202 09:33:05.196028 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hlbft\" (UniqueName: \"kubernetes.io/projected/667bdcf6-12ac-4bfc-b652-48ce7aca1fc2-kube-api-access-hlbft\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-nx4kb\" (UID: \"667bdcf6-12ac-4bfc-b652-48ce7aca1fc2\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-nx4kb" Feb 02 09:33:05 crc kubenswrapper[4764]: I0202 09:33:05.196391 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/667bdcf6-12ac-4bfc-b652-48ce7aca1fc2-ssh-key-openstack-edpm-ipam\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-nx4kb\" (UID: \"667bdcf6-12ac-4bfc-b652-48ce7aca1fc2\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-nx4kb" Feb 02 09:33:05 crc kubenswrapper[4764]: I0202 09:33:05.196451 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/667bdcf6-12ac-4bfc-b652-48ce7aca1fc2-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-nx4kb\" (UID: \"667bdcf6-12ac-4bfc-b652-48ce7aca1fc2\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-nx4kb" Feb 02 09:33:05 crc kubenswrapper[4764]: I0202 09:33:05.208714 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/667bdcf6-12ac-4bfc-b652-48ce7aca1fc2-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-nx4kb\" (UID: \"667bdcf6-12ac-4bfc-b652-48ce7aca1fc2\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-nx4kb" Feb 02 09:33:05 crc kubenswrapper[4764]: I0202 09:33:05.210464 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/667bdcf6-12ac-4bfc-b652-48ce7aca1fc2-ssh-key-openstack-edpm-ipam\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-nx4kb\" (UID: \"667bdcf6-12ac-4bfc-b652-48ce7aca1fc2\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-nx4kb" Feb 02 09:33:05 crc kubenswrapper[4764]: I0202 09:33:05.219153 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hlbft\" (UniqueName: \"kubernetes.io/projected/667bdcf6-12ac-4bfc-b652-48ce7aca1fc2-kube-api-access-hlbft\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-nx4kb\" (UID: \"667bdcf6-12ac-4bfc-b652-48ce7aca1fc2\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-nx4kb" Feb 02 09:33:05 crc kubenswrapper[4764]: I0202 09:33:05.349640 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-nx4kb" Feb 02 09:33:05 crc kubenswrapper[4764]: I0202 09:33:05.925422 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-nx4kb"] Feb 02 09:33:06 crc kubenswrapper[4764]: I0202 09:33:06.896446 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-nx4kb" event={"ID":"667bdcf6-12ac-4bfc-b652-48ce7aca1fc2","Type":"ContainerStarted","Data":"3782d7a113fa14a8d1b7a9b1c1ff7a6a278ec65b1543752bd6307e2d11046199"} Feb 02 09:33:06 crc kubenswrapper[4764]: I0202 09:33:06.896865 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-nx4kb" event={"ID":"667bdcf6-12ac-4bfc-b652-48ce7aca1fc2","Type":"ContainerStarted","Data":"952bbf482325a75cc8cc7b8af067e5f6cdb90d9142dcc57c129a1780e469aaca"} Feb 02 09:33:06 crc kubenswrapper[4764]: I0202 09:33:06.917048 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-nx4kb" podStartSLOduration=2.444133549 podStartE2EDuration="2.917019607s" podCreationTimestamp="2026-02-02 09:33:04 +0000 UTC" firstStartedPulling="2026-02-02 09:33:05.938388218 +0000 UTC m=+1548.872112306" lastFinishedPulling="2026-02-02 09:33:06.411274236 +0000 UTC m=+1549.344998364" observedRunningTime="2026-02-02 09:33:06.911474186 +0000 UTC m=+1549.845198274" watchObservedRunningTime="2026-02-02 09:33:06.917019607 +0000 UTC m=+1549.850743705" Feb 02 09:33:13 crc kubenswrapper[4764]: I0202 09:33:13.522742 4764 patch_prober.go:28] interesting pod/machine-config-daemon-4ndm4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 09:33:13 crc kubenswrapper[4764]: I0202 09:33:13.523266 4764 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 09:33:13 crc kubenswrapper[4764]: I0202 09:33:13.523309 4764 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" Feb 02 09:33:13 crc kubenswrapper[4764]: I0202 09:33:13.524996 4764 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6db43ab36406cdd4dbe61d7d8688257aba4271b8601a5ba5594675ae309c761d"} pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 02 09:33:13 crc kubenswrapper[4764]: I0202 09:33:13.525073 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" containerName="machine-config-daemon" containerID="cri-o://6db43ab36406cdd4dbe61d7d8688257aba4271b8601a5ba5594675ae309c761d" gracePeriod=600 Feb 02 09:33:13 crc kubenswrapper[4764]: E0202 09:33:13.646754 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 09:33:13 crc kubenswrapper[4764]: I0202 09:33:13.963731 4764 generic.go:334] "Generic (PLEG): container finished" podID="d192f670-9f9d-4539-9641-e4bed73acdd4" containerID="6db43ab36406cdd4dbe61d7d8688257aba4271b8601a5ba5594675ae309c761d" exitCode=0 Feb 02 09:33:13 crc kubenswrapper[4764]: I0202 09:33:13.963773 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" event={"ID":"d192f670-9f9d-4539-9641-e4bed73acdd4","Type":"ContainerDied","Data":"6db43ab36406cdd4dbe61d7d8688257aba4271b8601a5ba5594675ae309c761d"} Feb 02 09:33:13 crc kubenswrapper[4764]: I0202 09:33:13.963809 4764 scope.go:117] "RemoveContainer" containerID="7e0e1c40ccad424f2dce09dc2a68260202cf2abc147b119996f40fbb4f3a453b" Feb 02 09:33:13 crc kubenswrapper[4764]: I0202 09:33:13.964548 4764 scope.go:117] "RemoveContainer" containerID="6db43ab36406cdd4dbe61d7d8688257aba4271b8601a5ba5594675ae309c761d" Feb 02 09:33:13 crc kubenswrapper[4764]: E0202 09:33:13.964906 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 09:33:24 crc kubenswrapper[4764]: I0202 09:33:24.825725 4764 scope.go:117] "RemoveContainer" containerID="6db43ab36406cdd4dbe61d7d8688257aba4271b8601a5ba5594675ae309c761d" Feb 02 09:33:24 crc kubenswrapper[4764]: E0202 09:33:24.826621 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 09:33:30 crc kubenswrapper[4764]: I0202 09:33:30.347102 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-ks9d2"] Feb 02 09:33:30 crc kubenswrapper[4764]: I0202 09:33:30.349991 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ks9d2" Feb 02 09:33:30 crc kubenswrapper[4764]: I0202 09:33:30.359848 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-ks9d2"] Feb 02 09:33:30 crc kubenswrapper[4764]: I0202 09:33:30.511000 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-98l9f\" (UniqueName: \"kubernetes.io/projected/f297bd47-08cb-4c18-b7c0-9d9f89a5bc37-kube-api-access-98l9f\") pod \"redhat-marketplace-ks9d2\" (UID: \"f297bd47-08cb-4c18-b7c0-9d9f89a5bc37\") " pod="openshift-marketplace/redhat-marketplace-ks9d2" Feb 02 09:33:30 crc kubenswrapper[4764]: I0202 09:33:30.511116 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f297bd47-08cb-4c18-b7c0-9d9f89a5bc37-catalog-content\") pod \"redhat-marketplace-ks9d2\" (UID: \"f297bd47-08cb-4c18-b7c0-9d9f89a5bc37\") " pod="openshift-marketplace/redhat-marketplace-ks9d2" Feb 02 09:33:30 crc kubenswrapper[4764]: I0202 09:33:30.511148 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f297bd47-08cb-4c18-b7c0-9d9f89a5bc37-utilities\") pod \"redhat-marketplace-ks9d2\" (UID: \"f297bd47-08cb-4c18-b7c0-9d9f89a5bc37\") " pod="openshift-marketplace/redhat-marketplace-ks9d2" Feb 02 09:33:30 crc kubenswrapper[4764]: I0202 09:33:30.613278 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f297bd47-08cb-4c18-b7c0-9d9f89a5bc37-catalog-content\") pod \"redhat-marketplace-ks9d2\" (UID: \"f297bd47-08cb-4c18-b7c0-9d9f89a5bc37\") " pod="openshift-marketplace/redhat-marketplace-ks9d2" Feb 02 09:33:30 crc kubenswrapper[4764]: I0202 09:33:30.613634 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f297bd47-08cb-4c18-b7c0-9d9f89a5bc37-utilities\") pod \"redhat-marketplace-ks9d2\" (UID: \"f297bd47-08cb-4c18-b7c0-9d9f89a5bc37\") " pod="openshift-marketplace/redhat-marketplace-ks9d2" Feb 02 09:33:30 crc kubenswrapper[4764]: I0202 09:33:30.613844 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f297bd47-08cb-4c18-b7c0-9d9f89a5bc37-catalog-content\") pod \"redhat-marketplace-ks9d2\" (UID: \"f297bd47-08cb-4c18-b7c0-9d9f89a5bc37\") " pod="openshift-marketplace/redhat-marketplace-ks9d2" Feb 02 09:33:30 crc kubenswrapper[4764]: I0202 09:33:30.613924 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f297bd47-08cb-4c18-b7c0-9d9f89a5bc37-utilities\") pod \"redhat-marketplace-ks9d2\" (UID: \"f297bd47-08cb-4c18-b7c0-9d9f89a5bc37\") " pod="openshift-marketplace/redhat-marketplace-ks9d2" Feb 02 09:33:30 crc kubenswrapper[4764]: I0202 09:33:30.614148 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-98l9f\" (UniqueName: \"kubernetes.io/projected/f297bd47-08cb-4c18-b7c0-9d9f89a5bc37-kube-api-access-98l9f\") pod \"redhat-marketplace-ks9d2\" (UID: \"f297bd47-08cb-4c18-b7c0-9d9f89a5bc37\") " pod="openshift-marketplace/redhat-marketplace-ks9d2" Feb 02 09:33:30 crc kubenswrapper[4764]: I0202 09:33:30.640687 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-98l9f\" (UniqueName: \"kubernetes.io/projected/f297bd47-08cb-4c18-b7c0-9d9f89a5bc37-kube-api-access-98l9f\") pod \"redhat-marketplace-ks9d2\" (UID: \"f297bd47-08cb-4c18-b7c0-9d9f89a5bc37\") " pod="openshift-marketplace/redhat-marketplace-ks9d2" Feb 02 09:33:30 crc kubenswrapper[4764]: I0202 09:33:30.669828 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ks9d2" Feb 02 09:33:31 crc kubenswrapper[4764]: I0202 09:33:31.176236 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-ks9d2"] Feb 02 09:33:31 crc kubenswrapper[4764]: W0202 09:33:31.182725 4764 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf297bd47_08cb_4c18_b7c0_9d9f89a5bc37.slice/crio-6d65ae54dc4570a3eeb3a6c68f4f054723097624884f0fea8cb151e084d57357 WatchSource:0}: Error finding container 6d65ae54dc4570a3eeb3a6c68f4f054723097624884f0fea8cb151e084d57357: Status 404 returned error can't find the container with id 6d65ae54dc4570a3eeb3a6c68f4f054723097624884f0fea8cb151e084d57357 Feb 02 09:33:32 crc kubenswrapper[4764]: I0202 09:33:32.149111 4764 generic.go:334] "Generic (PLEG): container finished" podID="f297bd47-08cb-4c18-b7c0-9d9f89a5bc37" containerID="93f3d113771762e245560898eab84f02779254f39fc84e859bd587f97498005e" exitCode=0 Feb 02 09:33:32 crc kubenswrapper[4764]: I0202 09:33:32.149215 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ks9d2" event={"ID":"f297bd47-08cb-4c18-b7c0-9d9f89a5bc37","Type":"ContainerDied","Data":"93f3d113771762e245560898eab84f02779254f39fc84e859bd587f97498005e"} Feb 02 09:33:32 crc kubenswrapper[4764]: I0202 09:33:32.150172 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ks9d2" event={"ID":"f297bd47-08cb-4c18-b7c0-9d9f89a5bc37","Type":"ContainerStarted","Data":"6d65ae54dc4570a3eeb3a6c68f4f054723097624884f0fea8cb151e084d57357"} Feb 02 09:33:32 crc kubenswrapper[4764]: I0202 09:33:32.152189 4764 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 02 09:33:33 crc kubenswrapper[4764]: I0202 09:33:33.161736 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ks9d2" event={"ID":"f297bd47-08cb-4c18-b7c0-9d9f89a5bc37","Type":"ContainerStarted","Data":"d39f88b257b2050abe30248ec340ba7441c2b219f2434cb7cb7a7f64edf882a2"} Feb 02 09:33:34 crc kubenswrapper[4764]: I0202 09:33:34.171809 4764 generic.go:334] "Generic (PLEG): container finished" podID="f297bd47-08cb-4c18-b7c0-9d9f89a5bc37" containerID="d39f88b257b2050abe30248ec340ba7441c2b219f2434cb7cb7a7f64edf882a2" exitCode=0 Feb 02 09:33:34 crc kubenswrapper[4764]: I0202 09:33:34.171850 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ks9d2" event={"ID":"f297bd47-08cb-4c18-b7c0-9d9f89a5bc37","Type":"ContainerDied","Data":"d39f88b257b2050abe30248ec340ba7441c2b219f2434cb7cb7a7f64edf882a2"} Feb 02 09:33:35 crc kubenswrapper[4764]: I0202 09:33:35.185786 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ks9d2" event={"ID":"f297bd47-08cb-4c18-b7c0-9d9f89a5bc37","Type":"ContainerStarted","Data":"ff883f715e8b38411ae92428609cc783561d9d34f10c1d54d0b2de1e9b124f89"} Feb 02 09:33:35 crc kubenswrapper[4764]: I0202 09:33:35.210725 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-ks9d2" podStartSLOduration=2.794796653 podStartE2EDuration="5.210709085s" podCreationTimestamp="2026-02-02 09:33:30 +0000 UTC" firstStartedPulling="2026-02-02 09:33:32.151597181 +0000 UTC m=+1575.085321309" lastFinishedPulling="2026-02-02 09:33:34.567509643 +0000 UTC m=+1577.501233741" observedRunningTime="2026-02-02 09:33:35.205663548 +0000 UTC m=+1578.139387636" watchObservedRunningTime="2026-02-02 09:33:35.210709085 +0000 UTC m=+1578.144433173" Feb 02 09:33:35 crc kubenswrapper[4764]: I0202 09:33:35.829750 4764 scope.go:117] "RemoveContainer" containerID="6db43ab36406cdd4dbe61d7d8688257aba4271b8601a5ba5594675ae309c761d" Feb 02 09:33:35 crc kubenswrapper[4764]: E0202 09:33:35.830504 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 09:33:40 crc kubenswrapper[4764]: I0202 09:33:40.670485 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-ks9d2" Feb 02 09:33:40 crc kubenswrapper[4764]: I0202 09:33:40.671062 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-ks9d2" Feb 02 09:33:40 crc kubenswrapper[4764]: I0202 09:33:40.717361 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-ks9d2" Feb 02 09:33:41 crc kubenswrapper[4764]: I0202 09:33:41.305843 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-ks9d2" Feb 02 09:33:41 crc kubenswrapper[4764]: I0202 09:33:41.362589 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-ks9d2"] Feb 02 09:33:43 crc kubenswrapper[4764]: I0202 09:33:43.252055 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-ks9d2" podUID="f297bd47-08cb-4c18-b7c0-9d9f89a5bc37" containerName="registry-server" containerID="cri-o://ff883f715e8b38411ae92428609cc783561d9d34f10c1d54d0b2de1e9b124f89" gracePeriod=2 Feb 02 09:33:44 crc kubenswrapper[4764]: I0202 09:33:44.250715 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ks9d2" Feb 02 09:33:44 crc kubenswrapper[4764]: I0202 09:33:44.264484 4764 generic.go:334] "Generic (PLEG): container finished" podID="f297bd47-08cb-4c18-b7c0-9d9f89a5bc37" containerID="ff883f715e8b38411ae92428609cc783561d9d34f10c1d54d0b2de1e9b124f89" exitCode=0 Feb 02 09:33:44 crc kubenswrapper[4764]: I0202 09:33:44.264539 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ks9d2" event={"ID":"f297bd47-08cb-4c18-b7c0-9d9f89a5bc37","Type":"ContainerDied","Data":"ff883f715e8b38411ae92428609cc783561d9d34f10c1d54d0b2de1e9b124f89"} Feb 02 09:33:44 crc kubenswrapper[4764]: I0202 09:33:44.264572 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ks9d2" event={"ID":"f297bd47-08cb-4c18-b7c0-9d9f89a5bc37","Type":"ContainerDied","Data":"6d65ae54dc4570a3eeb3a6c68f4f054723097624884f0fea8cb151e084d57357"} Feb 02 09:33:44 crc kubenswrapper[4764]: I0202 09:33:44.264597 4764 scope.go:117] "RemoveContainer" containerID="ff883f715e8b38411ae92428609cc783561d9d34f10c1d54d0b2de1e9b124f89" Feb 02 09:33:44 crc kubenswrapper[4764]: I0202 09:33:44.264870 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ks9d2" Feb 02 09:33:44 crc kubenswrapper[4764]: I0202 09:33:44.303670 4764 scope.go:117] "RemoveContainer" containerID="d39f88b257b2050abe30248ec340ba7441c2b219f2434cb7cb7a7f64edf882a2" Feb 02 09:33:44 crc kubenswrapper[4764]: I0202 09:33:44.339075 4764 scope.go:117] "RemoveContainer" containerID="93f3d113771762e245560898eab84f02779254f39fc84e859bd587f97498005e" Feb 02 09:33:44 crc kubenswrapper[4764]: I0202 09:33:44.370228 4764 scope.go:117] "RemoveContainer" containerID="ff883f715e8b38411ae92428609cc783561d9d34f10c1d54d0b2de1e9b124f89" Feb 02 09:33:44 crc kubenswrapper[4764]: E0202 09:33:44.371022 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ff883f715e8b38411ae92428609cc783561d9d34f10c1d54d0b2de1e9b124f89\": container with ID starting with ff883f715e8b38411ae92428609cc783561d9d34f10c1d54d0b2de1e9b124f89 not found: ID does not exist" containerID="ff883f715e8b38411ae92428609cc783561d9d34f10c1d54d0b2de1e9b124f89" Feb 02 09:33:44 crc kubenswrapper[4764]: I0202 09:33:44.371055 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff883f715e8b38411ae92428609cc783561d9d34f10c1d54d0b2de1e9b124f89"} err="failed to get container status \"ff883f715e8b38411ae92428609cc783561d9d34f10c1d54d0b2de1e9b124f89\": rpc error: code = NotFound desc = could not find container \"ff883f715e8b38411ae92428609cc783561d9d34f10c1d54d0b2de1e9b124f89\": container with ID starting with ff883f715e8b38411ae92428609cc783561d9d34f10c1d54d0b2de1e9b124f89 not found: ID does not exist" Feb 02 09:33:44 crc kubenswrapper[4764]: I0202 09:33:44.371073 4764 scope.go:117] "RemoveContainer" containerID="d39f88b257b2050abe30248ec340ba7441c2b219f2434cb7cb7a7f64edf882a2" Feb 02 09:33:44 crc kubenswrapper[4764]: E0202 09:33:44.371487 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d39f88b257b2050abe30248ec340ba7441c2b219f2434cb7cb7a7f64edf882a2\": container with ID starting with d39f88b257b2050abe30248ec340ba7441c2b219f2434cb7cb7a7f64edf882a2 not found: ID does not exist" containerID="d39f88b257b2050abe30248ec340ba7441c2b219f2434cb7cb7a7f64edf882a2" Feb 02 09:33:44 crc kubenswrapper[4764]: I0202 09:33:44.371535 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d39f88b257b2050abe30248ec340ba7441c2b219f2434cb7cb7a7f64edf882a2"} err="failed to get container status \"d39f88b257b2050abe30248ec340ba7441c2b219f2434cb7cb7a7f64edf882a2\": rpc error: code = NotFound desc = could not find container \"d39f88b257b2050abe30248ec340ba7441c2b219f2434cb7cb7a7f64edf882a2\": container with ID starting with d39f88b257b2050abe30248ec340ba7441c2b219f2434cb7cb7a7f64edf882a2 not found: ID does not exist" Feb 02 09:33:44 crc kubenswrapper[4764]: I0202 09:33:44.371564 4764 scope.go:117] "RemoveContainer" containerID="93f3d113771762e245560898eab84f02779254f39fc84e859bd587f97498005e" Feb 02 09:33:44 crc kubenswrapper[4764]: E0202 09:33:44.371905 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"93f3d113771762e245560898eab84f02779254f39fc84e859bd587f97498005e\": container with ID starting with 93f3d113771762e245560898eab84f02779254f39fc84e859bd587f97498005e not found: ID does not exist" containerID="93f3d113771762e245560898eab84f02779254f39fc84e859bd587f97498005e" Feb 02 09:33:44 crc kubenswrapper[4764]: I0202 09:33:44.371958 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"93f3d113771762e245560898eab84f02779254f39fc84e859bd587f97498005e"} err="failed to get container status \"93f3d113771762e245560898eab84f02779254f39fc84e859bd587f97498005e\": rpc error: code = NotFound desc = could not find container \"93f3d113771762e245560898eab84f02779254f39fc84e859bd587f97498005e\": container with ID starting with 93f3d113771762e245560898eab84f02779254f39fc84e859bd587f97498005e not found: ID does not exist" Feb 02 09:33:44 crc kubenswrapper[4764]: I0202 09:33:44.408247 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f297bd47-08cb-4c18-b7c0-9d9f89a5bc37-utilities\") pod \"f297bd47-08cb-4c18-b7c0-9d9f89a5bc37\" (UID: \"f297bd47-08cb-4c18-b7c0-9d9f89a5bc37\") " Feb 02 09:33:44 crc kubenswrapper[4764]: I0202 09:33:44.408318 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f297bd47-08cb-4c18-b7c0-9d9f89a5bc37-catalog-content\") pod \"f297bd47-08cb-4c18-b7c0-9d9f89a5bc37\" (UID: \"f297bd47-08cb-4c18-b7c0-9d9f89a5bc37\") " Feb 02 09:33:44 crc kubenswrapper[4764]: I0202 09:33:44.408532 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-98l9f\" (UniqueName: \"kubernetes.io/projected/f297bd47-08cb-4c18-b7c0-9d9f89a5bc37-kube-api-access-98l9f\") pod \"f297bd47-08cb-4c18-b7c0-9d9f89a5bc37\" (UID: \"f297bd47-08cb-4c18-b7c0-9d9f89a5bc37\") " Feb 02 09:33:44 crc kubenswrapper[4764]: I0202 09:33:44.410889 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f297bd47-08cb-4c18-b7c0-9d9f89a5bc37-utilities" (OuterVolumeSpecName: "utilities") pod "f297bd47-08cb-4c18-b7c0-9d9f89a5bc37" (UID: "f297bd47-08cb-4c18-b7c0-9d9f89a5bc37"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 09:33:44 crc kubenswrapper[4764]: I0202 09:33:44.421730 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f297bd47-08cb-4c18-b7c0-9d9f89a5bc37-kube-api-access-98l9f" (OuterVolumeSpecName: "kube-api-access-98l9f") pod "f297bd47-08cb-4c18-b7c0-9d9f89a5bc37" (UID: "f297bd47-08cb-4c18-b7c0-9d9f89a5bc37"). InnerVolumeSpecName "kube-api-access-98l9f". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:33:44 crc kubenswrapper[4764]: I0202 09:33:44.451432 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f297bd47-08cb-4c18-b7c0-9d9f89a5bc37-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f297bd47-08cb-4c18-b7c0-9d9f89a5bc37" (UID: "f297bd47-08cb-4c18-b7c0-9d9f89a5bc37"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 09:33:44 crc kubenswrapper[4764]: I0202 09:33:44.511649 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-98l9f\" (UniqueName: \"kubernetes.io/projected/f297bd47-08cb-4c18-b7c0-9d9f89a5bc37-kube-api-access-98l9f\") on node \"crc\" DevicePath \"\"" Feb 02 09:33:44 crc kubenswrapper[4764]: I0202 09:33:44.511676 4764 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f297bd47-08cb-4c18-b7c0-9d9f89a5bc37-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 09:33:44 crc kubenswrapper[4764]: I0202 09:33:44.511686 4764 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f297bd47-08cb-4c18-b7c0-9d9f89a5bc37-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 09:33:44 crc kubenswrapper[4764]: I0202 09:33:44.612314 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-ks9d2"] Feb 02 09:33:44 crc kubenswrapper[4764]: I0202 09:33:44.623894 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-ks9d2"] Feb 02 09:33:45 crc kubenswrapper[4764]: I0202 09:33:45.844242 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f297bd47-08cb-4c18-b7c0-9d9f89a5bc37" path="/var/lib/kubelet/pods/f297bd47-08cb-4c18-b7c0-9d9f89a5bc37/volumes" Feb 02 09:33:49 crc kubenswrapper[4764]: I0202 09:33:49.825703 4764 scope.go:117] "RemoveContainer" containerID="6db43ab36406cdd4dbe61d7d8688257aba4271b8601a5ba5594675ae309c761d" Feb 02 09:33:49 crc kubenswrapper[4764]: E0202 09:33:49.826425 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 09:33:57 crc kubenswrapper[4764]: I0202 09:33:57.203073 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-h55pp"] Feb 02 09:33:57 crc kubenswrapper[4764]: E0202 09:33:57.204990 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f297bd47-08cb-4c18-b7c0-9d9f89a5bc37" containerName="extract-utilities" Feb 02 09:33:57 crc kubenswrapper[4764]: I0202 09:33:57.205075 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="f297bd47-08cb-4c18-b7c0-9d9f89a5bc37" containerName="extract-utilities" Feb 02 09:33:57 crc kubenswrapper[4764]: E0202 09:33:57.205142 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f297bd47-08cb-4c18-b7c0-9d9f89a5bc37" containerName="registry-server" Feb 02 09:33:57 crc kubenswrapper[4764]: I0202 09:33:57.205211 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="f297bd47-08cb-4c18-b7c0-9d9f89a5bc37" containerName="registry-server" Feb 02 09:33:57 crc kubenswrapper[4764]: E0202 09:33:57.205295 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f297bd47-08cb-4c18-b7c0-9d9f89a5bc37" containerName="extract-content" Feb 02 09:33:57 crc kubenswrapper[4764]: I0202 09:33:57.205366 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="f297bd47-08cb-4c18-b7c0-9d9f89a5bc37" containerName="extract-content" Feb 02 09:33:57 crc kubenswrapper[4764]: I0202 09:33:57.205653 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="f297bd47-08cb-4c18-b7c0-9d9f89a5bc37" containerName="registry-server" Feb 02 09:33:57 crc kubenswrapper[4764]: I0202 09:33:57.206954 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-h55pp" Feb 02 09:33:57 crc kubenswrapper[4764]: I0202 09:33:57.225669 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-h55pp"] Feb 02 09:33:57 crc kubenswrapper[4764]: I0202 09:33:57.340031 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b90c4f10-f497-40ab-a6c7-599395789302-utilities\") pod \"certified-operators-h55pp\" (UID: \"b90c4f10-f497-40ab-a6c7-599395789302\") " pod="openshift-marketplace/certified-operators-h55pp" Feb 02 09:33:57 crc kubenswrapper[4764]: I0202 09:33:57.340403 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b6h2d\" (UniqueName: \"kubernetes.io/projected/b90c4f10-f497-40ab-a6c7-599395789302-kube-api-access-b6h2d\") pod \"certified-operators-h55pp\" (UID: \"b90c4f10-f497-40ab-a6c7-599395789302\") " pod="openshift-marketplace/certified-operators-h55pp" Feb 02 09:33:57 crc kubenswrapper[4764]: I0202 09:33:57.340612 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b90c4f10-f497-40ab-a6c7-599395789302-catalog-content\") pod \"certified-operators-h55pp\" (UID: \"b90c4f10-f497-40ab-a6c7-599395789302\") " pod="openshift-marketplace/certified-operators-h55pp" Feb 02 09:33:57 crc kubenswrapper[4764]: I0202 09:33:57.442688 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b90c4f10-f497-40ab-a6c7-599395789302-utilities\") pod \"certified-operators-h55pp\" (UID: \"b90c4f10-f497-40ab-a6c7-599395789302\") " pod="openshift-marketplace/certified-operators-h55pp" Feb 02 09:33:57 crc kubenswrapper[4764]: I0202 09:33:57.442812 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b6h2d\" (UniqueName: \"kubernetes.io/projected/b90c4f10-f497-40ab-a6c7-599395789302-kube-api-access-b6h2d\") pod \"certified-operators-h55pp\" (UID: \"b90c4f10-f497-40ab-a6c7-599395789302\") " pod="openshift-marketplace/certified-operators-h55pp" Feb 02 09:33:57 crc kubenswrapper[4764]: I0202 09:33:57.442863 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b90c4f10-f497-40ab-a6c7-599395789302-catalog-content\") pod \"certified-operators-h55pp\" (UID: \"b90c4f10-f497-40ab-a6c7-599395789302\") " pod="openshift-marketplace/certified-operators-h55pp" Feb 02 09:33:57 crc kubenswrapper[4764]: I0202 09:33:57.443197 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b90c4f10-f497-40ab-a6c7-599395789302-utilities\") pod \"certified-operators-h55pp\" (UID: \"b90c4f10-f497-40ab-a6c7-599395789302\") " pod="openshift-marketplace/certified-operators-h55pp" Feb 02 09:33:57 crc kubenswrapper[4764]: I0202 09:33:57.443334 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b90c4f10-f497-40ab-a6c7-599395789302-catalog-content\") pod \"certified-operators-h55pp\" (UID: \"b90c4f10-f497-40ab-a6c7-599395789302\") " pod="openshift-marketplace/certified-operators-h55pp" Feb 02 09:33:57 crc kubenswrapper[4764]: I0202 09:33:57.472957 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b6h2d\" (UniqueName: \"kubernetes.io/projected/b90c4f10-f497-40ab-a6c7-599395789302-kube-api-access-b6h2d\") pod \"certified-operators-h55pp\" (UID: \"b90c4f10-f497-40ab-a6c7-599395789302\") " pod="openshift-marketplace/certified-operators-h55pp" Feb 02 09:33:57 crc kubenswrapper[4764]: I0202 09:33:57.526198 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-h55pp" Feb 02 09:33:58 crc kubenswrapper[4764]: I0202 09:33:58.039261 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-h55pp"] Feb 02 09:33:58 crc kubenswrapper[4764]: I0202 09:33:58.405187 4764 generic.go:334] "Generic (PLEG): container finished" podID="b90c4f10-f497-40ab-a6c7-599395789302" containerID="0bce34f191ce18fc6a3c32b29f7c4f8ed428f8c87a09179d11f76dca95456fdf" exitCode=0 Feb 02 09:33:58 crc kubenswrapper[4764]: I0202 09:33:58.405264 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h55pp" event={"ID":"b90c4f10-f497-40ab-a6c7-599395789302","Type":"ContainerDied","Data":"0bce34f191ce18fc6a3c32b29f7c4f8ed428f8c87a09179d11f76dca95456fdf"} Feb 02 09:33:58 crc kubenswrapper[4764]: I0202 09:33:58.405452 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h55pp" event={"ID":"b90c4f10-f497-40ab-a6c7-599395789302","Type":"ContainerStarted","Data":"134d11f1ac2274e3a10edd87aaf92ad2672e9c8bd213aae7ea7b41630021fc51"} Feb 02 09:33:59 crc kubenswrapper[4764]: I0202 09:33:59.414331 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h55pp" event={"ID":"b90c4f10-f497-40ab-a6c7-599395789302","Type":"ContainerStarted","Data":"b8efec5a74edc100e02f75b6da78fac2d97705d943e1e43af89d11fa22dfefa9"} Feb 02 09:34:01 crc kubenswrapper[4764]: I0202 09:34:01.493116 4764 generic.go:334] "Generic (PLEG): container finished" podID="b90c4f10-f497-40ab-a6c7-599395789302" containerID="b8efec5a74edc100e02f75b6da78fac2d97705d943e1e43af89d11fa22dfefa9" exitCode=0 Feb 02 09:34:01 crc kubenswrapper[4764]: I0202 09:34:01.493434 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h55pp" event={"ID":"b90c4f10-f497-40ab-a6c7-599395789302","Type":"ContainerDied","Data":"b8efec5a74edc100e02f75b6da78fac2d97705d943e1e43af89d11fa22dfefa9"} Feb 02 09:34:02 crc kubenswrapper[4764]: I0202 09:34:02.504403 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h55pp" event={"ID":"b90c4f10-f497-40ab-a6c7-599395789302","Type":"ContainerStarted","Data":"e0217193819fa3477312d7a6e94e9d4233e15c19a131e58a5977bc5e8099fa0c"} Feb 02 09:34:02 crc kubenswrapper[4764]: I0202 09:34:02.527382 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-h55pp" podStartSLOduration=1.923938026 podStartE2EDuration="5.527348678s" podCreationTimestamp="2026-02-02 09:33:57 +0000 UTC" firstStartedPulling="2026-02-02 09:33:58.408163323 +0000 UTC m=+1601.341887411" lastFinishedPulling="2026-02-02 09:34:02.011573955 +0000 UTC m=+1604.945298063" observedRunningTime="2026-02-02 09:34:02.522164417 +0000 UTC m=+1605.455888505" watchObservedRunningTime="2026-02-02 09:34:02.527348678 +0000 UTC m=+1605.461072766" Feb 02 09:34:04 crc kubenswrapper[4764]: I0202 09:34:04.825468 4764 scope.go:117] "RemoveContainer" containerID="6db43ab36406cdd4dbe61d7d8688257aba4271b8601a5ba5594675ae309c761d" Feb 02 09:34:04 crc kubenswrapper[4764]: E0202 09:34:04.826077 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 09:34:07 crc kubenswrapper[4764]: I0202 09:34:07.527117 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-h55pp" Feb 02 09:34:07 crc kubenswrapper[4764]: I0202 09:34:07.527743 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-h55pp" Feb 02 09:34:07 crc kubenswrapper[4764]: I0202 09:34:07.577784 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-h55pp" Feb 02 09:34:07 crc kubenswrapper[4764]: I0202 09:34:07.640169 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-h55pp" Feb 02 09:34:07 crc kubenswrapper[4764]: I0202 09:34:07.821574 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-h55pp"] Feb 02 09:34:09 crc kubenswrapper[4764]: I0202 09:34:09.564463 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-h55pp" podUID="b90c4f10-f497-40ab-a6c7-599395789302" containerName="registry-server" containerID="cri-o://e0217193819fa3477312d7a6e94e9d4233e15c19a131e58a5977bc5e8099fa0c" gracePeriod=2 Feb 02 09:34:09 crc kubenswrapper[4764]: I0202 09:34:09.963243 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-h55pp" Feb 02 09:34:10 crc kubenswrapper[4764]: I0202 09:34:10.086266 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b90c4f10-f497-40ab-a6c7-599395789302-utilities\") pod \"b90c4f10-f497-40ab-a6c7-599395789302\" (UID: \"b90c4f10-f497-40ab-a6c7-599395789302\") " Feb 02 09:34:10 crc kubenswrapper[4764]: I0202 09:34:10.086416 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b90c4f10-f497-40ab-a6c7-599395789302-catalog-content\") pod \"b90c4f10-f497-40ab-a6c7-599395789302\" (UID: \"b90c4f10-f497-40ab-a6c7-599395789302\") " Feb 02 09:34:10 crc kubenswrapper[4764]: I0202 09:34:10.086448 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b6h2d\" (UniqueName: \"kubernetes.io/projected/b90c4f10-f497-40ab-a6c7-599395789302-kube-api-access-b6h2d\") pod \"b90c4f10-f497-40ab-a6c7-599395789302\" (UID: \"b90c4f10-f497-40ab-a6c7-599395789302\") " Feb 02 09:34:10 crc kubenswrapper[4764]: I0202 09:34:10.087701 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b90c4f10-f497-40ab-a6c7-599395789302-utilities" (OuterVolumeSpecName: "utilities") pod "b90c4f10-f497-40ab-a6c7-599395789302" (UID: "b90c4f10-f497-40ab-a6c7-599395789302"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 09:34:10 crc kubenswrapper[4764]: I0202 09:34:10.105157 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b90c4f10-f497-40ab-a6c7-599395789302-kube-api-access-b6h2d" (OuterVolumeSpecName: "kube-api-access-b6h2d") pod "b90c4f10-f497-40ab-a6c7-599395789302" (UID: "b90c4f10-f497-40ab-a6c7-599395789302"). InnerVolumeSpecName "kube-api-access-b6h2d". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:34:10 crc kubenswrapper[4764]: I0202 09:34:10.137859 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b90c4f10-f497-40ab-a6c7-599395789302-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b90c4f10-f497-40ab-a6c7-599395789302" (UID: "b90c4f10-f497-40ab-a6c7-599395789302"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 09:34:10 crc kubenswrapper[4764]: I0202 09:34:10.189482 4764 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b90c4f10-f497-40ab-a6c7-599395789302-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 09:34:10 crc kubenswrapper[4764]: I0202 09:34:10.189533 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b6h2d\" (UniqueName: \"kubernetes.io/projected/b90c4f10-f497-40ab-a6c7-599395789302-kube-api-access-b6h2d\") on node \"crc\" DevicePath \"\"" Feb 02 09:34:10 crc kubenswrapper[4764]: I0202 09:34:10.189547 4764 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b90c4f10-f497-40ab-a6c7-599395789302-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 09:34:10 crc kubenswrapper[4764]: I0202 09:34:10.587102 4764 generic.go:334] "Generic (PLEG): container finished" podID="b90c4f10-f497-40ab-a6c7-599395789302" containerID="e0217193819fa3477312d7a6e94e9d4233e15c19a131e58a5977bc5e8099fa0c" exitCode=0 Feb 02 09:34:10 crc kubenswrapper[4764]: I0202 09:34:10.587152 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h55pp" event={"ID":"b90c4f10-f497-40ab-a6c7-599395789302","Type":"ContainerDied","Data":"e0217193819fa3477312d7a6e94e9d4233e15c19a131e58a5977bc5e8099fa0c"} Feb 02 09:34:10 crc kubenswrapper[4764]: I0202 09:34:10.587222 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h55pp" event={"ID":"b90c4f10-f497-40ab-a6c7-599395789302","Type":"ContainerDied","Data":"134d11f1ac2274e3a10edd87aaf92ad2672e9c8bd213aae7ea7b41630021fc51"} Feb 02 09:34:10 crc kubenswrapper[4764]: I0202 09:34:10.587222 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-h55pp" Feb 02 09:34:10 crc kubenswrapper[4764]: I0202 09:34:10.587243 4764 scope.go:117] "RemoveContainer" containerID="e0217193819fa3477312d7a6e94e9d4233e15c19a131e58a5977bc5e8099fa0c" Feb 02 09:34:10 crc kubenswrapper[4764]: I0202 09:34:10.627147 4764 scope.go:117] "RemoveContainer" containerID="b8efec5a74edc100e02f75b6da78fac2d97705d943e1e43af89d11fa22dfefa9" Feb 02 09:34:10 crc kubenswrapper[4764]: I0202 09:34:10.631021 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-h55pp"] Feb 02 09:34:10 crc kubenswrapper[4764]: I0202 09:34:10.638462 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-h55pp"] Feb 02 09:34:10 crc kubenswrapper[4764]: I0202 09:34:10.650004 4764 scope.go:117] "RemoveContainer" containerID="0bce34f191ce18fc6a3c32b29f7c4f8ed428f8c87a09179d11f76dca95456fdf" Feb 02 09:34:10 crc kubenswrapper[4764]: I0202 09:34:10.704538 4764 scope.go:117] "RemoveContainer" containerID="e0217193819fa3477312d7a6e94e9d4233e15c19a131e58a5977bc5e8099fa0c" Feb 02 09:34:10 crc kubenswrapper[4764]: E0202 09:34:10.705102 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e0217193819fa3477312d7a6e94e9d4233e15c19a131e58a5977bc5e8099fa0c\": container with ID starting with e0217193819fa3477312d7a6e94e9d4233e15c19a131e58a5977bc5e8099fa0c not found: ID does not exist" containerID="e0217193819fa3477312d7a6e94e9d4233e15c19a131e58a5977bc5e8099fa0c" Feb 02 09:34:10 crc kubenswrapper[4764]: I0202 09:34:10.705230 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e0217193819fa3477312d7a6e94e9d4233e15c19a131e58a5977bc5e8099fa0c"} err="failed to get container status \"e0217193819fa3477312d7a6e94e9d4233e15c19a131e58a5977bc5e8099fa0c\": rpc error: code = NotFound desc = could not find container \"e0217193819fa3477312d7a6e94e9d4233e15c19a131e58a5977bc5e8099fa0c\": container with ID starting with e0217193819fa3477312d7a6e94e9d4233e15c19a131e58a5977bc5e8099fa0c not found: ID does not exist" Feb 02 09:34:10 crc kubenswrapper[4764]: I0202 09:34:10.705255 4764 scope.go:117] "RemoveContainer" containerID="b8efec5a74edc100e02f75b6da78fac2d97705d943e1e43af89d11fa22dfefa9" Feb 02 09:34:10 crc kubenswrapper[4764]: E0202 09:34:10.705494 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b8efec5a74edc100e02f75b6da78fac2d97705d943e1e43af89d11fa22dfefa9\": container with ID starting with b8efec5a74edc100e02f75b6da78fac2d97705d943e1e43af89d11fa22dfefa9 not found: ID does not exist" containerID="b8efec5a74edc100e02f75b6da78fac2d97705d943e1e43af89d11fa22dfefa9" Feb 02 09:34:10 crc kubenswrapper[4764]: I0202 09:34:10.705525 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b8efec5a74edc100e02f75b6da78fac2d97705d943e1e43af89d11fa22dfefa9"} err="failed to get container status \"b8efec5a74edc100e02f75b6da78fac2d97705d943e1e43af89d11fa22dfefa9\": rpc error: code = NotFound desc = could not find container \"b8efec5a74edc100e02f75b6da78fac2d97705d943e1e43af89d11fa22dfefa9\": container with ID starting with b8efec5a74edc100e02f75b6da78fac2d97705d943e1e43af89d11fa22dfefa9 not found: ID does not exist" Feb 02 09:34:10 crc kubenswrapper[4764]: I0202 09:34:10.705549 4764 scope.go:117] "RemoveContainer" containerID="0bce34f191ce18fc6a3c32b29f7c4f8ed428f8c87a09179d11f76dca95456fdf" Feb 02 09:34:10 crc kubenswrapper[4764]: E0202 09:34:10.705759 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0bce34f191ce18fc6a3c32b29f7c4f8ed428f8c87a09179d11f76dca95456fdf\": container with ID starting with 0bce34f191ce18fc6a3c32b29f7c4f8ed428f8c87a09179d11f76dca95456fdf not found: ID does not exist" containerID="0bce34f191ce18fc6a3c32b29f7c4f8ed428f8c87a09179d11f76dca95456fdf" Feb 02 09:34:10 crc kubenswrapper[4764]: I0202 09:34:10.705786 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0bce34f191ce18fc6a3c32b29f7c4f8ed428f8c87a09179d11f76dca95456fdf"} err="failed to get container status \"0bce34f191ce18fc6a3c32b29f7c4f8ed428f8c87a09179d11f76dca95456fdf\": rpc error: code = NotFound desc = could not find container \"0bce34f191ce18fc6a3c32b29f7c4f8ed428f8c87a09179d11f76dca95456fdf\": container with ID starting with 0bce34f191ce18fc6a3c32b29f7c4f8ed428f8c87a09179d11f76dca95456fdf not found: ID does not exist" Feb 02 09:34:11 crc kubenswrapper[4764]: I0202 09:34:11.065630 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-48c4-account-create-update-lvzbb"] Feb 02 09:34:11 crc kubenswrapper[4764]: I0202 09:34:11.081590 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-2a71-account-create-update-6bfns"] Feb 02 09:34:11 crc kubenswrapper[4764]: I0202 09:34:11.092980 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-wqk4n"] Feb 02 09:34:11 crc kubenswrapper[4764]: I0202 09:34:11.098696 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-2d92n"] Feb 02 09:34:11 crc kubenswrapper[4764]: I0202 09:34:11.105517 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-2a71-account-create-update-6bfns"] Feb 02 09:34:11 crc kubenswrapper[4764]: I0202 09:34:11.113455 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-wqk4n"] Feb 02 09:34:11 crc kubenswrapper[4764]: I0202 09:34:11.120041 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-2d92n"] Feb 02 09:34:11 crc kubenswrapper[4764]: I0202 09:34:11.126678 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-48c4-account-create-update-lvzbb"] Feb 02 09:34:11 crc kubenswrapper[4764]: I0202 09:34:11.840412 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1a85da62-5c86-4f54-9d14-b80c3974189a" path="/var/lib/kubelet/pods/1a85da62-5c86-4f54-9d14-b80c3974189a/volumes" Feb 02 09:34:11 crc kubenswrapper[4764]: I0202 09:34:11.841424 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9c4c3400-5c5d-4d8b-82a9-c8ca34abe7c0" path="/var/lib/kubelet/pods/9c4c3400-5c5d-4d8b-82a9-c8ca34abe7c0/volumes" Feb 02 09:34:11 crc kubenswrapper[4764]: I0202 09:34:11.842303 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a3a49c31-c4c9-415b-89d9-bee375775e38" path="/var/lib/kubelet/pods/a3a49c31-c4c9-415b-89d9-bee375775e38/volumes" Feb 02 09:34:11 crc kubenswrapper[4764]: I0202 09:34:11.843173 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b0e0b60e-0289-4700-b747-a6442ca6e34a" path="/var/lib/kubelet/pods/b0e0b60e-0289-4700-b747-a6442ca6e34a/volumes" Feb 02 09:34:11 crc kubenswrapper[4764]: I0202 09:34:11.844771 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b90c4f10-f497-40ab-a6c7-599395789302" path="/var/lib/kubelet/pods/b90c4f10-f497-40ab-a6c7-599395789302/volumes" Feb 02 09:34:15 crc kubenswrapper[4764]: I0202 09:34:15.647924 4764 generic.go:334] "Generic (PLEG): container finished" podID="667bdcf6-12ac-4bfc-b652-48ce7aca1fc2" containerID="3782d7a113fa14a8d1b7a9b1c1ff7a6a278ec65b1543752bd6307e2d11046199" exitCode=0 Feb 02 09:34:15 crc kubenswrapper[4764]: I0202 09:34:15.648100 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-nx4kb" event={"ID":"667bdcf6-12ac-4bfc-b652-48ce7aca1fc2","Type":"ContainerDied","Data":"3782d7a113fa14a8d1b7a9b1c1ff7a6a278ec65b1543752bd6307e2d11046199"} Feb 02 09:34:16 crc kubenswrapper[4764]: I0202 09:34:16.044914 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-z55kp"] Feb 02 09:34:16 crc kubenswrapper[4764]: I0202 09:34:16.060822 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-a0fd-account-create-update-qgrmb"] Feb 02 09:34:16 crc kubenswrapper[4764]: I0202 09:34:16.072058 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-a0fd-account-create-update-qgrmb"] Feb 02 09:34:16 crc kubenswrapper[4764]: I0202 09:34:16.082309 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-z55kp"] Feb 02 09:34:17 crc kubenswrapper[4764]: I0202 09:34:17.160722 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-nx4kb" Feb 02 09:34:17 crc kubenswrapper[4764]: I0202 09:34:17.326240 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/667bdcf6-12ac-4bfc-b652-48ce7aca1fc2-ssh-key-openstack-edpm-ipam\") pod \"667bdcf6-12ac-4bfc-b652-48ce7aca1fc2\" (UID: \"667bdcf6-12ac-4bfc-b652-48ce7aca1fc2\") " Feb 02 09:34:17 crc kubenswrapper[4764]: I0202 09:34:17.326386 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/667bdcf6-12ac-4bfc-b652-48ce7aca1fc2-inventory\") pod \"667bdcf6-12ac-4bfc-b652-48ce7aca1fc2\" (UID: \"667bdcf6-12ac-4bfc-b652-48ce7aca1fc2\") " Feb 02 09:34:17 crc kubenswrapper[4764]: I0202 09:34:17.326437 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hlbft\" (UniqueName: \"kubernetes.io/projected/667bdcf6-12ac-4bfc-b652-48ce7aca1fc2-kube-api-access-hlbft\") pod \"667bdcf6-12ac-4bfc-b652-48ce7aca1fc2\" (UID: \"667bdcf6-12ac-4bfc-b652-48ce7aca1fc2\") " Feb 02 09:34:17 crc kubenswrapper[4764]: I0202 09:34:17.333737 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/667bdcf6-12ac-4bfc-b652-48ce7aca1fc2-kube-api-access-hlbft" (OuterVolumeSpecName: "kube-api-access-hlbft") pod "667bdcf6-12ac-4bfc-b652-48ce7aca1fc2" (UID: "667bdcf6-12ac-4bfc-b652-48ce7aca1fc2"). InnerVolumeSpecName "kube-api-access-hlbft". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:34:17 crc kubenswrapper[4764]: I0202 09:34:17.356483 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/667bdcf6-12ac-4bfc-b652-48ce7aca1fc2-inventory" (OuterVolumeSpecName: "inventory") pod "667bdcf6-12ac-4bfc-b652-48ce7aca1fc2" (UID: "667bdcf6-12ac-4bfc-b652-48ce7aca1fc2"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:34:17 crc kubenswrapper[4764]: I0202 09:34:17.376737 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/667bdcf6-12ac-4bfc-b652-48ce7aca1fc2-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "667bdcf6-12ac-4bfc-b652-48ce7aca1fc2" (UID: "667bdcf6-12ac-4bfc-b652-48ce7aca1fc2"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:34:17 crc kubenswrapper[4764]: I0202 09:34:17.428863 4764 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/667bdcf6-12ac-4bfc-b652-48ce7aca1fc2-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 02 09:34:17 crc kubenswrapper[4764]: I0202 09:34:17.428902 4764 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/667bdcf6-12ac-4bfc-b652-48ce7aca1fc2-inventory\") on node \"crc\" DevicePath \"\"" Feb 02 09:34:17 crc kubenswrapper[4764]: I0202 09:34:17.428916 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hlbft\" (UniqueName: \"kubernetes.io/projected/667bdcf6-12ac-4bfc-b652-48ce7aca1fc2-kube-api-access-hlbft\") on node \"crc\" DevicePath \"\"" Feb 02 09:34:17 crc kubenswrapper[4764]: I0202 09:34:17.680221 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-nx4kb" event={"ID":"667bdcf6-12ac-4bfc-b652-48ce7aca1fc2","Type":"ContainerDied","Data":"952bbf482325a75cc8cc7b8af067e5f6cdb90d9142dcc57c129a1780e469aaca"} Feb 02 09:34:17 crc kubenswrapper[4764]: I0202 09:34:17.680473 4764 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="952bbf482325a75cc8cc7b8af067e5f6cdb90d9142dcc57c129a1780e469aaca" Feb 02 09:34:17 crc kubenswrapper[4764]: I0202 09:34:17.680331 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-nx4kb" Feb 02 09:34:17 crc kubenswrapper[4764]: I0202 09:34:17.779950 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-d928p"] Feb 02 09:34:17 crc kubenswrapper[4764]: E0202 09:34:17.780464 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b90c4f10-f497-40ab-a6c7-599395789302" containerName="registry-server" Feb 02 09:34:17 crc kubenswrapper[4764]: I0202 09:34:17.780492 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="b90c4f10-f497-40ab-a6c7-599395789302" containerName="registry-server" Feb 02 09:34:17 crc kubenswrapper[4764]: E0202 09:34:17.780517 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b90c4f10-f497-40ab-a6c7-599395789302" containerName="extract-utilities" Feb 02 09:34:17 crc kubenswrapper[4764]: I0202 09:34:17.780526 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="b90c4f10-f497-40ab-a6c7-599395789302" containerName="extract-utilities" Feb 02 09:34:17 crc kubenswrapper[4764]: E0202 09:34:17.780544 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="667bdcf6-12ac-4bfc-b652-48ce7aca1fc2" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Feb 02 09:34:17 crc kubenswrapper[4764]: I0202 09:34:17.780554 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="667bdcf6-12ac-4bfc-b652-48ce7aca1fc2" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Feb 02 09:34:17 crc kubenswrapper[4764]: E0202 09:34:17.780584 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b90c4f10-f497-40ab-a6c7-599395789302" containerName="extract-content" Feb 02 09:34:17 crc kubenswrapper[4764]: I0202 09:34:17.780592 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="b90c4f10-f497-40ab-a6c7-599395789302" containerName="extract-content" Feb 02 09:34:17 crc kubenswrapper[4764]: I0202 09:34:17.780791 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="667bdcf6-12ac-4bfc-b652-48ce7aca1fc2" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Feb 02 09:34:17 crc kubenswrapper[4764]: I0202 09:34:17.780822 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="b90c4f10-f497-40ab-a6c7-599395789302" containerName="registry-server" Feb 02 09:34:17 crc kubenswrapper[4764]: I0202 09:34:17.781658 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-d928p" Feb 02 09:34:17 crc kubenswrapper[4764]: I0202 09:34:17.789467 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 02 09:34:17 crc kubenswrapper[4764]: I0202 09:34:17.789516 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 02 09:34:17 crc kubenswrapper[4764]: I0202 09:34:17.789714 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 02 09:34:17 crc kubenswrapper[4764]: I0202 09:34:17.789841 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-xksq5" Feb 02 09:34:17 crc kubenswrapper[4764]: I0202 09:34:17.791372 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-d928p"] Feb 02 09:34:17 crc kubenswrapper[4764]: I0202 09:34:17.835832 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="52dfc490-3882-4fb5-9c6f-93253a7f794b" path="/var/lib/kubelet/pods/52dfc490-3882-4fb5-9c6f-93253a7f794b/volumes" Feb 02 09:34:17 crc kubenswrapper[4764]: I0202 09:34:17.836386 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="80593eb5-7787-484f-b3a5-8cd856f0a6b9" path="/var/lib/kubelet/pods/80593eb5-7787-484f-b3a5-8cd856f0a6b9/volumes" Feb 02 09:34:17 crc kubenswrapper[4764]: I0202 09:34:17.946223 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a4f5b290-7d44-4c9b-bffb-78a78f8bd546-ssh-key-openstack-edpm-ipam\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-d928p\" (UID: \"a4f5b290-7d44-4c9b-bffb-78a78f8bd546\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-d928p" Feb 02 09:34:17 crc kubenswrapper[4764]: I0202 09:34:17.946350 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4ljht\" (UniqueName: \"kubernetes.io/projected/a4f5b290-7d44-4c9b-bffb-78a78f8bd546-kube-api-access-4ljht\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-d928p\" (UID: \"a4f5b290-7d44-4c9b-bffb-78a78f8bd546\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-d928p" Feb 02 09:34:17 crc kubenswrapper[4764]: I0202 09:34:17.946400 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a4f5b290-7d44-4c9b-bffb-78a78f8bd546-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-d928p\" (UID: \"a4f5b290-7d44-4c9b-bffb-78a78f8bd546\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-d928p" Feb 02 09:34:18 crc kubenswrapper[4764]: I0202 09:34:18.034517 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/root-account-create-update-rd2c2"] Feb 02 09:34:18 crc kubenswrapper[4764]: I0202 09:34:18.048523 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a4f5b290-7d44-4c9b-bffb-78a78f8bd546-ssh-key-openstack-edpm-ipam\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-d928p\" (UID: \"a4f5b290-7d44-4c9b-bffb-78a78f8bd546\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-d928p" Feb 02 09:34:18 crc kubenswrapper[4764]: I0202 09:34:18.048663 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4ljht\" (UniqueName: \"kubernetes.io/projected/a4f5b290-7d44-4c9b-bffb-78a78f8bd546-kube-api-access-4ljht\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-d928p\" (UID: \"a4f5b290-7d44-4c9b-bffb-78a78f8bd546\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-d928p" Feb 02 09:34:18 crc kubenswrapper[4764]: I0202 09:34:18.048723 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a4f5b290-7d44-4c9b-bffb-78a78f8bd546-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-d928p\" (UID: \"a4f5b290-7d44-4c9b-bffb-78a78f8bd546\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-d928p" Feb 02 09:34:18 crc kubenswrapper[4764]: I0202 09:34:18.050702 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/root-account-create-update-rd2c2"] Feb 02 09:34:18 crc kubenswrapper[4764]: I0202 09:34:18.053613 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a4f5b290-7d44-4c9b-bffb-78a78f8bd546-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-d928p\" (UID: \"a4f5b290-7d44-4c9b-bffb-78a78f8bd546\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-d928p" Feb 02 09:34:18 crc kubenswrapper[4764]: I0202 09:34:18.059210 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a4f5b290-7d44-4c9b-bffb-78a78f8bd546-ssh-key-openstack-edpm-ipam\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-d928p\" (UID: \"a4f5b290-7d44-4c9b-bffb-78a78f8bd546\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-d928p" Feb 02 09:34:18 crc kubenswrapper[4764]: I0202 09:34:18.068339 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4ljht\" (UniqueName: \"kubernetes.io/projected/a4f5b290-7d44-4c9b-bffb-78a78f8bd546-kube-api-access-4ljht\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-d928p\" (UID: \"a4f5b290-7d44-4c9b-bffb-78a78f8bd546\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-d928p" Feb 02 09:34:18 crc kubenswrapper[4764]: I0202 09:34:18.110855 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-d928p" Feb 02 09:34:18 crc kubenswrapper[4764]: I0202 09:34:18.656268 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-d928p"] Feb 02 09:34:18 crc kubenswrapper[4764]: I0202 09:34:18.692652 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-d928p" event={"ID":"a4f5b290-7d44-4c9b-bffb-78a78f8bd546","Type":"ContainerStarted","Data":"e4b5dc9b6ddf905f25eac2c9a47671408d00b29ccf734ca76d437c83a93e8416"} Feb 02 09:34:19 crc kubenswrapper[4764]: I0202 09:34:19.710187 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-d928p" event={"ID":"a4f5b290-7d44-4c9b-bffb-78a78f8bd546","Type":"ContainerStarted","Data":"f9857c237746b859c2245f63119218034d063a2161209dbe4c8439af6a1eaf49"} Feb 02 09:34:19 crc kubenswrapper[4764]: I0202 09:34:19.747377 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-d928p" podStartSLOduration=2.273046591 podStartE2EDuration="2.747359679s" podCreationTimestamp="2026-02-02 09:34:17 +0000 UTC" firstStartedPulling="2026-02-02 09:34:18.67154213 +0000 UTC m=+1621.605266218" lastFinishedPulling="2026-02-02 09:34:19.145855208 +0000 UTC m=+1622.079579306" observedRunningTime="2026-02-02 09:34:19.733223555 +0000 UTC m=+1622.666947683" watchObservedRunningTime="2026-02-02 09:34:19.747359679 +0000 UTC m=+1622.681083777" Feb 02 09:34:19 crc kubenswrapper[4764]: I0202 09:34:19.826046 4764 scope.go:117] "RemoveContainer" containerID="6db43ab36406cdd4dbe61d7d8688257aba4271b8601a5ba5594675ae309c761d" Feb 02 09:34:19 crc kubenswrapper[4764]: E0202 09:34:19.826477 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 09:34:19 crc kubenswrapper[4764]: I0202 09:34:19.843748 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9a6a24ad-01bc-4766-bb16-b40605a07ae6" path="/var/lib/kubelet/pods/9a6a24ad-01bc-4766-bb16-b40605a07ae6/volumes" Feb 02 09:34:24 crc kubenswrapper[4764]: I0202 09:34:24.768054 4764 generic.go:334] "Generic (PLEG): container finished" podID="a4f5b290-7d44-4c9b-bffb-78a78f8bd546" containerID="f9857c237746b859c2245f63119218034d063a2161209dbe4c8439af6a1eaf49" exitCode=0 Feb 02 09:34:24 crc kubenswrapper[4764]: I0202 09:34:24.768138 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-d928p" event={"ID":"a4f5b290-7d44-4c9b-bffb-78a78f8bd546","Type":"ContainerDied","Data":"f9857c237746b859c2245f63119218034d063a2161209dbe4c8439af6a1eaf49"} Feb 02 09:34:24 crc kubenswrapper[4764]: I0202 09:34:24.927549 4764 scope.go:117] "RemoveContainer" containerID="8b7fb4bd5dbf9110d437990b0859eb46aa316b79e92b75be85900e53ef6305d1" Feb 02 09:34:24 crc kubenswrapper[4764]: I0202 09:34:24.956634 4764 scope.go:117] "RemoveContainer" containerID="eeac11c8826e4af748480f8b6e34d466eccf10b88248d850ef138bc625c07bec" Feb 02 09:34:25 crc kubenswrapper[4764]: I0202 09:34:25.006692 4764 scope.go:117] "RemoveContainer" containerID="7e4be9945a21765d2e816b5d7ef51e7f14627efe60631690e6d27f83860286f3" Feb 02 09:34:25 crc kubenswrapper[4764]: I0202 09:34:25.047600 4764 scope.go:117] "RemoveContainer" containerID="d62d6a087221f0d57aacf9b33844eb86a5a5c129fd54d5dd9da5ebf675b4c402" Feb 02 09:34:25 crc kubenswrapper[4764]: I0202 09:34:25.078097 4764 scope.go:117] "RemoveContainer" containerID="00e343bf0be6a97406e0c63cd0a3dc73dccc3aa1062c62938cb7d13b3623dd5b" Feb 02 09:34:25 crc kubenswrapper[4764]: I0202 09:34:25.117241 4764 scope.go:117] "RemoveContainer" containerID="15a867576bfef1a3a727c912d74f23f8a035f461961a78d5ad32e4e7c8833fc0" Feb 02 09:34:25 crc kubenswrapper[4764]: I0202 09:34:25.170539 4764 scope.go:117] "RemoveContainer" containerID="884c3cc99e06d5ffc876d00aabdecbb9303f422c3cafc35b2045e7f55876c9c3" Feb 02 09:34:26 crc kubenswrapper[4764]: I0202 09:34:26.169548 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-d928p" Feb 02 09:34:26 crc kubenswrapper[4764]: I0202 09:34:26.315343 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4ljht\" (UniqueName: \"kubernetes.io/projected/a4f5b290-7d44-4c9b-bffb-78a78f8bd546-kube-api-access-4ljht\") pod \"a4f5b290-7d44-4c9b-bffb-78a78f8bd546\" (UID: \"a4f5b290-7d44-4c9b-bffb-78a78f8bd546\") " Feb 02 09:34:26 crc kubenswrapper[4764]: I0202 09:34:26.315755 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a4f5b290-7d44-4c9b-bffb-78a78f8bd546-ssh-key-openstack-edpm-ipam\") pod \"a4f5b290-7d44-4c9b-bffb-78a78f8bd546\" (UID: \"a4f5b290-7d44-4c9b-bffb-78a78f8bd546\") " Feb 02 09:34:26 crc kubenswrapper[4764]: I0202 09:34:26.316425 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a4f5b290-7d44-4c9b-bffb-78a78f8bd546-inventory\") pod \"a4f5b290-7d44-4c9b-bffb-78a78f8bd546\" (UID: \"a4f5b290-7d44-4c9b-bffb-78a78f8bd546\") " Feb 02 09:34:26 crc kubenswrapper[4764]: I0202 09:34:26.324433 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a4f5b290-7d44-4c9b-bffb-78a78f8bd546-kube-api-access-4ljht" (OuterVolumeSpecName: "kube-api-access-4ljht") pod "a4f5b290-7d44-4c9b-bffb-78a78f8bd546" (UID: "a4f5b290-7d44-4c9b-bffb-78a78f8bd546"). InnerVolumeSpecName "kube-api-access-4ljht". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:34:26 crc kubenswrapper[4764]: I0202 09:34:26.349389 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a4f5b290-7d44-4c9b-bffb-78a78f8bd546-inventory" (OuterVolumeSpecName: "inventory") pod "a4f5b290-7d44-4c9b-bffb-78a78f8bd546" (UID: "a4f5b290-7d44-4c9b-bffb-78a78f8bd546"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:34:26 crc kubenswrapper[4764]: I0202 09:34:26.350216 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a4f5b290-7d44-4c9b-bffb-78a78f8bd546-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "a4f5b290-7d44-4c9b-bffb-78a78f8bd546" (UID: "a4f5b290-7d44-4c9b-bffb-78a78f8bd546"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:34:26 crc kubenswrapper[4764]: I0202 09:34:26.418491 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4ljht\" (UniqueName: \"kubernetes.io/projected/a4f5b290-7d44-4c9b-bffb-78a78f8bd546-kube-api-access-4ljht\") on node \"crc\" DevicePath \"\"" Feb 02 09:34:26 crc kubenswrapper[4764]: I0202 09:34:26.418538 4764 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a4f5b290-7d44-4c9b-bffb-78a78f8bd546-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 02 09:34:26 crc kubenswrapper[4764]: I0202 09:34:26.418557 4764 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a4f5b290-7d44-4c9b-bffb-78a78f8bd546-inventory\") on node \"crc\" DevicePath \"\"" Feb 02 09:34:26 crc kubenswrapper[4764]: I0202 09:34:26.793395 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-d928p" event={"ID":"a4f5b290-7d44-4c9b-bffb-78a78f8bd546","Type":"ContainerDied","Data":"e4b5dc9b6ddf905f25eac2c9a47671408d00b29ccf734ca76d437c83a93e8416"} Feb 02 09:34:26 crc kubenswrapper[4764]: I0202 09:34:26.793447 4764 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e4b5dc9b6ddf905f25eac2c9a47671408d00b29ccf734ca76d437c83a93e8416" Feb 02 09:34:26 crc kubenswrapper[4764]: I0202 09:34:26.793518 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-d928p" Feb 02 09:34:26 crc kubenswrapper[4764]: I0202 09:34:26.874721 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-ck2zz"] Feb 02 09:34:26 crc kubenswrapper[4764]: E0202 09:34:26.875288 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4f5b290-7d44-4c9b-bffb-78a78f8bd546" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Feb 02 09:34:26 crc kubenswrapper[4764]: I0202 09:34:26.875321 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4f5b290-7d44-4c9b-bffb-78a78f8bd546" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Feb 02 09:34:26 crc kubenswrapper[4764]: I0202 09:34:26.875612 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="a4f5b290-7d44-4c9b-bffb-78a78f8bd546" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Feb 02 09:34:26 crc kubenswrapper[4764]: I0202 09:34:26.877196 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-ck2zz" Feb 02 09:34:26 crc kubenswrapper[4764]: I0202 09:34:26.879326 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 02 09:34:26 crc kubenswrapper[4764]: I0202 09:34:26.879346 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-xksq5" Feb 02 09:34:26 crc kubenswrapper[4764]: I0202 09:34:26.881041 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 02 09:34:26 crc kubenswrapper[4764]: I0202 09:34:26.889830 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 02 09:34:26 crc kubenswrapper[4764]: I0202 09:34:26.903065 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-ck2zz"] Feb 02 09:34:26 crc kubenswrapper[4764]: E0202 09:34:26.959581 4764 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda4f5b290_7d44_4c9b_bffb_78a78f8bd546.slice\": RecentStats: unable to find data in memory cache]" Feb 02 09:34:27 crc kubenswrapper[4764]: I0202 09:34:27.032182 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-95h4g\" (UniqueName: \"kubernetes.io/projected/c66a8ab1-6fbb-4301-a4ce-34ddad011ee2-kube-api-access-95h4g\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-ck2zz\" (UID: \"c66a8ab1-6fbb-4301-a4ce-34ddad011ee2\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-ck2zz" Feb 02 09:34:27 crc kubenswrapper[4764]: I0202 09:34:27.032443 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c66a8ab1-6fbb-4301-a4ce-34ddad011ee2-ssh-key-openstack-edpm-ipam\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-ck2zz\" (UID: \"c66a8ab1-6fbb-4301-a4ce-34ddad011ee2\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-ck2zz" Feb 02 09:34:27 crc kubenswrapper[4764]: I0202 09:34:27.032548 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c66a8ab1-6fbb-4301-a4ce-34ddad011ee2-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-ck2zz\" (UID: \"c66a8ab1-6fbb-4301-a4ce-34ddad011ee2\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-ck2zz" Feb 02 09:34:27 crc kubenswrapper[4764]: I0202 09:34:27.135132 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-95h4g\" (UniqueName: \"kubernetes.io/projected/c66a8ab1-6fbb-4301-a4ce-34ddad011ee2-kube-api-access-95h4g\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-ck2zz\" (UID: \"c66a8ab1-6fbb-4301-a4ce-34ddad011ee2\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-ck2zz" Feb 02 09:34:27 crc kubenswrapper[4764]: I0202 09:34:27.135342 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c66a8ab1-6fbb-4301-a4ce-34ddad011ee2-ssh-key-openstack-edpm-ipam\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-ck2zz\" (UID: \"c66a8ab1-6fbb-4301-a4ce-34ddad011ee2\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-ck2zz" Feb 02 09:34:27 crc kubenswrapper[4764]: I0202 09:34:27.136302 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c66a8ab1-6fbb-4301-a4ce-34ddad011ee2-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-ck2zz\" (UID: \"c66a8ab1-6fbb-4301-a4ce-34ddad011ee2\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-ck2zz" Feb 02 09:34:27 crc kubenswrapper[4764]: I0202 09:34:27.144571 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c66a8ab1-6fbb-4301-a4ce-34ddad011ee2-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-ck2zz\" (UID: \"c66a8ab1-6fbb-4301-a4ce-34ddad011ee2\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-ck2zz" Feb 02 09:34:27 crc kubenswrapper[4764]: I0202 09:34:27.144730 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c66a8ab1-6fbb-4301-a4ce-34ddad011ee2-ssh-key-openstack-edpm-ipam\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-ck2zz\" (UID: \"c66a8ab1-6fbb-4301-a4ce-34ddad011ee2\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-ck2zz" Feb 02 09:34:27 crc kubenswrapper[4764]: I0202 09:34:27.163360 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-95h4g\" (UniqueName: \"kubernetes.io/projected/c66a8ab1-6fbb-4301-a4ce-34ddad011ee2-kube-api-access-95h4g\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-ck2zz\" (UID: \"c66a8ab1-6fbb-4301-a4ce-34ddad011ee2\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-ck2zz" Feb 02 09:34:27 crc kubenswrapper[4764]: I0202 09:34:27.197738 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-ck2zz" Feb 02 09:34:27 crc kubenswrapper[4764]: I0202 09:34:27.843421 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-ck2zz"] Feb 02 09:34:28 crc kubenswrapper[4764]: I0202 09:34:28.812512 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-ck2zz" event={"ID":"c66a8ab1-6fbb-4301-a4ce-34ddad011ee2","Type":"ContainerStarted","Data":"a358cb778023f9178a5758199b3a678d46fa14535321922c333a5d5a68a31834"} Feb 02 09:34:28 crc kubenswrapper[4764]: I0202 09:34:28.812857 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-ck2zz" event={"ID":"c66a8ab1-6fbb-4301-a4ce-34ddad011ee2","Type":"ContainerStarted","Data":"0315669b91b1605e6e0723ffc66351deaa1db1514dd2c338d8614856fb261256"} Feb 02 09:34:28 crc kubenswrapper[4764]: I0202 09:34:28.838234 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-ck2zz" podStartSLOduration=2.360477953 podStartE2EDuration="2.838204203s" podCreationTimestamp="2026-02-02 09:34:26 +0000 UTC" firstStartedPulling="2026-02-02 09:34:27.820245237 +0000 UTC m=+1630.753969345" lastFinishedPulling="2026-02-02 09:34:28.297971477 +0000 UTC m=+1631.231695595" observedRunningTime="2026-02-02 09:34:28.832336924 +0000 UTC m=+1631.766061042" watchObservedRunningTime="2026-02-02 09:34:28.838204203 +0000 UTC m=+1631.771928331" Feb 02 09:34:32 crc kubenswrapper[4764]: I0202 09:34:32.825816 4764 scope.go:117] "RemoveContainer" containerID="6db43ab36406cdd4dbe61d7d8688257aba4271b8601a5ba5594675ae309c761d" Feb 02 09:34:32 crc kubenswrapper[4764]: E0202 09:34:32.826569 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 09:34:41 crc kubenswrapper[4764]: I0202 09:34:41.045605 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-wzb8q"] Feb 02 09:34:41 crc kubenswrapper[4764]: I0202 09:34:41.058219 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-wzb8q"] Feb 02 09:34:41 crc kubenswrapper[4764]: I0202 09:34:41.839858 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b81f3f76-ea90-43bb-a084-c4631433b0da" path="/var/lib/kubelet/pods/b81f3f76-ea90-43bb-a084-c4631433b0da/volumes" Feb 02 09:34:45 crc kubenswrapper[4764]: I0202 09:34:45.829102 4764 scope.go:117] "RemoveContainer" containerID="6db43ab36406cdd4dbe61d7d8688257aba4271b8601a5ba5594675ae309c761d" Feb 02 09:34:45 crc kubenswrapper[4764]: E0202 09:34:45.829810 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 09:34:52 crc kubenswrapper[4764]: I0202 09:34:52.080073 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-873f-account-create-update-n74bq"] Feb 02 09:34:52 crc kubenswrapper[4764]: I0202 09:34:52.101116 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-bqq2t"] Feb 02 09:34:52 crc kubenswrapper[4764]: I0202 09:34:52.108311 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-66d2-account-create-update-6fznc"] Feb 02 09:34:52 crc kubenswrapper[4764]: I0202 09:34:52.116459 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-bqq2t"] Feb 02 09:34:52 crc kubenswrapper[4764]: I0202 09:34:52.123943 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-873f-account-create-update-n74bq"] Feb 02 09:34:52 crc kubenswrapper[4764]: I0202 09:34:52.131907 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-nbgkq"] Feb 02 09:34:52 crc kubenswrapper[4764]: I0202 09:34:52.141605 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-66d2-account-create-update-6fznc"] Feb 02 09:34:52 crc kubenswrapper[4764]: I0202 09:34:52.150486 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-nbgkq"] Feb 02 09:34:52 crc kubenswrapper[4764]: I0202 09:34:52.158256 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-4fxzb"] Feb 02 09:34:52 crc kubenswrapper[4764]: I0202 09:34:52.168361 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-0818-account-create-update-rkmtz"] Feb 02 09:34:52 crc kubenswrapper[4764]: I0202 09:34:52.175851 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-0818-account-create-update-rkmtz"] Feb 02 09:34:52 crc kubenswrapper[4764]: I0202 09:34:52.182159 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-4fxzb"] Feb 02 09:34:53 crc kubenswrapper[4764]: I0202 09:34:53.838038 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1709efed-3308-4801-a4b4-efd067c6ce90" path="/var/lib/kubelet/pods/1709efed-3308-4801-a4b4-efd067c6ce90/volumes" Feb 02 09:34:53 crc kubenswrapper[4764]: I0202 09:34:53.841059 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="467a028d-ccfb-4d13-84d5-b77cdd1e96ac" path="/var/lib/kubelet/pods/467a028d-ccfb-4d13-84d5-b77cdd1e96ac/volumes" Feb 02 09:34:53 crc kubenswrapper[4764]: I0202 09:34:53.842796 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49eeb3bb-263f-452e-a2dc-9c24c0fe62b1" path="/var/lib/kubelet/pods/49eeb3bb-263f-452e-a2dc-9c24c0fe62b1/volumes" Feb 02 09:34:53 crc kubenswrapper[4764]: I0202 09:34:53.844228 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb792d3-0d3c-44fc-8d1f-50e9763773dc" path="/var/lib/kubelet/pods/4bb792d3-0d3c-44fc-8d1f-50e9763773dc/volumes" Feb 02 09:34:53 crc kubenswrapper[4764]: I0202 09:34:53.846413 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a2004c87-c998-4892-ba8f-adeeeb31e746" path="/var/lib/kubelet/pods/a2004c87-c998-4892-ba8f-adeeeb31e746/volumes" Feb 02 09:34:53 crc kubenswrapper[4764]: I0202 09:34:53.847693 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a3915739-a4ee-4cdd-a27c-58e518e0dade" path="/var/lib/kubelet/pods/a3915739-a4ee-4cdd-a27c-58e518e0dade/volumes" Feb 02 09:34:57 crc kubenswrapper[4764]: I0202 09:34:57.044622 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-6k5q7"] Feb 02 09:34:57 crc kubenswrapper[4764]: I0202 09:34:57.055873 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-6k5q7"] Feb 02 09:34:57 crc kubenswrapper[4764]: I0202 09:34:57.837693 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fd7663ce-b995-4503-99d9-d5d22952db2c" path="/var/lib/kubelet/pods/fd7663ce-b995-4503-99d9-d5d22952db2c/volumes" Feb 02 09:34:59 crc kubenswrapper[4764]: I0202 09:34:59.825322 4764 scope.go:117] "RemoveContainer" containerID="6db43ab36406cdd4dbe61d7d8688257aba4271b8601a5ba5594675ae309c761d" Feb 02 09:34:59 crc kubenswrapper[4764]: E0202 09:34:59.825890 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 09:35:08 crc kubenswrapper[4764]: I0202 09:35:08.216799 4764 generic.go:334] "Generic (PLEG): container finished" podID="c66a8ab1-6fbb-4301-a4ce-34ddad011ee2" containerID="a358cb778023f9178a5758199b3a678d46fa14535321922c333a5d5a68a31834" exitCode=0 Feb 02 09:35:08 crc kubenswrapper[4764]: I0202 09:35:08.217514 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-ck2zz" event={"ID":"c66a8ab1-6fbb-4301-a4ce-34ddad011ee2","Type":"ContainerDied","Data":"a358cb778023f9178a5758199b3a678d46fa14535321922c333a5d5a68a31834"} Feb 02 09:35:09 crc kubenswrapper[4764]: I0202 09:35:09.677866 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-ck2zz" Feb 02 09:35:09 crc kubenswrapper[4764]: I0202 09:35:09.860509 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c66a8ab1-6fbb-4301-a4ce-34ddad011ee2-inventory\") pod \"c66a8ab1-6fbb-4301-a4ce-34ddad011ee2\" (UID: \"c66a8ab1-6fbb-4301-a4ce-34ddad011ee2\") " Feb 02 09:35:09 crc kubenswrapper[4764]: I0202 09:35:09.860682 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-95h4g\" (UniqueName: \"kubernetes.io/projected/c66a8ab1-6fbb-4301-a4ce-34ddad011ee2-kube-api-access-95h4g\") pod \"c66a8ab1-6fbb-4301-a4ce-34ddad011ee2\" (UID: \"c66a8ab1-6fbb-4301-a4ce-34ddad011ee2\") " Feb 02 09:35:09 crc kubenswrapper[4764]: I0202 09:35:09.860754 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c66a8ab1-6fbb-4301-a4ce-34ddad011ee2-ssh-key-openstack-edpm-ipam\") pod \"c66a8ab1-6fbb-4301-a4ce-34ddad011ee2\" (UID: \"c66a8ab1-6fbb-4301-a4ce-34ddad011ee2\") " Feb 02 09:35:09 crc kubenswrapper[4764]: I0202 09:35:09.868731 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c66a8ab1-6fbb-4301-a4ce-34ddad011ee2-kube-api-access-95h4g" (OuterVolumeSpecName: "kube-api-access-95h4g") pod "c66a8ab1-6fbb-4301-a4ce-34ddad011ee2" (UID: "c66a8ab1-6fbb-4301-a4ce-34ddad011ee2"). InnerVolumeSpecName "kube-api-access-95h4g". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:35:09 crc kubenswrapper[4764]: I0202 09:35:09.894848 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c66a8ab1-6fbb-4301-a4ce-34ddad011ee2-inventory" (OuterVolumeSpecName: "inventory") pod "c66a8ab1-6fbb-4301-a4ce-34ddad011ee2" (UID: "c66a8ab1-6fbb-4301-a4ce-34ddad011ee2"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:35:09 crc kubenswrapper[4764]: I0202 09:35:09.897151 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c66a8ab1-6fbb-4301-a4ce-34ddad011ee2-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "c66a8ab1-6fbb-4301-a4ce-34ddad011ee2" (UID: "c66a8ab1-6fbb-4301-a4ce-34ddad011ee2"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:35:09 crc kubenswrapper[4764]: I0202 09:35:09.963100 4764 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c66a8ab1-6fbb-4301-a4ce-34ddad011ee2-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 02 09:35:09 crc kubenswrapper[4764]: I0202 09:35:09.963350 4764 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c66a8ab1-6fbb-4301-a4ce-34ddad011ee2-inventory\") on node \"crc\" DevicePath \"\"" Feb 02 09:35:09 crc kubenswrapper[4764]: I0202 09:35:09.963444 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-95h4g\" (UniqueName: \"kubernetes.io/projected/c66a8ab1-6fbb-4301-a4ce-34ddad011ee2-kube-api-access-95h4g\") on node \"crc\" DevicePath \"\"" Feb 02 09:35:10 crc kubenswrapper[4764]: I0202 09:35:10.238041 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-ck2zz" event={"ID":"c66a8ab1-6fbb-4301-a4ce-34ddad011ee2","Type":"ContainerDied","Data":"0315669b91b1605e6e0723ffc66351deaa1db1514dd2c338d8614856fb261256"} Feb 02 09:35:10 crc kubenswrapper[4764]: I0202 09:35:10.238076 4764 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0315669b91b1605e6e0723ffc66351deaa1db1514dd2c338d8614856fb261256" Feb 02 09:35:10 crc kubenswrapper[4764]: I0202 09:35:10.238323 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-ck2zz" Feb 02 09:35:10 crc kubenswrapper[4764]: I0202 09:35:10.328898 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-j7qmc"] Feb 02 09:35:10 crc kubenswrapper[4764]: E0202 09:35:10.329387 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c66a8ab1-6fbb-4301-a4ce-34ddad011ee2" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Feb 02 09:35:10 crc kubenswrapper[4764]: I0202 09:35:10.329424 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="c66a8ab1-6fbb-4301-a4ce-34ddad011ee2" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Feb 02 09:35:10 crc kubenswrapper[4764]: I0202 09:35:10.329651 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="c66a8ab1-6fbb-4301-a4ce-34ddad011ee2" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Feb 02 09:35:10 crc kubenswrapper[4764]: I0202 09:35:10.330440 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-j7qmc" Feb 02 09:35:10 crc kubenswrapper[4764]: I0202 09:35:10.333139 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 02 09:35:10 crc kubenswrapper[4764]: I0202 09:35:10.333246 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-xksq5" Feb 02 09:35:10 crc kubenswrapper[4764]: I0202 09:35:10.333469 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 02 09:35:10 crc kubenswrapper[4764]: I0202 09:35:10.341957 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 02 09:35:10 crc kubenswrapper[4764]: I0202 09:35:10.346586 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-j7qmc"] Feb 02 09:35:10 crc kubenswrapper[4764]: I0202 09:35:10.474572 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/82028751-8b0c-4bab-b6a1-2542acd6e650-ssh-key-openstack-edpm-ipam\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-j7qmc\" (UID: \"82028751-8b0c-4bab-b6a1-2542acd6e650\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-j7qmc" Feb 02 09:35:10 crc kubenswrapper[4764]: I0202 09:35:10.475050 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jr59m\" (UniqueName: \"kubernetes.io/projected/82028751-8b0c-4bab-b6a1-2542acd6e650-kube-api-access-jr59m\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-j7qmc\" (UID: \"82028751-8b0c-4bab-b6a1-2542acd6e650\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-j7qmc" Feb 02 09:35:10 crc kubenswrapper[4764]: I0202 09:35:10.475196 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/82028751-8b0c-4bab-b6a1-2542acd6e650-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-j7qmc\" (UID: \"82028751-8b0c-4bab-b6a1-2542acd6e650\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-j7qmc" Feb 02 09:35:10 crc kubenswrapper[4764]: I0202 09:35:10.577173 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/82028751-8b0c-4bab-b6a1-2542acd6e650-ssh-key-openstack-edpm-ipam\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-j7qmc\" (UID: \"82028751-8b0c-4bab-b6a1-2542acd6e650\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-j7qmc" Feb 02 09:35:10 crc kubenswrapper[4764]: I0202 09:35:10.577227 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jr59m\" (UniqueName: \"kubernetes.io/projected/82028751-8b0c-4bab-b6a1-2542acd6e650-kube-api-access-jr59m\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-j7qmc\" (UID: \"82028751-8b0c-4bab-b6a1-2542acd6e650\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-j7qmc" Feb 02 09:35:10 crc kubenswrapper[4764]: I0202 09:35:10.577302 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/82028751-8b0c-4bab-b6a1-2542acd6e650-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-j7qmc\" (UID: \"82028751-8b0c-4bab-b6a1-2542acd6e650\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-j7qmc" Feb 02 09:35:10 crc kubenswrapper[4764]: I0202 09:35:10.583631 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/82028751-8b0c-4bab-b6a1-2542acd6e650-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-j7qmc\" (UID: \"82028751-8b0c-4bab-b6a1-2542acd6e650\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-j7qmc" Feb 02 09:35:10 crc kubenswrapper[4764]: I0202 09:35:10.583690 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/82028751-8b0c-4bab-b6a1-2542acd6e650-ssh-key-openstack-edpm-ipam\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-j7qmc\" (UID: \"82028751-8b0c-4bab-b6a1-2542acd6e650\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-j7qmc" Feb 02 09:35:10 crc kubenswrapper[4764]: I0202 09:35:10.593319 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jr59m\" (UniqueName: \"kubernetes.io/projected/82028751-8b0c-4bab-b6a1-2542acd6e650-kube-api-access-jr59m\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-j7qmc\" (UID: \"82028751-8b0c-4bab-b6a1-2542acd6e650\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-j7qmc" Feb 02 09:35:10 crc kubenswrapper[4764]: I0202 09:35:10.674096 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-j7qmc" Feb 02 09:35:11 crc kubenswrapper[4764]: W0202 09:35:11.224081 4764 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod82028751_8b0c_4bab_b6a1_2542acd6e650.slice/crio-9e6e6740bdf5b7e924c4d581a32641349669fa85035e6bfb172afb3c9a085a62 WatchSource:0}: Error finding container 9e6e6740bdf5b7e924c4d581a32641349669fa85035e6bfb172afb3c9a085a62: Status 404 returned error can't find the container with id 9e6e6740bdf5b7e924c4d581a32641349669fa85035e6bfb172afb3c9a085a62 Feb 02 09:35:11 crc kubenswrapper[4764]: I0202 09:35:11.232327 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-j7qmc"] Feb 02 09:35:11 crc kubenswrapper[4764]: I0202 09:35:11.249576 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-j7qmc" event={"ID":"82028751-8b0c-4bab-b6a1-2542acd6e650","Type":"ContainerStarted","Data":"9e6e6740bdf5b7e924c4d581a32641349669fa85035e6bfb172afb3c9a085a62"} Feb 02 09:35:12 crc kubenswrapper[4764]: I0202 09:35:12.259492 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-j7qmc" event={"ID":"82028751-8b0c-4bab-b6a1-2542acd6e650","Type":"ContainerStarted","Data":"3dbab190576e3e5b1dd4bf491ee4ec587fedd9b0a05df873f446d73c04191ead"} Feb 02 09:35:12 crc kubenswrapper[4764]: I0202 09:35:12.295385 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-j7qmc" podStartSLOduration=1.871713845 podStartE2EDuration="2.295369636s" podCreationTimestamp="2026-02-02 09:35:10 +0000 UTC" firstStartedPulling="2026-02-02 09:35:11.227556066 +0000 UTC m=+1674.161280154" lastFinishedPulling="2026-02-02 09:35:11.651211847 +0000 UTC m=+1674.584935945" observedRunningTime="2026-02-02 09:35:12.286259659 +0000 UTC m=+1675.219983777" watchObservedRunningTime="2026-02-02 09:35:12.295369636 +0000 UTC m=+1675.229093724" Feb 02 09:35:13 crc kubenswrapper[4764]: I0202 09:35:13.825737 4764 scope.go:117] "RemoveContainer" containerID="6db43ab36406cdd4dbe61d7d8688257aba4271b8601a5ba5594675ae309c761d" Feb 02 09:35:13 crc kubenswrapper[4764]: E0202 09:35:13.826003 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 09:35:16 crc kubenswrapper[4764]: I0202 09:35:16.312652 4764 generic.go:334] "Generic (PLEG): container finished" podID="82028751-8b0c-4bab-b6a1-2542acd6e650" containerID="3dbab190576e3e5b1dd4bf491ee4ec587fedd9b0a05df873f446d73c04191ead" exitCode=0 Feb 02 09:35:16 crc kubenswrapper[4764]: I0202 09:35:16.313334 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-j7qmc" event={"ID":"82028751-8b0c-4bab-b6a1-2542acd6e650","Type":"ContainerDied","Data":"3dbab190576e3e5b1dd4bf491ee4ec587fedd9b0a05df873f446d73c04191ead"} Feb 02 09:35:17 crc kubenswrapper[4764]: I0202 09:35:17.749547 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-j7qmc" Feb 02 09:35:17 crc kubenswrapper[4764]: I0202 09:35:17.844414 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jr59m\" (UniqueName: \"kubernetes.io/projected/82028751-8b0c-4bab-b6a1-2542acd6e650-kube-api-access-jr59m\") pod \"82028751-8b0c-4bab-b6a1-2542acd6e650\" (UID: \"82028751-8b0c-4bab-b6a1-2542acd6e650\") " Feb 02 09:35:17 crc kubenswrapper[4764]: I0202 09:35:17.844830 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/82028751-8b0c-4bab-b6a1-2542acd6e650-inventory\") pod \"82028751-8b0c-4bab-b6a1-2542acd6e650\" (UID: \"82028751-8b0c-4bab-b6a1-2542acd6e650\") " Feb 02 09:35:17 crc kubenswrapper[4764]: I0202 09:35:17.844915 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/82028751-8b0c-4bab-b6a1-2542acd6e650-ssh-key-openstack-edpm-ipam\") pod \"82028751-8b0c-4bab-b6a1-2542acd6e650\" (UID: \"82028751-8b0c-4bab-b6a1-2542acd6e650\") " Feb 02 09:35:17 crc kubenswrapper[4764]: I0202 09:35:17.861454 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/82028751-8b0c-4bab-b6a1-2542acd6e650-kube-api-access-jr59m" (OuterVolumeSpecName: "kube-api-access-jr59m") pod "82028751-8b0c-4bab-b6a1-2542acd6e650" (UID: "82028751-8b0c-4bab-b6a1-2542acd6e650"). InnerVolumeSpecName "kube-api-access-jr59m". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:35:17 crc kubenswrapper[4764]: I0202 09:35:17.874602 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82028751-8b0c-4bab-b6a1-2542acd6e650-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "82028751-8b0c-4bab-b6a1-2542acd6e650" (UID: "82028751-8b0c-4bab-b6a1-2542acd6e650"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:35:17 crc kubenswrapper[4764]: I0202 09:35:17.877202 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82028751-8b0c-4bab-b6a1-2542acd6e650-inventory" (OuterVolumeSpecName: "inventory") pod "82028751-8b0c-4bab-b6a1-2542acd6e650" (UID: "82028751-8b0c-4bab-b6a1-2542acd6e650"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:35:17 crc kubenswrapper[4764]: I0202 09:35:17.946395 4764 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/82028751-8b0c-4bab-b6a1-2542acd6e650-inventory\") on node \"crc\" DevicePath \"\"" Feb 02 09:35:17 crc kubenswrapper[4764]: I0202 09:35:17.947194 4764 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/82028751-8b0c-4bab-b6a1-2542acd6e650-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 02 09:35:17 crc kubenswrapper[4764]: I0202 09:35:17.947322 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jr59m\" (UniqueName: \"kubernetes.io/projected/82028751-8b0c-4bab-b6a1-2542acd6e650-kube-api-access-jr59m\") on node \"crc\" DevicePath \"\"" Feb 02 09:35:18 crc kubenswrapper[4764]: I0202 09:35:18.340316 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-j7qmc" event={"ID":"82028751-8b0c-4bab-b6a1-2542acd6e650","Type":"ContainerDied","Data":"9e6e6740bdf5b7e924c4d581a32641349669fa85035e6bfb172afb3c9a085a62"} Feb 02 09:35:18 crc kubenswrapper[4764]: I0202 09:35:18.340368 4764 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9e6e6740bdf5b7e924c4d581a32641349669fa85035e6bfb172afb3c9a085a62" Feb 02 09:35:18 crc kubenswrapper[4764]: I0202 09:35:18.340982 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-j7qmc" Feb 02 09:35:18 crc kubenswrapper[4764]: I0202 09:35:18.425808 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-8rfqp"] Feb 02 09:35:18 crc kubenswrapper[4764]: E0202 09:35:18.426435 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82028751-8b0c-4bab-b6a1-2542acd6e650" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Feb 02 09:35:18 crc kubenswrapper[4764]: I0202 09:35:18.426451 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="82028751-8b0c-4bab-b6a1-2542acd6e650" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Feb 02 09:35:18 crc kubenswrapper[4764]: I0202 09:35:18.426620 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="82028751-8b0c-4bab-b6a1-2542acd6e650" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Feb 02 09:35:18 crc kubenswrapper[4764]: I0202 09:35:18.428247 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-8rfqp" Feb 02 09:35:18 crc kubenswrapper[4764]: I0202 09:35:18.430913 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 02 09:35:18 crc kubenswrapper[4764]: I0202 09:35:18.431151 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 02 09:35:18 crc kubenswrapper[4764]: I0202 09:35:18.431467 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 02 09:35:18 crc kubenswrapper[4764]: I0202 09:35:18.431985 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-xksq5" Feb 02 09:35:18 crc kubenswrapper[4764]: I0202 09:35:18.444427 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-8rfqp"] Feb 02 09:35:18 crc kubenswrapper[4764]: I0202 09:35:18.461677 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vtqlk\" (UniqueName: \"kubernetes.io/projected/203f8fcc-2055-40f5-8ef6-7d00d658157b-kube-api-access-vtqlk\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-8rfqp\" (UID: \"203f8fcc-2055-40f5-8ef6-7d00d658157b\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-8rfqp" Feb 02 09:35:18 crc kubenswrapper[4764]: I0202 09:35:18.461878 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/203f8fcc-2055-40f5-8ef6-7d00d658157b-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-8rfqp\" (UID: \"203f8fcc-2055-40f5-8ef6-7d00d658157b\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-8rfqp" Feb 02 09:35:18 crc kubenswrapper[4764]: I0202 09:35:18.462007 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/203f8fcc-2055-40f5-8ef6-7d00d658157b-ssh-key-openstack-edpm-ipam\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-8rfqp\" (UID: \"203f8fcc-2055-40f5-8ef6-7d00d658157b\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-8rfqp" Feb 02 09:35:18 crc kubenswrapper[4764]: I0202 09:35:18.562916 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vtqlk\" (UniqueName: \"kubernetes.io/projected/203f8fcc-2055-40f5-8ef6-7d00d658157b-kube-api-access-vtqlk\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-8rfqp\" (UID: \"203f8fcc-2055-40f5-8ef6-7d00d658157b\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-8rfqp" Feb 02 09:35:18 crc kubenswrapper[4764]: I0202 09:35:18.563022 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/203f8fcc-2055-40f5-8ef6-7d00d658157b-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-8rfqp\" (UID: \"203f8fcc-2055-40f5-8ef6-7d00d658157b\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-8rfqp" Feb 02 09:35:18 crc kubenswrapper[4764]: I0202 09:35:18.563078 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/203f8fcc-2055-40f5-8ef6-7d00d658157b-ssh-key-openstack-edpm-ipam\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-8rfqp\" (UID: \"203f8fcc-2055-40f5-8ef6-7d00d658157b\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-8rfqp" Feb 02 09:35:18 crc kubenswrapper[4764]: I0202 09:35:18.566804 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/203f8fcc-2055-40f5-8ef6-7d00d658157b-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-8rfqp\" (UID: \"203f8fcc-2055-40f5-8ef6-7d00d658157b\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-8rfqp" Feb 02 09:35:18 crc kubenswrapper[4764]: I0202 09:35:18.573217 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/203f8fcc-2055-40f5-8ef6-7d00d658157b-ssh-key-openstack-edpm-ipam\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-8rfqp\" (UID: \"203f8fcc-2055-40f5-8ef6-7d00d658157b\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-8rfqp" Feb 02 09:35:18 crc kubenswrapper[4764]: I0202 09:35:18.579099 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vtqlk\" (UniqueName: \"kubernetes.io/projected/203f8fcc-2055-40f5-8ef6-7d00d658157b-kube-api-access-vtqlk\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-8rfqp\" (UID: \"203f8fcc-2055-40f5-8ef6-7d00d658157b\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-8rfqp" Feb 02 09:35:18 crc kubenswrapper[4764]: I0202 09:35:18.748970 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-8rfqp" Feb 02 09:35:19 crc kubenswrapper[4764]: W0202 09:35:19.264453 4764 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod203f8fcc_2055_40f5_8ef6_7d00d658157b.slice/crio-ee6e04d692eccb1aa1ff2aff1cba2254a6f49735cc36d0f36a6be729da23aad9 WatchSource:0}: Error finding container ee6e04d692eccb1aa1ff2aff1cba2254a6f49735cc36d0f36a6be729da23aad9: Status 404 returned error can't find the container with id ee6e04d692eccb1aa1ff2aff1cba2254a6f49735cc36d0f36a6be729da23aad9 Feb 02 09:35:19 crc kubenswrapper[4764]: I0202 09:35:19.268010 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-8rfqp"] Feb 02 09:35:19 crc kubenswrapper[4764]: I0202 09:35:19.348898 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-8rfqp" event={"ID":"203f8fcc-2055-40f5-8ef6-7d00d658157b","Type":"ContainerStarted","Data":"ee6e04d692eccb1aa1ff2aff1cba2254a6f49735cc36d0f36a6be729da23aad9"} Feb 02 09:35:20 crc kubenswrapper[4764]: I0202 09:35:20.363903 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-8rfqp" event={"ID":"203f8fcc-2055-40f5-8ef6-7d00d658157b","Type":"ContainerStarted","Data":"6940e5f96de57f57e75f49942e77472a14857af61522b9d9c0e931feb7e6a97e"} Feb 02 09:35:20 crc kubenswrapper[4764]: I0202 09:35:20.397177 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-8rfqp" podStartSLOduration=1.87673553 podStartE2EDuration="2.397150379s" podCreationTimestamp="2026-02-02 09:35:18 +0000 UTC" firstStartedPulling="2026-02-02 09:35:19.266889312 +0000 UTC m=+1682.200613440" lastFinishedPulling="2026-02-02 09:35:19.787304161 +0000 UTC m=+1682.721028289" observedRunningTime="2026-02-02 09:35:20.389010118 +0000 UTC m=+1683.322734276" watchObservedRunningTime="2026-02-02 09:35:20.397150379 +0000 UTC m=+1683.330874497" Feb 02 09:35:23 crc kubenswrapper[4764]: I0202 09:35:23.050271 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-p7g2s"] Feb 02 09:35:23 crc kubenswrapper[4764]: I0202 09:35:23.064449 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-p7g2s"] Feb 02 09:35:23 crc kubenswrapper[4764]: I0202 09:35:23.845434 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fd30e6c5-1c0f-4ee4-9df1-c23b745ec136" path="/var/lib/kubelet/pods/fd30e6c5-1c0f-4ee4-9df1-c23b745ec136/volumes" Feb 02 09:35:25 crc kubenswrapper[4764]: I0202 09:35:25.351197 4764 scope.go:117] "RemoveContainer" containerID="c626eaf56455d96ba04df17db01428f27600844910ce734771f5704cb5969915" Feb 02 09:35:25 crc kubenswrapper[4764]: I0202 09:35:25.384126 4764 scope.go:117] "RemoveContainer" containerID="af83f130e1580a8eb2b2162ba0adf21c90b40709512ef5e33a6c9de59adaa512" Feb 02 09:35:25 crc kubenswrapper[4764]: I0202 09:35:25.462369 4764 scope.go:117] "RemoveContainer" containerID="2bafaf1ec4086fb317564b3ab7647ca03464edcb1c64eb49e00c153bf3d7d09f" Feb 02 09:35:25 crc kubenswrapper[4764]: I0202 09:35:25.487915 4764 scope.go:117] "RemoveContainer" containerID="adb69964a6b9f244142facf775fd53dfc1a686481ec2d2b42286a7edf9112752" Feb 02 09:35:25 crc kubenswrapper[4764]: I0202 09:35:25.535926 4764 scope.go:117] "RemoveContainer" containerID="5543885f98fc6fc86034790d4953d84f4d1c24207aeca5e8f52d8e0eb431f891" Feb 02 09:35:25 crc kubenswrapper[4764]: I0202 09:35:25.565144 4764 scope.go:117] "RemoveContainer" containerID="0a9aa80c6f0babc8f62ada7c7954125d50f6816eafcd5d7a65e1b84be7a050bd" Feb 02 09:35:25 crc kubenswrapper[4764]: I0202 09:35:25.598919 4764 scope.go:117] "RemoveContainer" containerID="f10b287949161d05b18bc14b4c82c7334d4201f0215296bfcdbf79060570f7c7" Feb 02 09:35:25 crc kubenswrapper[4764]: I0202 09:35:25.617327 4764 scope.go:117] "RemoveContainer" containerID="3cfba45ffa4b9101660134caf144daceeb7817ff8fbd85821de2c11047da9cc9" Feb 02 09:35:25 crc kubenswrapper[4764]: I0202 09:35:25.636374 4764 scope.go:117] "RemoveContainer" containerID="c48c0eb8f08401abc4475fcbb3a0ef88bad3296f2474f680e07a7865ce799f47" Feb 02 09:35:27 crc kubenswrapper[4764]: I0202 09:35:27.840842 4764 scope.go:117] "RemoveContainer" containerID="6db43ab36406cdd4dbe61d7d8688257aba4271b8601a5ba5594675ae309c761d" Feb 02 09:35:27 crc kubenswrapper[4764]: E0202 09:35:27.841639 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 09:35:32 crc kubenswrapper[4764]: I0202 09:35:32.065832 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-229dq"] Feb 02 09:35:32 crc kubenswrapper[4764]: I0202 09:35:32.078751 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-sm8gf"] Feb 02 09:35:32 crc kubenswrapper[4764]: I0202 09:35:32.087360 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-229dq"] Feb 02 09:35:32 crc kubenswrapper[4764]: I0202 09:35:32.097251 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-sm8gf"] Feb 02 09:35:33 crc kubenswrapper[4764]: I0202 09:35:33.836651 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2f7f0eb3-55c5-4dbc-8636-eba9a31b0a27" path="/var/lib/kubelet/pods/2f7f0eb3-55c5-4dbc-8636-eba9a31b0a27/volumes" Feb 02 09:35:33 crc kubenswrapper[4764]: I0202 09:35:33.837273 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e61fa54a-f23a-409c-923f-522e04a8099f" path="/var/lib/kubelet/pods/e61fa54a-f23a-409c-923f-522e04a8099f/volumes" Feb 02 09:35:36 crc kubenswrapper[4764]: I0202 09:35:36.036730 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-phkvk"] Feb 02 09:35:36 crc kubenswrapper[4764]: I0202 09:35:36.047674 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-phkvk"] Feb 02 09:35:37 crc kubenswrapper[4764]: I0202 09:35:37.837545 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4537c8c1-f5dc-4974-9885-527d6c85e100" path="/var/lib/kubelet/pods/4537c8c1-f5dc-4974-9885-527d6c85e100/volumes" Feb 02 09:35:38 crc kubenswrapper[4764]: I0202 09:35:38.825681 4764 scope.go:117] "RemoveContainer" containerID="6db43ab36406cdd4dbe61d7d8688257aba4271b8601a5ba5594675ae309c761d" Feb 02 09:35:38 crc kubenswrapper[4764]: E0202 09:35:38.826227 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 09:35:50 crc kubenswrapper[4764]: I0202 09:35:50.044762 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-dkrnx"] Feb 02 09:35:50 crc kubenswrapper[4764]: I0202 09:35:50.061817 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-dkrnx"] Feb 02 09:35:51 crc kubenswrapper[4764]: I0202 09:35:51.838911 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6b3d3314-2f16-4dc3-aa52-5f7f8576813a" path="/var/lib/kubelet/pods/6b3d3314-2f16-4dc3-aa52-5f7f8576813a/volumes" Feb 02 09:35:53 crc kubenswrapper[4764]: I0202 09:35:53.826162 4764 scope.go:117] "RemoveContainer" containerID="6db43ab36406cdd4dbe61d7d8688257aba4271b8601a5ba5594675ae309c761d" Feb 02 09:35:53 crc kubenswrapper[4764]: E0202 09:35:53.826664 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 09:36:04 crc kubenswrapper[4764]: I0202 09:36:04.825998 4764 scope.go:117] "RemoveContainer" containerID="6db43ab36406cdd4dbe61d7d8688257aba4271b8601a5ba5594675ae309c761d" Feb 02 09:36:04 crc kubenswrapper[4764]: E0202 09:36:04.826784 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 09:36:10 crc kubenswrapper[4764]: I0202 09:36:10.820511 4764 generic.go:334] "Generic (PLEG): container finished" podID="203f8fcc-2055-40f5-8ef6-7d00d658157b" containerID="6940e5f96de57f57e75f49942e77472a14857af61522b9d9c0e931feb7e6a97e" exitCode=0 Feb 02 09:36:10 crc kubenswrapper[4764]: I0202 09:36:10.820669 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-8rfqp" event={"ID":"203f8fcc-2055-40f5-8ef6-7d00d658157b","Type":"ContainerDied","Data":"6940e5f96de57f57e75f49942e77472a14857af61522b9d9c0e931feb7e6a97e"} Feb 02 09:36:12 crc kubenswrapper[4764]: I0202 09:36:12.230074 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-8rfqp" Feb 02 09:36:12 crc kubenswrapper[4764]: I0202 09:36:12.407394 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vtqlk\" (UniqueName: \"kubernetes.io/projected/203f8fcc-2055-40f5-8ef6-7d00d658157b-kube-api-access-vtqlk\") pod \"203f8fcc-2055-40f5-8ef6-7d00d658157b\" (UID: \"203f8fcc-2055-40f5-8ef6-7d00d658157b\") " Feb 02 09:36:12 crc kubenswrapper[4764]: I0202 09:36:12.407512 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/203f8fcc-2055-40f5-8ef6-7d00d658157b-inventory\") pod \"203f8fcc-2055-40f5-8ef6-7d00d658157b\" (UID: \"203f8fcc-2055-40f5-8ef6-7d00d658157b\") " Feb 02 09:36:12 crc kubenswrapper[4764]: I0202 09:36:12.407754 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/203f8fcc-2055-40f5-8ef6-7d00d658157b-ssh-key-openstack-edpm-ipam\") pod \"203f8fcc-2055-40f5-8ef6-7d00d658157b\" (UID: \"203f8fcc-2055-40f5-8ef6-7d00d658157b\") " Feb 02 09:36:12 crc kubenswrapper[4764]: I0202 09:36:12.419171 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/203f8fcc-2055-40f5-8ef6-7d00d658157b-kube-api-access-vtqlk" (OuterVolumeSpecName: "kube-api-access-vtqlk") pod "203f8fcc-2055-40f5-8ef6-7d00d658157b" (UID: "203f8fcc-2055-40f5-8ef6-7d00d658157b"). InnerVolumeSpecName "kube-api-access-vtqlk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:36:12 crc kubenswrapper[4764]: I0202 09:36:12.440396 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/203f8fcc-2055-40f5-8ef6-7d00d658157b-inventory" (OuterVolumeSpecName: "inventory") pod "203f8fcc-2055-40f5-8ef6-7d00d658157b" (UID: "203f8fcc-2055-40f5-8ef6-7d00d658157b"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:36:12 crc kubenswrapper[4764]: I0202 09:36:12.510084 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vtqlk\" (UniqueName: \"kubernetes.io/projected/203f8fcc-2055-40f5-8ef6-7d00d658157b-kube-api-access-vtqlk\") on node \"crc\" DevicePath \"\"" Feb 02 09:36:12 crc kubenswrapper[4764]: I0202 09:36:12.510109 4764 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/203f8fcc-2055-40f5-8ef6-7d00d658157b-inventory\") on node \"crc\" DevicePath \"\"" Feb 02 09:36:12 crc kubenswrapper[4764]: I0202 09:36:12.630326 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/203f8fcc-2055-40f5-8ef6-7d00d658157b-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "203f8fcc-2055-40f5-8ef6-7d00d658157b" (UID: "203f8fcc-2055-40f5-8ef6-7d00d658157b"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:36:12 crc kubenswrapper[4764]: I0202 09:36:12.714276 4764 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/203f8fcc-2055-40f5-8ef6-7d00d658157b-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 02 09:36:12 crc kubenswrapper[4764]: I0202 09:36:12.848214 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-8rfqp" event={"ID":"203f8fcc-2055-40f5-8ef6-7d00d658157b","Type":"ContainerDied","Data":"ee6e04d692eccb1aa1ff2aff1cba2254a6f49735cc36d0f36a6be729da23aad9"} Feb 02 09:36:12 crc kubenswrapper[4764]: I0202 09:36:12.848254 4764 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ee6e04d692eccb1aa1ff2aff1cba2254a6f49735cc36d0f36a6be729da23aad9" Feb 02 09:36:12 crc kubenswrapper[4764]: I0202 09:36:12.848305 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-8rfqp" Feb 02 09:36:12 crc kubenswrapper[4764]: I0202 09:36:12.948587 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-rcpbs"] Feb 02 09:36:12 crc kubenswrapper[4764]: E0202 09:36:12.949201 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="203f8fcc-2055-40f5-8ef6-7d00d658157b" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Feb 02 09:36:12 crc kubenswrapper[4764]: I0202 09:36:12.949275 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="203f8fcc-2055-40f5-8ef6-7d00d658157b" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Feb 02 09:36:12 crc kubenswrapper[4764]: I0202 09:36:12.949503 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="203f8fcc-2055-40f5-8ef6-7d00d658157b" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Feb 02 09:36:12 crc kubenswrapper[4764]: I0202 09:36:12.950111 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-rcpbs" Feb 02 09:36:12 crc kubenswrapper[4764]: I0202 09:36:12.961673 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 02 09:36:12 crc kubenswrapper[4764]: I0202 09:36:12.961972 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 02 09:36:12 crc kubenswrapper[4764]: I0202 09:36:12.965314 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 02 09:36:12 crc kubenswrapper[4764]: I0202 09:36:12.966815 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-xksq5" Feb 02 09:36:12 crc kubenswrapper[4764]: I0202 09:36:12.983429 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-rcpbs"] Feb 02 09:36:13 crc kubenswrapper[4764]: I0202 09:36:13.021054 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/9b9af3ad-4a64-4996-bcdc-220df06db737-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-rcpbs\" (UID: \"9b9af3ad-4a64-4996-bcdc-220df06db737\") " pod="openstack/ssh-known-hosts-edpm-deployment-rcpbs" Feb 02 09:36:13 crc kubenswrapper[4764]: I0202 09:36:13.021163 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-24gg4\" (UniqueName: \"kubernetes.io/projected/9b9af3ad-4a64-4996-bcdc-220df06db737-kube-api-access-24gg4\") pod \"ssh-known-hosts-edpm-deployment-rcpbs\" (UID: \"9b9af3ad-4a64-4996-bcdc-220df06db737\") " pod="openstack/ssh-known-hosts-edpm-deployment-rcpbs" Feb 02 09:36:13 crc kubenswrapper[4764]: I0202 09:36:13.021257 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/9b9af3ad-4a64-4996-bcdc-220df06db737-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-rcpbs\" (UID: \"9b9af3ad-4a64-4996-bcdc-220df06db737\") " pod="openstack/ssh-known-hosts-edpm-deployment-rcpbs" Feb 02 09:36:13 crc kubenswrapper[4764]: I0202 09:36:13.123525 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/9b9af3ad-4a64-4996-bcdc-220df06db737-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-rcpbs\" (UID: \"9b9af3ad-4a64-4996-bcdc-220df06db737\") " pod="openstack/ssh-known-hosts-edpm-deployment-rcpbs" Feb 02 09:36:13 crc kubenswrapper[4764]: I0202 09:36:13.123596 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-24gg4\" (UniqueName: \"kubernetes.io/projected/9b9af3ad-4a64-4996-bcdc-220df06db737-kube-api-access-24gg4\") pod \"ssh-known-hosts-edpm-deployment-rcpbs\" (UID: \"9b9af3ad-4a64-4996-bcdc-220df06db737\") " pod="openstack/ssh-known-hosts-edpm-deployment-rcpbs" Feb 02 09:36:13 crc kubenswrapper[4764]: I0202 09:36:13.123669 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/9b9af3ad-4a64-4996-bcdc-220df06db737-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-rcpbs\" (UID: \"9b9af3ad-4a64-4996-bcdc-220df06db737\") " pod="openstack/ssh-known-hosts-edpm-deployment-rcpbs" Feb 02 09:36:13 crc kubenswrapper[4764]: I0202 09:36:13.129257 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/9b9af3ad-4a64-4996-bcdc-220df06db737-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-rcpbs\" (UID: \"9b9af3ad-4a64-4996-bcdc-220df06db737\") " pod="openstack/ssh-known-hosts-edpm-deployment-rcpbs" Feb 02 09:36:13 crc kubenswrapper[4764]: I0202 09:36:13.129293 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/9b9af3ad-4a64-4996-bcdc-220df06db737-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-rcpbs\" (UID: \"9b9af3ad-4a64-4996-bcdc-220df06db737\") " pod="openstack/ssh-known-hosts-edpm-deployment-rcpbs" Feb 02 09:36:13 crc kubenswrapper[4764]: I0202 09:36:13.156608 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-24gg4\" (UniqueName: \"kubernetes.io/projected/9b9af3ad-4a64-4996-bcdc-220df06db737-kube-api-access-24gg4\") pod \"ssh-known-hosts-edpm-deployment-rcpbs\" (UID: \"9b9af3ad-4a64-4996-bcdc-220df06db737\") " pod="openstack/ssh-known-hosts-edpm-deployment-rcpbs" Feb 02 09:36:13 crc kubenswrapper[4764]: I0202 09:36:13.269060 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-rcpbs" Feb 02 09:36:13 crc kubenswrapper[4764]: I0202 09:36:13.884174 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-rcpbs"] Feb 02 09:36:14 crc kubenswrapper[4764]: I0202 09:36:14.890165 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-rcpbs" event={"ID":"9b9af3ad-4a64-4996-bcdc-220df06db737","Type":"ContainerStarted","Data":"2a30b88b0273d89dc757e5d0af4296f1781e0e52d78980006db3cb7f8b222a2a"} Feb 02 09:36:14 crc kubenswrapper[4764]: I0202 09:36:14.890541 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-rcpbs" event={"ID":"9b9af3ad-4a64-4996-bcdc-220df06db737","Type":"ContainerStarted","Data":"f1ce82e3deea0dc4ff29de926468e7cf74fe0da30b67ef0981bb0aeb81698433"} Feb 02 09:36:14 crc kubenswrapper[4764]: I0202 09:36:14.911084 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-rcpbs" podStartSLOduration=2.420451263 podStartE2EDuration="2.911057383s" podCreationTimestamp="2026-02-02 09:36:12 +0000 UTC" firstStartedPulling="2026-02-02 09:36:13.871997482 +0000 UTC m=+1736.805721570" lastFinishedPulling="2026-02-02 09:36:14.362603602 +0000 UTC m=+1737.296327690" observedRunningTime="2026-02-02 09:36:14.909517911 +0000 UTC m=+1737.843242019" watchObservedRunningTime="2026-02-02 09:36:14.911057383 +0000 UTC m=+1737.844781511" Feb 02 09:36:15 crc kubenswrapper[4764]: I0202 09:36:15.826839 4764 scope.go:117] "RemoveContainer" containerID="6db43ab36406cdd4dbe61d7d8688257aba4271b8601a5ba5594675ae309c761d" Feb 02 09:36:15 crc kubenswrapper[4764]: E0202 09:36:15.827335 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 09:36:21 crc kubenswrapper[4764]: I0202 09:36:21.946809 4764 generic.go:334] "Generic (PLEG): container finished" podID="9b9af3ad-4a64-4996-bcdc-220df06db737" containerID="2a30b88b0273d89dc757e5d0af4296f1781e0e52d78980006db3cb7f8b222a2a" exitCode=0 Feb 02 09:36:21 crc kubenswrapper[4764]: I0202 09:36:21.947100 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-rcpbs" event={"ID":"9b9af3ad-4a64-4996-bcdc-220df06db737","Type":"ContainerDied","Data":"2a30b88b0273d89dc757e5d0af4296f1781e0e52d78980006db3cb7f8b222a2a"} Feb 02 09:36:23 crc kubenswrapper[4764]: I0202 09:36:23.377241 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-rcpbs" Feb 02 09:36:23 crc kubenswrapper[4764]: I0202 09:36:23.517463 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/9b9af3ad-4a64-4996-bcdc-220df06db737-inventory-0\") pod \"9b9af3ad-4a64-4996-bcdc-220df06db737\" (UID: \"9b9af3ad-4a64-4996-bcdc-220df06db737\") " Feb 02 09:36:23 crc kubenswrapper[4764]: I0202 09:36:23.518161 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/9b9af3ad-4a64-4996-bcdc-220df06db737-ssh-key-openstack-edpm-ipam\") pod \"9b9af3ad-4a64-4996-bcdc-220df06db737\" (UID: \"9b9af3ad-4a64-4996-bcdc-220df06db737\") " Feb 02 09:36:23 crc kubenswrapper[4764]: I0202 09:36:23.518277 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-24gg4\" (UniqueName: \"kubernetes.io/projected/9b9af3ad-4a64-4996-bcdc-220df06db737-kube-api-access-24gg4\") pod \"9b9af3ad-4a64-4996-bcdc-220df06db737\" (UID: \"9b9af3ad-4a64-4996-bcdc-220df06db737\") " Feb 02 09:36:23 crc kubenswrapper[4764]: I0202 09:36:23.522595 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9b9af3ad-4a64-4996-bcdc-220df06db737-kube-api-access-24gg4" (OuterVolumeSpecName: "kube-api-access-24gg4") pod "9b9af3ad-4a64-4996-bcdc-220df06db737" (UID: "9b9af3ad-4a64-4996-bcdc-220df06db737"). InnerVolumeSpecName "kube-api-access-24gg4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:36:23 crc kubenswrapper[4764]: I0202 09:36:23.547135 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9b9af3ad-4a64-4996-bcdc-220df06db737-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "9b9af3ad-4a64-4996-bcdc-220df06db737" (UID: "9b9af3ad-4a64-4996-bcdc-220df06db737"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:36:23 crc kubenswrapper[4764]: I0202 09:36:23.550713 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9b9af3ad-4a64-4996-bcdc-220df06db737-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "9b9af3ad-4a64-4996-bcdc-220df06db737" (UID: "9b9af3ad-4a64-4996-bcdc-220df06db737"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:36:23 crc kubenswrapper[4764]: I0202 09:36:23.620826 4764 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/9b9af3ad-4a64-4996-bcdc-220df06db737-inventory-0\") on node \"crc\" DevicePath \"\"" Feb 02 09:36:23 crc kubenswrapper[4764]: I0202 09:36:23.620860 4764 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/9b9af3ad-4a64-4996-bcdc-220df06db737-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 02 09:36:23 crc kubenswrapper[4764]: I0202 09:36:23.620873 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-24gg4\" (UniqueName: \"kubernetes.io/projected/9b9af3ad-4a64-4996-bcdc-220df06db737-kube-api-access-24gg4\") on node \"crc\" DevicePath \"\"" Feb 02 09:36:23 crc kubenswrapper[4764]: I0202 09:36:23.971983 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-rcpbs" event={"ID":"9b9af3ad-4a64-4996-bcdc-220df06db737","Type":"ContainerDied","Data":"f1ce82e3deea0dc4ff29de926468e7cf74fe0da30b67ef0981bb0aeb81698433"} Feb 02 09:36:23 crc kubenswrapper[4764]: I0202 09:36:23.972040 4764 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f1ce82e3deea0dc4ff29de926468e7cf74fe0da30b67ef0981bb0aeb81698433" Feb 02 09:36:23 crc kubenswrapper[4764]: I0202 09:36:23.972112 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-rcpbs" Feb 02 09:36:24 crc kubenswrapper[4764]: I0202 09:36:24.057159 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-gklb7"] Feb 02 09:36:24 crc kubenswrapper[4764]: E0202 09:36:24.057578 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b9af3ad-4a64-4996-bcdc-220df06db737" containerName="ssh-known-hosts-edpm-deployment" Feb 02 09:36:24 crc kubenswrapper[4764]: I0202 09:36:24.057599 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b9af3ad-4a64-4996-bcdc-220df06db737" containerName="ssh-known-hosts-edpm-deployment" Feb 02 09:36:24 crc kubenswrapper[4764]: I0202 09:36:24.057828 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b9af3ad-4a64-4996-bcdc-220df06db737" containerName="ssh-known-hosts-edpm-deployment" Feb 02 09:36:24 crc kubenswrapper[4764]: I0202 09:36:24.058513 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-gklb7" Feb 02 09:36:24 crc kubenswrapper[4764]: I0202 09:36:24.060510 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 02 09:36:24 crc kubenswrapper[4764]: I0202 09:36:24.061904 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 02 09:36:24 crc kubenswrapper[4764]: I0202 09:36:24.062546 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-xksq5" Feb 02 09:36:24 crc kubenswrapper[4764]: I0202 09:36:24.063140 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 02 09:36:24 crc kubenswrapper[4764]: I0202 09:36:24.080154 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-gklb7"] Feb 02 09:36:24 crc kubenswrapper[4764]: I0202 09:36:24.131780 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/502ceb21-d3d1-4638-a565-2223d41fd1c0-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-gklb7\" (UID: \"502ceb21-d3d1-4638-a565-2223d41fd1c0\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-gklb7" Feb 02 09:36:24 crc kubenswrapper[4764]: I0202 09:36:24.131837 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/502ceb21-d3d1-4638-a565-2223d41fd1c0-ssh-key-openstack-edpm-ipam\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-gklb7\" (UID: \"502ceb21-d3d1-4638-a565-2223d41fd1c0\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-gklb7" Feb 02 09:36:24 crc kubenswrapper[4764]: I0202 09:36:24.131903 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8vbpf\" (UniqueName: \"kubernetes.io/projected/502ceb21-d3d1-4638-a565-2223d41fd1c0-kube-api-access-8vbpf\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-gklb7\" (UID: \"502ceb21-d3d1-4638-a565-2223d41fd1c0\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-gklb7" Feb 02 09:36:24 crc kubenswrapper[4764]: I0202 09:36:24.232928 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/502ceb21-d3d1-4638-a565-2223d41fd1c0-ssh-key-openstack-edpm-ipam\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-gklb7\" (UID: \"502ceb21-d3d1-4638-a565-2223d41fd1c0\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-gklb7" Feb 02 09:36:24 crc kubenswrapper[4764]: I0202 09:36:24.233052 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8vbpf\" (UniqueName: \"kubernetes.io/projected/502ceb21-d3d1-4638-a565-2223d41fd1c0-kube-api-access-8vbpf\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-gklb7\" (UID: \"502ceb21-d3d1-4638-a565-2223d41fd1c0\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-gklb7" Feb 02 09:36:24 crc kubenswrapper[4764]: I0202 09:36:24.233413 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/502ceb21-d3d1-4638-a565-2223d41fd1c0-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-gklb7\" (UID: \"502ceb21-d3d1-4638-a565-2223d41fd1c0\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-gklb7" Feb 02 09:36:24 crc kubenswrapper[4764]: I0202 09:36:24.237259 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/502ceb21-d3d1-4638-a565-2223d41fd1c0-ssh-key-openstack-edpm-ipam\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-gklb7\" (UID: \"502ceb21-d3d1-4638-a565-2223d41fd1c0\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-gklb7" Feb 02 09:36:24 crc kubenswrapper[4764]: I0202 09:36:24.242908 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/502ceb21-d3d1-4638-a565-2223d41fd1c0-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-gklb7\" (UID: \"502ceb21-d3d1-4638-a565-2223d41fd1c0\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-gklb7" Feb 02 09:36:24 crc kubenswrapper[4764]: I0202 09:36:24.254211 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8vbpf\" (UniqueName: \"kubernetes.io/projected/502ceb21-d3d1-4638-a565-2223d41fd1c0-kube-api-access-8vbpf\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-gklb7\" (UID: \"502ceb21-d3d1-4638-a565-2223d41fd1c0\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-gklb7" Feb 02 09:36:24 crc kubenswrapper[4764]: I0202 09:36:24.375186 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-gklb7" Feb 02 09:36:24 crc kubenswrapper[4764]: I0202 09:36:24.915114 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-gklb7"] Feb 02 09:36:24 crc kubenswrapper[4764]: W0202 09:36:24.931055 4764 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod502ceb21_d3d1_4638_a565_2223d41fd1c0.slice/crio-f6480ec994ddc262eee0b9fa900eb415bf36a08e53467a7ea3b732c15ae6867f WatchSource:0}: Error finding container f6480ec994ddc262eee0b9fa900eb415bf36a08e53467a7ea3b732c15ae6867f: Status 404 returned error can't find the container with id f6480ec994ddc262eee0b9fa900eb415bf36a08e53467a7ea3b732c15ae6867f Feb 02 09:36:24 crc kubenswrapper[4764]: I0202 09:36:24.982322 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-gklb7" event={"ID":"502ceb21-d3d1-4638-a565-2223d41fd1c0","Type":"ContainerStarted","Data":"f6480ec994ddc262eee0b9fa900eb415bf36a08e53467a7ea3b732c15ae6867f"} Feb 02 09:36:25 crc kubenswrapper[4764]: I0202 09:36:25.835791 4764 scope.go:117] "RemoveContainer" containerID="7c9ba2df7010e52fb30b7d6b65df5f1f4f2d3f54edd5411f7806169a53553d20" Feb 02 09:36:25 crc kubenswrapper[4764]: I0202 09:36:25.884882 4764 scope.go:117] "RemoveContainer" containerID="380944daad0b77acae8f51a0b218175b09a49e178b6975b54acdf9ed0e99a75b" Feb 02 09:36:25 crc kubenswrapper[4764]: I0202 09:36:25.914643 4764 scope.go:117] "RemoveContainer" containerID="3a7e2521eea6d5fe7739f7c62d9ec21dc8b57406acc029a0ee0d71a25fd330b2" Feb 02 09:36:25 crc kubenswrapper[4764]: I0202 09:36:25.959331 4764 scope.go:117] "RemoveContainer" containerID="d9716bc9e16c96c00ea61d05b88a5f3eb7f109dd1e6457854009a15ff57725d1" Feb 02 09:36:25 crc kubenswrapper[4764]: I0202 09:36:25.998918 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-gklb7" event={"ID":"502ceb21-d3d1-4638-a565-2223d41fd1c0","Type":"ContainerStarted","Data":"992b91746b103206d2c4c9472b0c41ddf30a14e060fded00f2f48b6d51734e8f"} Feb 02 09:36:26 crc kubenswrapper[4764]: I0202 09:36:26.037425 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-gklb7" podStartSLOduration=1.536599942 podStartE2EDuration="2.037382805s" podCreationTimestamp="2026-02-02 09:36:24 +0000 UTC" firstStartedPulling="2026-02-02 09:36:24.935189451 +0000 UTC m=+1747.868913539" lastFinishedPulling="2026-02-02 09:36:25.435972314 +0000 UTC m=+1748.369696402" observedRunningTime="2026-02-02 09:36:26.018755993 +0000 UTC m=+1748.952480071" watchObservedRunningTime="2026-02-02 09:36:26.037382805 +0000 UTC m=+1748.971106933" Feb 02 09:36:28 crc kubenswrapper[4764]: I0202 09:36:28.825023 4764 scope.go:117] "RemoveContainer" containerID="6db43ab36406cdd4dbe61d7d8688257aba4271b8601a5ba5594675ae309c761d" Feb 02 09:36:28 crc kubenswrapper[4764]: E0202 09:36:28.826582 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 09:36:34 crc kubenswrapper[4764]: I0202 09:36:34.069404 4764 generic.go:334] "Generic (PLEG): container finished" podID="502ceb21-d3d1-4638-a565-2223d41fd1c0" containerID="992b91746b103206d2c4c9472b0c41ddf30a14e060fded00f2f48b6d51734e8f" exitCode=0 Feb 02 09:36:34 crc kubenswrapper[4764]: I0202 09:36:34.069661 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-gklb7" event={"ID":"502ceb21-d3d1-4638-a565-2223d41fd1c0","Type":"ContainerDied","Data":"992b91746b103206d2c4c9472b0c41ddf30a14e060fded00f2f48b6d51734e8f"} Feb 02 09:36:35 crc kubenswrapper[4764]: I0202 09:36:35.520779 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-gklb7" Feb 02 09:36:35 crc kubenswrapper[4764]: I0202 09:36:35.646224 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/502ceb21-d3d1-4638-a565-2223d41fd1c0-ssh-key-openstack-edpm-ipam\") pod \"502ceb21-d3d1-4638-a565-2223d41fd1c0\" (UID: \"502ceb21-d3d1-4638-a565-2223d41fd1c0\") " Feb 02 09:36:35 crc kubenswrapper[4764]: I0202 09:36:35.646334 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/502ceb21-d3d1-4638-a565-2223d41fd1c0-inventory\") pod \"502ceb21-d3d1-4638-a565-2223d41fd1c0\" (UID: \"502ceb21-d3d1-4638-a565-2223d41fd1c0\") " Feb 02 09:36:35 crc kubenswrapper[4764]: I0202 09:36:35.646394 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8vbpf\" (UniqueName: \"kubernetes.io/projected/502ceb21-d3d1-4638-a565-2223d41fd1c0-kube-api-access-8vbpf\") pod \"502ceb21-d3d1-4638-a565-2223d41fd1c0\" (UID: \"502ceb21-d3d1-4638-a565-2223d41fd1c0\") " Feb 02 09:36:35 crc kubenswrapper[4764]: I0202 09:36:35.672174 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/502ceb21-d3d1-4638-a565-2223d41fd1c0-kube-api-access-8vbpf" (OuterVolumeSpecName: "kube-api-access-8vbpf") pod "502ceb21-d3d1-4638-a565-2223d41fd1c0" (UID: "502ceb21-d3d1-4638-a565-2223d41fd1c0"). InnerVolumeSpecName "kube-api-access-8vbpf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:36:35 crc kubenswrapper[4764]: I0202 09:36:35.681631 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/502ceb21-d3d1-4638-a565-2223d41fd1c0-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "502ceb21-d3d1-4638-a565-2223d41fd1c0" (UID: "502ceb21-d3d1-4638-a565-2223d41fd1c0"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:36:35 crc kubenswrapper[4764]: I0202 09:36:35.697703 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/502ceb21-d3d1-4638-a565-2223d41fd1c0-inventory" (OuterVolumeSpecName: "inventory") pod "502ceb21-d3d1-4638-a565-2223d41fd1c0" (UID: "502ceb21-d3d1-4638-a565-2223d41fd1c0"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:36:35 crc kubenswrapper[4764]: I0202 09:36:35.748227 4764 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/502ceb21-d3d1-4638-a565-2223d41fd1c0-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 02 09:36:35 crc kubenswrapper[4764]: I0202 09:36:35.748260 4764 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/502ceb21-d3d1-4638-a565-2223d41fd1c0-inventory\") on node \"crc\" DevicePath \"\"" Feb 02 09:36:35 crc kubenswrapper[4764]: I0202 09:36:35.748270 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8vbpf\" (UniqueName: \"kubernetes.io/projected/502ceb21-d3d1-4638-a565-2223d41fd1c0-kube-api-access-8vbpf\") on node \"crc\" DevicePath \"\"" Feb 02 09:36:36 crc kubenswrapper[4764]: I0202 09:36:36.090500 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-gklb7" event={"ID":"502ceb21-d3d1-4638-a565-2223d41fd1c0","Type":"ContainerDied","Data":"f6480ec994ddc262eee0b9fa900eb415bf36a08e53467a7ea3b732c15ae6867f"} Feb 02 09:36:36 crc kubenswrapper[4764]: I0202 09:36:36.090920 4764 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f6480ec994ddc262eee0b9fa900eb415bf36a08e53467a7ea3b732c15ae6867f" Feb 02 09:36:36 crc kubenswrapper[4764]: I0202 09:36:36.090576 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-gklb7" Feb 02 09:36:36 crc kubenswrapper[4764]: I0202 09:36:36.184150 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fzl68"] Feb 02 09:36:36 crc kubenswrapper[4764]: E0202 09:36:36.184550 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="502ceb21-d3d1-4638-a565-2223d41fd1c0" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Feb 02 09:36:36 crc kubenswrapper[4764]: I0202 09:36:36.184572 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="502ceb21-d3d1-4638-a565-2223d41fd1c0" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Feb 02 09:36:36 crc kubenswrapper[4764]: I0202 09:36:36.184793 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="502ceb21-d3d1-4638-a565-2223d41fd1c0" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Feb 02 09:36:36 crc kubenswrapper[4764]: I0202 09:36:36.185472 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fzl68" Feb 02 09:36:36 crc kubenswrapper[4764]: I0202 09:36:36.187736 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 02 09:36:36 crc kubenswrapper[4764]: I0202 09:36:36.187993 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 02 09:36:36 crc kubenswrapper[4764]: I0202 09:36:36.188114 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 02 09:36:36 crc kubenswrapper[4764]: I0202 09:36:36.188704 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-xksq5" Feb 02 09:36:36 crc kubenswrapper[4764]: I0202 09:36:36.210003 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fzl68"] Feb 02 09:36:36 crc kubenswrapper[4764]: I0202 09:36:36.360314 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/dbda7bef-a208-4766-ad9a-8cf38c77054a-ssh-key-openstack-edpm-ipam\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-fzl68\" (UID: \"dbda7bef-a208-4766-ad9a-8cf38c77054a\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fzl68" Feb 02 09:36:36 crc kubenswrapper[4764]: I0202 09:36:36.360386 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dbda7bef-a208-4766-ad9a-8cf38c77054a-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-fzl68\" (UID: \"dbda7bef-a208-4766-ad9a-8cf38c77054a\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fzl68" Feb 02 09:36:36 crc kubenswrapper[4764]: I0202 09:36:36.360495 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b82wc\" (UniqueName: \"kubernetes.io/projected/dbda7bef-a208-4766-ad9a-8cf38c77054a-kube-api-access-b82wc\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-fzl68\" (UID: \"dbda7bef-a208-4766-ad9a-8cf38c77054a\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fzl68" Feb 02 09:36:36 crc kubenswrapper[4764]: I0202 09:36:36.462278 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/dbda7bef-a208-4766-ad9a-8cf38c77054a-ssh-key-openstack-edpm-ipam\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-fzl68\" (UID: \"dbda7bef-a208-4766-ad9a-8cf38c77054a\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fzl68" Feb 02 09:36:36 crc kubenswrapper[4764]: I0202 09:36:36.462373 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dbda7bef-a208-4766-ad9a-8cf38c77054a-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-fzl68\" (UID: \"dbda7bef-a208-4766-ad9a-8cf38c77054a\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fzl68" Feb 02 09:36:36 crc kubenswrapper[4764]: I0202 09:36:36.462456 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b82wc\" (UniqueName: \"kubernetes.io/projected/dbda7bef-a208-4766-ad9a-8cf38c77054a-kube-api-access-b82wc\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-fzl68\" (UID: \"dbda7bef-a208-4766-ad9a-8cf38c77054a\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fzl68" Feb 02 09:36:36 crc kubenswrapper[4764]: I0202 09:36:36.469593 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/dbda7bef-a208-4766-ad9a-8cf38c77054a-ssh-key-openstack-edpm-ipam\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-fzl68\" (UID: \"dbda7bef-a208-4766-ad9a-8cf38c77054a\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fzl68" Feb 02 09:36:36 crc kubenswrapper[4764]: I0202 09:36:36.472792 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dbda7bef-a208-4766-ad9a-8cf38c77054a-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-fzl68\" (UID: \"dbda7bef-a208-4766-ad9a-8cf38c77054a\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fzl68" Feb 02 09:36:36 crc kubenswrapper[4764]: I0202 09:36:36.480444 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b82wc\" (UniqueName: \"kubernetes.io/projected/dbda7bef-a208-4766-ad9a-8cf38c77054a-kube-api-access-b82wc\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-fzl68\" (UID: \"dbda7bef-a208-4766-ad9a-8cf38c77054a\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fzl68" Feb 02 09:36:36 crc kubenswrapper[4764]: I0202 09:36:36.505781 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fzl68" Feb 02 09:36:37 crc kubenswrapper[4764]: I0202 09:36:37.084704 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fzl68"] Feb 02 09:36:37 crc kubenswrapper[4764]: I0202 09:36:37.099502 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fzl68" event={"ID":"dbda7bef-a208-4766-ad9a-8cf38c77054a","Type":"ContainerStarted","Data":"eeed649629527e0077945f6c1b625aa28c20f30051ebfd91f59a088753f2c75f"} Feb 02 09:36:38 crc kubenswrapper[4764]: I0202 09:36:38.109741 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fzl68" event={"ID":"dbda7bef-a208-4766-ad9a-8cf38c77054a","Type":"ContainerStarted","Data":"8a1dfe2f4d9180101b312e72b2b479bc433502cf047e61bd017d6d4e1c6a0273"} Feb 02 09:36:38 crc kubenswrapper[4764]: I0202 09:36:38.138242 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fzl68" podStartSLOduration=1.642855869 podStartE2EDuration="2.138224678s" podCreationTimestamp="2026-02-02 09:36:36 +0000 UTC" firstStartedPulling="2026-02-02 09:36:37.089045735 +0000 UTC m=+1760.022769823" lastFinishedPulling="2026-02-02 09:36:37.584414544 +0000 UTC m=+1760.518138632" observedRunningTime="2026-02-02 09:36:38.13186705 +0000 UTC m=+1761.065591148" watchObservedRunningTime="2026-02-02 09:36:38.138224678 +0000 UTC m=+1761.071948776" Feb 02 09:36:41 crc kubenswrapper[4764]: I0202 09:36:41.056550 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-scmkd"] Feb 02 09:36:41 crc kubenswrapper[4764]: I0202 09:36:41.066523 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-scmkd"] Feb 02 09:36:41 crc kubenswrapper[4764]: I0202 09:36:41.843699 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2ccd54a4-fbee-4ad6-b0eb-9d22fdb1eebe" path="/var/lib/kubelet/pods/2ccd54a4-fbee-4ad6-b0eb-9d22fdb1eebe/volumes" Feb 02 09:36:42 crc kubenswrapper[4764]: I0202 09:36:42.037179 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-39d3-account-create-update-mtxtf"] Feb 02 09:36:42 crc kubenswrapper[4764]: I0202 09:36:42.047203 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-448kc"] Feb 02 09:36:42 crc kubenswrapper[4764]: I0202 09:36:42.057033 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-91bb-account-create-update-gg7c4"] Feb 02 09:36:42 crc kubenswrapper[4764]: I0202 09:36:42.067268 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-39d3-account-create-update-mtxtf"] Feb 02 09:36:42 crc kubenswrapper[4764]: I0202 09:36:42.074102 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-8ms85"] Feb 02 09:36:42 crc kubenswrapper[4764]: I0202 09:36:42.080580 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-10df-account-create-update-gb8cd"] Feb 02 09:36:42 crc kubenswrapper[4764]: I0202 09:36:42.087646 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-448kc"] Feb 02 09:36:42 crc kubenswrapper[4764]: I0202 09:36:42.094759 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-91bb-account-create-update-gg7c4"] Feb 02 09:36:42 crc kubenswrapper[4764]: I0202 09:36:42.107062 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-10df-account-create-update-gb8cd"] Feb 02 09:36:42 crc kubenswrapper[4764]: I0202 09:36:42.114691 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-8ms85"] Feb 02 09:36:43 crc kubenswrapper[4764]: I0202 09:36:43.825530 4764 scope.go:117] "RemoveContainer" containerID="6db43ab36406cdd4dbe61d7d8688257aba4271b8601a5ba5594675ae309c761d" Feb 02 09:36:43 crc kubenswrapper[4764]: E0202 09:36:43.825978 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 09:36:43 crc kubenswrapper[4764]: I0202 09:36:43.835280 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ee0de70-5bfb-4b6e-974c-601071027218" path="/var/lib/kubelet/pods/3ee0de70-5bfb-4b6e-974c-601071027218/volumes" Feb 02 09:36:43 crc kubenswrapper[4764]: I0202 09:36:43.835851 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="64d8d838-0afe-4f39-ae93-027fea76ced1" path="/var/lib/kubelet/pods/64d8d838-0afe-4f39-ae93-027fea76ced1/volumes" Feb 02 09:36:43 crc kubenswrapper[4764]: I0202 09:36:43.836402 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="902c771b-8da4-4c38-911e-c8b531d076a8" path="/var/lib/kubelet/pods/902c771b-8da4-4c38-911e-c8b531d076a8/volumes" Feb 02 09:36:43 crc kubenswrapper[4764]: I0202 09:36:43.837069 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="afbe3865-2077-45c5-8a94-3d0a6774d8e2" path="/var/lib/kubelet/pods/afbe3865-2077-45c5-8a94-3d0a6774d8e2/volumes" Feb 02 09:36:43 crc kubenswrapper[4764]: I0202 09:36:43.838035 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d492e01f-38f0-4c76-b2aa-a177760a757f" path="/var/lib/kubelet/pods/d492e01f-38f0-4c76-b2aa-a177760a757f/volumes" Feb 02 09:36:48 crc kubenswrapper[4764]: I0202 09:36:48.211002 4764 generic.go:334] "Generic (PLEG): container finished" podID="dbda7bef-a208-4766-ad9a-8cf38c77054a" containerID="8a1dfe2f4d9180101b312e72b2b479bc433502cf047e61bd017d6d4e1c6a0273" exitCode=0 Feb 02 09:36:48 crc kubenswrapper[4764]: I0202 09:36:48.211093 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fzl68" event={"ID":"dbda7bef-a208-4766-ad9a-8cf38c77054a","Type":"ContainerDied","Data":"8a1dfe2f4d9180101b312e72b2b479bc433502cf047e61bd017d6d4e1c6a0273"} Feb 02 09:36:49 crc kubenswrapper[4764]: I0202 09:36:49.694042 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fzl68" Feb 02 09:36:49 crc kubenswrapper[4764]: I0202 09:36:49.797885 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b82wc\" (UniqueName: \"kubernetes.io/projected/dbda7bef-a208-4766-ad9a-8cf38c77054a-kube-api-access-b82wc\") pod \"dbda7bef-a208-4766-ad9a-8cf38c77054a\" (UID: \"dbda7bef-a208-4766-ad9a-8cf38c77054a\") " Feb 02 09:36:49 crc kubenswrapper[4764]: I0202 09:36:49.797955 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/dbda7bef-a208-4766-ad9a-8cf38c77054a-ssh-key-openstack-edpm-ipam\") pod \"dbda7bef-a208-4766-ad9a-8cf38c77054a\" (UID: \"dbda7bef-a208-4766-ad9a-8cf38c77054a\") " Feb 02 09:36:49 crc kubenswrapper[4764]: I0202 09:36:49.797980 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dbda7bef-a208-4766-ad9a-8cf38c77054a-inventory\") pod \"dbda7bef-a208-4766-ad9a-8cf38c77054a\" (UID: \"dbda7bef-a208-4766-ad9a-8cf38c77054a\") " Feb 02 09:36:49 crc kubenswrapper[4764]: I0202 09:36:49.804245 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dbda7bef-a208-4766-ad9a-8cf38c77054a-kube-api-access-b82wc" (OuterVolumeSpecName: "kube-api-access-b82wc") pod "dbda7bef-a208-4766-ad9a-8cf38c77054a" (UID: "dbda7bef-a208-4766-ad9a-8cf38c77054a"). InnerVolumeSpecName "kube-api-access-b82wc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:36:49 crc kubenswrapper[4764]: I0202 09:36:49.822363 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dbda7bef-a208-4766-ad9a-8cf38c77054a-inventory" (OuterVolumeSpecName: "inventory") pod "dbda7bef-a208-4766-ad9a-8cf38c77054a" (UID: "dbda7bef-a208-4766-ad9a-8cf38c77054a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:36:49 crc kubenswrapper[4764]: I0202 09:36:49.848993 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dbda7bef-a208-4766-ad9a-8cf38c77054a-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "dbda7bef-a208-4766-ad9a-8cf38c77054a" (UID: "dbda7bef-a208-4766-ad9a-8cf38c77054a"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:36:49 crc kubenswrapper[4764]: I0202 09:36:49.900018 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b82wc\" (UniqueName: \"kubernetes.io/projected/dbda7bef-a208-4766-ad9a-8cf38c77054a-kube-api-access-b82wc\") on node \"crc\" DevicePath \"\"" Feb 02 09:36:49 crc kubenswrapper[4764]: I0202 09:36:49.900052 4764 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/dbda7bef-a208-4766-ad9a-8cf38c77054a-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 02 09:36:49 crc kubenswrapper[4764]: I0202 09:36:49.900066 4764 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dbda7bef-a208-4766-ad9a-8cf38c77054a-inventory\") on node \"crc\" DevicePath \"\"" Feb 02 09:36:50 crc kubenswrapper[4764]: I0202 09:36:50.233743 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fzl68" event={"ID":"dbda7bef-a208-4766-ad9a-8cf38c77054a","Type":"ContainerDied","Data":"eeed649629527e0077945f6c1b625aa28c20f30051ebfd91f59a088753f2c75f"} Feb 02 09:36:50 crc kubenswrapper[4764]: I0202 09:36:50.234144 4764 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eeed649629527e0077945f6c1b625aa28c20f30051ebfd91f59a088753f2c75f" Feb 02 09:36:50 crc kubenswrapper[4764]: I0202 09:36:50.233829 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fzl68" Feb 02 09:36:58 crc kubenswrapper[4764]: I0202 09:36:58.825662 4764 scope.go:117] "RemoveContainer" containerID="6db43ab36406cdd4dbe61d7d8688257aba4271b8601a5ba5594675ae309c761d" Feb 02 09:36:58 crc kubenswrapper[4764]: E0202 09:36:58.826870 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 09:37:10 crc kubenswrapper[4764]: I0202 09:37:10.046799 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-bnkbg"] Feb 02 09:37:10 crc kubenswrapper[4764]: I0202 09:37:10.060431 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-bnkbg"] Feb 02 09:37:11 crc kubenswrapper[4764]: I0202 09:37:11.826215 4764 scope.go:117] "RemoveContainer" containerID="6db43ab36406cdd4dbe61d7d8688257aba4271b8601a5ba5594675ae309c761d" Feb 02 09:37:11 crc kubenswrapper[4764]: E0202 09:37:11.826683 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 09:37:11 crc kubenswrapper[4764]: I0202 09:37:11.838000 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="40a1e8e4-e9f7-4108-858e-ff65c0becc8d" path="/var/lib/kubelet/pods/40a1e8e4-e9f7-4108-858e-ff65c0becc8d/volumes" Feb 02 09:37:22 crc kubenswrapper[4764]: I0202 09:37:22.825073 4764 scope.go:117] "RemoveContainer" containerID="6db43ab36406cdd4dbe61d7d8688257aba4271b8601a5ba5594675ae309c761d" Feb 02 09:37:22 crc kubenswrapper[4764]: E0202 09:37:22.825691 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 09:37:26 crc kubenswrapper[4764]: I0202 09:37:26.053509 4764 scope.go:117] "RemoveContainer" containerID="f3f1a2804c26e261d58dd4afe0a1d559aee3da232f720be90cb093d535a10712" Feb 02 09:37:26 crc kubenswrapper[4764]: I0202 09:37:26.087284 4764 scope.go:117] "RemoveContainer" containerID="e7f0299d2bd51585893ffb5380d548111de8ffb0dd7db4b97a394a0df7f69d58" Feb 02 09:37:26 crc kubenswrapper[4764]: I0202 09:37:26.162898 4764 scope.go:117] "RemoveContainer" containerID="5db23b665031218d72259cf4646733ed63a4bd51ab8afc325096c5343d72c977" Feb 02 09:37:26 crc kubenswrapper[4764]: I0202 09:37:26.208189 4764 scope.go:117] "RemoveContainer" containerID="7fbdf340596d9e601a0c1ec9ec2dbca538a4261351156b563a3e08f1a090259f" Feb 02 09:37:26 crc kubenswrapper[4764]: I0202 09:37:26.256672 4764 scope.go:117] "RemoveContainer" containerID="5e0a384e8eeb871467c9fe900e3e0979df01eba7e41cc26e321f6df83fb5eae8" Feb 02 09:37:26 crc kubenswrapper[4764]: I0202 09:37:26.309399 4764 scope.go:117] "RemoveContainer" containerID="08540cf734e603fff5d5dcb78a4f068507e9eb3462471ce00d9d5cfaf8926d7d" Feb 02 09:37:26 crc kubenswrapper[4764]: I0202 09:37:26.343568 4764 scope.go:117] "RemoveContainer" containerID="2afeea0c0b612522569ddb9572b7ff6d5cf8c43140c52fc6c0f8275474893368" Feb 02 09:37:33 crc kubenswrapper[4764]: I0202 09:37:33.042653 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-zfgz6"] Feb 02 09:37:33 crc kubenswrapper[4764]: I0202 09:37:33.050515 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-zfgz6"] Feb 02 09:37:33 crc kubenswrapper[4764]: I0202 09:37:33.845404 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6408cd2c-fa48-4c79-a6dd-558008245a0f" path="/var/lib/kubelet/pods/6408cd2c-fa48-4c79-a6dd-558008245a0f/volumes" Feb 02 09:37:34 crc kubenswrapper[4764]: I0202 09:37:34.037380 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-9zzqm"] Feb 02 09:37:34 crc kubenswrapper[4764]: I0202 09:37:34.045354 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-9zzqm"] Feb 02 09:37:34 crc kubenswrapper[4764]: I0202 09:37:34.826478 4764 scope.go:117] "RemoveContainer" containerID="6db43ab36406cdd4dbe61d7d8688257aba4271b8601a5ba5594675ae309c761d" Feb 02 09:37:34 crc kubenswrapper[4764]: E0202 09:37:34.826732 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 09:37:35 crc kubenswrapper[4764]: I0202 09:37:35.839251 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="91e8d389-437f-4eed-b6b1-2b64802f9bc9" path="/var/lib/kubelet/pods/91e8d389-437f-4eed-b6b1-2b64802f9bc9/volumes" Feb 02 09:37:47 crc kubenswrapper[4764]: I0202 09:37:47.834094 4764 scope.go:117] "RemoveContainer" containerID="6db43ab36406cdd4dbe61d7d8688257aba4271b8601a5ba5594675ae309c761d" Feb 02 09:37:47 crc kubenswrapper[4764]: E0202 09:37:47.835456 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 09:37:58 crc kubenswrapper[4764]: I0202 09:37:58.826191 4764 scope.go:117] "RemoveContainer" containerID="6db43ab36406cdd4dbe61d7d8688257aba4271b8601a5ba5594675ae309c761d" Feb 02 09:37:58 crc kubenswrapper[4764]: E0202 09:37:58.829096 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 09:38:11 crc kubenswrapper[4764]: I0202 09:38:11.833442 4764 scope.go:117] "RemoveContainer" containerID="6db43ab36406cdd4dbe61d7d8688257aba4271b8601a5ba5594675ae309c761d" Feb 02 09:38:11 crc kubenswrapper[4764]: E0202 09:38:11.834642 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 09:38:17 crc kubenswrapper[4764]: I0202 09:38:17.076219 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-q58jt"] Feb 02 09:38:17 crc kubenswrapper[4764]: I0202 09:38:17.092074 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-q58jt"] Feb 02 09:38:17 crc kubenswrapper[4764]: I0202 09:38:17.839463 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f83a9caf-7a27-4adf-90ea-072142c914f9" path="/var/lib/kubelet/pods/f83a9caf-7a27-4adf-90ea-072142c914f9/volumes" Feb 02 09:38:24 crc kubenswrapper[4764]: I0202 09:38:24.825336 4764 scope.go:117] "RemoveContainer" containerID="6db43ab36406cdd4dbe61d7d8688257aba4271b8601a5ba5594675ae309c761d" Feb 02 09:38:25 crc kubenswrapper[4764]: I0202 09:38:25.199062 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" event={"ID":"d192f670-9f9d-4539-9641-e4bed73acdd4","Type":"ContainerStarted","Data":"7a73d717f76ca801e8516faf5325e4c71beae4270ab4dde30751c8baf713bdb2"} Feb 02 09:38:26 crc kubenswrapper[4764]: I0202 09:38:26.448665 4764 scope.go:117] "RemoveContainer" containerID="f26b51118aaf57ee0b36ac05fa18a18fb719c48c913ef9f25bad65b10713ba4a" Feb 02 09:38:26 crc kubenswrapper[4764]: I0202 09:38:26.501796 4764 scope.go:117] "RemoveContainer" containerID="0567ac860308c61ce0ead95bb217f119d788a67221475b28b7de06c2f202c35a" Feb 02 09:38:26 crc kubenswrapper[4764]: I0202 09:38:26.558267 4764 scope.go:117] "RemoveContainer" containerID="11f32a8c2d61df5ff6d2656eb4ba0d2fe0897343eea8c23f62baa26ab81563c5" Feb 02 09:39:36 crc kubenswrapper[4764]: I0202 09:39:36.870132 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-mk82r"] Feb 02 09:39:36 crc kubenswrapper[4764]: E0202 09:39:36.871062 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dbda7bef-a208-4766-ad9a-8cf38c77054a" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Feb 02 09:39:36 crc kubenswrapper[4764]: I0202 09:39:36.871080 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="dbda7bef-a208-4766-ad9a-8cf38c77054a" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Feb 02 09:39:36 crc kubenswrapper[4764]: I0202 09:39:36.871286 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="dbda7bef-a208-4766-ad9a-8cf38c77054a" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Feb 02 09:39:36 crc kubenswrapper[4764]: I0202 09:39:36.872911 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mk82r" Feb 02 09:39:36 crc kubenswrapper[4764]: I0202 09:39:36.894159 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mk82r"] Feb 02 09:39:37 crc kubenswrapper[4764]: I0202 09:39:37.068699 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/753a3e0b-5938-434e-b331-89296fd5712f-utilities\") pod \"community-operators-mk82r\" (UID: \"753a3e0b-5938-434e-b331-89296fd5712f\") " pod="openshift-marketplace/community-operators-mk82r" Feb 02 09:39:37 crc kubenswrapper[4764]: I0202 09:39:37.069094 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5fgf2\" (UniqueName: \"kubernetes.io/projected/753a3e0b-5938-434e-b331-89296fd5712f-kube-api-access-5fgf2\") pod \"community-operators-mk82r\" (UID: \"753a3e0b-5938-434e-b331-89296fd5712f\") " pod="openshift-marketplace/community-operators-mk82r" Feb 02 09:39:37 crc kubenswrapper[4764]: I0202 09:39:37.069544 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/753a3e0b-5938-434e-b331-89296fd5712f-catalog-content\") pod \"community-operators-mk82r\" (UID: \"753a3e0b-5938-434e-b331-89296fd5712f\") " pod="openshift-marketplace/community-operators-mk82r" Feb 02 09:39:37 crc kubenswrapper[4764]: I0202 09:39:37.171395 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/753a3e0b-5938-434e-b331-89296fd5712f-utilities\") pod \"community-operators-mk82r\" (UID: \"753a3e0b-5938-434e-b331-89296fd5712f\") " pod="openshift-marketplace/community-operators-mk82r" Feb 02 09:39:37 crc kubenswrapper[4764]: I0202 09:39:37.171539 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5fgf2\" (UniqueName: \"kubernetes.io/projected/753a3e0b-5938-434e-b331-89296fd5712f-kube-api-access-5fgf2\") pod \"community-operators-mk82r\" (UID: \"753a3e0b-5938-434e-b331-89296fd5712f\") " pod="openshift-marketplace/community-operators-mk82r" Feb 02 09:39:37 crc kubenswrapper[4764]: I0202 09:39:37.171564 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/753a3e0b-5938-434e-b331-89296fd5712f-catalog-content\") pod \"community-operators-mk82r\" (UID: \"753a3e0b-5938-434e-b331-89296fd5712f\") " pod="openshift-marketplace/community-operators-mk82r" Feb 02 09:39:37 crc kubenswrapper[4764]: I0202 09:39:37.172079 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/753a3e0b-5938-434e-b331-89296fd5712f-utilities\") pod \"community-operators-mk82r\" (UID: \"753a3e0b-5938-434e-b331-89296fd5712f\") " pod="openshift-marketplace/community-operators-mk82r" Feb 02 09:39:37 crc kubenswrapper[4764]: I0202 09:39:37.172100 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/753a3e0b-5938-434e-b331-89296fd5712f-catalog-content\") pod \"community-operators-mk82r\" (UID: \"753a3e0b-5938-434e-b331-89296fd5712f\") " pod="openshift-marketplace/community-operators-mk82r" Feb 02 09:39:37 crc kubenswrapper[4764]: I0202 09:39:37.198149 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5fgf2\" (UniqueName: \"kubernetes.io/projected/753a3e0b-5938-434e-b331-89296fd5712f-kube-api-access-5fgf2\") pod \"community-operators-mk82r\" (UID: \"753a3e0b-5938-434e-b331-89296fd5712f\") " pod="openshift-marketplace/community-operators-mk82r" Feb 02 09:39:37 crc kubenswrapper[4764]: I0202 09:39:37.212054 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mk82r" Feb 02 09:39:37 crc kubenswrapper[4764]: I0202 09:39:37.698681 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mk82r"] Feb 02 09:39:37 crc kubenswrapper[4764]: W0202 09:39:37.711101 4764 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod753a3e0b_5938_434e_b331_89296fd5712f.slice/crio-c655261a03d04955520e3c702e05769755998ed2f24aa419f2afd1960f96f258 WatchSource:0}: Error finding container c655261a03d04955520e3c702e05769755998ed2f24aa419f2afd1960f96f258: Status 404 returned error can't find the container with id c655261a03d04955520e3c702e05769755998ed2f24aa419f2afd1960f96f258 Feb 02 09:39:37 crc kubenswrapper[4764]: I0202 09:39:37.954547 4764 generic.go:334] "Generic (PLEG): container finished" podID="753a3e0b-5938-434e-b331-89296fd5712f" containerID="12454c0b57840d04f4b4c6a5cb7ab85d84dec3712411c8a98e523afe16500b44" exitCode=0 Feb 02 09:39:37 crc kubenswrapper[4764]: I0202 09:39:37.954803 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mk82r" event={"ID":"753a3e0b-5938-434e-b331-89296fd5712f","Type":"ContainerDied","Data":"12454c0b57840d04f4b4c6a5cb7ab85d84dec3712411c8a98e523afe16500b44"} Feb 02 09:39:37 crc kubenswrapper[4764]: I0202 09:39:37.954872 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mk82r" event={"ID":"753a3e0b-5938-434e-b331-89296fd5712f","Type":"ContainerStarted","Data":"c655261a03d04955520e3c702e05769755998ed2f24aa419f2afd1960f96f258"} Feb 02 09:39:37 crc kubenswrapper[4764]: I0202 09:39:37.957667 4764 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 02 09:39:38 crc kubenswrapper[4764]: I0202 09:39:38.963268 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mk82r" event={"ID":"753a3e0b-5938-434e-b331-89296fd5712f","Type":"ContainerStarted","Data":"616c6ac93ae9fbe1c15911f2e31ee8c56961eac91d859a34aaa798dba901ae09"} Feb 02 09:39:40 crc kubenswrapper[4764]: I0202 09:39:40.983244 4764 generic.go:334] "Generic (PLEG): container finished" podID="753a3e0b-5938-434e-b331-89296fd5712f" containerID="616c6ac93ae9fbe1c15911f2e31ee8c56961eac91d859a34aaa798dba901ae09" exitCode=0 Feb 02 09:39:40 crc kubenswrapper[4764]: I0202 09:39:40.983316 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mk82r" event={"ID":"753a3e0b-5938-434e-b331-89296fd5712f","Type":"ContainerDied","Data":"616c6ac93ae9fbe1c15911f2e31ee8c56961eac91d859a34aaa798dba901ae09"} Feb 02 09:39:41 crc kubenswrapper[4764]: I0202 09:39:41.993981 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mk82r" event={"ID":"753a3e0b-5938-434e-b331-89296fd5712f","Type":"ContainerStarted","Data":"ba6181469d98dbca2ec367209bbd218b0a9e05f28fac1ca2b69cf44499eff435"} Feb 02 09:39:42 crc kubenswrapper[4764]: I0202 09:39:42.035349 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-mk82r" podStartSLOduration=2.57787444 podStartE2EDuration="6.035324294s" podCreationTimestamp="2026-02-02 09:39:36 +0000 UTC" firstStartedPulling="2026-02-02 09:39:37.957242882 +0000 UTC m=+1940.890967010" lastFinishedPulling="2026-02-02 09:39:41.414692776 +0000 UTC m=+1944.348416864" observedRunningTime="2026-02-02 09:39:42.023145873 +0000 UTC m=+1944.956869981" watchObservedRunningTime="2026-02-02 09:39:42.035324294 +0000 UTC m=+1944.969048392" Feb 02 09:39:47 crc kubenswrapper[4764]: I0202 09:39:47.213212 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-mk82r" Feb 02 09:39:47 crc kubenswrapper[4764]: I0202 09:39:47.213739 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-mk82r" Feb 02 09:39:47 crc kubenswrapper[4764]: I0202 09:39:47.277418 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-mk82r" Feb 02 09:39:48 crc kubenswrapper[4764]: I0202 09:39:48.127980 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-mk82r" Feb 02 09:39:48 crc kubenswrapper[4764]: I0202 09:39:48.181823 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-mk82r"] Feb 02 09:39:50 crc kubenswrapper[4764]: I0202 09:39:50.079036 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-mk82r" podUID="753a3e0b-5938-434e-b331-89296fd5712f" containerName="registry-server" containerID="cri-o://ba6181469d98dbca2ec367209bbd218b0a9e05f28fac1ca2b69cf44499eff435" gracePeriod=2 Feb 02 09:39:50 crc kubenswrapper[4764]: I0202 09:39:50.515684 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mk82r" Feb 02 09:39:50 crc kubenswrapper[4764]: I0202 09:39:50.683355 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/753a3e0b-5938-434e-b331-89296fd5712f-catalog-content\") pod \"753a3e0b-5938-434e-b331-89296fd5712f\" (UID: \"753a3e0b-5938-434e-b331-89296fd5712f\") " Feb 02 09:39:50 crc kubenswrapper[4764]: I0202 09:39:50.683488 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5fgf2\" (UniqueName: \"kubernetes.io/projected/753a3e0b-5938-434e-b331-89296fd5712f-kube-api-access-5fgf2\") pod \"753a3e0b-5938-434e-b331-89296fd5712f\" (UID: \"753a3e0b-5938-434e-b331-89296fd5712f\") " Feb 02 09:39:50 crc kubenswrapper[4764]: I0202 09:39:50.683529 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/753a3e0b-5938-434e-b331-89296fd5712f-utilities\") pod \"753a3e0b-5938-434e-b331-89296fd5712f\" (UID: \"753a3e0b-5938-434e-b331-89296fd5712f\") " Feb 02 09:39:50 crc kubenswrapper[4764]: I0202 09:39:50.684475 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/753a3e0b-5938-434e-b331-89296fd5712f-utilities" (OuterVolumeSpecName: "utilities") pod "753a3e0b-5938-434e-b331-89296fd5712f" (UID: "753a3e0b-5938-434e-b331-89296fd5712f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 09:39:50 crc kubenswrapper[4764]: I0202 09:39:50.690150 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/753a3e0b-5938-434e-b331-89296fd5712f-kube-api-access-5fgf2" (OuterVolumeSpecName: "kube-api-access-5fgf2") pod "753a3e0b-5938-434e-b331-89296fd5712f" (UID: "753a3e0b-5938-434e-b331-89296fd5712f"). InnerVolumeSpecName "kube-api-access-5fgf2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:39:50 crc kubenswrapper[4764]: I0202 09:39:50.738811 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/753a3e0b-5938-434e-b331-89296fd5712f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "753a3e0b-5938-434e-b331-89296fd5712f" (UID: "753a3e0b-5938-434e-b331-89296fd5712f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 09:39:50 crc kubenswrapper[4764]: I0202 09:39:50.785646 4764 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/753a3e0b-5938-434e-b331-89296fd5712f-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 09:39:50 crc kubenswrapper[4764]: I0202 09:39:50.785729 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5fgf2\" (UniqueName: \"kubernetes.io/projected/753a3e0b-5938-434e-b331-89296fd5712f-kube-api-access-5fgf2\") on node \"crc\" DevicePath \"\"" Feb 02 09:39:50 crc kubenswrapper[4764]: I0202 09:39:50.785745 4764 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/753a3e0b-5938-434e-b331-89296fd5712f-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 09:39:51 crc kubenswrapper[4764]: I0202 09:39:51.086774 4764 generic.go:334] "Generic (PLEG): container finished" podID="753a3e0b-5938-434e-b331-89296fd5712f" containerID="ba6181469d98dbca2ec367209bbd218b0a9e05f28fac1ca2b69cf44499eff435" exitCode=0 Feb 02 09:39:51 crc kubenswrapper[4764]: I0202 09:39:51.086842 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mk82r" event={"ID":"753a3e0b-5938-434e-b331-89296fd5712f","Type":"ContainerDied","Data":"ba6181469d98dbca2ec367209bbd218b0a9e05f28fac1ca2b69cf44499eff435"} Feb 02 09:39:51 crc kubenswrapper[4764]: I0202 09:39:51.087129 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mk82r" event={"ID":"753a3e0b-5938-434e-b331-89296fd5712f","Type":"ContainerDied","Data":"c655261a03d04955520e3c702e05769755998ed2f24aa419f2afd1960f96f258"} Feb 02 09:39:51 crc kubenswrapper[4764]: I0202 09:39:51.087155 4764 scope.go:117] "RemoveContainer" containerID="ba6181469d98dbca2ec367209bbd218b0a9e05f28fac1ca2b69cf44499eff435" Feb 02 09:39:51 crc kubenswrapper[4764]: I0202 09:39:51.086869 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mk82r" Feb 02 09:39:51 crc kubenswrapper[4764]: I0202 09:39:51.106661 4764 scope.go:117] "RemoveContainer" containerID="616c6ac93ae9fbe1c15911f2e31ee8c56961eac91d859a34aaa798dba901ae09" Feb 02 09:39:51 crc kubenswrapper[4764]: I0202 09:39:51.134904 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-mk82r"] Feb 02 09:39:51 crc kubenswrapper[4764]: I0202 09:39:51.138382 4764 scope.go:117] "RemoveContainer" containerID="12454c0b57840d04f4b4c6a5cb7ab85d84dec3712411c8a98e523afe16500b44" Feb 02 09:39:51 crc kubenswrapper[4764]: I0202 09:39:51.146382 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-mk82r"] Feb 02 09:39:51 crc kubenswrapper[4764]: I0202 09:39:51.176800 4764 scope.go:117] "RemoveContainer" containerID="ba6181469d98dbca2ec367209bbd218b0a9e05f28fac1ca2b69cf44499eff435" Feb 02 09:39:51 crc kubenswrapper[4764]: E0202 09:39:51.177660 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ba6181469d98dbca2ec367209bbd218b0a9e05f28fac1ca2b69cf44499eff435\": container with ID starting with ba6181469d98dbca2ec367209bbd218b0a9e05f28fac1ca2b69cf44499eff435 not found: ID does not exist" containerID="ba6181469d98dbca2ec367209bbd218b0a9e05f28fac1ca2b69cf44499eff435" Feb 02 09:39:51 crc kubenswrapper[4764]: I0202 09:39:51.177710 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ba6181469d98dbca2ec367209bbd218b0a9e05f28fac1ca2b69cf44499eff435"} err="failed to get container status \"ba6181469d98dbca2ec367209bbd218b0a9e05f28fac1ca2b69cf44499eff435\": rpc error: code = NotFound desc = could not find container \"ba6181469d98dbca2ec367209bbd218b0a9e05f28fac1ca2b69cf44499eff435\": container with ID starting with ba6181469d98dbca2ec367209bbd218b0a9e05f28fac1ca2b69cf44499eff435 not found: ID does not exist" Feb 02 09:39:51 crc kubenswrapper[4764]: I0202 09:39:51.177741 4764 scope.go:117] "RemoveContainer" containerID="616c6ac93ae9fbe1c15911f2e31ee8c56961eac91d859a34aaa798dba901ae09" Feb 02 09:39:51 crc kubenswrapper[4764]: E0202 09:39:51.178207 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"616c6ac93ae9fbe1c15911f2e31ee8c56961eac91d859a34aaa798dba901ae09\": container with ID starting with 616c6ac93ae9fbe1c15911f2e31ee8c56961eac91d859a34aaa798dba901ae09 not found: ID does not exist" containerID="616c6ac93ae9fbe1c15911f2e31ee8c56961eac91d859a34aaa798dba901ae09" Feb 02 09:39:51 crc kubenswrapper[4764]: I0202 09:39:51.178233 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"616c6ac93ae9fbe1c15911f2e31ee8c56961eac91d859a34aaa798dba901ae09"} err="failed to get container status \"616c6ac93ae9fbe1c15911f2e31ee8c56961eac91d859a34aaa798dba901ae09\": rpc error: code = NotFound desc = could not find container \"616c6ac93ae9fbe1c15911f2e31ee8c56961eac91d859a34aaa798dba901ae09\": container with ID starting with 616c6ac93ae9fbe1c15911f2e31ee8c56961eac91d859a34aaa798dba901ae09 not found: ID does not exist" Feb 02 09:39:51 crc kubenswrapper[4764]: I0202 09:39:51.178253 4764 scope.go:117] "RemoveContainer" containerID="12454c0b57840d04f4b4c6a5cb7ab85d84dec3712411c8a98e523afe16500b44" Feb 02 09:39:51 crc kubenswrapper[4764]: E0202 09:39:51.178771 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"12454c0b57840d04f4b4c6a5cb7ab85d84dec3712411c8a98e523afe16500b44\": container with ID starting with 12454c0b57840d04f4b4c6a5cb7ab85d84dec3712411c8a98e523afe16500b44 not found: ID does not exist" containerID="12454c0b57840d04f4b4c6a5cb7ab85d84dec3712411c8a98e523afe16500b44" Feb 02 09:39:51 crc kubenswrapper[4764]: I0202 09:39:51.178794 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"12454c0b57840d04f4b4c6a5cb7ab85d84dec3712411c8a98e523afe16500b44"} err="failed to get container status \"12454c0b57840d04f4b4c6a5cb7ab85d84dec3712411c8a98e523afe16500b44\": rpc error: code = NotFound desc = could not find container \"12454c0b57840d04f4b4c6a5cb7ab85d84dec3712411c8a98e523afe16500b44\": container with ID starting with 12454c0b57840d04f4b4c6a5cb7ab85d84dec3712411c8a98e523afe16500b44 not found: ID does not exist" Feb 02 09:39:51 crc kubenswrapper[4764]: I0202 09:39:51.835468 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="753a3e0b-5938-434e-b331-89296fd5712f" path="/var/lib/kubelet/pods/753a3e0b-5938-434e-b331-89296fd5712f/volumes" Feb 02 09:40:43 crc kubenswrapper[4764]: I0202 09:40:43.523011 4764 patch_prober.go:28] interesting pod/machine-config-daemon-4ndm4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 09:40:43 crc kubenswrapper[4764]: I0202 09:40:43.523874 4764 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 09:41:13 crc kubenswrapper[4764]: I0202 09:41:13.523047 4764 patch_prober.go:28] interesting pod/machine-config-daemon-4ndm4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 09:41:13 crc kubenswrapper[4764]: I0202 09:41:13.524375 4764 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 09:41:31 crc kubenswrapper[4764]: I0202 09:41:31.294855 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-bzbm2"] Feb 02 09:41:31 crc kubenswrapper[4764]: E0202 09:41:31.295737 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="753a3e0b-5938-434e-b331-89296fd5712f" containerName="registry-server" Feb 02 09:41:31 crc kubenswrapper[4764]: I0202 09:41:31.295754 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="753a3e0b-5938-434e-b331-89296fd5712f" containerName="registry-server" Feb 02 09:41:31 crc kubenswrapper[4764]: E0202 09:41:31.295773 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="753a3e0b-5938-434e-b331-89296fd5712f" containerName="extract-utilities" Feb 02 09:41:31 crc kubenswrapper[4764]: I0202 09:41:31.295779 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="753a3e0b-5938-434e-b331-89296fd5712f" containerName="extract-utilities" Feb 02 09:41:31 crc kubenswrapper[4764]: E0202 09:41:31.295797 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="753a3e0b-5938-434e-b331-89296fd5712f" containerName="extract-content" Feb 02 09:41:31 crc kubenswrapper[4764]: I0202 09:41:31.295802 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="753a3e0b-5938-434e-b331-89296fd5712f" containerName="extract-content" Feb 02 09:41:31 crc kubenswrapper[4764]: I0202 09:41:31.295988 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="753a3e0b-5938-434e-b331-89296fd5712f" containerName="registry-server" Feb 02 09:41:31 crc kubenswrapper[4764]: I0202 09:41:31.299088 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bzbm2" Feb 02 09:41:31 crc kubenswrapper[4764]: I0202 09:41:31.313224 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bzbm2"] Feb 02 09:41:31 crc kubenswrapper[4764]: I0202 09:41:31.355186 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/993d9be2-0891-4680-a0f9-702bb07aa5b3-utilities\") pod \"redhat-operators-bzbm2\" (UID: \"993d9be2-0891-4680-a0f9-702bb07aa5b3\") " pod="openshift-marketplace/redhat-operators-bzbm2" Feb 02 09:41:31 crc kubenswrapper[4764]: I0202 09:41:31.355290 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pr6t9\" (UniqueName: \"kubernetes.io/projected/993d9be2-0891-4680-a0f9-702bb07aa5b3-kube-api-access-pr6t9\") pod \"redhat-operators-bzbm2\" (UID: \"993d9be2-0891-4680-a0f9-702bb07aa5b3\") " pod="openshift-marketplace/redhat-operators-bzbm2" Feb 02 09:41:31 crc kubenswrapper[4764]: I0202 09:41:31.355345 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/993d9be2-0891-4680-a0f9-702bb07aa5b3-catalog-content\") pod \"redhat-operators-bzbm2\" (UID: \"993d9be2-0891-4680-a0f9-702bb07aa5b3\") " pod="openshift-marketplace/redhat-operators-bzbm2" Feb 02 09:41:31 crc kubenswrapper[4764]: I0202 09:41:31.456870 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/993d9be2-0891-4680-a0f9-702bb07aa5b3-utilities\") pod \"redhat-operators-bzbm2\" (UID: \"993d9be2-0891-4680-a0f9-702bb07aa5b3\") " pod="openshift-marketplace/redhat-operators-bzbm2" Feb 02 09:41:31 crc kubenswrapper[4764]: I0202 09:41:31.457234 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pr6t9\" (UniqueName: \"kubernetes.io/projected/993d9be2-0891-4680-a0f9-702bb07aa5b3-kube-api-access-pr6t9\") pod \"redhat-operators-bzbm2\" (UID: \"993d9be2-0891-4680-a0f9-702bb07aa5b3\") " pod="openshift-marketplace/redhat-operators-bzbm2" Feb 02 09:41:31 crc kubenswrapper[4764]: I0202 09:41:31.457394 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/993d9be2-0891-4680-a0f9-702bb07aa5b3-catalog-content\") pod \"redhat-operators-bzbm2\" (UID: \"993d9be2-0891-4680-a0f9-702bb07aa5b3\") " pod="openshift-marketplace/redhat-operators-bzbm2" Feb 02 09:41:31 crc kubenswrapper[4764]: I0202 09:41:31.457424 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/993d9be2-0891-4680-a0f9-702bb07aa5b3-utilities\") pod \"redhat-operators-bzbm2\" (UID: \"993d9be2-0891-4680-a0f9-702bb07aa5b3\") " pod="openshift-marketplace/redhat-operators-bzbm2" Feb 02 09:41:31 crc kubenswrapper[4764]: I0202 09:41:31.457674 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/993d9be2-0891-4680-a0f9-702bb07aa5b3-catalog-content\") pod \"redhat-operators-bzbm2\" (UID: \"993d9be2-0891-4680-a0f9-702bb07aa5b3\") " pod="openshift-marketplace/redhat-operators-bzbm2" Feb 02 09:41:31 crc kubenswrapper[4764]: I0202 09:41:31.492166 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pr6t9\" (UniqueName: \"kubernetes.io/projected/993d9be2-0891-4680-a0f9-702bb07aa5b3-kube-api-access-pr6t9\") pod \"redhat-operators-bzbm2\" (UID: \"993d9be2-0891-4680-a0f9-702bb07aa5b3\") " pod="openshift-marketplace/redhat-operators-bzbm2" Feb 02 09:41:31 crc kubenswrapper[4764]: I0202 09:41:31.622789 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bzbm2" Feb 02 09:41:32 crc kubenswrapper[4764]: I0202 09:41:32.178787 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bzbm2"] Feb 02 09:41:32 crc kubenswrapper[4764]: W0202 09:41:32.185041 4764 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod993d9be2_0891_4680_a0f9_702bb07aa5b3.slice/crio-b52fbaa26eb077d0a6f5412c2d063e8d06c31e91506d5ce454a792d2e2bd7edc WatchSource:0}: Error finding container b52fbaa26eb077d0a6f5412c2d063e8d06c31e91506d5ce454a792d2e2bd7edc: Status 404 returned error can't find the container with id b52fbaa26eb077d0a6f5412c2d063e8d06c31e91506d5ce454a792d2e2bd7edc Feb 02 09:41:33 crc kubenswrapper[4764]: I0202 09:41:33.148582 4764 generic.go:334] "Generic (PLEG): container finished" podID="993d9be2-0891-4680-a0f9-702bb07aa5b3" containerID="d4711358de4b369d0913995858fe499153a6c2ae764ee62a3939d9db9a168533" exitCode=0 Feb 02 09:41:33 crc kubenswrapper[4764]: I0202 09:41:33.148652 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bzbm2" event={"ID":"993d9be2-0891-4680-a0f9-702bb07aa5b3","Type":"ContainerDied","Data":"d4711358de4b369d0913995858fe499153a6c2ae764ee62a3939d9db9a168533"} Feb 02 09:41:33 crc kubenswrapper[4764]: I0202 09:41:33.148704 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bzbm2" event={"ID":"993d9be2-0891-4680-a0f9-702bb07aa5b3","Type":"ContainerStarted","Data":"b52fbaa26eb077d0a6f5412c2d063e8d06c31e91506d5ce454a792d2e2bd7edc"} Feb 02 09:41:34 crc kubenswrapper[4764]: I0202 09:41:34.170952 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bzbm2" event={"ID":"993d9be2-0891-4680-a0f9-702bb07aa5b3","Type":"ContainerStarted","Data":"a213d40fbda79a1198b3f6e1e7131adcef3f5a4def5fea28c01231f8b28b7a8a"} Feb 02 09:41:42 crc kubenswrapper[4764]: I0202 09:41:42.253354 4764 generic.go:334] "Generic (PLEG): container finished" podID="993d9be2-0891-4680-a0f9-702bb07aa5b3" containerID="a213d40fbda79a1198b3f6e1e7131adcef3f5a4def5fea28c01231f8b28b7a8a" exitCode=0 Feb 02 09:41:42 crc kubenswrapper[4764]: I0202 09:41:42.253440 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bzbm2" event={"ID":"993d9be2-0891-4680-a0f9-702bb07aa5b3","Type":"ContainerDied","Data":"a213d40fbda79a1198b3f6e1e7131adcef3f5a4def5fea28c01231f8b28b7a8a"} Feb 02 09:41:42 crc kubenswrapper[4764]: E0202 09:41:42.930133 4764 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.217:50864->38.102.83.217:34003: write tcp 38.102.83.217:50864->38.102.83.217:34003: write: broken pipe Feb 02 09:41:43 crc kubenswrapper[4764]: I0202 09:41:43.267551 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bzbm2" event={"ID":"993d9be2-0891-4680-a0f9-702bb07aa5b3","Type":"ContainerStarted","Data":"5cbad5620f17be7545a4087fce4cb01e5cbf2bd29fe11c813fa46b0531b70a81"} Feb 02 09:41:43 crc kubenswrapper[4764]: I0202 09:41:43.306213 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-bzbm2" podStartSLOduration=2.668590392 podStartE2EDuration="12.306184183s" podCreationTimestamp="2026-02-02 09:41:31 +0000 UTC" firstStartedPulling="2026-02-02 09:41:33.15049493 +0000 UTC m=+2056.084219018" lastFinishedPulling="2026-02-02 09:41:42.788088731 +0000 UTC m=+2065.721812809" observedRunningTime="2026-02-02 09:41:43.293239101 +0000 UTC m=+2066.226963199" watchObservedRunningTime="2026-02-02 09:41:43.306184183 +0000 UTC m=+2066.239908271" Feb 02 09:41:43 crc kubenswrapper[4764]: I0202 09:41:43.523116 4764 patch_prober.go:28] interesting pod/machine-config-daemon-4ndm4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 09:41:43 crc kubenswrapper[4764]: I0202 09:41:43.523194 4764 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 09:41:43 crc kubenswrapper[4764]: I0202 09:41:43.523252 4764 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" Feb 02 09:41:43 crc kubenswrapper[4764]: I0202 09:41:43.523979 4764 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7a73d717f76ca801e8516faf5325e4c71beae4270ab4dde30751c8baf713bdb2"} pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 02 09:41:43 crc kubenswrapper[4764]: I0202 09:41:43.524056 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" containerName="machine-config-daemon" containerID="cri-o://7a73d717f76ca801e8516faf5325e4c71beae4270ab4dde30751c8baf713bdb2" gracePeriod=600 Feb 02 09:41:44 crc kubenswrapper[4764]: I0202 09:41:44.279706 4764 generic.go:334] "Generic (PLEG): container finished" podID="d192f670-9f9d-4539-9641-e4bed73acdd4" containerID="7a73d717f76ca801e8516faf5325e4c71beae4270ab4dde30751c8baf713bdb2" exitCode=0 Feb 02 09:41:44 crc kubenswrapper[4764]: I0202 09:41:44.279775 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" event={"ID":"d192f670-9f9d-4539-9641-e4bed73acdd4","Type":"ContainerDied","Data":"7a73d717f76ca801e8516faf5325e4c71beae4270ab4dde30751c8baf713bdb2"} Feb 02 09:41:44 crc kubenswrapper[4764]: I0202 09:41:44.280032 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" event={"ID":"d192f670-9f9d-4539-9641-e4bed73acdd4","Type":"ContainerStarted","Data":"be8fa5800ec42d1c141765dcfc6111f3ef96b05b6e8d5a9feaf2baa219897bcf"} Feb 02 09:41:44 crc kubenswrapper[4764]: I0202 09:41:44.280050 4764 scope.go:117] "RemoveContainer" containerID="6db43ab36406cdd4dbe61d7d8688257aba4271b8601a5ba5594675ae309c761d" Feb 02 09:41:51 crc kubenswrapper[4764]: I0202 09:41:51.623896 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-bzbm2" Feb 02 09:41:51 crc kubenswrapper[4764]: I0202 09:41:51.624523 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-bzbm2" Feb 02 09:41:51 crc kubenswrapper[4764]: I0202 09:41:51.689159 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-bzbm2" Feb 02 09:41:52 crc kubenswrapper[4764]: I0202 09:41:52.420362 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-bzbm2" Feb 02 09:41:52 crc kubenswrapper[4764]: I0202 09:41:52.486538 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bzbm2"] Feb 02 09:41:54 crc kubenswrapper[4764]: I0202 09:41:54.370721 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-bzbm2" podUID="993d9be2-0891-4680-a0f9-702bb07aa5b3" containerName="registry-server" containerID="cri-o://5cbad5620f17be7545a4087fce4cb01e5cbf2bd29fe11c813fa46b0531b70a81" gracePeriod=2 Feb 02 09:41:54 crc kubenswrapper[4764]: I0202 09:41:54.840235 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bzbm2" Feb 02 09:41:55 crc kubenswrapper[4764]: I0202 09:41:55.024542 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pr6t9\" (UniqueName: \"kubernetes.io/projected/993d9be2-0891-4680-a0f9-702bb07aa5b3-kube-api-access-pr6t9\") pod \"993d9be2-0891-4680-a0f9-702bb07aa5b3\" (UID: \"993d9be2-0891-4680-a0f9-702bb07aa5b3\") " Feb 02 09:41:55 crc kubenswrapper[4764]: I0202 09:41:55.024724 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/993d9be2-0891-4680-a0f9-702bb07aa5b3-catalog-content\") pod \"993d9be2-0891-4680-a0f9-702bb07aa5b3\" (UID: \"993d9be2-0891-4680-a0f9-702bb07aa5b3\") " Feb 02 09:41:55 crc kubenswrapper[4764]: I0202 09:41:55.024901 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/993d9be2-0891-4680-a0f9-702bb07aa5b3-utilities\") pod \"993d9be2-0891-4680-a0f9-702bb07aa5b3\" (UID: \"993d9be2-0891-4680-a0f9-702bb07aa5b3\") " Feb 02 09:41:55 crc kubenswrapper[4764]: I0202 09:41:55.025613 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/993d9be2-0891-4680-a0f9-702bb07aa5b3-utilities" (OuterVolumeSpecName: "utilities") pod "993d9be2-0891-4680-a0f9-702bb07aa5b3" (UID: "993d9be2-0891-4680-a0f9-702bb07aa5b3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 09:41:55 crc kubenswrapper[4764]: I0202 09:41:55.025823 4764 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/993d9be2-0891-4680-a0f9-702bb07aa5b3-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 09:41:55 crc kubenswrapper[4764]: I0202 09:41:55.043347 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/993d9be2-0891-4680-a0f9-702bb07aa5b3-kube-api-access-pr6t9" (OuterVolumeSpecName: "kube-api-access-pr6t9") pod "993d9be2-0891-4680-a0f9-702bb07aa5b3" (UID: "993d9be2-0891-4680-a0f9-702bb07aa5b3"). InnerVolumeSpecName "kube-api-access-pr6t9". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:41:55 crc kubenswrapper[4764]: I0202 09:41:55.127357 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pr6t9\" (UniqueName: \"kubernetes.io/projected/993d9be2-0891-4680-a0f9-702bb07aa5b3-kube-api-access-pr6t9\") on node \"crc\" DevicePath \"\"" Feb 02 09:41:55 crc kubenswrapper[4764]: I0202 09:41:55.154492 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/993d9be2-0891-4680-a0f9-702bb07aa5b3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "993d9be2-0891-4680-a0f9-702bb07aa5b3" (UID: "993d9be2-0891-4680-a0f9-702bb07aa5b3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 09:41:55 crc kubenswrapper[4764]: I0202 09:41:55.228489 4764 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/993d9be2-0891-4680-a0f9-702bb07aa5b3-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 09:41:55 crc kubenswrapper[4764]: I0202 09:41:55.384365 4764 generic.go:334] "Generic (PLEG): container finished" podID="993d9be2-0891-4680-a0f9-702bb07aa5b3" containerID="5cbad5620f17be7545a4087fce4cb01e5cbf2bd29fe11c813fa46b0531b70a81" exitCode=0 Feb 02 09:41:55 crc kubenswrapper[4764]: I0202 09:41:55.384416 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bzbm2" event={"ID":"993d9be2-0891-4680-a0f9-702bb07aa5b3","Type":"ContainerDied","Data":"5cbad5620f17be7545a4087fce4cb01e5cbf2bd29fe11c813fa46b0531b70a81"} Feb 02 09:41:55 crc kubenswrapper[4764]: I0202 09:41:55.384425 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bzbm2" Feb 02 09:41:55 crc kubenswrapper[4764]: I0202 09:41:55.384452 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bzbm2" event={"ID":"993d9be2-0891-4680-a0f9-702bb07aa5b3","Type":"ContainerDied","Data":"b52fbaa26eb077d0a6f5412c2d063e8d06c31e91506d5ce454a792d2e2bd7edc"} Feb 02 09:41:55 crc kubenswrapper[4764]: I0202 09:41:55.384473 4764 scope.go:117] "RemoveContainer" containerID="5cbad5620f17be7545a4087fce4cb01e5cbf2bd29fe11c813fa46b0531b70a81" Feb 02 09:41:55 crc kubenswrapper[4764]: I0202 09:41:55.407622 4764 scope.go:117] "RemoveContainer" containerID="a213d40fbda79a1198b3f6e1e7131adcef3f5a4def5fea28c01231f8b28b7a8a" Feb 02 09:41:55 crc kubenswrapper[4764]: I0202 09:41:55.443752 4764 scope.go:117] "RemoveContainer" containerID="d4711358de4b369d0913995858fe499153a6c2ae764ee62a3939d9db9a168533" Feb 02 09:41:55 crc kubenswrapper[4764]: I0202 09:41:55.444197 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bzbm2"] Feb 02 09:41:55 crc kubenswrapper[4764]: I0202 09:41:55.458698 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-bzbm2"] Feb 02 09:41:55 crc kubenswrapper[4764]: I0202 09:41:55.472479 4764 scope.go:117] "RemoveContainer" containerID="5cbad5620f17be7545a4087fce4cb01e5cbf2bd29fe11c813fa46b0531b70a81" Feb 02 09:41:55 crc kubenswrapper[4764]: E0202 09:41:55.473062 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5cbad5620f17be7545a4087fce4cb01e5cbf2bd29fe11c813fa46b0531b70a81\": container with ID starting with 5cbad5620f17be7545a4087fce4cb01e5cbf2bd29fe11c813fa46b0531b70a81 not found: ID does not exist" containerID="5cbad5620f17be7545a4087fce4cb01e5cbf2bd29fe11c813fa46b0531b70a81" Feb 02 09:41:55 crc kubenswrapper[4764]: I0202 09:41:55.473111 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5cbad5620f17be7545a4087fce4cb01e5cbf2bd29fe11c813fa46b0531b70a81"} err="failed to get container status \"5cbad5620f17be7545a4087fce4cb01e5cbf2bd29fe11c813fa46b0531b70a81\": rpc error: code = NotFound desc = could not find container \"5cbad5620f17be7545a4087fce4cb01e5cbf2bd29fe11c813fa46b0531b70a81\": container with ID starting with 5cbad5620f17be7545a4087fce4cb01e5cbf2bd29fe11c813fa46b0531b70a81 not found: ID does not exist" Feb 02 09:41:55 crc kubenswrapper[4764]: I0202 09:41:55.473136 4764 scope.go:117] "RemoveContainer" containerID="a213d40fbda79a1198b3f6e1e7131adcef3f5a4def5fea28c01231f8b28b7a8a" Feb 02 09:41:55 crc kubenswrapper[4764]: E0202 09:41:55.473402 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a213d40fbda79a1198b3f6e1e7131adcef3f5a4def5fea28c01231f8b28b7a8a\": container with ID starting with a213d40fbda79a1198b3f6e1e7131adcef3f5a4def5fea28c01231f8b28b7a8a not found: ID does not exist" containerID="a213d40fbda79a1198b3f6e1e7131adcef3f5a4def5fea28c01231f8b28b7a8a" Feb 02 09:41:55 crc kubenswrapper[4764]: I0202 09:41:55.473433 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a213d40fbda79a1198b3f6e1e7131adcef3f5a4def5fea28c01231f8b28b7a8a"} err="failed to get container status \"a213d40fbda79a1198b3f6e1e7131adcef3f5a4def5fea28c01231f8b28b7a8a\": rpc error: code = NotFound desc = could not find container \"a213d40fbda79a1198b3f6e1e7131adcef3f5a4def5fea28c01231f8b28b7a8a\": container with ID starting with a213d40fbda79a1198b3f6e1e7131adcef3f5a4def5fea28c01231f8b28b7a8a not found: ID does not exist" Feb 02 09:41:55 crc kubenswrapper[4764]: I0202 09:41:55.473452 4764 scope.go:117] "RemoveContainer" containerID="d4711358de4b369d0913995858fe499153a6c2ae764ee62a3939d9db9a168533" Feb 02 09:41:55 crc kubenswrapper[4764]: E0202 09:41:55.473752 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d4711358de4b369d0913995858fe499153a6c2ae764ee62a3939d9db9a168533\": container with ID starting with d4711358de4b369d0913995858fe499153a6c2ae764ee62a3939d9db9a168533 not found: ID does not exist" containerID="d4711358de4b369d0913995858fe499153a6c2ae764ee62a3939d9db9a168533" Feb 02 09:41:55 crc kubenswrapper[4764]: I0202 09:41:55.473783 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d4711358de4b369d0913995858fe499153a6c2ae764ee62a3939d9db9a168533"} err="failed to get container status \"d4711358de4b369d0913995858fe499153a6c2ae764ee62a3939d9db9a168533\": rpc error: code = NotFound desc = could not find container \"d4711358de4b369d0913995858fe499153a6c2ae764ee62a3939d9db9a168533\": container with ID starting with d4711358de4b369d0913995858fe499153a6c2ae764ee62a3939d9db9a168533 not found: ID does not exist" Feb 02 09:41:55 crc kubenswrapper[4764]: I0202 09:41:55.839684 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="993d9be2-0891-4680-a0f9-702bb07aa5b3" path="/var/lib/kubelet/pods/993d9be2-0891-4680-a0f9-702bb07aa5b3/volumes" Feb 02 09:42:05 crc kubenswrapper[4764]: I0202 09:42:05.117209 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5qvdd"] Feb 02 09:42:05 crc kubenswrapper[4764]: I0202 09:42:05.127547 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-j7qmc"] Feb 02 09:42:05 crc kubenswrapper[4764]: I0202 09:42:05.148060 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5qvdd"] Feb 02 09:42:05 crc kubenswrapper[4764]: I0202 09:42:05.152477 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-j7qmc"] Feb 02 09:42:05 crc kubenswrapper[4764]: I0202 09:42:05.164810 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-rcpbs"] Feb 02 09:42:05 crc kubenswrapper[4764]: I0202 09:42:05.172852 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gv65q"] Feb 02 09:42:05 crc kubenswrapper[4764]: I0202 09:42:05.186116 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-8rfqp"] Feb 02 09:42:05 crc kubenswrapper[4764]: I0202 09:42:05.195000 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-8rfqp"] Feb 02 09:42:05 crc kubenswrapper[4764]: I0202 09:42:05.206182 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-rcpbs"] Feb 02 09:42:05 crc kubenswrapper[4764]: I0202 09:42:05.211149 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gv65q"] Feb 02 09:42:05 crc kubenswrapper[4764]: I0202 09:42:05.222273 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-nx4kb"] Feb 02 09:42:05 crc kubenswrapper[4764]: I0202 09:42:05.226117 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-gklb7"] Feb 02 09:42:05 crc kubenswrapper[4764]: I0202 09:42:05.233726 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fzl68"] Feb 02 09:42:05 crc kubenswrapper[4764]: I0202 09:42:05.243765 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-ck2zz"] Feb 02 09:42:05 crc kubenswrapper[4764]: I0202 09:42:05.249996 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-d928p"] Feb 02 09:42:05 crc kubenswrapper[4764]: I0202 09:42:05.259982 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-nx4kb"] Feb 02 09:42:05 crc kubenswrapper[4764]: I0202 09:42:05.265498 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-gklb7"] Feb 02 09:42:05 crc kubenswrapper[4764]: I0202 09:42:05.277351 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-d928p"] Feb 02 09:42:05 crc kubenswrapper[4764]: I0202 09:42:05.286301 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fzl68"] Feb 02 09:42:05 crc kubenswrapper[4764]: I0202 09:42:05.292574 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-ck2zz"] Feb 02 09:42:05 crc kubenswrapper[4764]: I0202 09:42:05.836028 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="203f8fcc-2055-40f5-8ef6-7d00d658157b" path="/var/lib/kubelet/pods/203f8fcc-2055-40f5-8ef6-7d00d658157b/volumes" Feb 02 09:42:05 crc kubenswrapper[4764]: I0202 09:42:05.836635 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="502ceb21-d3d1-4638-a565-2223d41fd1c0" path="/var/lib/kubelet/pods/502ceb21-d3d1-4638-a565-2223d41fd1c0/volumes" Feb 02 09:42:05 crc kubenswrapper[4764]: I0202 09:42:05.837200 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="667bdcf6-12ac-4bfc-b652-48ce7aca1fc2" path="/var/lib/kubelet/pods/667bdcf6-12ac-4bfc-b652-48ce7aca1fc2/volumes" Feb 02 09:42:05 crc kubenswrapper[4764]: I0202 09:42:05.837720 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="82028751-8b0c-4bab-b6a1-2542acd6e650" path="/var/lib/kubelet/pods/82028751-8b0c-4bab-b6a1-2542acd6e650/volumes" Feb 02 09:42:05 crc kubenswrapper[4764]: I0202 09:42:05.838760 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9b9af3ad-4a64-4996-bcdc-220df06db737" path="/var/lib/kubelet/pods/9b9af3ad-4a64-4996-bcdc-220df06db737/volumes" Feb 02 09:42:05 crc kubenswrapper[4764]: I0202 09:42:05.839313 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a4f5b290-7d44-4c9b-bffb-78a78f8bd546" path="/var/lib/kubelet/pods/a4f5b290-7d44-4c9b-bffb-78a78f8bd546/volumes" Feb 02 09:42:05 crc kubenswrapper[4764]: I0202 09:42:05.839860 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b60a0abc-6093-4ab4-b8ac-5bd6220d1533" path="/var/lib/kubelet/pods/b60a0abc-6093-4ab4-b8ac-5bd6220d1533/volumes" Feb 02 09:42:05 crc kubenswrapper[4764]: I0202 09:42:05.840874 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c66a8ab1-6fbb-4301-a4ce-34ddad011ee2" path="/var/lib/kubelet/pods/c66a8ab1-6fbb-4301-a4ce-34ddad011ee2/volumes" Feb 02 09:42:05 crc kubenswrapper[4764]: I0202 09:42:05.841477 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d04cc0fd-f303-47b2-a8d4-6efd0a591b77" path="/var/lib/kubelet/pods/d04cc0fd-f303-47b2-a8d4-6efd0a591b77/volumes" Feb 02 09:42:05 crc kubenswrapper[4764]: I0202 09:42:05.842015 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dbda7bef-a208-4766-ad9a-8cf38c77054a" path="/var/lib/kubelet/pods/dbda7bef-a208-4766-ad9a-8cf38c77054a/volumes" Feb 02 09:42:18 crc kubenswrapper[4764]: I0202 09:42:18.024285 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8f7dl"] Feb 02 09:42:18 crc kubenswrapper[4764]: E0202 09:42:18.025163 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="993d9be2-0891-4680-a0f9-702bb07aa5b3" containerName="extract-content" Feb 02 09:42:18 crc kubenswrapper[4764]: I0202 09:42:18.025177 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="993d9be2-0891-4680-a0f9-702bb07aa5b3" containerName="extract-content" Feb 02 09:42:18 crc kubenswrapper[4764]: E0202 09:42:18.025193 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="993d9be2-0891-4680-a0f9-702bb07aa5b3" containerName="extract-utilities" Feb 02 09:42:18 crc kubenswrapper[4764]: I0202 09:42:18.025200 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="993d9be2-0891-4680-a0f9-702bb07aa5b3" containerName="extract-utilities" Feb 02 09:42:18 crc kubenswrapper[4764]: E0202 09:42:18.025215 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="993d9be2-0891-4680-a0f9-702bb07aa5b3" containerName="registry-server" Feb 02 09:42:18 crc kubenswrapper[4764]: I0202 09:42:18.025221 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="993d9be2-0891-4680-a0f9-702bb07aa5b3" containerName="registry-server" Feb 02 09:42:18 crc kubenswrapper[4764]: I0202 09:42:18.025390 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="993d9be2-0891-4680-a0f9-702bb07aa5b3" containerName="registry-server" Feb 02 09:42:18 crc kubenswrapper[4764]: I0202 09:42:18.026040 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8f7dl" Feb 02 09:42:18 crc kubenswrapper[4764]: I0202 09:42:18.029716 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 02 09:42:18 crc kubenswrapper[4764]: I0202 09:42:18.029892 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 02 09:42:18 crc kubenswrapper[4764]: I0202 09:42:18.030067 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-xksq5" Feb 02 09:42:18 crc kubenswrapper[4764]: I0202 09:42:18.030183 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Feb 02 09:42:18 crc kubenswrapper[4764]: I0202 09:42:18.041333 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 02 09:42:18 crc kubenswrapper[4764]: I0202 09:42:18.049822 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8f7dl"] Feb 02 09:42:18 crc kubenswrapper[4764]: I0202 09:42:18.210062 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9qd69\" (UniqueName: \"kubernetes.io/projected/feb92d24-e755-4837-8461-15e7e3a7b0e8-kube-api-access-9qd69\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-8f7dl\" (UID: \"feb92d24-e755-4837-8461-15e7e3a7b0e8\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8f7dl" Feb 02 09:42:18 crc kubenswrapper[4764]: I0202 09:42:18.210128 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/feb92d24-e755-4837-8461-15e7e3a7b0e8-ssh-key-openstack-edpm-ipam\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-8f7dl\" (UID: \"feb92d24-e755-4837-8461-15e7e3a7b0e8\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8f7dl" Feb 02 09:42:18 crc kubenswrapper[4764]: I0202 09:42:18.210323 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/feb92d24-e755-4837-8461-15e7e3a7b0e8-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-8f7dl\" (UID: \"feb92d24-e755-4837-8461-15e7e3a7b0e8\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8f7dl" Feb 02 09:42:18 crc kubenswrapper[4764]: I0202 09:42:18.210583 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/feb92d24-e755-4837-8461-15e7e3a7b0e8-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-8f7dl\" (UID: \"feb92d24-e755-4837-8461-15e7e3a7b0e8\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8f7dl" Feb 02 09:42:18 crc kubenswrapper[4764]: I0202 09:42:18.210766 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/feb92d24-e755-4837-8461-15e7e3a7b0e8-ceph\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-8f7dl\" (UID: \"feb92d24-e755-4837-8461-15e7e3a7b0e8\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8f7dl" Feb 02 09:42:18 crc kubenswrapper[4764]: I0202 09:42:18.312823 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/feb92d24-e755-4837-8461-15e7e3a7b0e8-ceph\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-8f7dl\" (UID: \"feb92d24-e755-4837-8461-15e7e3a7b0e8\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8f7dl" Feb 02 09:42:18 crc kubenswrapper[4764]: I0202 09:42:18.312875 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9qd69\" (UniqueName: \"kubernetes.io/projected/feb92d24-e755-4837-8461-15e7e3a7b0e8-kube-api-access-9qd69\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-8f7dl\" (UID: \"feb92d24-e755-4837-8461-15e7e3a7b0e8\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8f7dl" Feb 02 09:42:18 crc kubenswrapper[4764]: I0202 09:42:18.312921 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/feb92d24-e755-4837-8461-15e7e3a7b0e8-ssh-key-openstack-edpm-ipam\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-8f7dl\" (UID: \"feb92d24-e755-4837-8461-15e7e3a7b0e8\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8f7dl" Feb 02 09:42:18 crc kubenswrapper[4764]: I0202 09:42:18.312999 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/feb92d24-e755-4837-8461-15e7e3a7b0e8-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-8f7dl\" (UID: \"feb92d24-e755-4837-8461-15e7e3a7b0e8\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8f7dl" Feb 02 09:42:18 crc kubenswrapper[4764]: I0202 09:42:18.313087 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/feb92d24-e755-4837-8461-15e7e3a7b0e8-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-8f7dl\" (UID: \"feb92d24-e755-4837-8461-15e7e3a7b0e8\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8f7dl" Feb 02 09:42:18 crc kubenswrapper[4764]: I0202 09:42:18.320710 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/feb92d24-e755-4837-8461-15e7e3a7b0e8-ceph\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-8f7dl\" (UID: \"feb92d24-e755-4837-8461-15e7e3a7b0e8\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8f7dl" Feb 02 09:42:18 crc kubenswrapper[4764]: I0202 09:42:18.320848 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/feb92d24-e755-4837-8461-15e7e3a7b0e8-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-8f7dl\" (UID: \"feb92d24-e755-4837-8461-15e7e3a7b0e8\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8f7dl" Feb 02 09:42:18 crc kubenswrapper[4764]: I0202 09:42:18.324718 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/feb92d24-e755-4837-8461-15e7e3a7b0e8-ssh-key-openstack-edpm-ipam\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-8f7dl\" (UID: \"feb92d24-e755-4837-8461-15e7e3a7b0e8\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8f7dl" Feb 02 09:42:18 crc kubenswrapper[4764]: I0202 09:42:18.325255 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/feb92d24-e755-4837-8461-15e7e3a7b0e8-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-8f7dl\" (UID: \"feb92d24-e755-4837-8461-15e7e3a7b0e8\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8f7dl" Feb 02 09:42:18 crc kubenswrapper[4764]: I0202 09:42:18.351480 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9qd69\" (UniqueName: \"kubernetes.io/projected/feb92d24-e755-4837-8461-15e7e3a7b0e8-kube-api-access-9qd69\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-8f7dl\" (UID: \"feb92d24-e755-4837-8461-15e7e3a7b0e8\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8f7dl" Feb 02 09:42:18 crc kubenswrapper[4764]: I0202 09:42:18.648368 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8f7dl" Feb 02 09:42:19 crc kubenswrapper[4764]: I0202 09:42:19.221399 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8f7dl"] Feb 02 09:42:19 crc kubenswrapper[4764]: I0202 09:42:19.619821 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8f7dl" event={"ID":"feb92d24-e755-4837-8461-15e7e3a7b0e8","Type":"ContainerStarted","Data":"6b70a3e91b7836548e791a9488bac5286ef435617f54f3e65f5a6bc3c85b2a1d"} Feb 02 09:42:20 crc kubenswrapper[4764]: I0202 09:42:20.630731 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8f7dl" event={"ID":"feb92d24-e755-4837-8461-15e7e3a7b0e8","Type":"ContainerStarted","Data":"d054b6feac5f24fdbc1241dbf835fa21e9d41f9da2bba4f1ed71669be4af81f3"} Feb 02 09:42:26 crc kubenswrapper[4764]: I0202 09:42:26.740331 4764 scope.go:117] "RemoveContainer" containerID="f9857c237746b859c2245f63119218034d063a2161209dbe4c8439af6a1eaf49" Feb 02 09:42:26 crc kubenswrapper[4764]: I0202 09:42:26.776831 4764 scope.go:117] "RemoveContainer" containerID="f130cc11505e280b34eb9b6df346654a0fad23fde9ea5194446dd5fed996e091" Feb 02 09:42:26 crc kubenswrapper[4764]: I0202 09:42:26.814894 4764 scope.go:117] "RemoveContainer" containerID="6940e5f96de57f57e75f49942e77472a14857af61522b9d9c0e931feb7e6a97e" Feb 02 09:42:26 crc kubenswrapper[4764]: I0202 09:42:26.878354 4764 scope.go:117] "RemoveContainer" containerID="a358cb778023f9178a5758199b3a678d46fa14535321922c333a5d5a68a31834" Feb 02 09:42:26 crc kubenswrapper[4764]: I0202 09:42:26.947815 4764 scope.go:117] "RemoveContainer" containerID="3dbab190576e3e5b1dd4bf491ee4ec587fedd9b0a05df873f446d73c04191ead" Feb 02 09:42:26 crc kubenswrapper[4764]: I0202 09:42:26.997424 4764 scope.go:117] "RemoveContainer" containerID="3782d7a113fa14a8d1b7a9b1c1ff7a6a278ec65b1543752bd6307e2d11046199" Feb 02 09:42:27 crc kubenswrapper[4764]: I0202 09:42:27.060183 4764 scope.go:117] "RemoveContainer" containerID="72a3eb32dbfc5f69f955c69ab48f7b3367908aed17f2d9f5d6311123e797ccbc" Feb 02 09:42:27 crc kubenswrapper[4764]: I0202 09:42:27.149427 4764 scope.go:117] "RemoveContainer" containerID="2a30b88b0273d89dc757e5d0af4296f1781e0e52d78980006db3cb7f8b222a2a" Feb 02 09:42:27 crc kubenswrapper[4764]: I0202 09:42:27.183178 4764 scope.go:117] "RemoveContainer" containerID="992b91746b103206d2c4c9472b0c41ddf30a14e060fded00f2f48b6d51734e8f" Feb 02 09:42:33 crc kubenswrapper[4764]: I0202 09:42:33.746169 4764 generic.go:334] "Generic (PLEG): container finished" podID="feb92d24-e755-4837-8461-15e7e3a7b0e8" containerID="d054b6feac5f24fdbc1241dbf835fa21e9d41f9da2bba4f1ed71669be4af81f3" exitCode=0 Feb 02 09:42:33 crc kubenswrapper[4764]: I0202 09:42:33.746751 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8f7dl" event={"ID":"feb92d24-e755-4837-8461-15e7e3a7b0e8","Type":"ContainerDied","Data":"d054b6feac5f24fdbc1241dbf835fa21e9d41f9da2bba4f1ed71669be4af81f3"} Feb 02 09:42:35 crc kubenswrapper[4764]: I0202 09:42:35.232489 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8f7dl" Feb 02 09:42:35 crc kubenswrapper[4764]: I0202 09:42:35.270952 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9qd69\" (UniqueName: \"kubernetes.io/projected/feb92d24-e755-4837-8461-15e7e3a7b0e8-kube-api-access-9qd69\") pod \"feb92d24-e755-4837-8461-15e7e3a7b0e8\" (UID: \"feb92d24-e755-4837-8461-15e7e3a7b0e8\") " Feb 02 09:42:35 crc kubenswrapper[4764]: I0202 09:42:35.271100 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/feb92d24-e755-4837-8461-15e7e3a7b0e8-ssh-key-openstack-edpm-ipam\") pod \"feb92d24-e755-4837-8461-15e7e3a7b0e8\" (UID: \"feb92d24-e755-4837-8461-15e7e3a7b0e8\") " Feb 02 09:42:35 crc kubenswrapper[4764]: I0202 09:42:35.271129 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/feb92d24-e755-4837-8461-15e7e3a7b0e8-inventory\") pod \"feb92d24-e755-4837-8461-15e7e3a7b0e8\" (UID: \"feb92d24-e755-4837-8461-15e7e3a7b0e8\") " Feb 02 09:42:35 crc kubenswrapper[4764]: I0202 09:42:35.271163 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/feb92d24-e755-4837-8461-15e7e3a7b0e8-repo-setup-combined-ca-bundle\") pod \"feb92d24-e755-4837-8461-15e7e3a7b0e8\" (UID: \"feb92d24-e755-4837-8461-15e7e3a7b0e8\") " Feb 02 09:42:35 crc kubenswrapper[4764]: I0202 09:42:35.271220 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/feb92d24-e755-4837-8461-15e7e3a7b0e8-ceph\") pod \"feb92d24-e755-4837-8461-15e7e3a7b0e8\" (UID: \"feb92d24-e755-4837-8461-15e7e3a7b0e8\") " Feb 02 09:42:35 crc kubenswrapper[4764]: I0202 09:42:35.296765 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/feb92d24-e755-4837-8461-15e7e3a7b0e8-kube-api-access-9qd69" (OuterVolumeSpecName: "kube-api-access-9qd69") pod "feb92d24-e755-4837-8461-15e7e3a7b0e8" (UID: "feb92d24-e755-4837-8461-15e7e3a7b0e8"). InnerVolumeSpecName "kube-api-access-9qd69". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:42:35 crc kubenswrapper[4764]: I0202 09:42:35.299734 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/feb92d24-e755-4837-8461-15e7e3a7b0e8-ceph" (OuterVolumeSpecName: "ceph") pod "feb92d24-e755-4837-8461-15e7e3a7b0e8" (UID: "feb92d24-e755-4837-8461-15e7e3a7b0e8"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:42:35 crc kubenswrapper[4764]: I0202 09:42:35.299108 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/feb92d24-e755-4837-8461-15e7e3a7b0e8-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "feb92d24-e755-4837-8461-15e7e3a7b0e8" (UID: "feb92d24-e755-4837-8461-15e7e3a7b0e8"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:42:35 crc kubenswrapper[4764]: I0202 09:42:35.306966 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/feb92d24-e755-4837-8461-15e7e3a7b0e8-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "feb92d24-e755-4837-8461-15e7e3a7b0e8" (UID: "feb92d24-e755-4837-8461-15e7e3a7b0e8"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:42:35 crc kubenswrapper[4764]: I0202 09:42:35.313134 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/feb92d24-e755-4837-8461-15e7e3a7b0e8-inventory" (OuterVolumeSpecName: "inventory") pod "feb92d24-e755-4837-8461-15e7e3a7b0e8" (UID: "feb92d24-e755-4837-8461-15e7e3a7b0e8"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:42:35 crc kubenswrapper[4764]: I0202 09:42:35.372666 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9qd69\" (UniqueName: \"kubernetes.io/projected/feb92d24-e755-4837-8461-15e7e3a7b0e8-kube-api-access-9qd69\") on node \"crc\" DevicePath \"\"" Feb 02 09:42:35 crc kubenswrapper[4764]: I0202 09:42:35.372701 4764 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/feb92d24-e755-4837-8461-15e7e3a7b0e8-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 02 09:42:35 crc kubenswrapper[4764]: I0202 09:42:35.372712 4764 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/feb92d24-e755-4837-8461-15e7e3a7b0e8-inventory\") on node \"crc\" DevicePath \"\"" Feb 02 09:42:35 crc kubenswrapper[4764]: I0202 09:42:35.372722 4764 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/feb92d24-e755-4837-8461-15e7e3a7b0e8-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 09:42:35 crc kubenswrapper[4764]: I0202 09:42:35.372734 4764 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/feb92d24-e755-4837-8461-15e7e3a7b0e8-ceph\") on node \"crc\" DevicePath \"\"" Feb 02 09:42:35 crc kubenswrapper[4764]: I0202 09:42:35.765841 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8f7dl" event={"ID":"feb92d24-e755-4837-8461-15e7e3a7b0e8","Type":"ContainerDied","Data":"6b70a3e91b7836548e791a9488bac5286ef435617f54f3e65f5a6bc3c85b2a1d"} Feb 02 09:42:35 crc kubenswrapper[4764]: I0202 09:42:35.765890 4764 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6b70a3e91b7836548e791a9488bac5286ef435617f54f3e65f5a6bc3c85b2a1d" Feb 02 09:42:35 crc kubenswrapper[4764]: I0202 09:42:35.765927 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8f7dl" Feb 02 09:42:35 crc kubenswrapper[4764]: I0202 09:42:35.867760 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-d9x8m"] Feb 02 09:42:35 crc kubenswrapper[4764]: E0202 09:42:35.868261 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="feb92d24-e755-4837-8461-15e7e3a7b0e8" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Feb 02 09:42:35 crc kubenswrapper[4764]: I0202 09:42:35.868287 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="feb92d24-e755-4837-8461-15e7e3a7b0e8" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Feb 02 09:42:35 crc kubenswrapper[4764]: I0202 09:42:35.868519 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="feb92d24-e755-4837-8461-15e7e3a7b0e8" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Feb 02 09:42:35 crc kubenswrapper[4764]: I0202 09:42:35.869277 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-d9x8m" Feb 02 09:42:35 crc kubenswrapper[4764]: I0202 09:42:35.873121 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 02 09:42:35 crc kubenswrapper[4764]: I0202 09:42:35.873351 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 02 09:42:35 crc kubenswrapper[4764]: I0202 09:42:35.873530 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Feb 02 09:42:35 crc kubenswrapper[4764]: I0202 09:42:35.873690 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-xksq5" Feb 02 09:42:35 crc kubenswrapper[4764]: I0202 09:42:35.876019 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 02 09:42:35 crc kubenswrapper[4764]: I0202 09:42:35.891743 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-d9x8m"] Feb 02 09:42:35 crc kubenswrapper[4764]: I0202 09:42:35.980343 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af835a1d-09cc-49ca-979f-41eaf146e926-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-d9x8m\" (UID: \"af835a1d-09cc-49ca-979f-41eaf146e926\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-d9x8m" Feb 02 09:42:35 crc kubenswrapper[4764]: I0202 09:42:35.980672 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fn6s9\" (UniqueName: \"kubernetes.io/projected/af835a1d-09cc-49ca-979f-41eaf146e926-kube-api-access-fn6s9\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-d9x8m\" (UID: \"af835a1d-09cc-49ca-979f-41eaf146e926\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-d9x8m" Feb 02 09:42:35 crc kubenswrapper[4764]: I0202 09:42:35.980841 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/af835a1d-09cc-49ca-979f-41eaf146e926-ssh-key-openstack-edpm-ipam\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-d9x8m\" (UID: \"af835a1d-09cc-49ca-979f-41eaf146e926\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-d9x8m" Feb 02 09:42:35 crc kubenswrapper[4764]: I0202 09:42:35.981033 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/af835a1d-09cc-49ca-979f-41eaf146e926-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-d9x8m\" (UID: \"af835a1d-09cc-49ca-979f-41eaf146e926\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-d9x8m" Feb 02 09:42:35 crc kubenswrapper[4764]: I0202 09:42:35.981207 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/af835a1d-09cc-49ca-979f-41eaf146e926-ceph\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-d9x8m\" (UID: \"af835a1d-09cc-49ca-979f-41eaf146e926\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-d9x8m" Feb 02 09:42:36 crc kubenswrapper[4764]: I0202 09:42:36.083657 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af835a1d-09cc-49ca-979f-41eaf146e926-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-d9x8m\" (UID: \"af835a1d-09cc-49ca-979f-41eaf146e926\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-d9x8m" Feb 02 09:42:36 crc kubenswrapper[4764]: I0202 09:42:36.083748 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fn6s9\" (UniqueName: \"kubernetes.io/projected/af835a1d-09cc-49ca-979f-41eaf146e926-kube-api-access-fn6s9\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-d9x8m\" (UID: \"af835a1d-09cc-49ca-979f-41eaf146e926\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-d9x8m" Feb 02 09:42:36 crc kubenswrapper[4764]: I0202 09:42:36.083839 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/af835a1d-09cc-49ca-979f-41eaf146e926-ssh-key-openstack-edpm-ipam\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-d9x8m\" (UID: \"af835a1d-09cc-49ca-979f-41eaf146e926\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-d9x8m" Feb 02 09:42:36 crc kubenswrapper[4764]: I0202 09:42:36.083926 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/af835a1d-09cc-49ca-979f-41eaf146e926-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-d9x8m\" (UID: \"af835a1d-09cc-49ca-979f-41eaf146e926\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-d9x8m" Feb 02 09:42:36 crc kubenswrapper[4764]: I0202 09:42:36.084014 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/af835a1d-09cc-49ca-979f-41eaf146e926-ceph\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-d9x8m\" (UID: \"af835a1d-09cc-49ca-979f-41eaf146e926\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-d9x8m" Feb 02 09:42:36 crc kubenswrapper[4764]: I0202 09:42:36.094177 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af835a1d-09cc-49ca-979f-41eaf146e926-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-d9x8m\" (UID: \"af835a1d-09cc-49ca-979f-41eaf146e926\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-d9x8m" Feb 02 09:42:36 crc kubenswrapper[4764]: I0202 09:42:36.095442 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/af835a1d-09cc-49ca-979f-41eaf146e926-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-d9x8m\" (UID: \"af835a1d-09cc-49ca-979f-41eaf146e926\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-d9x8m" Feb 02 09:42:36 crc kubenswrapper[4764]: I0202 09:42:36.095811 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/af835a1d-09cc-49ca-979f-41eaf146e926-ssh-key-openstack-edpm-ipam\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-d9x8m\" (UID: \"af835a1d-09cc-49ca-979f-41eaf146e926\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-d9x8m" Feb 02 09:42:36 crc kubenswrapper[4764]: I0202 09:42:36.096377 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/af835a1d-09cc-49ca-979f-41eaf146e926-ceph\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-d9x8m\" (UID: \"af835a1d-09cc-49ca-979f-41eaf146e926\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-d9x8m" Feb 02 09:42:36 crc kubenswrapper[4764]: I0202 09:42:36.105022 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fn6s9\" (UniqueName: \"kubernetes.io/projected/af835a1d-09cc-49ca-979f-41eaf146e926-kube-api-access-fn6s9\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-d9x8m\" (UID: \"af835a1d-09cc-49ca-979f-41eaf146e926\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-d9x8m" Feb 02 09:42:36 crc kubenswrapper[4764]: I0202 09:42:36.192325 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-d9x8m" Feb 02 09:42:36 crc kubenswrapper[4764]: I0202 09:42:36.911402 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-d9x8m"] Feb 02 09:42:37 crc kubenswrapper[4764]: I0202 09:42:37.793270 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-d9x8m" event={"ID":"af835a1d-09cc-49ca-979f-41eaf146e926","Type":"ContainerStarted","Data":"11eea62e5a292de4fd143d20114bbdddfdf73d0ccfe3c45abc9e58813bd4e267"} Feb 02 09:42:37 crc kubenswrapper[4764]: I0202 09:42:37.793806 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-d9x8m" event={"ID":"af835a1d-09cc-49ca-979f-41eaf146e926","Type":"ContainerStarted","Data":"2bb25760321f502b97dd1c83efb0ca860f9417899597ef4ed76035522ae4cab0"} Feb 02 09:42:37 crc kubenswrapper[4764]: I0202 09:42:37.830279 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-d9x8m" podStartSLOduration=2.367569653 podStartE2EDuration="2.830252734s" podCreationTimestamp="2026-02-02 09:42:35 +0000 UTC" firstStartedPulling="2026-02-02 09:42:36.940731904 +0000 UTC m=+2119.874456002" lastFinishedPulling="2026-02-02 09:42:37.403414995 +0000 UTC m=+2120.337139083" observedRunningTime="2026-02-02 09:42:37.823408483 +0000 UTC m=+2120.757132571" watchObservedRunningTime="2026-02-02 09:42:37.830252734 +0000 UTC m=+2120.763976842" Feb 02 09:43:27 crc kubenswrapper[4764]: I0202 09:43:27.446691 4764 scope.go:117] "RemoveContainer" containerID="8a1dfe2f4d9180101b312e72b2b479bc433502cf047e61bd017d6d4e1c6a0273" Feb 02 09:43:43 crc kubenswrapper[4764]: I0202 09:43:43.522834 4764 patch_prober.go:28] interesting pod/machine-config-daemon-4ndm4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 09:43:43 crc kubenswrapper[4764]: I0202 09:43:43.523498 4764 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 09:43:59 crc kubenswrapper[4764]: I0202 09:43:59.121879 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-bwr6d"] Feb 02 09:43:59 crc kubenswrapper[4764]: I0202 09:43:59.125925 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bwr6d" Feb 02 09:43:59 crc kubenswrapper[4764]: I0202 09:43:59.157916 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bwr6d"] Feb 02 09:43:59 crc kubenswrapper[4764]: I0202 09:43:59.162588 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/82ca325a-895f-4a03-995f-96e014397404-utilities\") pod \"redhat-marketplace-bwr6d\" (UID: \"82ca325a-895f-4a03-995f-96e014397404\") " pod="openshift-marketplace/redhat-marketplace-bwr6d" Feb 02 09:43:59 crc kubenswrapper[4764]: I0202 09:43:59.162723 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kxh29\" (UniqueName: \"kubernetes.io/projected/82ca325a-895f-4a03-995f-96e014397404-kube-api-access-kxh29\") pod \"redhat-marketplace-bwr6d\" (UID: \"82ca325a-895f-4a03-995f-96e014397404\") " pod="openshift-marketplace/redhat-marketplace-bwr6d" Feb 02 09:43:59 crc kubenswrapper[4764]: I0202 09:43:59.162837 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/82ca325a-895f-4a03-995f-96e014397404-catalog-content\") pod \"redhat-marketplace-bwr6d\" (UID: \"82ca325a-895f-4a03-995f-96e014397404\") " pod="openshift-marketplace/redhat-marketplace-bwr6d" Feb 02 09:43:59 crc kubenswrapper[4764]: I0202 09:43:59.264256 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kxh29\" (UniqueName: \"kubernetes.io/projected/82ca325a-895f-4a03-995f-96e014397404-kube-api-access-kxh29\") pod \"redhat-marketplace-bwr6d\" (UID: \"82ca325a-895f-4a03-995f-96e014397404\") " pod="openshift-marketplace/redhat-marketplace-bwr6d" Feb 02 09:43:59 crc kubenswrapper[4764]: I0202 09:43:59.264355 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/82ca325a-895f-4a03-995f-96e014397404-catalog-content\") pod \"redhat-marketplace-bwr6d\" (UID: \"82ca325a-895f-4a03-995f-96e014397404\") " pod="openshift-marketplace/redhat-marketplace-bwr6d" Feb 02 09:43:59 crc kubenswrapper[4764]: I0202 09:43:59.264383 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/82ca325a-895f-4a03-995f-96e014397404-utilities\") pod \"redhat-marketplace-bwr6d\" (UID: \"82ca325a-895f-4a03-995f-96e014397404\") " pod="openshift-marketplace/redhat-marketplace-bwr6d" Feb 02 09:43:59 crc kubenswrapper[4764]: I0202 09:43:59.264844 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/82ca325a-895f-4a03-995f-96e014397404-utilities\") pod \"redhat-marketplace-bwr6d\" (UID: \"82ca325a-895f-4a03-995f-96e014397404\") " pod="openshift-marketplace/redhat-marketplace-bwr6d" Feb 02 09:43:59 crc kubenswrapper[4764]: I0202 09:43:59.265666 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/82ca325a-895f-4a03-995f-96e014397404-catalog-content\") pod \"redhat-marketplace-bwr6d\" (UID: \"82ca325a-895f-4a03-995f-96e014397404\") " pod="openshift-marketplace/redhat-marketplace-bwr6d" Feb 02 09:43:59 crc kubenswrapper[4764]: I0202 09:43:59.288070 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kxh29\" (UniqueName: \"kubernetes.io/projected/82ca325a-895f-4a03-995f-96e014397404-kube-api-access-kxh29\") pod \"redhat-marketplace-bwr6d\" (UID: \"82ca325a-895f-4a03-995f-96e014397404\") " pod="openshift-marketplace/redhat-marketplace-bwr6d" Feb 02 09:43:59 crc kubenswrapper[4764]: I0202 09:43:59.462903 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bwr6d" Feb 02 09:43:59 crc kubenswrapper[4764]: I0202 09:43:59.964781 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bwr6d"] Feb 02 09:44:00 crc kubenswrapper[4764]: E0202 09:44:00.283725 4764 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod82ca325a_895f_4a03_995f_96e014397404.slice/crio-conmon-45e1e9b41c01987825611bef92b5b2a2eea5390691a3abf51b0059eacf79b758.scope\": RecentStats: unable to find data in memory cache]" Feb 02 09:44:00 crc kubenswrapper[4764]: I0202 09:44:00.629075 4764 generic.go:334] "Generic (PLEG): container finished" podID="82ca325a-895f-4a03-995f-96e014397404" containerID="45e1e9b41c01987825611bef92b5b2a2eea5390691a3abf51b0059eacf79b758" exitCode=0 Feb 02 09:44:00 crc kubenswrapper[4764]: I0202 09:44:00.629381 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bwr6d" event={"ID":"82ca325a-895f-4a03-995f-96e014397404","Type":"ContainerDied","Data":"45e1e9b41c01987825611bef92b5b2a2eea5390691a3abf51b0059eacf79b758"} Feb 02 09:44:00 crc kubenswrapper[4764]: I0202 09:44:00.629487 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bwr6d" event={"ID":"82ca325a-895f-4a03-995f-96e014397404","Type":"ContainerStarted","Data":"ecffdb18d16536e368d11369fa51423039414174672355cea269f7052d8858fb"} Feb 02 09:44:02 crc kubenswrapper[4764]: I0202 09:44:02.653164 4764 generic.go:334] "Generic (PLEG): container finished" podID="82ca325a-895f-4a03-995f-96e014397404" containerID="3aead57b9d957646e7ce5f813fb988c5d0c776d92b393f4b3af0854c32b0c0ac" exitCode=0 Feb 02 09:44:02 crc kubenswrapper[4764]: I0202 09:44:02.653199 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bwr6d" event={"ID":"82ca325a-895f-4a03-995f-96e014397404","Type":"ContainerDied","Data":"3aead57b9d957646e7ce5f813fb988c5d0c776d92b393f4b3af0854c32b0c0ac"} Feb 02 09:44:03 crc kubenswrapper[4764]: I0202 09:44:03.664134 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bwr6d" event={"ID":"82ca325a-895f-4a03-995f-96e014397404","Type":"ContainerStarted","Data":"89ad388f461187e6f9393ac80d1602d626c029abd7041ef98a48a887a66c1b60"} Feb 02 09:44:03 crc kubenswrapper[4764]: I0202 09:44:03.695130 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-bwr6d" podStartSLOduration=2.288777434 podStartE2EDuration="4.695110555s" podCreationTimestamp="2026-02-02 09:43:59 +0000 UTC" firstStartedPulling="2026-02-02 09:44:00.63208821 +0000 UTC m=+2203.565812298" lastFinishedPulling="2026-02-02 09:44:03.038421331 +0000 UTC m=+2205.972145419" observedRunningTime="2026-02-02 09:44:03.688097536 +0000 UTC m=+2206.621821634" watchObservedRunningTime="2026-02-02 09:44:03.695110555 +0000 UTC m=+2206.628834653" Feb 02 09:44:04 crc kubenswrapper[4764]: I0202 09:44:04.516906 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-5x5h2"] Feb 02 09:44:04 crc kubenswrapper[4764]: I0202 09:44:04.518783 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5x5h2" Feb 02 09:44:04 crc kubenswrapper[4764]: I0202 09:44:04.538739 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5x5h2"] Feb 02 09:44:04 crc kubenswrapper[4764]: I0202 09:44:04.599022 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4ppff\" (UniqueName: \"kubernetes.io/projected/0331eb99-078e-4945-abb7-db9c1d76cfe3-kube-api-access-4ppff\") pod \"certified-operators-5x5h2\" (UID: \"0331eb99-078e-4945-abb7-db9c1d76cfe3\") " pod="openshift-marketplace/certified-operators-5x5h2" Feb 02 09:44:04 crc kubenswrapper[4764]: I0202 09:44:04.599176 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0331eb99-078e-4945-abb7-db9c1d76cfe3-utilities\") pod \"certified-operators-5x5h2\" (UID: \"0331eb99-078e-4945-abb7-db9c1d76cfe3\") " pod="openshift-marketplace/certified-operators-5x5h2" Feb 02 09:44:04 crc kubenswrapper[4764]: I0202 09:44:04.599243 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0331eb99-078e-4945-abb7-db9c1d76cfe3-catalog-content\") pod \"certified-operators-5x5h2\" (UID: \"0331eb99-078e-4945-abb7-db9c1d76cfe3\") " pod="openshift-marketplace/certified-operators-5x5h2" Feb 02 09:44:04 crc kubenswrapper[4764]: I0202 09:44:04.700697 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0331eb99-078e-4945-abb7-db9c1d76cfe3-utilities\") pod \"certified-operators-5x5h2\" (UID: \"0331eb99-078e-4945-abb7-db9c1d76cfe3\") " pod="openshift-marketplace/certified-operators-5x5h2" Feb 02 09:44:04 crc kubenswrapper[4764]: I0202 09:44:04.700763 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0331eb99-078e-4945-abb7-db9c1d76cfe3-catalog-content\") pod \"certified-operators-5x5h2\" (UID: \"0331eb99-078e-4945-abb7-db9c1d76cfe3\") " pod="openshift-marketplace/certified-operators-5x5h2" Feb 02 09:44:04 crc kubenswrapper[4764]: I0202 09:44:04.700863 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4ppff\" (UniqueName: \"kubernetes.io/projected/0331eb99-078e-4945-abb7-db9c1d76cfe3-kube-api-access-4ppff\") pod \"certified-operators-5x5h2\" (UID: \"0331eb99-078e-4945-abb7-db9c1d76cfe3\") " pod="openshift-marketplace/certified-operators-5x5h2" Feb 02 09:44:04 crc kubenswrapper[4764]: I0202 09:44:04.701124 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0331eb99-078e-4945-abb7-db9c1d76cfe3-utilities\") pod \"certified-operators-5x5h2\" (UID: \"0331eb99-078e-4945-abb7-db9c1d76cfe3\") " pod="openshift-marketplace/certified-operators-5x5h2" Feb 02 09:44:04 crc kubenswrapper[4764]: I0202 09:44:04.701542 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0331eb99-078e-4945-abb7-db9c1d76cfe3-catalog-content\") pod \"certified-operators-5x5h2\" (UID: \"0331eb99-078e-4945-abb7-db9c1d76cfe3\") " pod="openshift-marketplace/certified-operators-5x5h2" Feb 02 09:44:04 crc kubenswrapper[4764]: I0202 09:44:04.719405 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4ppff\" (UniqueName: \"kubernetes.io/projected/0331eb99-078e-4945-abb7-db9c1d76cfe3-kube-api-access-4ppff\") pod \"certified-operators-5x5h2\" (UID: \"0331eb99-078e-4945-abb7-db9c1d76cfe3\") " pod="openshift-marketplace/certified-operators-5x5h2" Feb 02 09:44:04 crc kubenswrapper[4764]: I0202 09:44:04.836731 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5x5h2" Feb 02 09:44:05 crc kubenswrapper[4764]: I0202 09:44:05.147755 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5x5h2"] Feb 02 09:44:05 crc kubenswrapper[4764]: I0202 09:44:05.680121 4764 generic.go:334] "Generic (PLEG): container finished" podID="0331eb99-078e-4945-abb7-db9c1d76cfe3" containerID="76cb20c772efc6ecc3f7d48912ae47b7c9a4eec7cfdac8f5c80ec640435a8fec" exitCode=0 Feb 02 09:44:05 crc kubenswrapper[4764]: I0202 09:44:05.680239 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5x5h2" event={"ID":"0331eb99-078e-4945-abb7-db9c1d76cfe3","Type":"ContainerDied","Data":"76cb20c772efc6ecc3f7d48912ae47b7c9a4eec7cfdac8f5c80ec640435a8fec"} Feb 02 09:44:05 crc kubenswrapper[4764]: I0202 09:44:05.680723 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5x5h2" event={"ID":"0331eb99-078e-4945-abb7-db9c1d76cfe3","Type":"ContainerStarted","Data":"263103d84402bf4f3a2941e588eb027b96e3d01ddc2339f262dcaa4079a008bc"} Feb 02 09:44:06 crc kubenswrapper[4764]: I0202 09:44:06.692154 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5x5h2" event={"ID":"0331eb99-078e-4945-abb7-db9c1d76cfe3","Type":"ContainerStarted","Data":"e5b5da8c7355cb8f4fcf01db5724af04ad0efa9c2ec750b3976d6592b8c37786"} Feb 02 09:44:08 crc kubenswrapper[4764]: I0202 09:44:08.712976 4764 generic.go:334] "Generic (PLEG): container finished" podID="0331eb99-078e-4945-abb7-db9c1d76cfe3" containerID="e5b5da8c7355cb8f4fcf01db5724af04ad0efa9c2ec750b3976d6592b8c37786" exitCode=0 Feb 02 09:44:08 crc kubenswrapper[4764]: I0202 09:44:08.713068 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5x5h2" event={"ID":"0331eb99-078e-4945-abb7-db9c1d76cfe3","Type":"ContainerDied","Data":"e5b5da8c7355cb8f4fcf01db5724af04ad0efa9c2ec750b3976d6592b8c37786"} Feb 02 09:44:09 crc kubenswrapper[4764]: I0202 09:44:09.464440 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-bwr6d" Feb 02 09:44:09 crc kubenswrapper[4764]: I0202 09:44:09.464778 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-bwr6d" Feb 02 09:44:09 crc kubenswrapper[4764]: I0202 09:44:09.530911 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-bwr6d" Feb 02 09:44:09 crc kubenswrapper[4764]: I0202 09:44:09.725323 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5x5h2" event={"ID":"0331eb99-078e-4945-abb7-db9c1d76cfe3","Type":"ContainerStarted","Data":"5dd8f895484886381b347a8fd3c953b05c563b9b855d129e7d3b0756d128755d"} Feb 02 09:44:09 crc kubenswrapper[4764]: I0202 09:44:09.744724 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-5x5h2" podStartSLOduration=2.278841078 podStartE2EDuration="5.744696168s" podCreationTimestamp="2026-02-02 09:44:04 +0000 UTC" firstStartedPulling="2026-02-02 09:44:05.681821364 +0000 UTC m=+2208.615545452" lastFinishedPulling="2026-02-02 09:44:09.147676454 +0000 UTC m=+2212.081400542" observedRunningTime="2026-02-02 09:44:09.742437647 +0000 UTC m=+2212.676161735" watchObservedRunningTime="2026-02-02 09:44:09.744696168 +0000 UTC m=+2212.678420256" Feb 02 09:44:09 crc kubenswrapper[4764]: I0202 09:44:09.786271 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-bwr6d" Feb 02 09:44:11 crc kubenswrapper[4764]: I0202 09:44:11.909839 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bwr6d"] Feb 02 09:44:11 crc kubenswrapper[4764]: I0202 09:44:11.911128 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-bwr6d" podUID="82ca325a-895f-4a03-995f-96e014397404" containerName="registry-server" containerID="cri-o://89ad388f461187e6f9393ac80d1602d626c029abd7041ef98a48a887a66c1b60" gracePeriod=2 Feb 02 09:44:12 crc kubenswrapper[4764]: I0202 09:44:12.764910 4764 generic.go:334] "Generic (PLEG): container finished" podID="82ca325a-895f-4a03-995f-96e014397404" containerID="89ad388f461187e6f9393ac80d1602d626c029abd7041ef98a48a887a66c1b60" exitCode=0 Feb 02 09:44:12 crc kubenswrapper[4764]: I0202 09:44:12.765024 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bwr6d" event={"ID":"82ca325a-895f-4a03-995f-96e014397404","Type":"ContainerDied","Data":"89ad388f461187e6f9393ac80d1602d626c029abd7041ef98a48a887a66c1b60"} Feb 02 09:44:13 crc kubenswrapper[4764]: I0202 09:44:13.150554 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bwr6d" Feb 02 09:44:13 crc kubenswrapper[4764]: I0202 09:44:13.170091 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/82ca325a-895f-4a03-995f-96e014397404-catalog-content\") pod \"82ca325a-895f-4a03-995f-96e014397404\" (UID: \"82ca325a-895f-4a03-995f-96e014397404\") " Feb 02 09:44:13 crc kubenswrapper[4764]: I0202 09:44:13.170127 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/82ca325a-895f-4a03-995f-96e014397404-utilities\") pod \"82ca325a-895f-4a03-995f-96e014397404\" (UID: \"82ca325a-895f-4a03-995f-96e014397404\") " Feb 02 09:44:13 crc kubenswrapper[4764]: I0202 09:44:13.170204 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kxh29\" (UniqueName: \"kubernetes.io/projected/82ca325a-895f-4a03-995f-96e014397404-kube-api-access-kxh29\") pod \"82ca325a-895f-4a03-995f-96e014397404\" (UID: \"82ca325a-895f-4a03-995f-96e014397404\") " Feb 02 09:44:13 crc kubenswrapper[4764]: I0202 09:44:13.171060 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/82ca325a-895f-4a03-995f-96e014397404-utilities" (OuterVolumeSpecName: "utilities") pod "82ca325a-895f-4a03-995f-96e014397404" (UID: "82ca325a-895f-4a03-995f-96e014397404"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 09:44:13 crc kubenswrapper[4764]: I0202 09:44:13.179243 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/82ca325a-895f-4a03-995f-96e014397404-kube-api-access-kxh29" (OuterVolumeSpecName: "kube-api-access-kxh29") pod "82ca325a-895f-4a03-995f-96e014397404" (UID: "82ca325a-895f-4a03-995f-96e014397404"). InnerVolumeSpecName "kube-api-access-kxh29". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:44:13 crc kubenswrapper[4764]: I0202 09:44:13.196486 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/82ca325a-895f-4a03-995f-96e014397404-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "82ca325a-895f-4a03-995f-96e014397404" (UID: "82ca325a-895f-4a03-995f-96e014397404"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 09:44:13 crc kubenswrapper[4764]: I0202 09:44:13.271705 4764 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/82ca325a-895f-4a03-995f-96e014397404-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 09:44:13 crc kubenswrapper[4764]: I0202 09:44:13.271741 4764 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/82ca325a-895f-4a03-995f-96e014397404-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 09:44:13 crc kubenswrapper[4764]: I0202 09:44:13.271753 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kxh29\" (UniqueName: \"kubernetes.io/projected/82ca325a-895f-4a03-995f-96e014397404-kube-api-access-kxh29\") on node \"crc\" DevicePath \"\"" Feb 02 09:44:13 crc kubenswrapper[4764]: I0202 09:44:13.522740 4764 patch_prober.go:28] interesting pod/machine-config-daemon-4ndm4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 09:44:13 crc kubenswrapper[4764]: I0202 09:44:13.522796 4764 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 09:44:13 crc kubenswrapper[4764]: I0202 09:44:13.778450 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bwr6d" event={"ID":"82ca325a-895f-4a03-995f-96e014397404","Type":"ContainerDied","Data":"ecffdb18d16536e368d11369fa51423039414174672355cea269f7052d8858fb"} Feb 02 09:44:13 crc kubenswrapper[4764]: I0202 09:44:13.778565 4764 scope.go:117] "RemoveContainer" containerID="89ad388f461187e6f9393ac80d1602d626c029abd7041ef98a48a887a66c1b60" Feb 02 09:44:13 crc kubenswrapper[4764]: I0202 09:44:13.778591 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bwr6d" Feb 02 09:44:13 crc kubenswrapper[4764]: I0202 09:44:13.809402 4764 scope.go:117] "RemoveContainer" containerID="3aead57b9d957646e7ce5f813fb988c5d0c776d92b393f4b3af0854c32b0c0ac" Feb 02 09:44:13 crc kubenswrapper[4764]: I0202 09:44:13.851440 4764 scope.go:117] "RemoveContainer" containerID="45e1e9b41c01987825611bef92b5b2a2eea5390691a3abf51b0059eacf79b758" Feb 02 09:44:13 crc kubenswrapper[4764]: I0202 09:44:13.863257 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bwr6d"] Feb 02 09:44:13 crc kubenswrapper[4764]: I0202 09:44:13.864962 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-bwr6d"] Feb 02 09:44:14 crc kubenswrapper[4764]: I0202 09:44:14.837440 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-5x5h2" Feb 02 09:44:14 crc kubenswrapper[4764]: I0202 09:44:14.837498 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-5x5h2" Feb 02 09:44:14 crc kubenswrapper[4764]: I0202 09:44:14.891693 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-5x5h2" Feb 02 09:44:15 crc kubenswrapper[4764]: I0202 09:44:15.836870 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="82ca325a-895f-4a03-995f-96e014397404" path="/var/lib/kubelet/pods/82ca325a-895f-4a03-995f-96e014397404/volumes" Feb 02 09:44:15 crc kubenswrapper[4764]: I0202 09:44:15.860649 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-5x5h2" Feb 02 09:44:16 crc kubenswrapper[4764]: I0202 09:44:16.505344 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5x5h2"] Feb 02 09:44:17 crc kubenswrapper[4764]: I0202 09:44:17.813314 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-5x5h2" podUID="0331eb99-078e-4945-abb7-db9c1d76cfe3" containerName="registry-server" containerID="cri-o://5dd8f895484886381b347a8fd3c953b05c563b9b855d129e7d3b0756d128755d" gracePeriod=2 Feb 02 09:44:18 crc kubenswrapper[4764]: I0202 09:44:18.279395 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5x5h2" Feb 02 09:44:18 crc kubenswrapper[4764]: I0202 09:44:18.465273 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0331eb99-078e-4945-abb7-db9c1d76cfe3-catalog-content\") pod \"0331eb99-078e-4945-abb7-db9c1d76cfe3\" (UID: \"0331eb99-078e-4945-abb7-db9c1d76cfe3\") " Feb 02 09:44:18 crc kubenswrapper[4764]: I0202 09:44:18.465420 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4ppff\" (UniqueName: \"kubernetes.io/projected/0331eb99-078e-4945-abb7-db9c1d76cfe3-kube-api-access-4ppff\") pod \"0331eb99-078e-4945-abb7-db9c1d76cfe3\" (UID: \"0331eb99-078e-4945-abb7-db9c1d76cfe3\") " Feb 02 09:44:18 crc kubenswrapper[4764]: I0202 09:44:18.465451 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0331eb99-078e-4945-abb7-db9c1d76cfe3-utilities\") pod \"0331eb99-078e-4945-abb7-db9c1d76cfe3\" (UID: \"0331eb99-078e-4945-abb7-db9c1d76cfe3\") " Feb 02 09:44:18 crc kubenswrapper[4764]: I0202 09:44:18.466968 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0331eb99-078e-4945-abb7-db9c1d76cfe3-utilities" (OuterVolumeSpecName: "utilities") pod "0331eb99-078e-4945-abb7-db9c1d76cfe3" (UID: "0331eb99-078e-4945-abb7-db9c1d76cfe3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 09:44:18 crc kubenswrapper[4764]: I0202 09:44:18.473590 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0331eb99-078e-4945-abb7-db9c1d76cfe3-kube-api-access-4ppff" (OuterVolumeSpecName: "kube-api-access-4ppff") pod "0331eb99-078e-4945-abb7-db9c1d76cfe3" (UID: "0331eb99-078e-4945-abb7-db9c1d76cfe3"). InnerVolumeSpecName "kube-api-access-4ppff". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:44:18 crc kubenswrapper[4764]: I0202 09:44:18.515280 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0331eb99-078e-4945-abb7-db9c1d76cfe3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0331eb99-078e-4945-abb7-db9c1d76cfe3" (UID: "0331eb99-078e-4945-abb7-db9c1d76cfe3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 09:44:18 crc kubenswrapper[4764]: I0202 09:44:18.567382 4764 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0331eb99-078e-4945-abb7-db9c1d76cfe3-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 09:44:18 crc kubenswrapper[4764]: I0202 09:44:18.567411 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4ppff\" (UniqueName: \"kubernetes.io/projected/0331eb99-078e-4945-abb7-db9c1d76cfe3-kube-api-access-4ppff\") on node \"crc\" DevicePath \"\"" Feb 02 09:44:18 crc kubenswrapper[4764]: I0202 09:44:18.567422 4764 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0331eb99-078e-4945-abb7-db9c1d76cfe3-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 09:44:18 crc kubenswrapper[4764]: I0202 09:44:18.824911 4764 generic.go:334] "Generic (PLEG): container finished" podID="0331eb99-078e-4945-abb7-db9c1d76cfe3" containerID="5dd8f895484886381b347a8fd3c953b05c563b9b855d129e7d3b0756d128755d" exitCode=0 Feb 02 09:44:18 crc kubenswrapper[4764]: I0202 09:44:18.825040 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5x5h2" event={"ID":"0331eb99-078e-4945-abb7-db9c1d76cfe3","Type":"ContainerDied","Data":"5dd8f895484886381b347a8fd3c953b05c563b9b855d129e7d3b0756d128755d"} Feb 02 09:44:18 crc kubenswrapper[4764]: I0202 09:44:18.825240 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5x5h2" Feb 02 09:44:18 crc kubenswrapper[4764]: I0202 09:44:18.829994 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5x5h2" event={"ID":"0331eb99-078e-4945-abb7-db9c1d76cfe3","Type":"ContainerDied","Data":"263103d84402bf4f3a2941e588eb027b96e3d01ddc2339f262dcaa4079a008bc"} Feb 02 09:44:18 crc kubenswrapper[4764]: I0202 09:44:18.830056 4764 scope.go:117] "RemoveContainer" containerID="5dd8f895484886381b347a8fd3c953b05c563b9b855d129e7d3b0756d128755d" Feb 02 09:44:18 crc kubenswrapper[4764]: I0202 09:44:18.855276 4764 scope.go:117] "RemoveContainer" containerID="e5b5da8c7355cb8f4fcf01db5724af04ad0efa9c2ec750b3976d6592b8c37786" Feb 02 09:44:18 crc kubenswrapper[4764]: I0202 09:44:18.892723 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5x5h2"] Feb 02 09:44:18 crc kubenswrapper[4764]: I0202 09:44:18.904658 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-5x5h2"] Feb 02 09:44:18 crc kubenswrapper[4764]: I0202 09:44:18.904685 4764 scope.go:117] "RemoveContainer" containerID="76cb20c772efc6ecc3f7d48912ae47b7c9a4eec7cfdac8f5c80ec640435a8fec" Feb 02 09:44:18 crc kubenswrapper[4764]: I0202 09:44:18.940006 4764 scope.go:117] "RemoveContainer" containerID="5dd8f895484886381b347a8fd3c953b05c563b9b855d129e7d3b0756d128755d" Feb 02 09:44:18 crc kubenswrapper[4764]: E0202 09:44:18.940471 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5dd8f895484886381b347a8fd3c953b05c563b9b855d129e7d3b0756d128755d\": container with ID starting with 5dd8f895484886381b347a8fd3c953b05c563b9b855d129e7d3b0756d128755d not found: ID does not exist" containerID="5dd8f895484886381b347a8fd3c953b05c563b9b855d129e7d3b0756d128755d" Feb 02 09:44:18 crc kubenswrapper[4764]: I0202 09:44:18.940529 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5dd8f895484886381b347a8fd3c953b05c563b9b855d129e7d3b0756d128755d"} err="failed to get container status \"5dd8f895484886381b347a8fd3c953b05c563b9b855d129e7d3b0756d128755d\": rpc error: code = NotFound desc = could not find container \"5dd8f895484886381b347a8fd3c953b05c563b9b855d129e7d3b0756d128755d\": container with ID starting with 5dd8f895484886381b347a8fd3c953b05c563b9b855d129e7d3b0756d128755d not found: ID does not exist" Feb 02 09:44:18 crc kubenswrapper[4764]: I0202 09:44:18.940566 4764 scope.go:117] "RemoveContainer" containerID="e5b5da8c7355cb8f4fcf01db5724af04ad0efa9c2ec750b3976d6592b8c37786" Feb 02 09:44:18 crc kubenswrapper[4764]: E0202 09:44:18.940856 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e5b5da8c7355cb8f4fcf01db5724af04ad0efa9c2ec750b3976d6592b8c37786\": container with ID starting with e5b5da8c7355cb8f4fcf01db5724af04ad0efa9c2ec750b3976d6592b8c37786 not found: ID does not exist" containerID="e5b5da8c7355cb8f4fcf01db5724af04ad0efa9c2ec750b3976d6592b8c37786" Feb 02 09:44:18 crc kubenswrapper[4764]: I0202 09:44:18.940887 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e5b5da8c7355cb8f4fcf01db5724af04ad0efa9c2ec750b3976d6592b8c37786"} err="failed to get container status \"e5b5da8c7355cb8f4fcf01db5724af04ad0efa9c2ec750b3976d6592b8c37786\": rpc error: code = NotFound desc = could not find container \"e5b5da8c7355cb8f4fcf01db5724af04ad0efa9c2ec750b3976d6592b8c37786\": container with ID starting with e5b5da8c7355cb8f4fcf01db5724af04ad0efa9c2ec750b3976d6592b8c37786 not found: ID does not exist" Feb 02 09:44:18 crc kubenswrapper[4764]: I0202 09:44:18.940902 4764 scope.go:117] "RemoveContainer" containerID="76cb20c772efc6ecc3f7d48912ae47b7c9a4eec7cfdac8f5c80ec640435a8fec" Feb 02 09:44:18 crc kubenswrapper[4764]: E0202 09:44:18.941153 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"76cb20c772efc6ecc3f7d48912ae47b7c9a4eec7cfdac8f5c80ec640435a8fec\": container with ID starting with 76cb20c772efc6ecc3f7d48912ae47b7c9a4eec7cfdac8f5c80ec640435a8fec not found: ID does not exist" containerID="76cb20c772efc6ecc3f7d48912ae47b7c9a4eec7cfdac8f5c80ec640435a8fec" Feb 02 09:44:18 crc kubenswrapper[4764]: I0202 09:44:18.941186 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"76cb20c772efc6ecc3f7d48912ae47b7c9a4eec7cfdac8f5c80ec640435a8fec"} err="failed to get container status \"76cb20c772efc6ecc3f7d48912ae47b7c9a4eec7cfdac8f5c80ec640435a8fec\": rpc error: code = NotFound desc = could not find container \"76cb20c772efc6ecc3f7d48912ae47b7c9a4eec7cfdac8f5c80ec640435a8fec\": container with ID starting with 76cb20c772efc6ecc3f7d48912ae47b7c9a4eec7cfdac8f5c80ec640435a8fec not found: ID does not exist" Feb 02 09:44:19 crc kubenswrapper[4764]: I0202 09:44:19.838487 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0331eb99-078e-4945-abb7-db9c1d76cfe3" path="/var/lib/kubelet/pods/0331eb99-078e-4945-abb7-db9c1d76cfe3/volumes" Feb 02 09:44:20 crc kubenswrapper[4764]: E0202 09:44:20.759045 4764 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0331eb99_078e_4945_abb7_db9c1d76cfe3.slice/crio-5dd8f895484886381b347a8fd3c953b05c563b9b855d129e7d3b0756d128755d.scope\": RecentStats: unable to find data in memory cache]" Feb 02 09:44:23 crc kubenswrapper[4764]: I0202 09:44:23.886930 4764 generic.go:334] "Generic (PLEG): container finished" podID="af835a1d-09cc-49ca-979f-41eaf146e926" containerID="11eea62e5a292de4fd143d20114bbdddfdf73d0ccfe3c45abc9e58813bd4e267" exitCode=0 Feb 02 09:44:23 crc kubenswrapper[4764]: I0202 09:44:23.887144 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-d9x8m" event={"ID":"af835a1d-09cc-49ca-979f-41eaf146e926","Type":"ContainerDied","Data":"11eea62e5a292de4fd143d20114bbdddfdf73d0ccfe3c45abc9e58813bd4e267"} Feb 02 09:44:25 crc kubenswrapper[4764]: I0202 09:44:25.395609 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-d9x8m" Feb 02 09:44:25 crc kubenswrapper[4764]: I0202 09:44:25.402507 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af835a1d-09cc-49ca-979f-41eaf146e926-bootstrap-combined-ca-bundle\") pod \"af835a1d-09cc-49ca-979f-41eaf146e926\" (UID: \"af835a1d-09cc-49ca-979f-41eaf146e926\") " Feb 02 09:44:25 crc kubenswrapper[4764]: I0202 09:44:25.402578 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/af835a1d-09cc-49ca-979f-41eaf146e926-inventory\") pod \"af835a1d-09cc-49ca-979f-41eaf146e926\" (UID: \"af835a1d-09cc-49ca-979f-41eaf146e926\") " Feb 02 09:44:25 crc kubenswrapper[4764]: I0202 09:44:25.402631 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/af835a1d-09cc-49ca-979f-41eaf146e926-ssh-key-openstack-edpm-ipam\") pod \"af835a1d-09cc-49ca-979f-41eaf146e926\" (UID: \"af835a1d-09cc-49ca-979f-41eaf146e926\") " Feb 02 09:44:25 crc kubenswrapper[4764]: I0202 09:44:25.402674 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fn6s9\" (UniqueName: \"kubernetes.io/projected/af835a1d-09cc-49ca-979f-41eaf146e926-kube-api-access-fn6s9\") pod \"af835a1d-09cc-49ca-979f-41eaf146e926\" (UID: \"af835a1d-09cc-49ca-979f-41eaf146e926\") " Feb 02 09:44:25 crc kubenswrapper[4764]: I0202 09:44:25.402706 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/af835a1d-09cc-49ca-979f-41eaf146e926-ceph\") pod \"af835a1d-09cc-49ca-979f-41eaf146e926\" (UID: \"af835a1d-09cc-49ca-979f-41eaf146e926\") " Feb 02 09:44:25 crc kubenswrapper[4764]: I0202 09:44:25.409660 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af835a1d-09cc-49ca-979f-41eaf146e926-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "af835a1d-09cc-49ca-979f-41eaf146e926" (UID: "af835a1d-09cc-49ca-979f-41eaf146e926"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:44:25 crc kubenswrapper[4764]: I0202 09:44:25.411511 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af835a1d-09cc-49ca-979f-41eaf146e926-ceph" (OuterVolumeSpecName: "ceph") pod "af835a1d-09cc-49ca-979f-41eaf146e926" (UID: "af835a1d-09cc-49ca-979f-41eaf146e926"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:44:25 crc kubenswrapper[4764]: I0202 09:44:25.413416 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af835a1d-09cc-49ca-979f-41eaf146e926-kube-api-access-fn6s9" (OuterVolumeSpecName: "kube-api-access-fn6s9") pod "af835a1d-09cc-49ca-979f-41eaf146e926" (UID: "af835a1d-09cc-49ca-979f-41eaf146e926"). InnerVolumeSpecName "kube-api-access-fn6s9". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:44:25 crc kubenswrapper[4764]: I0202 09:44:25.441756 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af835a1d-09cc-49ca-979f-41eaf146e926-inventory" (OuterVolumeSpecName: "inventory") pod "af835a1d-09cc-49ca-979f-41eaf146e926" (UID: "af835a1d-09cc-49ca-979f-41eaf146e926"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:44:25 crc kubenswrapper[4764]: I0202 09:44:25.442457 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af835a1d-09cc-49ca-979f-41eaf146e926-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "af835a1d-09cc-49ca-979f-41eaf146e926" (UID: "af835a1d-09cc-49ca-979f-41eaf146e926"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:44:25 crc kubenswrapper[4764]: I0202 09:44:25.504060 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fn6s9\" (UniqueName: \"kubernetes.io/projected/af835a1d-09cc-49ca-979f-41eaf146e926-kube-api-access-fn6s9\") on node \"crc\" DevicePath \"\"" Feb 02 09:44:25 crc kubenswrapper[4764]: I0202 09:44:25.504084 4764 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/af835a1d-09cc-49ca-979f-41eaf146e926-ceph\") on node \"crc\" DevicePath \"\"" Feb 02 09:44:25 crc kubenswrapper[4764]: I0202 09:44:25.504096 4764 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af835a1d-09cc-49ca-979f-41eaf146e926-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 09:44:25 crc kubenswrapper[4764]: I0202 09:44:25.504106 4764 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/af835a1d-09cc-49ca-979f-41eaf146e926-inventory\") on node \"crc\" DevicePath \"\"" Feb 02 09:44:25 crc kubenswrapper[4764]: I0202 09:44:25.504114 4764 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/af835a1d-09cc-49ca-979f-41eaf146e926-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 02 09:44:25 crc kubenswrapper[4764]: I0202 09:44:25.909507 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-d9x8m" event={"ID":"af835a1d-09cc-49ca-979f-41eaf146e926","Type":"ContainerDied","Data":"2bb25760321f502b97dd1c83efb0ca860f9417899597ef4ed76035522ae4cab0"} Feb 02 09:44:25 crc kubenswrapper[4764]: I0202 09:44:25.909562 4764 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2bb25760321f502b97dd1c83efb0ca860f9417899597ef4ed76035522ae4cab0" Feb 02 09:44:25 crc kubenswrapper[4764]: I0202 09:44:25.909640 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-d9x8m" Feb 02 09:44:26 crc kubenswrapper[4764]: I0202 09:44:26.046130 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-rsd54"] Feb 02 09:44:26 crc kubenswrapper[4764]: E0202 09:44:26.046669 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82ca325a-895f-4a03-995f-96e014397404" containerName="extract-content" Feb 02 09:44:26 crc kubenswrapper[4764]: I0202 09:44:26.046713 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="82ca325a-895f-4a03-995f-96e014397404" containerName="extract-content" Feb 02 09:44:26 crc kubenswrapper[4764]: E0202 09:44:26.046752 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0331eb99-078e-4945-abb7-db9c1d76cfe3" containerName="extract-content" Feb 02 09:44:26 crc kubenswrapper[4764]: I0202 09:44:26.046767 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="0331eb99-078e-4945-abb7-db9c1d76cfe3" containerName="extract-content" Feb 02 09:44:26 crc kubenswrapper[4764]: E0202 09:44:26.046798 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82ca325a-895f-4a03-995f-96e014397404" containerName="registry-server" Feb 02 09:44:26 crc kubenswrapper[4764]: I0202 09:44:26.046811 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="82ca325a-895f-4a03-995f-96e014397404" containerName="registry-server" Feb 02 09:44:26 crc kubenswrapper[4764]: E0202 09:44:26.046855 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0331eb99-078e-4945-abb7-db9c1d76cfe3" containerName="registry-server" Feb 02 09:44:26 crc kubenswrapper[4764]: I0202 09:44:26.046867 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="0331eb99-078e-4945-abb7-db9c1d76cfe3" containerName="registry-server" Feb 02 09:44:26 crc kubenswrapper[4764]: E0202 09:44:26.046891 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af835a1d-09cc-49ca-979f-41eaf146e926" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Feb 02 09:44:26 crc kubenswrapper[4764]: I0202 09:44:26.046903 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="af835a1d-09cc-49ca-979f-41eaf146e926" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Feb 02 09:44:26 crc kubenswrapper[4764]: E0202 09:44:26.046967 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82ca325a-895f-4a03-995f-96e014397404" containerName="extract-utilities" Feb 02 09:44:26 crc kubenswrapper[4764]: I0202 09:44:26.046981 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="82ca325a-895f-4a03-995f-96e014397404" containerName="extract-utilities" Feb 02 09:44:26 crc kubenswrapper[4764]: E0202 09:44:26.047003 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0331eb99-078e-4945-abb7-db9c1d76cfe3" containerName="extract-utilities" Feb 02 09:44:26 crc kubenswrapper[4764]: I0202 09:44:26.047015 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="0331eb99-078e-4945-abb7-db9c1d76cfe3" containerName="extract-utilities" Feb 02 09:44:26 crc kubenswrapper[4764]: I0202 09:44:26.047306 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="0331eb99-078e-4945-abb7-db9c1d76cfe3" containerName="registry-server" Feb 02 09:44:26 crc kubenswrapper[4764]: I0202 09:44:26.047333 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="82ca325a-895f-4a03-995f-96e014397404" containerName="registry-server" Feb 02 09:44:26 crc kubenswrapper[4764]: I0202 09:44:26.047361 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="af835a1d-09cc-49ca-979f-41eaf146e926" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Feb 02 09:44:26 crc kubenswrapper[4764]: I0202 09:44:26.048300 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-rsd54" Feb 02 09:44:26 crc kubenswrapper[4764]: I0202 09:44:26.057679 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 02 09:44:26 crc kubenswrapper[4764]: I0202 09:44:26.057903 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-xksq5" Feb 02 09:44:26 crc kubenswrapper[4764]: I0202 09:44:26.059102 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-rsd54"] Feb 02 09:44:26 crc kubenswrapper[4764]: I0202 09:44:26.059365 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Feb 02 09:44:26 crc kubenswrapper[4764]: I0202 09:44:26.060485 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 02 09:44:26 crc kubenswrapper[4764]: I0202 09:44:26.061224 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 02 09:44:26 crc kubenswrapper[4764]: I0202 09:44:26.117040 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5a0ee3ef-d01f-4961-8362-8b6c4d3f37fc-ssh-key-openstack-edpm-ipam\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-rsd54\" (UID: \"5a0ee3ef-d01f-4961-8362-8b6c4d3f37fc\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-rsd54" Feb 02 09:44:26 crc kubenswrapper[4764]: I0202 09:44:26.117122 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5a0ee3ef-d01f-4961-8362-8b6c4d3f37fc-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-rsd54\" (UID: \"5a0ee3ef-d01f-4961-8362-8b6c4d3f37fc\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-rsd54" Feb 02 09:44:26 crc kubenswrapper[4764]: I0202 09:44:26.117219 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j657r\" (UniqueName: \"kubernetes.io/projected/5a0ee3ef-d01f-4961-8362-8b6c4d3f37fc-kube-api-access-j657r\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-rsd54\" (UID: \"5a0ee3ef-d01f-4961-8362-8b6c4d3f37fc\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-rsd54" Feb 02 09:44:26 crc kubenswrapper[4764]: I0202 09:44:26.117275 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/5a0ee3ef-d01f-4961-8362-8b6c4d3f37fc-ceph\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-rsd54\" (UID: \"5a0ee3ef-d01f-4961-8362-8b6c4d3f37fc\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-rsd54" Feb 02 09:44:26 crc kubenswrapper[4764]: I0202 09:44:26.218650 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/5a0ee3ef-d01f-4961-8362-8b6c4d3f37fc-ceph\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-rsd54\" (UID: \"5a0ee3ef-d01f-4961-8362-8b6c4d3f37fc\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-rsd54" Feb 02 09:44:26 crc kubenswrapper[4764]: I0202 09:44:26.220115 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5a0ee3ef-d01f-4961-8362-8b6c4d3f37fc-ssh-key-openstack-edpm-ipam\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-rsd54\" (UID: \"5a0ee3ef-d01f-4961-8362-8b6c4d3f37fc\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-rsd54" Feb 02 09:44:26 crc kubenswrapper[4764]: I0202 09:44:26.220266 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5a0ee3ef-d01f-4961-8362-8b6c4d3f37fc-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-rsd54\" (UID: \"5a0ee3ef-d01f-4961-8362-8b6c4d3f37fc\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-rsd54" Feb 02 09:44:26 crc kubenswrapper[4764]: I0202 09:44:26.220327 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j657r\" (UniqueName: \"kubernetes.io/projected/5a0ee3ef-d01f-4961-8362-8b6c4d3f37fc-kube-api-access-j657r\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-rsd54\" (UID: \"5a0ee3ef-d01f-4961-8362-8b6c4d3f37fc\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-rsd54" Feb 02 09:44:26 crc kubenswrapper[4764]: I0202 09:44:26.225577 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5a0ee3ef-d01f-4961-8362-8b6c4d3f37fc-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-rsd54\" (UID: \"5a0ee3ef-d01f-4961-8362-8b6c4d3f37fc\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-rsd54" Feb 02 09:44:26 crc kubenswrapper[4764]: I0202 09:44:26.226082 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/5a0ee3ef-d01f-4961-8362-8b6c4d3f37fc-ceph\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-rsd54\" (UID: \"5a0ee3ef-d01f-4961-8362-8b6c4d3f37fc\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-rsd54" Feb 02 09:44:26 crc kubenswrapper[4764]: I0202 09:44:26.239728 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5a0ee3ef-d01f-4961-8362-8b6c4d3f37fc-ssh-key-openstack-edpm-ipam\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-rsd54\" (UID: \"5a0ee3ef-d01f-4961-8362-8b6c4d3f37fc\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-rsd54" Feb 02 09:44:26 crc kubenswrapper[4764]: I0202 09:44:26.251969 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j657r\" (UniqueName: \"kubernetes.io/projected/5a0ee3ef-d01f-4961-8362-8b6c4d3f37fc-kube-api-access-j657r\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-rsd54\" (UID: \"5a0ee3ef-d01f-4961-8362-8b6c4d3f37fc\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-rsd54" Feb 02 09:44:26 crc kubenswrapper[4764]: I0202 09:44:26.367179 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-rsd54" Feb 02 09:44:26 crc kubenswrapper[4764]: I0202 09:44:26.960608 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-rsd54"] Feb 02 09:44:27 crc kubenswrapper[4764]: I0202 09:44:27.926728 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-rsd54" event={"ID":"5a0ee3ef-d01f-4961-8362-8b6c4d3f37fc","Type":"ContainerStarted","Data":"1411039eb4320bf542c7e62ea7798d553c8161b97c37c21e25583fa779d517aa"} Feb 02 09:44:27 crc kubenswrapper[4764]: I0202 09:44:27.926783 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-rsd54" event={"ID":"5a0ee3ef-d01f-4961-8362-8b6c4d3f37fc","Type":"ContainerStarted","Data":"371604b0a0d95aa2fd7e5f7dc36984f3956bc7d32cecfb2262971ca2bb4013e4"} Feb 02 09:44:30 crc kubenswrapper[4764]: E0202 09:44:30.970382 4764 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0331eb99_078e_4945_abb7_db9c1d76cfe3.slice/crio-5dd8f895484886381b347a8fd3c953b05c563b9b855d129e7d3b0756d128755d.scope\": RecentStats: unable to find data in memory cache]" Feb 02 09:44:41 crc kubenswrapper[4764]: E0202 09:44:41.180872 4764 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0331eb99_078e_4945_abb7_db9c1d76cfe3.slice/crio-5dd8f895484886381b347a8fd3c953b05c563b9b855d129e7d3b0756d128755d.scope\": RecentStats: unable to find data in memory cache]" Feb 02 09:44:43 crc kubenswrapper[4764]: I0202 09:44:43.523169 4764 patch_prober.go:28] interesting pod/machine-config-daemon-4ndm4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 09:44:43 crc kubenswrapper[4764]: I0202 09:44:43.523444 4764 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 09:44:43 crc kubenswrapper[4764]: I0202 09:44:43.523480 4764 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" Feb 02 09:44:43 crc kubenswrapper[4764]: I0202 09:44:43.524156 4764 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"be8fa5800ec42d1c141765dcfc6111f3ef96b05b6e8d5a9feaf2baa219897bcf"} pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 02 09:44:43 crc kubenswrapper[4764]: I0202 09:44:43.524199 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" containerName="machine-config-daemon" containerID="cri-o://be8fa5800ec42d1c141765dcfc6111f3ef96b05b6e8d5a9feaf2baa219897bcf" gracePeriod=600 Feb 02 09:44:43 crc kubenswrapper[4764]: E0202 09:44:43.652490 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 09:44:44 crc kubenswrapper[4764]: I0202 09:44:44.059573 4764 generic.go:334] "Generic (PLEG): container finished" podID="d192f670-9f9d-4539-9641-e4bed73acdd4" containerID="be8fa5800ec42d1c141765dcfc6111f3ef96b05b6e8d5a9feaf2baa219897bcf" exitCode=0 Feb 02 09:44:44 crc kubenswrapper[4764]: I0202 09:44:44.059656 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" event={"ID":"d192f670-9f9d-4539-9641-e4bed73acdd4","Type":"ContainerDied","Data":"be8fa5800ec42d1c141765dcfc6111f3ef96b05b6e8d5a9feaf2baa219897bcf"} Feb 02 09:44:44 crc kubenswrapper[4764]: I0202 09:44:44.059982 4764 scope.go:117] "RemoveContainer" containerID="7a73d717f76ca801e8516faf5325e4c71beae4270ab4dde30751c8baf713bdb2" Feb 02 09:44:44 crc kubenswrapper[4764]: I0202 09:44:44.060362 4764 scope.go:117] "RemoveContainer" containerID="be8fa5800ec42d1c141765dcfc6111f3ef96b05b6e8d5a9feaf2baa219897bcf" Feb 02 09:44:44 crc kubenswrapper[4764]: E0202 09:44:44.060606 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 09:44:44 crc kubenswrapper[4764]: I0202 09:44:44.082651 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-rsd54" podStartSLOduration=17.651670991 podStartE2EDuration="18.082628926s" podCreationTimestamp="2026-02-02 09:44:26 +0000 UTC" firstStartedPulling="2026-02-02 09:44:26.952583157 +0000 UTC m=+2229.886307245" lastFinishedPulling="2026-02-02 09:44:27.383541052 +0000 UTC m=+2230.317265180" observedRunningTime="2026-02-02 09:44:27.957395331 +0000 UTC m=+2230.891119459" watchObservedRunningTime="2026-02-02 09:44:44.082628926 +0000 UTC m=+2247.016353024" Feb 02 09:44:51 crc kubenswrapper[4764]: E0202 09:44:51.473057 4764 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0331eb99_078e_4945_abb7_db9c1d76cfe3.slice/crio-5dd8f895484886381b347a8fd3c953b05c563b9b855d129e7d3b0756d128755d.scope\": RecentStats: unable to find data in memory cache]" Feb 02 09:44:54 crc kubenswrapper[4764]: I0202 09:44:54.150006 4764 generic.go:334] "Generic (PLEG): container finished" podID="5a0ee3ef-d01f-4961-8362-8b6c4d3f37fc" containerID="1411039eb4320bf542c7e62ea7798d553c8161b97c37c21e25583fa779d517aa" exitCode=0 Feb 02 09:44:54 crc kubenswrapper[4764]: I0202 09:44:54.150105 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-rsd54" event={"ID":"5a0ee3ef-d01f-4961-8362-8b6c4d3f37fc","Type":"ContainerDied","Data":"1411039eb4320bf542c7e62ea7798d553c8161b97c37c21e25583fa779d517aa"} Feb 02 09:44:54 crc kubenswrapper[4764]: I0202 09:44:54.825883 4764 scope.go:117] "RemoveContainer" containerID="be8fa5800ec42d1c141765dcfc6111f3ef96b05b6e8d5a9feaf2baa219897bcf" Feb 02 09:44:54 crc kubenswrapper[4764]: E0202 09:44:54.826540 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 09:44:55 crc kubenswrapper[4764]: I0202 09:44:55.586866 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-rsd54" Feb 02 09:44:55 crc kubenswrapper[4764]: I0202 09:44:55.690846 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5a0ee3ef-d01f-4961-8362-8b6c4d3f37fc-ssh-key-openstack-edpm-ipam\") pod \"5a0ee3ef-d01f-4961-8362-8b6c4d3f37fc\" (UID: \"5a0ee3ef-d01f-4961-8362-8b6c4d3f37fc\") " Feb 02 09:44:55 crc kubenswrapper[4764]: I0202 09:44:55.690969 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5a0ee3ef-d01f-4961-8362-8b6c4d3f37fc-inventory\") pod \"5a0ee3ef-d01f-4961-8362-8b6c4d3f37fc\" (UID: \"5a0ee3ef-d01f-4961-8362-8b6c4d3f37fc\") " Feb 02 09:44:55 crc kubenswrapper[4764]: I0202 09:44:55.691043 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j657r\" (UniqueName: \"kubernetes.io/projected/5a0ee3ef-d01f-4961-8362-8b6c4d3f37fc-kube-api-access-j657r\") pod \"5a0ee3ef-d01f-4961-8362-8b6c4d3f37fc\" (UID: \"5a0ee3ef-d01f-4961-8362-8b6c4d3f37fc\") " Feb 02 09:44:55 crc kubenswrapper[4764]: I0202 09:44:55.691111 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/5a0ee3ef-d01f-4961-8362-8b6c4d3f37fc-ceph\") pod \"5a0ee3ef-d01f-4961-8362-8b6c4d3f37fc\" (UID: \"5a0ee3ef-d01f-4961-8362-8b6c4d3f37fc\") " Feb 02 09:44:55 crc kubenswrapper[4764]: I0202 09:44:55.698330 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5a0ee3ef-d01f-4961-8362-8b6c4d3f37fc-kube-api-access-j657r" (OuterVolumeSpecName: "kube-api-access-j657r") pod "5a0ee3ef-d01f-4961-8362-8b6c4d3f37fc" (UID: "5a0ee3ef-d01f-4961-8362-8b6c4d3f37fc"). InnerVolumeSpecName "kube-api-access-j657r". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:44:55 crc kubenswrapper[4764]: I0202 09:44:55.718662 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a0ee3ef-d01f-4961-8362-8b6c4d3f37fc-ceph" (OuterVolumeSpecName: "ceph") pod "5a0ee3ef-d01f-4961-8362-8b6c4d3f37fc" (UID: "5a0ee3ef-d01f-4961-8362-8b6c4d3f37fc"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:44:55 crc kubenswrapper[4764]: I0202 09:44:55.724085 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a0ee3ef-d01f-4961-8362-8b6c4d3f37fc-inventory" (OuterVolumeSpecName: "inventory") pod "5a0ee3ef-d01f-4961-8362-8b6c4d3f37fc" (UID: "5a0ee3ef-d01f-4961-8362-8b6c4d3f37fc"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:44:55 crc kubenswrapper[4764]: I0202 09:44:55.728547 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a0ee3ef-d01f-4961-8362-8b6c4d3f37fc-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "5a0ee3ef-d01f-4961-8362-8b6c4d3f37fc" (UID: "5a0ee3ef-d01f-4961-8362-8b6c4d3f37fc"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:44:55 crc kubenswrapper[4764]: I0202 09:44:55.793852 4764 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5a0ee3ef-d01f-4961-8362-8b6c4d3f37fc-inventory\") on node \"crc\" DevicePath \"\"" Feb 02 09:44:55 crc kubenswrapper[4764]: I0202 09:44:55.793921 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j657r\" (UniqueName: \"kubernetes.io/projected/5a0ee3ef-d01f-4961-8362-8b6c4d3f37fc-kube-api-access-j657r\") on node \"crc\" DevicePath \"\"" Feb 02 09:44:55 crc kubenswrapper[4764]: I0202 09:44:55.794144 4764 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/5a0ee3ef-d01f-4961-8362-8b6c4d3f37fc-ceph\") on node \"crc\" DevicePath \"\"" Feb 02 09:44:55 crc kubenswrapper[4764]: I0202 09:44:55.794159 4764 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5a0ee3ef-d01f-4961-8362-8b6c4d3f37fc-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 02 09:44:56 crc kubenswrapper[4764]: I0202 09:44:56.176099 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-rsd54" event={"ID":"5a0ee3ef-d01f-4961-8362-8b6c4d3f37fc","Type":"ContainerDied","Data":"371604b0a0d95aa2fd7e5f7dc36984f3956bc7d32cecfb2262971ca2bb4013e4"} Feb 02 09:44:56 crc kubenswrapper[4764]: I0202 09:44:56.176683 4764 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="371604b0a0d95aa2fd7e5f7dc36984f3956bc7d32cecfb2262971ca2bb4013e4" Feb 02 09:44:56 crc kubenswrapper[4764]: I0202 09:44:56.176201 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-rsd54" Feb 02 09:44:56 crc kubenswrapper[4764]: I0202 09:44:56.302016 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-nwdxw"] Feb 02 09:44:56 crc kubenswrapper[4764]: E0202 09:44:56.302495 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a0ee3ef-d01f-4961-8362-8b6c4d3f37fc" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Feb 02 09:44:56 crc kubenswrapper[4764]: I0202 09:44:56.302518 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a0ee3ef-d01f-4961-8362-8b6c4d3f37fc" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Feb 02 09:44:56 crc kubenswrapper[4764]: I0202 09:44:56.302786 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a0ee3ef-d01f-4961-8362-8b6c4d3f37fc" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Feb 02 09:44:56 crc kubenswrapper[4764]: I0202 09:44:56.303682 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-nwdxw" Feb 02 09:44:56 crc kubenswrapper[4764]: I0202 09:44:56.308541 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 02 09:44:56 crc kubenswrapper[4764]: I0202 09:44:56.308755 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 02 09:44:56 crc kubenswrapper[4764]: I0202 09:44:56.311256 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2v4tg\" (UniqueName: \"kubernetes.io/projected/ddae0daa-153b-425b-92dd-d6c315ace4e5-kube-api-access-2v4tg\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-nwdxw\" (UID: \"ddae0daa-153b-425b-92dd-d6c315ace4e5\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-nwdxw" Feb 02 09:44:56 crc kubenswrapper[4764]: I0202 09:44:56.311337 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/ddae0daa-153b-425b-92dd-d6c315ace4e5-ssh-key-openstack-edpm-ipam\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-nwdxw\" (UID: \"ddae0daa-153b-425b-92dd-d6c315ace4e5\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-nwdxw" Feb 02 09:44:56 crc kubenswrapper[4764]: I0202 09:44:56.311356 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-xksq5" Feb 02 09:44:56 crc kubenswrapper[4764]: I0202 09:44:56.311423 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ddae0daa-153b-425b-92dd-d6c315ace4e5-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-nwdxw\" (UID: \"ddae0daa-153b-425b-92dd-d6c315ace4e5\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-nwdxw" Feb 02 09:44:56 crc kubenswrapper[4764]: I0202 09:44:56.311460 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ddae0daa-153b-425b-92dd-d6c315ace4e5-ceph\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-nwdxw\" (UID: \"ddae0daa-153b-425b-92dd-d6c315ace4e5\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-nwdxw" Feb 02 09:44:56 crc kubenswrapper[4764]: I0202 09:44:56.313034 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 02 09:44:56 crc kubenswrapper[4764]: I0202 09:44:56.324264 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Feb 02 09:44:56 crc kubenswrapper[4764]: I0202 09:44:56.340663 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-nwdxw"] Feb 02 09:44:56 crc kubenswrapper[4764]: I0202 09:44:56.412897 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2v4tg\" (UniqueName: \"kubernetes.io/projected/ddae0daa-153b-425b-92dd-d6c315ace4e5-kube-api-access-2v4tg\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-nwdxw\" (UID: \"ddae0daa-153b-425b-92dd-d6c315ace4e5\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-nwdxw" Feb 02 09:44:56 crc kubenswrapper[4764]: I0202 09:44:56.412988 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/ddae0daa-153b-425b-92dd-d6c315ace4e5-ssh-key-openstack-edpm-ipam\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-nwdxw\" (UID: \"ddae0daa-153b-425b-92dd-d6c315ace4e5\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-nwdxw" Feb 02 09:44:56 crc kubenswrapper[4764]: I0202 09:44:56.413042 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ddae0daa-153b-425b-92dd-d6c315ace4e5-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-nwdxw\" (UID: \"ddae0daa-153b-425b-92dd-d6c315ace4e5\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-nwdxw" Feb 02 09:44:56 crc kubenswrapper[4764]: I0202 09:44:56.413071 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ddae0daa-153b-425b-92dd-d6c315ace4e5-ceph\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-nwdxw\" (UID: \"ddae0daa-153b-425b-92dd-d6c315ace4e5\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-nwdxw" Feb 02 09:44:56 crc kubenswrapper[4764]: I0202 09:44:56.422439 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ddae0daa-153b-425b-92dd-d6c315ace4e5-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-nwdxw\" (UID: \"ddae0daa-153b-425b-92dd-d6c315ace4e5\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-nwdxw" Feb 02 09:44:56 crc kubenswrapper[4764]: I0202 09:44:56.427613 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ddae0daa-153b-425b-92dd-d6c315ace4e5-ceph\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-nwdxw\" (UID: \"ddae0daa-153b-425b-92dd-d6c315ace4e5\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-nwdxw" Feb 02 09:44:56 crc kubenswrapper[4764]: I0202 09:44:56.428268 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/ddae0daa-153b-425b-92dd-d6c315ace4e5-ssh-key-openstack-edpm-ipam\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-nwdxw\" (UID: \"ddae0daa-153b-425b-92dd-d6c315ace4e5\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-nwdxw" Feb 02 09:44:56 crc kubenswrapper[4764]: I0202 09:44:56.430741 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2v4tg\" (UniqueName: \"kubernetes.io/projected/ddae0daa-153b-425b-92dd-d6c315ace4e5-kube-api-access-2v4tg\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-nwdxw\" (UID: \"ddae0daa-153b-425b-92dd-d6c315ace4e5\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-nwdxw" Feb 02 09:44:56 crc kubenswrapper[4764]: I0202 09:44:56.636757 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-nwdxw" Feb 02 09:44:57 crc kubenswrapper[4764]: I0202 09:44:57.226270 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-nwdxw"] Feb 02 09:44:57 crc kubenswrapper[4764]: I0202 09:44:57.243190 4764 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 02 09:44:58 crc kubenswrapper[4764]: I0202 09:44:58.193317 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-nwdxw" event={"ID":"ddae0daa-153b-425b-92dd-d6c315ace4e5","Type":"ContainerStarted","Data":"b0ce157a48fe0ad3a7bd1c2aeab75782b05dccb163f2930e30696f850ac7cb44"} Feb 02 09:44:58 crc kubenswrapper[4764]: I0202 09:44:58.193694 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-nwdxw" event={"ID":"ddae0daa-153b-425b-92dd-d6c315ace4e5","Type":"ContainerStarted","Data":"c095fbc0616871fd3d237339522410fdde50b077ae41d7534924eed6654637ff"} Feb 02 09:44:58 crc kubenswrapper[4764]: I0202 09:44:58.216749 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-nwdxw" podStartSLOduration=1.7645471640000001 podStartE2EDuration="2.216730582s" podCreationTimestamp="2026-02-02 09:44:56 +0000 UTC" firstStartedPulling="2026-02-02 09:44:57.242964005 +0000 UTC m=+2260.176688093" lastFinishedPulling="2026-02-02 09:44:57.695147423 +0000 UTC m=+2260.628871511" observedRunningTime="2026-02-02 09:44:58.207581695 +0000 UTC m=+2261.141305783" watchObservedRunningTime="2026-02-02 09:44:58.216730582 +0000 UTC m=+2261.150454670" Feb 02 09:45:00 crc kubenswrapper[4764]: I0202 09:45:00.143037 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29500425-ppfcg"] Feb 02 09:45:00 crc kubenswrapper[4764]: I0202 09:45:00.145013 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29500425-ppfcg" Feb 02 09:45:00 crc kubenswrapper[4764]: I0202 09:45:00.153578 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 02 09:45:00 crc kubenswrapper[4764]: I0202 09:45:00.154340 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 02 09:45:00 crc kubenswrapper[4764]: I0202 09:45:00.156450 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29500425-ppfcg"] Feb 02 09:45:00 crc kubenswrapper[4764]: I0202 09:45:00.296639 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vpjc4\" (UniqueName: \"kubernetes.io/projected/237b5e52-e331-43b5-9d2d-1ec2c8e90854-kube-api-access-vpjc4\") pod \"collect-profiles-29500425-ppfcg\" (UID: \"237b5e52-e331-43b5-9d2d-1ec2c8e90854\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500425-ppfcg" Feb 02 09:45:00 crc kubenswrapper[4764]: I0202 09:45:00.296718 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/237b5e52-e331-43b5-9d2d-1ec2c8e90854-config-volume\") pod \"collect-profiles-29500425-ppfcg\" (UID: \"237b5e52-e331-43b5-9d2d-1ec2c8e90854\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500425-ppfcg" Feb 02 09:45:00 crc kubenswrapper[4764]: I0202 09:45:00.296774 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/237b5e52-e331-43b5-9d2d-1ec2c8e90854-secret-volume\") pod \"collect-profiles-29500425-ppfcg\" (UID: \"237b5e52-e331-43b5-9d2d-1ec2c8e90854\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500425-ppfcg" Feb 02 09:45:00 crc kubenswrapper[4764]: I0202 09:45:00.398704 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vpjc4\" (UniqueName: \"kubernetes.io/projected/237b5e52-e331-43b5-9d2d-1ec2c8e90854-kube-api-access-vpjc4\") pod \"collect-profiles-29500425-ppfcg\" (UID: \"237b5e52-e331-43b5-9d2d-1ec2c8e90854\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500425-ppfcg" Feb 02 09:45:00 crc kubenswrapper[4764]: I0202 09:45:00.398804 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/237b5e52-e331-43b5-9d2d-1ec2c8e90854-config-volume\") pod \"collect-profiles-29500425-ppfcg\" (UID: \"237b5e52-e331-43b5-9d2d-1ec2c8e90854\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500425-ppfcg" Feb 02 09:45:00 crc kubenswrapper[4764]: I0202 09:45:00.398886 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/237b5e52-e331-43b5-9d2d-1ec2c8e90854-secret-volume\") pod \"collect-profiles-29500425-ppfcg\" (UID: \"237b5e52-e331-43b5-9d2d-1ec2c8e90854\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500425-ppfcg" Feb 02 09:45:00 crc kubenswrapper[4764]: I0202 09:45:00.400283 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/237b5e52-e331-43b5-9d2d-1ec2c8e90854-config-volume\") pod \"collect-profiles-29500425-ppfcg\" (UID: \"237b5e52-e331-43b5-9d2d-1ec2c8e90854\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500425-ppfcg" Feb 02 09:45:00 crc kubenswrapper[4764]: I0202 09:45:00.414578 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/237b5e52-e331-43b5-9d2d-1ec2c8e90854-secret-volume\") pod \"collect-profiles-29500425-ppfcg\" (UID: \"237b5e52-e331-43b5-9d2d-1ec2c8e90854\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500425-ppfcg" Feb 02 09:45:00 crc kubenswrapper[4764]: I0202 09:45:00.417470 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vpjc4\" (UniqueName: \"kubernetes.io/projected/237b5e52-e331-43b5-9d2d-1ec2c8e90854-kube-api-access-vpjc4\") pod \"collect-profiles-29500425-ppfcg\" (UID: \"237b5e52-e331-43b5-9d2d-1ec2c8e90854\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500425-ppfcg" Feb 02 09:45:00 crc kubenswrapper[4764]: I0202 09:45:00.477993 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29500425-ppfcg" Feb 02 09:45:00 crc kubenswrapper[4764]: I0202 09:45:00.721691 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29500425-ppfcg"] Feb 02 09:45:01 crc kubenswrapper[4764]: I0202 09:45:01.225340 4764 generic.go:334] "Generic (PLEG): container finished" podID="237b5e52-e331-43b5-9d2d-1ec2c8e90854" containerID="395d3dca5cb98df00c029c52d20ebbda1fe995a4adede18c47a7d36365157315" exitCode=0 Feb 02 09:45:01 crc kubenswrapper[4764]: I0202 09:45:01.225523 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29500425-ppfcg" event={"ID":"237b5e52-e331-43b5-9d2d-1ec2c8e90854","Type":"ContainerDied","Data":"395d3dca5cb98df00c029c52d20ebbda1fe995a4adede18c47a7d36365157315"} Feb 02 09:45:01 crc kubenswrapper[4764]: I0202 09:45:01.225740 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29500425-ppfcg" event={"ID":"237b5e52-e331-43b5-9d2d-1ec2c8e90854","Type":"ContainerStarted","Data":"3fe0ed5c307a06de1776257d53f82491c1ac521c88f34fe003abb94f263b7bf4"} Feb 02 09:45:01 crc kubenswrapper[4764]: E0202 09:45:01.699493 4764 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0331eb99_078e_4945_abb7_db9c1d76cfe3.slice/crio-5dd8f895484886381b347a8fd3c953b05c563b9b855d129e7d3b0756d128755d.scope\": RecentStats: unable to find data in memory cache]" Feb 02 09:45:02 crc kubenswrapper[4764]: I0202 09:45:02.641458 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29500425-ppfcg" Feb 02 09:45:02 crc kubenswrapper[4764]: I0202 09:45:02.663785 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/237b5e52-e331-43b5-9d2d-1ec2c8e90854-secret-volume\") pod \"237b5e52-e331-43b5-9d2d-1ec2c8e90854\" (UID: \"237b5e52-e331-43b5-9d2d-1ec2c8e90854\") " Feb 02 09:45:02 crc kubenswrapper[4764]: I0202 09:45:02.664456 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/237b5e52-e331-43b5-9d2d-1ec2c8e90854-config-volume\") pod \"237b5e52-e331-43b5-9d2d-1ec2c8e90854\" (UID: \"237b5e52-e331-43b5-9d2d-1ec2c8e90854\") " Feb 02 09:45:02 crc kubenswrapper[4764]: I0202 09:45:02.664580 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vpjc4\" (UniqueName: \"kubernetes.io/projected/237b5e52-e331-43b5-9d2d-1ec2c8e90854-kube-api-access-vpjc4\") pod \"237b5e52-e331-43b5-9d2d-1ec2c8e90854\" (UID: \"237b5e52-e331-43b5-9d2d-1ec2c8e90854\") " Feb 02 09:45:02 crc kubenswrapper[4764]: I0202 09:45:02.665555 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/237b5e52-e331-43b5-9d2d-1ec2c8e90854-config-volume" (OuterVolumeSpecName: "config-volume") pod "237b5e52-e331-43b5-9d2d-1ec2c8e90854" (UID: "237b5e52-e331-43b5-9d2d-1ec2c8e90854"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:45:02 crc kubenswrapper[4764]: I0202 09:45:02.669308 4764 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/237b5e52-e331-43b5-9d2d-1ec2c8e90854-config-volume\") on node \"crc\" DevicePath \"\"" Feb 02 09:45:02 crc kubenswrapper[4764]: I0202 09:45:02.671562 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/237b5e52-e331-43b5-9d2d-1ec2c8e90854-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "237b5e52-e331-43b5-9d2d-1ec2c8e90854" (UID: "237b5e52-e331-43b5-9d2d-1ec2c8e90854"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:45:02 crc kubenswrapper[4764]: I0202 09:45:02.683569 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/237b5e52-e331-43b5-9d2d-1ec2c8e90854-kube-api-access-vpjc4" (OuterVolumeSpecName: "kube-api-access-vpjc4") pod "237b5e52-e331-43b5-9d2d-1ec2c8e90854" (UID: "237b5e52-e331-43b5-9d2d-1ec2c8e90854"). InnerVolumeSpecName "kube-api-access-vpjc4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:45:02 crc kubenswrapper[4764]: I0202 09:45:02.772228 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vpjc4\" (UniqueName: \"kubernetes.io/projected/237b5e52-e331-43b5-9d2d-1ec2c8e90854-kube-api-access-vpjc4\") on node \"crc\" DevicePath \"\"" Feb 02 09:45:02 crc kubenswrapper[4764]: I0202 09:45:02.772355 4764 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/237b5e52-e331-43b5-9d2d-1ec2c8e90854-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 02 09:45:03 crc kubenswrapper[4764]: I0202 09:45:03.246642 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29500425-ppfcg" Feb 02 09:45:03 crc kubenswrapper[4764]: I0202 09:45:03.246637 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29500425-ppfcg" event={"ID":"237b5e52-e331-43b5-9d2d-1ec2c8e90854","Type":"ContainerDied","Data":"3fe0ed5c307a06de1776257d53f82491c1ac521c88f34fe003abb94f263b7bf4"} Feb 02 09:45:03 crc kubenswrapper[4764]: I0202 09:45:03.247068 4764 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3fe0ed5c307a06de1776257d53f82491c1ac521c88f34fe003abb94f263b7bf4" Feb 02 09:45:03 crc kubenswrapper[4764]: I0202 09:45:03.250014 4764 generic.go:334] "Generic (PLEG): container finished" podID="ddae0daa-153b-425b-92dd-d6c315ace4e5" containerID="b0ce157a48fe0ad3a7bd1c2aeab75782b05dccb163f2930e30696f850ac7cb44" exitCode=0 Feb 02 09:45:03 crc kubenswrapper[4764]: I0202 09:45:03.250054 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-nwdxw" event={"ID":"ddae0daa-153b-425b-92dd-d6c315ace4e5","Type":"ContainerDied","Data":"b0ce157a48fe0ad3a7bd1c2aeab75782b05dccb163f2930e30696f850ac7cb44"} Feb 02 09:45:03 crc kubenswrapper[4764]: I0202 09:45:03.730130 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29500380-2gwrp"] Feb 02 09:45:03 crc kubenswrapper[4764]: I0202 09:45:03.739041 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29500380-2gwrp"] Feb 02 09:45:03 crc kubenswrapper[4764]: I0202 09:45:03.859700 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="006ce277-7c6b-49a5-9e98-0bf7caee30f9" path="/var/lib/kubelet/pods/006ce277-7c6b-49a5-9e98-0bf7caee30f9/volumes" Feb 02 09:45:04 crc kubenswrapper[4764]: I0202 09:45:04.687857 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-nwdxw" Feb 02 09:45:04 crc kubenswrapper[4764]: I0202 09:45:04.714614 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ddae0daa-153b-425b-92dd-d6c315ace4e5-ceph\") pod \"ddae0daa-153b-425b-92dd-d6c315ace4e5\" (UID: \"ddae0daa-153b-425b-92dd-d6c315ace4e5\") " Feb 02 09:45:04 crc kubenswrapper[4764]: I0202 09:45:04.714828 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ddae0daa-153b-425b-92dd-d6c315ace4e5-inventory\") pod \"ddae0daa-153b-425b-92dd-d6c315ace4e5\" (UID: \"ddae0daa-153b-425b-92dd-d6c315ace4e5\") " Feb 02 09:45:04 crc kubenswrapper[4764]: I0202 09:45:04.715050 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2v4tg\" (UniqueName: \"kubernetes.io/projected/ddae0daa-153b-425b-92dd-d6c315ace4e5-kube-api-access-2v4tg\") pod \"ddae0daa-153b-425b-92dd-d6c315ace4e5\" (UID: \"ddae0daa-153b-425b-92dd-d6c315ace4e5\") " Feb 02 09:45:04 crc kubenswrapper[4764]: I0202 09:45:04.715164 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/ddae0daa-153b-425b-92dd-d6c315ace4e5-ssh-key-openstack-edpm-ipam\") pod \"ddae0daa-153b-425b-92dd-d6c315ace4e5\" (UID: \"ddae0daa-153b-425b-92dd-d6c315ace4e5\") " Feb 02 09:45:04 crc kubenswrapper[4764]: I0202 09:45:04.729084 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ddae0daa-153b-425b-92dd-d6c315ace4e5-ceph" (OuterVolumeSpecName: "ceph") pod "ddae0daa-153b-425b-92dd-d6c315ace4e5" (UID: "ddae0daa-153b-425b-92dd-d6c315ace4e5"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:45:04 crc kubenswrapper[4764]: I0202 09:45:04.743892 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ddae0daa-153b-425b-92dd-d6c315ace4e5-kube-api-access-2v4tg" (OuterVolumeSpecName: "kube-api-access-2v4tg") pod "ddae0daa-153b-425b-92dd-d6c315ace4e5" (UID: "ddae0daa-153b-425b-92dd-d6c315ace4e5"). InnerVolumeSpecName "kube-api-access-2v4tg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:45:04 crc kubenswrapper[4764]: I0202 09:45:04.750705 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ddae0daa-153b-425b-92dd-d6c315ace4e5-inventory" (OuterVolumeSpecName: "inventory") pod "ddae0daa-153b-425b-92dd-d6c315ace4e5" (UID: "ddae0daa-153b-425b-92dd-d6c315ace4e5"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:45:04 crc kubenswrapper[4764]: I0202 09:45:04.766348 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ddae0daa-153b-425b-92dd-d6c315ace4e5-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "ddae0daa-153b-425b-92dd-d6c315ace4e5" (UID: "ddae0daa-153b-425b-92dd-d6c315ace4e5"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:45:04 crc kubenswrapper[4764]: I0202 09:45:04.819325 4764 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ddae0daa-153b-425b-92dd-d6c315ace4e5-inventory\") on node \"crc\" DevicePath \"\"" Feb 02 09:45:04 crc kubenswrapper[4764]: I0202 09:45:04.819736 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2v4tg\" (UniqueName: \"kubernetes.io/projected/ddae0daa-153b-425b-92dd-d6c315ace4e5-kube-api-access-2v4tg\") on node \"crc\" DevicePath \"\"" Feb 02 09:45:04 crc kubenswrapper[4764]: I0202 09:45:04.819821 4764 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/ddae0daa-153b-425b-92dd-d6c315ace4e5-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 02 09:45:04 crc kubenswrapper[4764]: I0202 09:45:04.819897 4764 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ddae0daa-153b-425b-92dd-d6c315ace4e5-ceph\") on node \"crc\" DevicePath \"\"" Feb 02 09:45:05 crc kubenswrapper[4764]: I0202 09:45:05.276120 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-nwdxw" event={"ID":"ddae0daa-153b-425b-92dd-d6c315ace4e5","Type":"ContainerDied","Data":"c095fbc0616871fd3d237339522410fdde50b077ae41d7534924eed6654637ff"} Feb 02 09:45:05 crc kubenswrapper[4764]: I0202 09:45:05.276179 4764 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c095fbc0616871fd3d237339522410fdde50b077ae41d7534924eed6654637ff" Feb 02 09:45:05 crc kubenswrapper[4764]: I0202 09:45:05.276256 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-nwdxw" Feb 02 09:45:05 crc kubenswrapper[4764]: I0202 09:45:05.382983 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-5469f"] Feb 02 09:45:05 crc kubenswrapper[4764]: E0202 09:45:05.383389 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="237b5e52-e331-43b5-9d2d-1ec2c8e90854" containerName="collect-profiles" Feb 02 09:45:05 crc kubenswrapper[4764]: I0202 09:45:05.383407 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="237b5e52-e331-43b5-9d2d-1ec2c8e90854" containerName="collect-profiles" Feb 02 09:45:05 crc kubenswrapper[4764]: E0202 09:45:05.383422 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ddae0daa-153b-425b-92dd-d6c315ace4e5" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Feb 02 09:45:05 crc kubenswrapper[4764]: I0202 09:45:05.383431 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="ddae0daa-153b-425b-92dd-d6c315ace4e5" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Feb 02 09:45:05 crc kubenswrapper[4764]: I0202 09:45:05.383608 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="237b5e52-e331-43b5-9d2d-1ec2c8e90854" containerName="collect-profiles" Feb 02 09:45:05 crc kubenswrapper[4764]: I0202 09:45:05.383641 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="ddae0daa-153b-425b-92dd-d6c315ace4e5" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Feb 02 09:45:05 crc kubenswrapper[4764]: I0202 09:45:05.384310 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-5469f" Feb 02 09:45:05 crc kubenswrapper[4764]: I0202 09:45:05.387222 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Feb 02 09:45:05 crc kubenswrapper[4764]: I0202 09:45:05.387876 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 02 09:45:05 crc kubenswrapper[4764]: I0202 09:45:05.395572 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-xksq5" Feb 02 09:45:05 crc kubenswrapper[4764]: I0202 09:45:05.395572 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 02 09:45:05 crc kubenswrapper[4764]: I0202 09:45:05.395715 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 02 09:45:05 crc kubenswrapper[4764]: I0202 09:45:05.400378 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-5469f"] Feb 02 09:45:05 crc kubenswrapper[4764]: I0202 09:45:05.432001 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f2nlw\" (UniqueName: \"kubernetes.io/projected/6b2831d8-cc8c-4ded-99e1-45ad94ff9776-kube-api-access-f2nlw\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-5469f\" (UID: \"6b2831d8-cc8c-4ded-99e1-45ad94ff9776\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-5469f" Feb 02 09:45:05 crc kubenswrapper[4764]: I0202 09:45:05.432079 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6b2831d8-cc8c-4ded-99e1-45ad94ff9776-ceph\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-5469f\" (UID: \"6b2831d8-cc8c-4ded-99e1-45ad94ff9776\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-5469f" Feb 02 09:45:05 crc kubenswrapper[4764]: I0202 09:45:05.432125 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6b2831d8-cc8c-4ded-99e1-45ad94ff9776-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-5469f\" (UID: \"6b2831d8-cc8c-4ded-99e1-45ad94ff9776\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-5469f" Feb 02 09:45:05 crc kubenswrapper[4764]: I0202 09:45:05.432482 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/6b2831d8-cc8c-4ded-99e1-45ad94ff9776-ssh-key-openstack-edpm-ipam\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-5469f\" (UID: \"6b2831d8-cc8c-4ded-99e1-45ad94ff9776\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-5469f" Feb 02 09:45:05 crc kubenswrapper[4764]: I0202 09:45:05.535918 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/6b2831d8-cc8c-4ded-99e1-45ad94ff9776-ssh-key-openstack-edpm-ipam\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-5469f\" (UID: \"6b2831d8-cc8c-4ded-99e1-45ad94ff9776\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-5469f" Feb 02 09:45:05 crc kubenswrapper[4764]: I0202 09:45:05.536088 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f2nlw\" (UniqueName: \"kubernetes.io/projected/6b2831d8-cc8c-4ded-99e1-45ad94ff9776-kube-api-access-f2nlw\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-5469f\" (UID: \"6b2831d8-cc8c-4ded-99e1-45ad94ff9776\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-5469f" Feb 02 09:45:05 crc kubenswrapper[4764]: I0202 09:45:05.536132 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6b2831d8-cc8c-4ded-99e1-45ad94ff9776-ceph\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-5469f\" (UID: \"6b2831d8-cc8c-4ded-99e1-45ad94ff9776\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-5469f" Feb 02 09:45:05 crc kubenswrapper[4764]: I0202 09:45:05.536173 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6b2831d8-cc8c-4ded-99e1-45ad94ff9776-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-5469f\" (UID: \"6b2831d8-cc8c-4ded-99e1-45ad94ff9776\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-5469f" Feb 02 09:45:05 crc kubenswrapper[4764]: I0202 09:45:05.540195 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6b2831d8-cc8c-4ded-99e1-45ad94ff9776-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-5469f\" (UID: \"6b2831d8-cc8c-4ded-99e1-45ad94ff9776\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-5469f" Feb 02 09:45:05 crc kubenswrapper[4764]: I0202 09:45:05.540313 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/6b2831d8-cc8c-4ded-99e1-45ad94ff9776-ssh-key-openstack-edpm-ipam\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-5469f\" (UID: \"6b2831d8-cc8c-4ded-99e1-45ad94ff9776\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-5469f" Feb 02 09:45:05 crc kubenswrapper[4764]: I0202 09:45:05.540567 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6b2831d8-cc8c-4ded-99e1-45ad94ff9776-ceph\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-5469f\" (UID: \"6b2831d8-cc8c-4ded-99e1-45ad94ff9776\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-5469f" Feb 02 09:45:05 crc kubenswrapper[4764]: I0202 09:45:05.557806 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f2nlw\" (UniqueName: \"kubernetes.io/projected/6b2831d8-cc8c-4ded-99e1-45ad94ff9776-kube-api-access-f2nlw\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-5469f\" (UID: \"6b2831d8-cc8c-4ded-99e1-45ad94ff9776\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-5469f" Feb 02 09:45:05 crc kubenswrapper[4764]: I0202 09:45:05.712782 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-5469f" Feb 02 09:45:06 crc kubenswrapper[4764]: I0202 09:45:06.315719 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-5469f"] Feb 02 09:45:06 crc kubenswrapper[4764]: W0202 09:45:06.324326 4764 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6b2831d8_cc8c_4ded_99e1_45ad94ff9776.slice/crio-38a99869ae0648a0d0234fb32d15a9290537bf6bdf4ac8c021d4265aa4efda27 WatchSource:0}: Error finding container 38a99869ae0648a0d0234fb32d15a9290537bf6bdf4ac8c021d4265aa4efda27: Status 404 returned error can't find the container with id 38a99869ae0648a0d0234fb32d15a9290537bf6bdf4ac8c021d4265aa4efda27 Feb 02 09:45:07 crc kubenswrapper[4764]: I0202 09:45:07.298623 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-5469f" event={"ID":"6b2831d8-cc8c-4ded-99e1-45ad94ff9776","Type":"ContainerStarted","Data":"a0c10fa367b087873fe8629a20acdbc0ad61b8cc57059e7afa859a04540f79ed"} Feb 02 09:45:07 crc kubenswrapper[4764]: I0202 09:45:07.299261 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-5469f" event={"ID":"6b2831d8-cc8c-4ded-99e1-45ad94ff9776","Type":"ContainerStarted","Data":"38a99869ae0648a0d0234fb32d15a9290537bf6bdf4ac8c021d4265aa4efda27"} Feb 02 09:45:09 crc kubenswrapper[4764]: I0202 09:45:09.826458 4764 scope.go:117] "RemoveContainer" containerID="be8fa5800ec42d1c141765dcfc6111f3ef96b05b6e8d5a9feaf2baa219897bcf" Feb 02 09:45:09 crc kubenswrapper[4764]: E0202 09:45:09.827334 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 09:45:11 crc kubenswrapper[4764]: E0202 09:45:11.934580 4764 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0331eb99_078e_4945_abb7_db9c1d76cfe3.slice/crio-5dd8f895484886381b347a8fd3c953b05c563b9b855d129e7d3b0756d128755d.scope\": RecentStats: unable to find data in memory cache]" Feb 02 09:45:17 crc kubenswrapper[4764]: E0202 09:45:17.855218 4764 fsHandler.go:119] failed to collect filesystem stats - rootDiskErr: could not stat "/var/lib/containers/storage/overlay/f2fd00b1a91030aea515aea03fc65a5ca3ec9ae90b035b81aea8213ece99e636/diff" to get inode usage: stat /var/lib/containers/storage/overlay/f2fd00b1a91030aea515aea03fc65a5ca3ec9ae90b035b81aea8213ece99e636/diff: no such file or directory, extraDiskErr: could not stat "/var/log/pods/openshift-marketplace_certified-operators-5x5h2_0331eb99-078e-4945-abb7-db9c1d76cfe3/registry-server/0.log" to get inode usage: stat /var/log/pods/openshift-marketplace_certified-operators-5x5h2_0331eb99-078e-4945-abb7-db9c1d76cfe3/registry-server/0.log: no such file or directory Feb 02 09:45:20 crc kubenswrapper[4764]: I0202 09:45:20.828700 4764 scope.go:117] "RemoveContainer" containerID="be8fa5800ec42d1c141765dcfc6111f3ef96b05b6e8d5a9feaf2baa219897bcf" Feb 02 09:45:20 crc kubenswrapper[4764]: E0202 09:45:20.829485 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 09:45:27 crc kubenswrapper[4764]: I0202 09:45:27.613667 4764 scope.go:117] "RemoveContainer" containerID="3a0f35f6d6c22683f1379ec4e011d9fc24fb03375e1e4b79536567e348c406c7" Feb 02 09:45:31 crc kubenswrapper[4764]: I0202 09:45:31.826381 4764 scope.go:117] "RemoveContainer" containerID="be8fa5800ec42d1c141765dcfc6111f3ef96b05b6e8d5a9feaf2baa219897bcf" Feb 02 09:45:31 crc kubenswrapper[4764]: E0202 09:45:31.827557 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 09:45:42 crc kubenswrapper[4764]: I0202 09:45:42.825268 4764 scope.go:117] "RemoveContainer" containerID="be8fa5800ec42d1c141765dcfc6111f3ef96b05b6e8d5a9feaf2baa219897bcf" Feb 02 09:45:42 crc kubenswrapper[4764]: E0202 09:45:42.825968 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 09:45:48 crc kubenswrapper[4764]: I0202 09:45:48.684686 4764 generic.go:334] "Generic (PLEG): container finished" podID="6b2831d8-cc8c-4ded-99e1-45ad94ff9776" containerID="a0c10fa367b087873fe8629a20acdbc0ad61b8cc57059e7afa859a04540f79ed" exitCode=0 Feb 02 09:45:48 crc kubenswrapper[4764]: I0202 09:45:48.684922 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-5469f" event={"ID":"6b2831d8-cc8c-4ded-99e1-45ad94ff9776","Type":"ContainerDied","Data":"a0c10fa367b087873fe8629a20acdbc0ad61b8cc57059e7afa859a04540f79ed"} Feb 02 09:45:50 crc kubenswrapper[4764]: I0202 09:45:50.101691 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-5469f" Feb 02 09:45:50 crc kubenswrapper[4764]: I0202 09:45:50.252005 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6b2831d8-cc8c-4ded-99e1-45ad94ff9776-ceph\") pod \"6b2831d8-cc8c-4ded-99e1-45ad94ff9776\" (UID: \"6b2831d8-cc8c-4ded-99e1-45ad94ff9776\") " Feb 02 09:45:50 crc kubenswrapper[4764]: I0202 09:45:50.252206 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6b2831d8-cc8c-4ded-99e1-45ad94ff9776-inventory\") pod \"6b2831d8-cc8c-4ded-99e1-45ad94ff9776\" (UID: \"6b2831d8-cc8c-4ded-99e1-45ad94ff9776\") " Feb 02 09:45:50 crc kubenswrapper[4764]: I0202 09:45:50.253824 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f2nlw\" (UniqueName: \"kubernetes.io/projected/6b2831d8-cc8c-4ded-99e1-45ad94ff9776-kube-api-access-f2nlw\") pod \"6b2831d8-cc8c-4ded-99e1-45ad94ff9776\" (UID: \"6b2831d8-cc8c-4ded-99e1-45ad94ff9776\") " Feb 02 09:45:50 crc kubenswrapper[4764]: I0202 09:45:50.253989 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/6b2831d8-cc8c-4ded-99e1-45ad94ff9776-ssh-key-openstack-edpm-ipam\") pod \"6b2831d8-cc8c-4ded-99e1-45ad94ff9776\" (UID: \"6b2831d8-cc8c-4ded-99e1-45ad94ff9776\") " Feb 02 09:45:50 crc kubenswrapper[4764]: I0202 09:45:50.258132 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b2831d8-cc8c-4ded-99e1-45ad94ff9776-ceph" (OuterVolumeSpecName: "ceph") pod "6b2831d8-cc8c-4ded-99e1-45ad94ff9776" (UID: "6b2831d8-cc8c-4ded-99e1-45ad94ff9776"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:45:50 crc kubenswrapper[4764]: I0202 09:45:50.269930 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6b2831d8-cc8c-4ded-99e1-45ad94ff9776-kube-api-access-f2nlw" (OuterVolumeSpecName: "kube-api-access-f2nlw") pod "6b2831d8-cc8c-4ded-99e1-45ad94ff9776" (UID: "6b2831d8-cc8c-4ded-99e1-45ad94ff9776"). InnerVolumeSpecName "kube-api-access-f2nlw". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:45:50 crc kubenswrapper[4764]: I0202 09:45:50.284886 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b2831d8-cc8c-4ded-99e1-45ad94ff9776-inventory" (OuterVolumeSpecName: "inventory") pod "6b2831d8-cc8c-4ded-99e1-45ad94ff9776" (UID: "6b2831d8-cc8c-4ded-99e1-45ad94ff9776"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:45:50 crc kubenswrapper[4764]: I0202 09:45:50.294505 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b2831d8-cc8c-4ded-99e1-45ad94ff9776-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "6b2831d8-cc8c-4ded-99e1-45ad94ff9776" (UID: "6b2831d8-cc8c-4ded-99e1-45ad94ff9776"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:45:50 crc kubenswrapper[4764]: I0202 09:45:50.355879 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f2nlw\" (UniqueName: \"kubernetes.io/projected/6b2831d8-cc8c-4ded-99e1-45ad94ff9776-kube-api-access-f2nlw\") on node \"crc\" DevicePath \"\"" Feb 02 09:45:50 crc kubenswrapper[4764]: I0202 09:45:50.355927 4764 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/6b2831d8-cc8c-4ded-99e1-45ad94ff9776-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 02 09:45:50 crc kubenswrapper[4764]: I0202 09:45:50.355961 4764 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6b2831d8-cc8c-4ded-99e1-45ad94ff9776-ceph\") on node \"crc\" DevicePath \"\"" Feb 02 09:45:50 crc kubenswrapper[4764]: I0202 09:45:50.355981 4764 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6b2831d8-cc8c-4ded-99e1-45ad94ff9776-inventory\") on node \"crc\" DevicePath \"\"" Feb 02 09:45:50 crc kubenswrapper[4764]: I0202 09:45:50.708228 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-5469f" event={"ID":"6b2831d8-cc8c-4ded-99e1-45ad94ff9776","Type":"ContainerDied","Data":"38a99869ae0648a0d0234fb32d15a9290537bf6bdf4ac8c021d4265aa4efda27"} Feb 02 09:45:50 crc kubenswrapper[4764]: I0202 09:45:50.708289 4764 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="38a99869ae0648a0d0234fb32d15a9290537bf6bdf4ac8c021d4265aa4efda27" Feb 02 09:45:50 crc kubenswrapper[4764]: I0202 09:45:50.708351 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-5469f" Feb 02 09:45:50 crc kubenswrapper[4764]: I0202 09:45:50.831156 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-d4lm7"] Feb 02 09:45:50 crc kubenswrapper[4764]: E0202 09:45:50.831503 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b2831d8-cc8c-4ded-99e1-45ad94ff9776" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Feb 02 09:45:50 crc kubenswrapper[4764]: I0202 09:45:50.831516 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b2831d8-cc8c-4ded-99e1-45ad94ff9776" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Feb 02 09:45:50 crc kubenswrapper[4764]: I0202 09:45:50.831702 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b2831d8-cc8c-4ded-99e1-45ad94ff9776" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Feb 02 09:45:50 crc kubenswrapper[4764]: I0202 09:45:50.832255 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-d4lm7" Feb 02 09:45:50 crc kubenswrapper[4764]: I0202 09:45:50.836275 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Feb 02 09:45:50 crc kubenswrapper[4764]: I0202 09:45:50.839606 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 02 09:45:50 crc kubenswrapper[4764]: I0202 09:45:50.839758 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-xksq5" Feb 02 09:45:50 crc kubenswrapper[4764]: I0202 09:45:50.842827 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 02 09:45:50 crc kubenswrapper[4764]: I0202 09:45:50.858204 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 02 09:45:50 crc kubenswrapper[4764]: I0202 09:45:50.869445 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-d4lm7"] Feb 02 09:45:50 crc kubenswrapper[4764]: I0202 09:45:50.967003 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/329c5f09-ca39-4ced-a6b6-02db68c98510-ssh-key-openstack-edpm-ipam\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-d4lm7\" (UID: \"329c5f09-ca39-4ced-a6b6-02db68c98510\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-d4lm7" Feb 02 09:45:50 crc kubenswrapper[4764]: I0202 09:45:50.967134 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/329c5f09-ca39-4ced-a6b6-02db68c98510-ceph\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-d4lm7\" (UID: \"329c5f09-ca39-4ced-a6b6-02db68c98510\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-d4lm7" Feb 02 09:45:50 crc kubenswrapper[4764]: I0202 09:45:50.967156 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-djlmv\" (UniqueName: \"kubernetes.io/projected/329c5f09-ca39-4ced-a6b6-02db68c98510-kube-api-access-djlmv\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-d4lm7\" (UID: \"329c5f09-ca39-4ced-a6b6-02db68c98510\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-d4lm7" Feb 02 09:45:50 crc kubenswrapper[4764]: I0202 09:45:50.967187 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/329c5f09-ca39-4ced-a6b6-02db68c98510-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-d4lm7\" (UID: \"329c5f09-ca39-4ced-a6b6-02db68c98510\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-d4lm7" Feb 02 09:45:51 crc kubenswrapper[4764]: I0202 09:45:51.068780 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/329c5f09-ca39-4ced-a6b6-02db68c98510-ceph\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-d4lm7\" (UID: \"329c5f09-ca39-4ced-a6b6-02db68c98510\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-d4lm7" Feb 02 09:45:51 crc kubenswrapper[4764]: I0202 09:45:51.068837 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-djlmv\" (UniqueName: \"kubernetes.io/projected/329c5f09-ca39-4ced-a6b6-02db68c98510-kube-api-access-djlmv\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-d4lm7\" (UID: \"329c5f09-ca39-4ced-a6b6-02db68c98510\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-d4lm7" Feb 02 09:45:51 crc kubenswrapper[4764]: I0202 09:45:51.068871 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/329c5f09-ca39-4ced-a6b6-02db68c98510-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-d4lm7\" (UID: \"329c5f09-ca39-4ced-a6b6-02db68c98510\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-d4lm7" Feb 02 09:45:51 crc kubenswrapper[4764]: I0202 09:45:51.068956 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/329c5f09-ca39-4ced-a6b6-02db68c98510-ssh-key-openstack-edpm-ipam\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-d4lm7\" (UID: \"329c5f09-ca39-4ced-a6b6-02db68c98510\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-d4lm7" Feb 02 09:45:51 crc kubenswrapper[4764]: I0202 09:45:51.073987 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/329c5f09-ca39-4ced-a6b6-02db68c98510-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-d4lm7\" (UID: \"329c5f09-ca39-4ced-a6b6-02db68c98510\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-d4lm7" Feb 02 09:45:51 crc kubenswrapper[4764]: I0202 09:45:51.074245 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/329c5f09-ca39-4ced-a6b6-02db68c98510-ssh-key-openstack-edpm-ipam\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-d4lm7\" (UID: \"329c5f09-ca39-4ced-a6b6-02db68c98510\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-d4lm7" Feb 02 09:45:51 crc kubenswrapper[4764]: I0202 09:45:51.076616 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/329c5f09-ca39-4ced-a6b6-02db68c98510-ceph\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-d4lm7\" (UID: \"329c5f09-ca39-4ced-a6b6-02db68c98510\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-d4lm7" Feb 02 09:45:51 crc kubenswrapper[4764]: I0202 09:45:51.090828 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-djlmv\" (UniqueName: \"kubernetes.io/projected/329c5f09-ca39-4ced-a6b6-02db68c98510-kube-api-access-djlmv\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-d4lm7\" (UID: \"329c5f09-ca39-4ced-a6b6-02db68c98510\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-d4lm7" Feb 02 09:45:51 crc kubenswrapper[4764]: I0202 09:45:51.147469 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-d4lm7" Feb 02 09:45:51 crc kubenswrapper[4764]: W0202 09:45:51.460473 4764 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod329c5f09_ca39_4ced_a6b6_02db68c98510.slice/crio-f817fc0c47989af2f0ab43388bf1b6ff565b98b2d06f9a7a77cd7fecd74debd8 WatchSource:0}: Error finding container f817fc0c47989af2f0ab43388bf1b6ff565b98b2d06f9a7a77cd7fecd74debd8: Status 404 returned error can't find the container with id f817fc0c47989af2f0ab43388bf1b6ff565b98b2d06f9a7a77cd7fecd74debd8 Feb 02 09:45:51 crc kubenswrapper[4764]: I0202 09:45:51.465138 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-d4lm7"] Feb 02 09:45:51 crc kubenswrapper[4764]: I0202 09:45:51.719672 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-d4lm7" event={"ID":"329c5f09-ca39-4ced-a6b6-02db68c98510","Type":"ContainerStarted","Data":"f817fc0c47989af2f0ab43388bf1b6ff565b98b2d06f9a7a77cd7fecd74debd8"} Feb 02 09:45:52 crc kubenswrapper[4764]: I0202 09:45:52.737002 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-d4lm7" event={"ID":"329c5f09-ca39-4ced-a6b6-02db68c98510","Type":"ContainerStarted","Data":"a5d1e420c7c75664573f3c0083ae9b5c77e936fe283237307224fea5cb54ce64"} Feb 02 09:45:52 crc kubenswrapper[4764]: I0202 09:45:52.785262 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-d4lm7" podStartSLOduration=2.345381236 podStartE2EDuration="2.785242411s" podCreationTimestamp="2026-02-02 09:45:50 +0000 UTC" firstStartedPulling="2026-02-02 09:45:51.466162192 +0000 UTC m=+2314.399886300" lastFinishedPulling="2026-02-02 09:45:51.906023357 +0000 UTC m=+2314.839747475" observedRunningTime="2026-02-02 09:45:52.767565759 +0000 UTC m=+2315.701289867" watchObservedRunningTime="2026-02-02 09:45:52.785242411 +0000 UTC m=+2315.718966519" Feb 02 09:45:55 crc kubenswrapper[4764]: I0202 09:45:55.826504 4764 scope.go:117] "RemoveContainer" containerID="be8fa5800ec42d1c141765dcfc6111f3ef96b05b6e8d5a9feaf2baa219897bcf" Feb 02 09:45:55 crc kubenswrapper[4764]: E0202 09:45:55.827697 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 09:45:56 crc kubenswrapper[4764]: I0202 09:45:56.777946 4764 generic.go:334] "Generic (PLEG): container finished" podID="329c5f09-ca39-4ced-a6b6-02db68c98510" containerID="a5d1e420c7c75664573f3c0083ae9b5c77e936fe283237307224fea5cb54ce64" exitCode=0 Feb 02 09:45:56 crc kubenswrapper[4764]: I0202 09:45:56.777998 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-d4lm7" event={"ID":"329c5f09-ca39-4ced-a6b6-02db68c98510","Type":"ContainerDied","Data":"a5d1e420c7c75664573f3c0083ae9b5c77e936fe283237307224fea5cb54ce64"} Feb 02 09:45:58 crc kubenswrapper[4764]: I0202 09:45:58.135455 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-d4lm7" Feb 02 09:45:58 crc kubenswrapper[4764]: I0202 09:45:58.320213 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/329c5f09-ca39-4ced-a6b6-02db68c98510-ssh-key-openstack-edpm-ipam\") pod \"329c5f09-ca39-4ced-a6b6-02db68c98510\" (UID: \"329c5f09-ca39-4ced-a6b6-02db68c98510\") " Feb 02 09:45:58 crc kubenswrapper[4764]: I0202 09:45:58.320668 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/329c5f09-ca39-4ced-a6b6-02db68c98510-ceph\") pod \"329c5f09-ca39-4ced-a6b6-02db68c98510\" (UID: \"329c5f09-ca39-4ced-a6b6-02db68c98510\") " Feb 02 09:45:58 crc kubenswrapper[4764]: I0202 09:45:58.321747 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/329c5f09-ca39-4ced-a6b6-02db68c98510-inventory\") pod \"329c5f09-ca39-4ced-a6b6-02db68c98510\" (UID: \"329c5f09-ca39-4ced-a6b6-02db68c98510\") " Feb 02 09:45:58 crc kubenswrapper[4764]: I0202 09:45:58.321822 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-djlmv\" (UniqueName: \"kubernetes.io/projected/329c5f09-ca39-4ced-a6b6-02db68c98510-kube-api-access-djlmv\") pod \"329c5f09-ca39-4ced-a6b6-02db68c98510\" (UID: \"329c5f09-ca39-4ced-a6b6-02db68c98510\") " Feb 02 09:45:58 crc kubenswrapper[4764]: I0202 09:45:58.339896 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/329c5f09-ca39-4ced-a6b6-02db68c98510-kube-api-access-djlmv" (OuterVolumeSpecName: "kube-api-access-djlmv") pod "329c5f09-ca39-4ced-a6b6-02db68c98510" (UID: "329c5f09-ca39-4ced-a6b6-02db68c98510"). InnerVolumeSpecName "kube-api-access-djlmv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:45:58 crc kubenswrapper[4764]: I0202 09:45:58.342998 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/329c5f09-ca39-4ced-a6b6-02db68c98510-ceph" (OuterVolumeSpecName: "ceph") pod "329c5f09-ca39-4ced-a6b6-02db68c98510" (UID: "329c5f09-ca39-4ced-a6b6-02db68c98510"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:45:58 crc kubenswrapper[4764]: I0202 09:45:58.354566 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/329c5f09-ca39-4ced-a6b6-02db68c98510-inventory" (OuterVolumeSpecName: "inventory") pod "329c5f09-ca39-4ced-a6b6-02db68c98510" (UID: "329c5f09-ca39-4ced-a6b6-02db68c98510"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:45:58 crc kubenswrapper[4764]: I0202 09:45:58.359776 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/329c5f09-ca39-4ced-a6b6-02db68c98510-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "329c5f09-ca39-4ced-a6b6-02db68c98510" (UID: "329c5f09-ca39-4ced-a6b6-02db68c98510"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:45:58 crc kubenswrapper[4764]: I0202 09:45:58.424084 4764 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/329c5f09-ca39-4ced-a6b6-02db68c98510-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 02 09:45:58 crc kubenswrapper[4764]: I0202 09:45:58.425350 4764 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/329c5f09-ca39-4ced-a6b6-02db68c98510-ceph\") on node \"crc\" DevicePath \"\"" Feb 02 09:45:58 crc kubenswrapper[4764]: I0202 09:45:58.425388 4764 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/329c5f09-ca39-4ced-a6b6-02db68c98510-inventory\") on node \"crc\" DevicePath \"\"" Feb 02 09:45:58 crc kubenswrapper[4764]: I0202 09:45:58.425402 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-djlmv\" (UniqueName: \"kubernetes.io/projected/329c5f09-ca39-4ced-a6b6-02db68c98510-kube-api-access-djlmv\") on node \"crc\" DevicePath \"\"" Feb 02 09:45:58 crc kubenswrapper[4764]: I0202 09:45:58.803664 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-d4lm7" event={"ID":"329c5f09-ca39-4ced-a6b6-02db68c98510","Type":"ContainerDied","Data":"f817fc0c47989af2f0ab43388bf1b6ff565b98b2d06f9a7a77cd7fecd74debd8"} Feb 02 09:45:58 crc kubenswrapper[4764]: I0202 09:45:58.803732 4764 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f817fc0c47989af2f0ab43388bf1b6ff565b98b2d06f9a7a77cd7fecd74debd8" Feb 02 09:45:58 crc kubenswrapper[4764]: I0202 09:45:58.803819 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-d4lm7" Feb 02 09:45:58 crc kubenswrapper[4764]: I0202 09:45:58.896104 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-c4mpn"] Feb 02 09:45:58 crc kubenswrapper[4764]: E0202 09:45:58.896493 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="329c5f09-ca39-4ced-a6b6-02db68c98510" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Feb 02 09:45:58 crc kubenswrapper[4764]: I0202 09:45:58.896511 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="329c5f09-ca39-4ced-a6b6-02db68c98510" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Feb 02 09:45:58 crc kubenswrapper[4764]: I0202 09:45:58.896668 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="329c5f09-ca39-4ced-a6b6-02db68c98510" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Feb 02 09:45:58 crc kubenswrapper[4764]: I0202 09:45:58.897302 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-c4mpn" Feb 02 09:45:58 crc kubenswrapper[4764]: I0202 09:45:58.900215 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 02 09:45:58 crc kubenswrapper[4764]: I0202 09:45:58.900592 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Feb 02 09:45:58 crc kubenswrapper[4764]: I0202 09:45:58.900703 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 02 09:45:58 crc kubenswrapper[4764]: I0202 09:45:58.900737 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-xksq5" Feb 02 09:45:58 crc kubenswrapper[4764]: I0202 09:45:58.900860 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 02 09:45:58 crc kubenswrapper[4764]: I0202 09:45:58.908731 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-c4mpn"] Feb 02 09:45:58 crc kubenswrapper[4764]: I0202 09:45:58.936218 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c3b49d2a-ed47-45cb-9940-b5e87ff7a197-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-c4mpn\" (UID: \"c3b49d2a-ed47-45cb-9940-b5e87ff7a197\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-c4mpn" Feb 02 09:45:58 crc kubenswrapper[4764]: I0202 09:45:58.936419 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/c3b49d2a-ed47-45cb-9940-b5e87ff7a197-ceph\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-c4mpn\" (UID: \"c3b49d2a-ed47-45cb-9940-b5e87ff7a197\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-c4mpn" Feb 02 09:45:58 crc kubenswrapper[4764]: I0202 09:45:58.936491 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-98k6d\" (UniqueName: \"kubernetes.io/projected/c3b49d2a-ed47-45cb-9940-b5e87ff7a197-kube-api-access-98k6d\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-c4mpn\" (UID: \"c3b49d2a-ed47-45cb-9940-b5e87ff7a197\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-c4mpn" Feb 02 09:45:58 crc kubenswrapper[4764]: I0202 09:45:58.936590 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c3b49d2a-ed47-45cb-9940-b5e87ff7a197-ssh-key-openstack-edpm-ipam\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-c4mpn\" (UID: \"c3b49d2a-ed47-45cb-9940-b5e87ff7a197\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-c4mpn" Feb 02 09:45:59 crc kubenswrapper[4764]: I0202 09:45:59.037801 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c3b49d2a-ed47-45cb-9940-b5e87ff7a197-ssh-key-openstack-edpm-ipam\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-c4mpn\" (UID: \"c3b49d2a-ed47-45cb-9940-b5e87ff7a197\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-c4mpn" Feb 02 09:45:59 crc kubenswrapper[4764]: I0202 09:45:59.037882 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c3b49d2a-ed47-45cb-9940-b5e87ff7a197-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-c4mpn\" (UID: \"c3b49d2a-ed47-45cb-9940-b5e87ff7a197\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-c4mpn" Feb 02 09:45:59 crc kubenswrapper[4764]: I0202 09:45:59.037962 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/c3b49d2a-ed47-45cb-9940-b5e87ff7a197-ceph\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-c4mpn\" (UID: \"c3b49d2a-ed47-45cb-9940-b5e87ff7a197\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-c4mpn" Feb 02 09:45:59 crc kubenswrapper[4764]: I0202 09:45:59.037995 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-98k6d\" (UniqueName: \"kubernetes.io/projected/c3b49d2a-ed47-45cb-9940-b5e87ff7a197-kube-api-access-98k6d\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-c4mpn\" (UID: \"c3b49d2a-ed47-45cb-9940-b5e87ff7a197\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-c4mpn" Feb 02 09:45:59 crc kubenswrapper[4764]: I0202 09:45:59.041662 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c3b49d2a-ed47-45cb-9940-b5e87ff7a197-ssh-key-openstack-edpm-ipam\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-c4mpn\" (UID: \"c3b49d2a-ed47-45cb-9940-b5e87ff7a197\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-c4mpn" Feb 02 09:45:59 crc kubenswrapper[4764]: I0202 09:45:59.041673 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/c3b49d2a-ed47-45cb-9940-b5e87ff7a197-ceph\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-c4mpn\" (UID: \"c3b49d2a-ed47-45cb-9940-b5e87ff7a197\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-c4mpn" Feb 02 09:45:59 crc kubenswrapper[4764]: I0202 09:45:59.042216 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c3b49d2a-ed47-45cb-9940-b5e87ff7a197-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-c4mpn\" (UID: \"c3b49d2a-ed47-45cb-9940-b5e87ff7a197\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-c4mpn" Feb 02 09:45:59 crc kubenswrapper[4764]: I0202 09:45:59.057014 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-98k6d\" (UniqueName: \"kubernetes.io/projected/c3b49d2a-ed47-45cb-9940-b5e87ff7a197-kube-api-access-98k6d\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-c4mpn\" (UID: \"c3b49d2a-ed47-45cb-9940-b5e87ff7a197\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-c4mpn" Feb 02 09:45:59 crc kubenswrapper[4764]: I0202 09:45:59.219077 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-c4mpn" Feb 02 09:45:59 crc kubenswrapper[4764]: I0202 09:45:59.757171 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-c4mpn"] Feb 02 09:45:59 crc kubenswrapper[4764]: I0202 09:45:59.811070 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-c4mpn" event={"ID":"c3b49d2a-ed47-45cb-9940-b5e87ff7a197","Type":"ContainerStarted","Data":"644cf2913a2c186f6af6286634d614839ab6508cfde5cd050fe5bf6d78378b9d"} Feb 02 09:46:00 crc kubenswrapper[4764]: I0202 09:46:00.839315 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-c4mpn" event={"ID":"c3b49d2a-ed47-45cb-9940-b5e87ff7a197","Type":"ContainerStarted","Data":"b91240df39dde43834542613139aa269e1cfed831f09265e15245058a8feb8e7"} Feb 02 09:46:00 crc kubenswrapper[4764]: I0202 09:46:00.865178 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-c4mpn" podStartSLOduration=2.367020885 podStartE2EDuration="2.865159456s" podCreationTimestamp="2026-02-02 09:45:58 +0000 UTC" firstStartedPulling="2026-02-02 09:45:59.762625507 +0000 UTC m=+2322.696349595" lastFinishedPulling="2026-02-02 09:46:00.260764078 +0000 UTC m=+2323.194488166" observedRunningTime="2026-02-02 09:46:00.859009011 +0000 UTC m=+2323.792733109" watchObservedRunningTime="2026-02-02 09:46:00.865159456 +0000 UTC m=+2323.798883544" Feb 02 09:46:06 crc kubenswrapper[4764]: I0202 09:46:06.826514 4764 scope.go:117] "RemoveContainer" containerID="be8fa5800ec42d1c141765dcfc6111f3ef96b05b6e8d5a9feaf2baa219897bcf" Feb 02 09:46:06 crc kubenswrapper[4764]: E0202 09:46:06.827761 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 09:46:17 crc kubenswrapper[4764]: I0202 09:46:17.825770 4764 scope.go:117] "RemoveContainer" containerID="be8fa5800ec42d1c141765dcfc6111f3ef96b05b6e8d5a9feaf2baa219897bcf" Feb 02 09:46:17 crc kubenswrapper[4764]: E0202 09:46:17.827801 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 09:46:29 crc kubenswrapper[4764]: I0202 09:46:29.825812 4764 scope.go:117] "RemoveContainer" containerID="be8fa5800ec42d1c141765dcfc6111f3ef96b05b6e8d5a9feaf2baa219897bcf" Feb 02 09:46:29 crc kubenswrapper[4764]: E0202 09:46:29.826432 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 09:46:43 crc kubenswrapper[4764]: I0202 09:46:43.825578 4764 scope.go:117] "RemoveContainer" containerID="be8fa5800ec42d1c141765dcfc6111f3ef96b05b6e8d5a9feaf2baa219897bcf" Feb 02 09:46:43 crc kubenswrapper[4764]: E0202 09:46:43.826204 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 09:46:49 crc kubenswrapper[4764]: I0202 09:46:49.238284 4764 generic.go:334] "Generic (PLEG): container finished" podID="c3b49d2a-ed47-45cb-9940-b5e87ff7a197" containerID="b91240df39dde43834542613139aa269e1cfed831f09265e15245058a8feb8e7" exitCode=0 Feb 02 09:46:49 crc kubenswrapper[4764]: I0202 09:46:49.238402 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-c4mpn" event={"ID":"c3b49d2a-ed47-45cb-9940-b5e87ff7a197","Type":"ContainerDied","Data":"b91240df39dde43834542613139aa269e1cfed831f09265e15245058a8feb8e7"} Feb 02 09:46:50 crc kubenswrapper[4764]: I0202 09:46:50.720522 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-c4mpn" Feb 02 09:46:50 crc kubenswrapper[4764]: I0202 09:46:50.799168 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-98k6d\" (UniqueName: \"kubernetes.io/projected/c3b49d2a-ed47-45cb-9940-b5e87ff7a197-kube-api-access-98k6d\") pod \"c3b49d2a-ed47-45cb-9940-b5e87ff7a197\" (UID: \"c3b49d2a-ed47-45cb-9940-b5e87ff7a197\") " Feb 02 09:46:50 crc kubenswrapper[4764]: I0202 09:46:50.799310 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c3b49d2a-ed47-45cb-9940-b5e87ff7a197-ssh-key-openstack-edpm-ipam\") pod \"c3b49d2a-ed47-45cb-9940-b5e87ff7a197\" (UID: \"c3b49d2a-ed47-45cb-9940-b5e87ff7a197\") " Feb 02 09:46:50 crc kubenswrapper[4764]: I0202 09:46:50.799447 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/c3b49d2a-ed47-45cb-9940-b5e87ff7a197-ceph\") pod \"c3b49d2a-ed47-45cb-9940-b5e87ff7a197\" (UID: \"c3b49d2a-ed47-45cb-9940-b5e87ff7a197\") " Feb 02 09:46:50 crc kubenswrapper[4764]: I0202 09:46:50.799519 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c3b49d2a-ed47-45cb-9940-b5e87ff7a197-inventory\") pod \"c3b49d2a-ed47-45cb-9940-b5e87ff7a197\" (UID: \"c3b49d2a-ed47-45cb-9940-b5e87ff7a197\") " Feb 02 09:46:50 crc kubenswrapper[4764]: I0202 09:46:50.823133 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c3b49d2a-ed47-45cb-9940-b5e87ff7a197-kube-api-access-98k6d" (OuterVolumeSpecName: "kube-api-access-98k6d") pod "c3b49d2a-ed47-45cb-9940-b5e87ff7a197" (UID: "c3b49d2a-ed47-45cb-9940-b5e87ff7a197"). InnerVolumeSpecName "kube-api-access-98k6d". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:46:50 crc kubenswrapper[4764]: I0202 09:46:50.824851 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c3b49d2a-ed47-45cb-9940-b5e87ff7a197-ceph" (OuterVolumeSpecName: "ceph") pod "c3b49d2a-ed47-45cb-9940-b5e87ff7a197" (UID: "c3b49d2a-ed47-45cb-9940-b5e87ff7a197"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:46:50 crc kubenswrapper[4764]: I0202 09:46:50.830085 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c3b49d2a-ed47-45cb-9940-b5e87ff7a197-inventory" (OuterVolumeSpecName: "inventory") pod "c3b49d2a-ed47-45cb-9940-b5e87ff7a197" (UID: "c3b49d2a-ed47-45cb-9940-b5e87ff7a197"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:46:50 crc kubenswrapper[4764]: I0202 09:46:50.846552 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c3b49d2a-ed47-45cb-9940-b5e87ff7a197-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "c3b49d2a-ed47-45cb-9940-b5e87ff7a197" (UID: "c3b49d2a-ed47-45cb-9940-b5e87ff7a197"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:46:50 crc kubenswrapper[4764]: I0202 09:46:50.903104 4764 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c3b49d2a-ed47-45cb-9940-b5e87ff7a197-inventory\") on node \"crc\" DevicePath \"\"" Feb 02 09:46:50 crc kubenswrapper[4764]: I0202 09:46:50.903521 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-98k6d\" (UniqueName: \"kubernetes.io/projected/c3b49d2a-ed47-45cb-9940-b5e87ff7a197-kube-api-access-98k6d\") on node \"crc\" DevicePath \"\"" Feb 02 09:46:50 crc kubenswrapper[4764]: I0202 09:46:50.903836 4764 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c3b49d2a-ed47-45cb-9940-b5e87ff7a197-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 02 09:46:50 crc kubenswrapper[4764]: I0202 09:46:50.904023 4764 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/c3b49d2a-ed47-45cb-9940-b5e87ff7a197-ceph\") on node \"crc\" DevicePath \"\"" Feb 02 09:46:51 crc kubenswrapper[4764]: I0202 09:46:51.260115 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-c4mpn" event={"ID":"c3b49d2a-ed47-45cb-9940-b5e87ff7a197","Type":"ContainerDied","Data":"644cf2913a2c186f6af6286634d614839ab6508cfde5cd050fe5bf6d78378b9d"} Feb 02 09:46:51 crc kubenswrapper[4764]: I0202 09:46:51.260370 4764 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="644cf2913a2c186f6af6286634d614839ab6508cfde5cd050fe5bf6d78378b9d" Feb 02 09:46:51 crc kubenswrapper[4764]: I0202 09:46:51.260313 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-c4mpn" Feb 02 09:46:51 crc kubenswrapper[4764]: I0202 09:46:51.372701 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-97dd2"] Feb 02 09:46:51 crc kubenswrapper[4764]: E0202 09:46:51.373068 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3b49d2a-ed47-45cb-9940-b5e87ff7a197" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Feb 02 09:46:51 crc kubenswrapper[4764]: I0202 09:46:51.373086 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3b49d2a-ed47-45cb-9940-b5e87ff7a197" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Feb 02 09:46:51 crc kubenswrapper[4764]: I0202 09:46:51.373268 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="c3b49d2a-ed47-45cb-9940-b5e87ff7a197" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Feb 02 09:46:51 crc kubenswrapper[4764]: I0202 09:46:51.373800 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-97dd2" Feb 02 09:46:51 crc kubenswrapper[4764]: I0202 09:46:51.379240 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Feb 02 09:46:51 crc kubenswrapper[4764]: I0202 09:46:51.379327 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 02 09:46:51 crc kubenswrapper[4764]: I0202 09:46:51.382015 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 02 09:46:51 crc kubenswrapper[4764]: I0202 09:46:51.382244 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 02 09:46:51 crc kubenswrapper[4764]: I0202 09:46:51.393293 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-xksq5" Feb 02 09:46:51 crc kubenswrapper[4764]: I0202 09:46:51.393963 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-97dd2"] Feb 02 09:46:51 crc kubenswrapper[4764]: I0202 09:46:51.516458 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/409c3e31-e6bf-409c-9bd8-7bcf433c0ce8-ceph\") pod \"ssh-known-hosts-edpm-deployment-97dd2\" (UID: \"409c3e31-e6bf-409c-9bd8-7bcf433c0ce8\") " pod="openstack/ssh-known-hosts-edpm-deployment-97dd2" Feb 02 09:46:51 crc kubenswrapper[4764]: I0202 09:46:51.516552 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/409c3e31-e6bf-409c-9bd8-7bcf433c0ce8-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-97dd2\" (UID: \"409c3e31-e6bf-409c-9bd8-7bcf433c0ce8\") " pod="openstack/ssh-known-hosts-edpm-deployment-97dd2" Feb 02 09:46:51 crc kubenswrapper[4764]: I0202 09:46:51.516613 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/409c3e31-e6bf-409c-9bd8-7bcf433c0ce8-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-97dd2\" (UID: \"409c3e31-e6bf-409c-9bd8-7bcf433c0ce8\") " pod="openstack/ssh-known-hosts-edpm-deployment-97dd2" Feb 02 09:46:51 crc kubenswrapper[4764]: I0202 09:46:51.516648 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bl4t6\" (UniqueName: \"kubernetes.io/projected/409c3e31-e6bf-409c-9bd8-7bcf433c0ce8-kube-api-access-bl4t6\") pod \"ssh-known-hosts-edpm-deployment-97dd2\" (UID: \"409c3e31-e6bf-409c-9bd8-7bcf433c0ce8\") " pod="openstack/ssh-known-hosts-edpm-deployment-97dd2" Feb 02 09:46:51 crc kubenswrapper[4764]: I0202 09:46:51.618021 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/409c3e31-e6bf-409c-9bd8-7bcf433c0ce8-ceph\") pod \"ssh-known-hosts-edpm-deployment-97dd2\" (UID: \"409c3e31-e6bf-409c-9bd8-7bcf433c0ce8\") " pod="openstack/ssh-known-hosts-edpm-deployment-97dd2" Feb 02 09:46:51 crc kubenswrapper[4764]: I0202 09:46:51.618150 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/409c3e31-e6bf-409c-9bd8-7bcf433c0ce8-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-97dd2\" (UID: \"409c3e31-e6bf-409c-9bd8-7bcf433c0ce8\") " pod="openstack/ssh-known-hosts-edpm-deployment-97dd2" Feb 02 09:46:51 crc kubenswrapper[4764]: I0202 09:46:51.618230 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/409c3e31-e6bf-409c-9bd8-7bcf433c0ce8-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-97dd2\" (UID: \"409c3e31-e6bf-409c-9bd8-7bcf433c0ce8\") " pod="openstack/ssh-known-hosts-edpm-deployment-97dd2" Feb 02 09:46:51 crc kubenswrapper[4764]: I0202 09:46:51.618278 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bl4t6\" (UniqueName: \"kubernetes.io/projected/409c3e31-e6bf-409c-9bd8-7bcf433c0ce8-kube-api-access-bl4t6\") pod \"ssh-known-hosts-edpm-deployment-97dd2\" (UID: \"409c3e31-e6bf-409c-9bd8-7bcf433c0ce8\") " pod="openstack/ssh-known-hosts-edpm-deployment-97dd2" Feb 02 09:46:51 crc kubenswrapper[4764]: I0202 09:46:51.624783 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/409c3e31-e6bf-409c-9bd8-7bcf433c0ce8-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-97dd2\" (UID: \"409c3e31-e6bf-409c-9bd8-7bcf433c0ce8\") " pod="openstack/ssh-known-hosts-edpm-deployment-97dd2" Feb 02 09:46:51 crc kubenswrapper[4764]: I0202 09:46:51.626085 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/409c3e31-e6bf-409c-9bd8-7bcf433c0ce8-ceph\") pod \"ssh-known-hosts-edpm-deployment-97dd2\" (UID: \"409c3e31-e6bf-409c-9bd8-7bcf433c0ce8\") " pod="openstack/ssh-known-hosts-edpm-deployment-97dd2" Feb 02 09:46:51 crc kubenswrapper[4764]: I0202 09:46:51.631539 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/409c3e31-e6bf-409c-9bd8-7bcf433c0ce8-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-97dd2\" (UID: \"409c3e31-e6bf-409c-9bd8-7bcf433c0ce8\") " pod="openstack/ssh-known-hosts-edpm-deployment-97dd2" Feb 02 09:46:51 crc kubenswrapper[4764]: I0202 09:46:51.646132 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bl4t6\" (UniqueName: \"kubernetes.io/projected/409c3e31-e6bf-409c-9bd8-7bcf433c0ce8-kube-api-access-bl4t6\") pod \"ssh-known-hosts-edpm-deployment-97dd2\" (UID: \"409c3e31-e6bf-409c-9bd8-7bcf433c0ce8\") " pod="openstack/ssh-known-hosts-edpm-deployment-97dd2" Feb 02 09:46:51 crc kubenswrapper[4764]: I0202 09:46:51.691219 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-97dd2" Feb 02 09:46:52 crc kubenswrapper[4764]: I0202 09:46:52.269305 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-97dd2"] Feb 02 09:46:53 crc kubenswrapper[4764]: I0202 09:46:53.279591 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-97dd2" event={"ID":"409c3e31-e6bf-409c-9bd8-7bcf433c0ce8","Type":"ContainerStarted","Data":"dd8da8cc452ac2a840909f7dfd6f57d0683c6d71f2fb92c5fe84d2772d8e37de"} Feb 02 09:46:53 crc kubenswrapper[4764]: I0202 09:46:53.280186 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-97dd2" event={"ID":"409c3e31-e6bf-409c-9bd8-7bcf433c0ce8","Type":"ContainerStarted","Data":"9a4e5a941b8d35e237c2dcee2b5ccee3397b86304f2972283cdaacb7736493e0"} Feb 02 09:46:53 crc kubenswrapper[4764]: I0202 09:46:53.314792 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-97dd2" podStartSLOduration=1.645631715 podStartE2EDuration="2.314774332s" podCreationTimestamp="2026-02-02 09:46:51 +0000 UTC" firstStartedPulling="2026-02-02 09:46:52.267410338 +0000 UTC m=+2375.201134426" lastFinishedPulling="2026-02-02 09:46:52.936552955 +0000 UTC m=+2375.870277043" observedRunningTime="2026-02-02 09:46:53.306572063 +0000 UTC m=+2376.240296171" watchObservedRunningTime="2026-02-02 09:46:53.314774332 +0000 UTC m=+2376.248498430" Feb 02 09:46:57 crc kubenswrapper[4764]: I0202 09:46:57.836714 4764 scope.go:117] "RemoveContainer" containerID="be8fa5800ec42d1c141765dcfc6111f3ef96b05b6e8d5a9feaf2baa219897bcf" Feb 02 09:46:57 crc kubenswrapper[4764]: E0202 09:46:57.837993 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 09:47:02 crc kubenswrapper[4764]: I0202 09:47:02.367061 4764 generic.go:334] "Generic (PLEG): container finished" podID="409c3e31-e6bf-409c-9bd8-7bcf433c0ce8" containerID="dd8da8cc452ac2a840909f7dfd6f57d0683c6d71f2fb92c5fe84d2772d8e37de" exitCode=0 Feb 02 09:47:02 crc kubenswrapper[4764]: I0202 09:47:02.367168 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-97dd2" event={"ID":"409c3e31-e6bf-409c-9bd8-7bcf433c0ce8","Type":"ContainerDied","Data":"dd8da8cc452ac2a840909f7dfd6f57d0683c6d71f2fb92c5fe84d2772d8e37de"} Feb 02 09:47:03 crc kubenswrapper[4764]: I0202 09:47:03.863593 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-97dd2" Feb 02 09:47:03 crc kubenswrapper[4764]: I0202 09:47:03.983888 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/409c3e31-e6bf-409c-9bd8-7bcf433c0ce8-ssh-key-openstack-edpm-ipam\") pod \"409c3e31-e6bf-409c-9bd8-7bcf433c0ce8\" (UID: \"409c3e31-e6bf-409c-9bd8-7bcf433c0ce8\") " Feb 02 09:47:03 crc kubenswrapper[4764]: I0202 09:47:03.984065 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/409c3e31-e6bf-409c-9bd8-7bcf433c0ce8-ceph\") pod \"409c3e31-e6bf-409c-9bd8-7bcf433c0ce8\" (UID: \"409c3e31-e6bf-409c-9bd8-7bcf433c0ce8\") " Feb 02 09:47:03 crc kubenswrapper[4764]: I0202 09:47:03.984171 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bl4t6\" (UniqueName: \"kubernetes.io/projected/409c3e31-e6bf-409c-9bd8-7bcf433c0ce8-kube-api-access-bl4t6\") pod \"409c3e31-e6bf-409c-9bd8-7bcf433c0ce8\" (UID: \"409c3e31-e6bf-409c-9bd8-7bcf433c0ce8\") " Feb 02 09:47:03 crc kubenswrapper[4764]: I0202 09:47:03.984234 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/409c3e31-e6bf-409c-9bd8-7bcf433c0ce8-inventory-0\") pod \"409c3e31-e6bf-409c-9bd8-7bcf433c0ce8\" (UID: \"409c3e31-e6bf-409c-9bd8-7bcf433c0ce8\") " Feb 02 09:47:03 crc kubenswrapper[4764]: I0202 09:47:03.992042 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/409c3e31-e6bf-409c-9bd8-7bcf433c0ce8-ceph" (OuterVolumeSpecName: "ceph") pod "409c3e31-e6bf-409c-9bd8-7bcf433c0ce8" (UID: "409c3e31-e6bf-409c-9bd8-7bcf433c0ce8"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:47:03 crc kubenswrapper[4764]: I0202 09:47:03.992058 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/409c3e31-e6bf-409c-9bd8-7bcf433c0ce8-kube-api-access-bl4t6" (OuterVolumeSpecName: "kube-api-access-bl4t6") pod "409c3e31-e6bf-409c-9bd8-7bcf433c0ce8" (UID: "409c3e31-e6bf-409c-9bd8-7bcf433c0ce8"). InnerVolumeSpecName "kube-api-access-bl4t6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:47:04 crc kubenswrapper[4764]: I0202 09:47:04.034473 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/409c3e31-e6bf-409c-9bd8-7bcf433c0ce8-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "409c3e31-e6bf-409c-9bd8-7bcf433c0ce8" (UID: "409c3e31-e6bf-409c-9bd8-7bcf433c0ce8"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:47:04 crc kubenswrapper[4764]: I0202 09:47:04.048907 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/409c3e31-e6bf-409c-9bd8-7bcf433c0ce8-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "409c3e31-e6bf-409c-9bd8-7bcf433c0ce8" (UID: "409c3e31-e6bf-409c-9bd8-7bcf433c0ce8"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:47:04 crc kubenswrapper[4764]: I0202 09:47:04.089482 4764 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/409c3e31-e6bf-409c-9bd8-7bcf433c0ce8-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 02 09:47:04 crc kubenswrapper[4764]: I0202 09:47:04.089881 4764 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/409c3e31-e6bf-409c-9bd8-7bcf433c0ce8-ceph\") on node \"crc\" DevicePath \"\"" Feb 02 09:47:04 crc kubenswrapper[4764]: I0202 09:47:04.089894 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bl4t6\" (UniqueName: \"kubernetes.io/projected/409c3e31-e6bf-409c-9bd8-7bcf433c0ce8-kube-api-access-bl4t6\") on node \"crc\" DevicePath \"\"" Feb 02 09:47:04 crc kubenswrapper[4764]: I0202 09:47:04.089904 4764 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/409c3e31-e6bf-409c-9bd8-7bcf433c0ce8-inventory-0\") on node \"crc\" DevicePath \"\"" Feb 02 09:47:04 crc kubenswrapper[4764]: I0202 09:47:04.395537 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-97dd2" event={"ID":"409c3e31-e6bf-409c-9bd8-7bcf433c0ce8","Type":"ContainerDied","Data":"9a4e5a941b8d35e237c2dcee2b5ccee3397b86304f2972283cdaacb7736493e0"} Feb 02 09:47:04 crc kubenswrapper[4764]: I0202 09:47:04.395816 4764 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9a4e5a941b8d35e237c2dcee2b5ccee3397b86304f2972283cdaacb7736493e0" Feb 02 09:47:04 crc kubenswrapper[4764]: I0202 09:47:04.395908 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-97dd2" Feb 02 09:47:04 crc kubenswrapper[4764]: I0202 09:47:04.530392 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-t5j4g"] Feb 02 09:47:04 crc kubenswrapper[4764]: E0202 09:47:04.531255 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="409c3e31-e6bf-409c-9bd8-7bcf433c0ce8" containerName="ssh-known-hosts-edpm-deployment" Feb 02 09:47:04 crc kubenswrapper[4764]: I0202 09:47:04.531292 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="409c3e31-e6bf-409c-9bd8-7bcf433c0ce8" containerName="ssh-known-hosts-edpm-deployment" Feb 02 09:47:04 crc kubenswrapper[4764]: I0202 09:47:04.531759 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="409c3e31-e6bf-409c-9bd8-7bcf433c0ce8" containerName="ssh-known-hosts-edpm-deployment" Feb 02 09:47:04 crc kubenswrapper[4764]: I0202 09:47:04.533783 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-t5j4g" Feb 02 09:47:04 crc kubenswrapper[4764]: I0202 09:47:04.536414 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Feb 02 09:47:04 crc kubenswrapper[4764]: I0202 09:47:04.539221 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 02 09:47:04 crc kubenswrapper[4764]: I0202 09:47:04.545548 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 02 09:47:04 crc kubenswrapper[4764]: I0202 09:47:04.545561 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 02 09:47:04 crc kubenswrapper[4764]: I0202 09:47:04.545906 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-xksq5" Feb 02 09:47:04 crc kubenswrapper[4764]: I0202 09:47:04.555559 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-t5j4g"] Feb 02 09:47:04 crc kubenswrapper[4764]: I0202 09:47:04.603137 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pb6xx\" (UniqueName: \"kubernetes.io/projected/8649507c-417e-40fc-9a67-b6d6831f3ab9-kube-api-access-pb6xx\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-t5j4g\" (UID: \"8649507c-417e-40fc-9a67-b6d6831f3ab9\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-t5j4g" Feb 02 09:47:04 crc kubenswrapper[4764]: I0202 09:47:04.603249 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8649507c-417e-40fc-9a67-b6d6831f3ab9-ceph\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-t5j4g\" (UID: \"8649507c-417e-40fc-9a67-b6d6831f3ab9\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-t5j4g" Feb 02 09:47:04 crc kubenswrapper[4764]: I0202 09:47:04.603279 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8649507c-417e-40fc-9a67-b6d6831f3ab9-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-t5j4g\" (UID: \"8649507c-417e-40fc-9a67-b6d6831f3ab9\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-t5j4g" Feb 02 09:47:04 crc kubenswrapper[4764]: I0202 09:47:04.603341 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8649507c-417e-40fc-9a67-b6d6831f3ab9-ssh-key-openstack-edpm-ipam\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-t5j4g\" (UID: \"8649507c-417e-40fc-9a67-b6d6831f3ab9\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-t5j4g" Feb 02 09:47:04 crc kubenswrapper[4764]: I0202 09:47:04.705495 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pb6xx\" (UniqueName: \"kubernetes.io/projected/8649507c-417e-40fc-9a67-b6d6831f3ab9-kube-api-access-pb6xx\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-t5j4g\" (UID: \"8649507c-417e-40fc-9a67-b6d6831f3ab9\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-t5j4g" Feb 02 09:47:04 crc kubenswrapper[4764]: I0202 09:47:04.705578 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8649507c-417e-40fc-9a67-b6d6831f3ab9-ceph\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-t5j4g\" (UID: \"8649507c-417e-40fc-9a67-b6d6831f3ab9\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-t5j4g" Feb 02 09:47:04 crc kubenswrapper[4764]: I0202 09:47:04.705604 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8649507c-417e-40fc-9a67-b6d6831f3ab9-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-t5j4g\" (UID: \"8649507c-417e-40fc-9a67-b6d6831f3ab9\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-t5j4g" Feb 02 09:47:04 crc kubenswrapper[4764]: I0202 09:47:04.705637 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8649507c-417e-40fc-9a67-b6d6831f3ab9-ssh-key-openstack-edpm-ipam\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-t5j4g\" (UID: \"8649507c-417e-40fc-9a67-b6d6831f3ab9\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-t5j4g" Feb 02 09:47:04 crc kubenswrapper[4764]: I0202 09:47:04.712495 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8649507c-417e-40fc-9a67-b6d6831f3ab9-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-t5j4g\" (UID: \"8649507c-417e-40fc-9a67-b6d6831f3ab9\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-t5j4g" Feb 02 09:47:04 crc kubenswrapper[4764]: I0202 09:47:04.712546 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8649507c-417e-40fc-9a67-b6d6831f3ab9-ssh-key-openstack-edpm-ipam\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-t5j4g\" (UID: \"8649507c-417e-40fc-9a67-b6d6831f3ab9\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-t5j4g" Feb 02 09:47:04 crc kubenswrapper[4764]: I0202 09:47:04.713910 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8649507c-417e-40fc-9a67-b6d6831f3ab9-ceph\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-t5j4g\" (UID: \"8649507c-417e-40fc-9a67-b6d6831f3ab9\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-t5j4g" Feb 02 09:47:04 crc kubenswrapper[4764]: I0202 09:47:04.729297 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pb6xx\" (UniqueName: \"kubernetes.io/projected/8649507c-417e-40fc-9a67-b6d6831f3ab9-kube-api-access-pb6xx\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-t5j4g\" (UID: \"8649507c-417e-40fc-9a67-b6d6831f3ab9\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-t5j4g" Feb 02 09:47:04 crc kubenswrapper[4764]: I0202 09:47:04.859656 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-t5j4g" Feb 02 09:47:05 crc kubenswrapper[4764]: I0202 09:47:05.444535 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-t5j4g"] Feb 02 09:47:06 crc kubenswrapper[4764]: I0202 09:47:06.414989 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-t5j4g" event={"ID":"8649507c-417e-40fc-9a67-b6d6831f3ab9","Type":"ContainerStarted","Data":"479c21af7f8bba97c873a61326a4037366b70d4f9a8316e9ccbfaf3717dc0265"} Feb 02 09:47:06 crc kubenswrapper[4764]: I0202 09:47:06.415467 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-t5j4g" event={"ID":"8649507c-417e-40fc-9a67-b6d6831f3ab9","Type":"ContainerStarted","Data":"a36ad3b5d0de02865c615715a8273ceaba329d24aa54afcc051a010d443b132a"} Feb 02 09:47:06 crc kubenswrapper[4764]: I0202 09:47:06.443196 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-t5j4g" podStartSLOduration=2.009464692 podStartE2EDuration="2.443177792s" podCreationTimestamp="2026-02-02 09:47:04 +0000 UTC" firstStartedPulling="2026-02-02 09:47:05.456075616 +0000 UTC m=+2388.389799724" lastFinishedPulling="2026-02-02 09:47:05.889788726 +0000 UTC m=+2388.823512824" observedRunningTime="2026-02-02 09:47:06.437230253 +0000 UTC m=+2389.370954341" watchObservedRunningTime="2026-02-02 09:47:06.443177792 +0000 UTC m=+2389.376901890" Feb 02 09:47:08 crc kubenswrapper[4764]: I0202 09:47:08.826133 4764 scope.go:117] "RemoveContainer" containerID="be8fa5800ec42d1c141765dcfc6111f3ef96b05b6e8d5a9feaf2baa219897bcf" Feb 02 09:47:08 crc kubenswrapper[4764]: E0202 09:47:08.826800 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 09:47:15 crc kubenswrapper[4764]: I0202 09:47:15.515261 4764 generic.go:334] "Generic (PLEG): container finished" podID="8649507c-417e-40fc-9a67-b6d6831f3ab9" containerID="479c21af7f8bba97c873a61326a4037366b70d4f9a8316e9ccbfaf3717dc0265" exitCode=0 Feb 02 09:47:15 crc kubenswrapper[4764]: I0202 09:47:15.515326 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-t5j4g" event={"ID":"8649507c-417e-40fc-9a67-b6d6831f3ab9","Type":"ContainerDied","Data":"479c21af7f8bba97c873a61326a4037366b70d4f9a8316e9ccbfaf3717dc0265"} Feb 02 09:47:16 crc kubenswrapper[4764]: I0202 09:47:16.992846 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-t5j4g" Feb 02 09:47:17 crc kubenswrapper[4764]: I0202 09:47:17.024895 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pb6xx\" (UniqueName: \"kubernetes.io/projected/8649507c-417e-40fc-9a67-b6d6831f3ab9-kube-api-access-pb6xx\") pod \"8649507c-417e-40fc-9a67-b6d6831f3ab9\" (UID: \"8649507c-417e-40fc-9a67-b6d6831f3ab9\") " Feb 02 09:47:17 crc kubenswrapper[4764]: I0202 09:47:17.025007 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8649507c-417e-40fc-9a67-b6d6831f3ab9-inventory\") pod \"8649507c-417e-40fc-9a67-b6d6831f3ab9\" (UID: \"8649507c-417e-40fc-9a67-b6d6831f3ab9\") " Feb 02 09:47:17 crc kubenswrapper[4764]: I0202 09:47:17.025127 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8649507c-417e-40fc-9a67-b6d6831f3ab9-ceph\") pod \"8649507c-417e-40fc-9a67-b6d6831f3ab9\" (UID: \"8649507c-417e-40fc-9a67-b6d6831f3ab9\") " Feb 02 09:47:17 crc kubenswrapper[4764]: I0202 09:47:17.025224 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8649507c-417e-40fc-9a67-b6d6831f3ab9-ssh-key-openstack-edpm-ipam\") pod \"8649507c-417e-40fc-9a67-b6d6831f3ab9\" (UID: \"8649507c-417e-40fc-9a67-b6d6831f3ab9\") " Feb 02 09:47:17 crc kubenswrapper[4764]: I0202 09:47:17.031390 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8649507c-417e-40fc-9a67-b6d6831f3ab9-ceph" (OuterVolumeSpecName: "ceph") pod "8649507c-417e-40fc-9a67-b6d6831f3ab9" (UID: "8649507c-417e-40fc-9a67-b6d6831f3ab9"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:47:17 crc kubenswrapper[4764]: I0202 09:47:17.043629 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8649507c-417e-40fc-9a67-b6d6831f3ab9-kube-api-access-pb6xx" (OuterVolumeSpecName: "kube-api-access-pb6xx") pod "8649507c-417e-40fc-9a67-b6d6831f3ab9" (UID: "8649507c-417e-40fc-9a67-b6d6831f3ab9"). InnerVolumeSpecName "kube-api-access-pb6xx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:47:17 crc kubenswrapper[4764]: I0202 09:47:17.057746 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8649507c-417e-40fc-9a67-b6d6831f3ab9-inventory" (OuterVolumeSpecName: "inventory") pod "8649507c-417e-40fc-9a67-b6d6831f3ab9" (UID: "8649507c-417e-40fc-9a67-b6d6831f3ab9"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:47:17 crc kubenswrapper[4764]: I0202 09:47:17.061533 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8649507c-417e-40fc-9a67-b6d6831f3ab9-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "8649507c-417e-40fc-9a67-b6d6831f3ab9" (UID: "8649507c-417e-40fc-9a67-b6d6831f3ab9"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:47:17 crc kubenswrapper[4764]: I0202 09:47:17.127034 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pb6xx\" (UniqueName: \"kubernetes.io/projected/8649507c-417e-40fc-9a67-b6d6831f3ab9-kube-api-access-pb6xx\") on node \"crc\" DevicePath \"\"" Feb 02 09:47:17 crc kubenswrapper[4764]: I0202 09:47:17.127066 4764 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8649507c-417e-40fc-9a67-b6d6831f3ab9-inventory\") on node \"crc\" DevicePath \"\"" Feb 02 09:47:17 crc kubenswrapper[4764]: I0202 09:47:17.127075 4764 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8649507c-417e-40fc-9a67-b6d6831f3ab9-ceph\") on node \"crc\" DevicePath \"\"" Feb 02 09:47:17 crc kubenswrapper[4764]: I0202 09:47:17.127084 4764 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8649507c-417e-40fc-9a67-b6d6831f3ab9-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 02 09:47:17 crc kubenswrapper[4764]: I0202 09:47:17.539971 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-t5j4g" event={"ID":"8649507c-417e-40fc-9a67-b6d6831f3ab9","Type":"ContainerDied","Data":"a36ad3b5d0de02865c615715a8273ceaba329d24aa54afcc051a010d443b132a"} Feb 02 09:47:17 crc kubenswrapper[4764]: I0202 09:47:17.540415 4764 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a36ad3b5d0de02865c615715a8273ceaba329d24aa54afcc051a010d443b132a" Feb 02 09:47:17 crc kubenswrapper[4764]: I0202 09:47:17.540078 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-t5j4g" Feb 02 09:47:17 crc kubenswrapper[4764]: I0202 09:47:17.712729 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-hrmzd"] Feb 02 09:47:17 crc kubenswrapper[4764]: E0202 09:47:17.713824 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8649507c-417e-40fc-9a67-b6d6831f3ab9" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Feb 02 09:47:17 crc kubenswrapper[4764]: I0202 09:47:17.713853 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="8649507c-417e-40fc-9a67-b6d6831f3ab9" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Feb 02 09:47:17 crc kubenswrapper[4764]: I0202 09:47:17.714182 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="8649507c-417e-40fc-9a67-b6d6831f3ab9" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Feb 02 09:47:17 crc kubenswrapper[4764]: I0202 09:47:17.715094 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-hrmzd" Feb 02 09:47:17 crc kubenswrapper[4764]: I0202 09:47:17.718567 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Feb 02 09:47:17 crc kubenswrapper[4764]: I0202 09:47:17.719038 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 02 09:47:17 crc kubenswrapper[4764]: I0202 09:47:17.719359 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-xksq5" Feb 02 09:47:17 crc kubenswrapper[4764]: I0202 09:47:17.719594 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 02 09:47:17 crc kubenswrapper[4764]: I0202 09:47:17.719594 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 02 09:47:17 crc kubenswrapper[4764]: I0202 09:47:17.726574 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-hrmzd"] Feb 02 09:47:17 crc kubenswrapper[4764]: I0202 09:47:17.758343 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a79016d5-22d2-4894-829c-7570224b6f69-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-hrmzd\" (UID: \"a79016d5-22d2-4894-829c-7570224b6f69\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-hrmzd" Feb 02 09:47:17 crc kubenswrapper[4764]: I0202 09:47:17.758459 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/a79016d5-22d2-4894-829c-7570224b6f69-ceph\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-hrmzd\" (UID: \"a79016d5-22d2-4894-829c-7570224b6f69\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-hrmzd" Feb 02 09:47:17 crc kubenswrapper[4764]: I0202 09:47:17.758559 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a79016d5-22d2-4894-829c-7570224b6f69-ssh-key-openstack-edpm-ipam\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-hrmzd\" (UID: \"a79016d5-22d2-4894-829c-7570224b6f69\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-hrmzd" Feb 02 09:47:17 crc kubenswrapper[4764]: I0202 09:47:17.758598 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qcm2c\" (UniqueName: \"kubernetes.io/projected/a79016d5-22d2-4894-829c-7570224b6f69-kube-api-access-qcm2c\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-hrmzd\" (UID: \"a79016d5-22d2-4894-829c-7570224b6f69\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-hrmzd" Feb 02 09:47:17 crc kubenswrapper[4764]: I0202 09:47:17.861263 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a79016d5-22d2-4894-829c-7570224b6f69-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-hrmzd\" (UID: \"a79016d5-22d2-4894-829c-7570224b6f69\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-hrmzd" Feb 02 09:47:17 crc kubenswrapper[4764]: I0202 09:47:17.861371 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/a79016d5-22d2-4894-829c-7570224b6f69-ceph\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-hrmzd\" (UID: \"a79016d5-22d2-4894-829c-7570224b6f69\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-hrmzd" Feb 02 09:47:17 crc kubenswrapper[4764]: I0202 09:47:17.861424 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a79016d5-22d2-4894-829c-7570224b6f69-ssh-key-openstack-edpm-ipam\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-hrmzd\" (UID: \"a79016d5-22d2-4894-829c-7570224b6f69\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-hrmzd" Feb 02 09:47:17 crc kubenswrapper[4764]: I0202 09:47:17.861461 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qcm2c\" (UniqueName: \"kubernetes.io/projected/a79016d5-22d2-4894-829c-7570224b6f69-kube-api-access-qcm2c\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-hrmzd\" (UID: \"a79016d5-22d2-4894-829c-7570224b6f69\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-hrmzd" Feb 02 09:47:17 crc kubenswrapper[4764]: I0202 09:47:17.863493 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Feb 02 09:47:17 crc kubenswrapper[4764]: I0202 09:47:17.863678 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 02 09:47:17 crc kubenswrapper[4764]: I0202 09:47:17.864578 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 02 09:47:17 crc kubenswrapper[4764]: I0202 09:47:17.878735 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a79016d5-22d2-4894-829c-7570224b6f69-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-hrmzd\" (UID: \"a79016d5-22d2-4894-829c-7570224b6f69\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-hrmzd" Feb 02 09:47:17 crc kubenswrapper[4764]: I0202 09:47:17.878760 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/a79016d5-22d2-4894-829c-7570224b6f69-ceph\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-hrmzd\" (UID: \"a79016d5-22d2-4894-829c-7570224b6f69\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-hrmzd" Feb 02 09:47:17 crc kubenswrapper[4764]: I0202 09:47:17.878802 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qcm2c\" (UniqueName: \"kubernetes.io/projected/a79016d5-22d2-4894-829c-7570224b6f69-kube-api-access-qcm2c\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-hrmzd\" (UID: \"a79016d5-22d2-4894-829c-7570224b6f69\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-hrmzd" Feb 02 09:47:17 crc kubenswrapper[4764]: I0202 09:47:17.882023 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a79016d5-22d2-4894-829c-7570224b6f69-ssh-key-openstack-edpm-ipam\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-hrmzd\" (UID: \"a79016d5-22d2-4894-829c-7570224b6f69\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-hrmzd" Feb 02 09:47:18 crc kubenswrapper[4764]: I0202 09:47:18.062476 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-xksq5" Feb 02 09:47:18 crc kubenswrapper[4764]: I0202 09:47:18.069241 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-hrmzd" Feb 02 09:47:18 crc kubenswrapper[4764]: I0202 09:47:18.444595 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-hrmzd"] Feb 02 09:47:18 crc kubenswrapper[4764]: I0202 09:47:18.548714 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-hrmzd" event={"ID":"a79016d5-22d2-4894-829c-7570224b6f69","Type":"ContainerStarted","Data":"672d2d45fa16e5ccb66284d69b67f0db84752e651108622675d7ae51d9507131"} Feb 02 09:47:18 crc kubenswrapper[4764]: I0202 09:47:18.883771 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 02 09:47:19 crc kubenswrapper[4764]: I0202 09:47:19.576806 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-hrmzd" event={"ID":"a79016d5-22d2-4894-829c-7570224b6f69","Type":"ContainerStarted","Data":"ac836ddc72b29c9da59bff922421472d3009fbcebe65cc573b1461110c70aaf6"} Feb 02 09:47:19 crc kubenswrapper[4764]: I0202 09:47:19.619412 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-hrmzd" podStartSLOduration=2.196433173 podStartE2EDuration="2.619370376s" podCreationTimestamp="2026-02-02 09:47:17 +0000 UTC" firstStartedPulling="2026-02-02 09:47:18.45622104 +0000 UTC m=+2401.389945128" lastFinishedPulling="2026-02-02 09:47:18.879158203 +0000 UTC m=+2401.812882331" observedRunningTime="2026-02-02 09:47:19.59966059 +0000 UTC m=+2402.533384678" watchObservedRunningTime="2026-02-02 09:47:19.619370376 +0000 UTC m=+2402.553094494" Feb 02 09:47:22 crc kubenswrapper[4764]: I0202 09:47:22.825415 4764 scope.go:117] "RemoveContainer" containerID="be8fa5800ec42d1c141765dcfc6111f3ef96b05b6e8d5a9feaf2baa219897bcf" Feb 02 09:47:22 crc kubenswrapper[4764]: E0202 09:47:22.826147 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 09:47:29 crc kubenswrapper[4764]: I0202 09:47:29.699964 4764 generic.go:334] "Generic (PLEG): container finished" podID="a79016d5-22d2-4894-829c-7570224b6f69" containerID="ac836ddc72b29c9da59bff922421472d3009fbcebe65cc573b1461110c70aaf6" exitCode=0 Feb 02 09:47:29 crc kubenswrapper[4764]: I0202 09:47:29.700015 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-hrmzd" event={"ID":"a79016d5-22d2-4894-829c-7570224b6f69","Type":"ContainerDied","Data":"ac836ddc72b29c9da59bff922421472d3009fbcebe65cc573b1461110c70aaf6"} Feb 02 09:47:31 crc kubenswrapper[4764]: I0202 09:47:31.135673 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-hrmzd" Feb 02 09:47:31 crc kubenswrapper[4764]: I0202 09:47:31.265284 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a79016d5-22d2-4894-829c-7570224b6f69-ssh-key-openstack-edpm-ipam\") pod \"a79016d5-22d2-4894-829c-7570224b6f69\" (UID: \"a79016d5-22d2-4894-829c-7570224b6f69\") " Feb 02 09:47:31 crc kubenswrapper[4764]: I0202 09:47:31.265721 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/a79016d5-22d2-4894-829c-7570224b6f69-ceph\") pod \"a79016d5-22d2-4894-829c-7570224b6f69\" (UID: \"a79016d5-22d2-4894-829c-7570224b6f69\") " Feb 02 09:47:31 crc kubenswrapper[4764]: I0202 09:47:31.265807 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a79016d5-22d2-4894-829c-7570224b6f69-inventory\") pod \"a79016d5-22d2-4894-829c-7570224b6f69\" (UID: \"a79016d5-22d2-4894-829c-7570224b6f69\") " Feb 02 09:47:31 crc kubenswrapper[4764]: I0202 09:47:31.265880 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qcm2c\" (UniqueName: \"kubernetes.io/projected/a79016d5-22d2-4894-829c-7570224b6f69-kube-api-access-qcm2c\") pod \"a79016d5-22d2-4894-829c-7570224b6f69\" (UID: \"a79016d5-22d2-4894-829c-7570224b6f69\") " Feb 02 09:47:31 crc kubenswrapper[4764]: I0202 09:47:31.271730 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a79016d5-22d2-4894-829c-7570224b6f69-ceph" (OuterVolumeSpecName: "ceph") pod "a79016d5-22d2-4894-829c-7570224b6f69" (UID: "a79016d5-22d2-4894-829c-7570224b6f69"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:47:31 crc kubenswrapper[4764]: I0202 09:47:31.277165 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a79016d5-22d2-4894-829c-7570224b6f69-kube-api-access-qcm2c" (OuterVolumeSpecName: "kube-api-access-qcm2c") pod "a79016d5-22d2-4894-829c-7570224b6f69" (UID: "a79016d5-22d2-4894-829c-7570224b6f69"). InnerVolumeSpecName "kube-api-access-qcm2c". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:47:31 crc kubenswrapper[4764]: I0202 09:47:31.298038 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a79016d5-22d2-4894-829c-7570224b6f69-inventory" (OuterVolumeSpecName: "inventory") pod "a79016d5-22d2-4894-829c-7570224b6f69" (UID: "a79016d5-22d2-4894-829c-7570224b6f69"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:47:31 crc kubenswrapper[4764]: I0202 09:47:31.324318 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a79016d5-22d2-4894-829c-7570224b6f69-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "a79016d5-22d2-4894-829c-7570224b6f69" (UID: "a79016d5-22d2-4894-829c-7570224b6f69"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:47:31 crc kubenswrapper[4764]: I0202 09:47:31.368229 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qcm2c\" (UniqueName: \"kubernetes.io/projected/a79016d5-22d2-4894-829c-7570224b6f69-kube-api-access-qcm2c\") on node \"crc\" DevicePath \"\"" Feb 02 09:47:31 crc kubenswrapper[4764]: I0202 09:47:31.368268 4764 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a79016d5-22d2-4894-829c-7570224b6f69-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 02 09:47:31 crc kubenswrapper[4764]: I0202 09:47:31.368284 4764 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/a79016d5-22d2-4894-829c-7570224b6f69-ceph\") on node \"crc\" DevicePath \"\"" Feb 02 09:47:31 crc kubenswrapper[4764]: I0202 09:47:31.368296 4764 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a79016d5-22d2-4894-829c-7570224b6f69-inventory\") on node \"crc\" DevicePath \"\"" Feb 02 09:47:31 crc kubenswrapper[4764]: I0202 09:47:31.727870 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-hrmzd" event={"ID":"a79016d5-22d2-4894-829c-7570224b6f69","Type":"ContainerDied","Data":"672d2d45fa16e5ccb66284d69b67f0db84752e651108622675d7ae51d9507131"} Feb 02 09:47:31 crc kubenswrapper[4764]: I0202 09:47:31.727929 4764 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="672d2d45fa16e5ccb66284d69b67f0db84752e651108622675d7ae51d9507131" Feb 02 09:47:31 crc kubenswrapper[4764]: I0202 09:47:31.727965 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-hrmzd" Feb 02 09:47:31 crc kubenswrapper[4764]: I0202 09:47:31.997072 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9q5kq"] Feb 02 09:47:31 crc kubenswrapper[4764]: E0202 09:47:31.997584 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a79016d5-22d2-4894-829c-7570224b6f69" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Feb 02 09:47:31 crc kubenswrapper[4764]: I0202 09:47:31.997613 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="a79016d5-22d2-4894-829c-7570224b6f69" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Feb 02 09:47:31 crc kubenswrapper[4764]: I0202 09:47:31.997867 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="a79016d5-22d2-4894-829c-7570224b6f69" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Feb 02 09:47:31 crc kubenswrapper[4764]: I0202 09:47:31.998583 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9q5kq" Feb 02 09:47:32 crc kubenswrapper[4764]: I0202 09:47:32.002411 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 02 09:47:32 crc kubenswrapper[4764]: I0202 09:47:32.005966 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 02 09:47:32 crc kubenswrapper[4764]: I0202 09:47:32.008679 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 02 09:47:32 crc kubenswrapper[4764]: I0202 09:47:32.008702 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-xksq5" Feb 02 09:47:32 crc kubenswrapper[4764]: I0202 09:47:32.008744 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Feb 02 09:47:32 crc kubenswrapper[4764]: I0202 09:47:32.008682 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Feb 02 09:47:32 crc kubenswrapper[4764]: I0202 09:47:32.009102 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Feb 02 09:47:32 crc kubenswrapper[4764]: I0202 09:47:32.009488 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Feb 02 09:47:32 crc kubenswrapper[4764]: I0202 09:47:32.023280 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9q5kq"] Feb 02 09:47:32 crc kubenswrapper[4764]: I0202 09:47:32.087115 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f2rf7\" (UniqueName: \"kubernetes.io/projected/7d388199-9f34-415c-b83c-d5f54d17d063-kube-api-access-f2rf7\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9q5kq\" (UID: \"7d388199-9f34-415c-b83c-d5f54d17d063\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9q5kq" Feb 02 09:47:32 crc kubenswrapper[4764]: I0202 09:47:32.087184 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d388199-9f34-415c-b83c-d5f54d17d063-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9q5kq\" (UID: \"7d388199-9f34-415c-b83c-d5f54d17d063\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9q5kq" Feb 02 09:47:32 crc kubenswrapper[4764]: I0202 09:47:32.087222 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/7d388199-9f34-415c-b83c-d5f54d17d063-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9q5kq\" (UID: \"7d388199-9f34-415c-b83c-d5f54d17d063\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9q5kq" Feb 02 09:47:32 crc kubenswrapper[4764]: I0202 09:47:32.087295 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d388199-9f34-415c-b83c-d5f54d17d063-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9q5kq\" (UID: \"7d388199-9f34-415c-b83c-d5f54d17d063\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9q5kq" Feb 02 09:47:32 crc kubenswrapper[4764]: I0202 09:47:32.087328 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d388199-9f34-415c-b83c-d5f54d17d063-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9q5kq\" (UID: \"7d388199-9f34-415c-b83c-d5f54d17d063\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9q5kq" Feb 02 09:47:32 crc kubenswrapper[4764]: I0202 09:47:32.087465 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7d388199-9f34-415c-b83c-d5f54d17d063-ceph\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9q5kq\" (UID: \"7d388199-9f34-415c-b83c-d5f54d17d063\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9q5kq" Feb 02 09:47:32 crc kubenswrapper[4764]: I0202 09:47:32.087500 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d388199-9f34-415c-b83c-d5f54d17d063-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9q5kq\" (UID: \"7d388199-9f34-415c-b83c-d5f54d17d063\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9q5kq" Feb 02 09:47:32 crc kubenswrapper[4764]: I0202 09:47:32.087530 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7d388199-9f34-415c-b83c-d5f54d17d063-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9q5kq\" (UID: \"7d388199-9f34-415c-b83c-d5f54d17d063\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9q5kq" Feb 02 09:47:32 crc kubenswrapper[4764]: I0202 09:47:32.087613 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d388199-9f34-415c-b83c-d5f54d17d063-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9q5kq\" (UID: \"7d388199-9f34-415c-b83c-d5f54d17d063\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9q5kq" Feb 02 09:47:32 crc kubenswrapper[4764]: I0202 09:47:32.087655 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/7d388199-9f34-415c-b83c-d5f54d17d063-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9q5kq\" (UID: \"7d388199-9f34-415c-b83c-d5f54d17d063\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9q5kq" Feb 02 09:47:32 crc kubenswrapper[4764]: I0202 09:47:32.087741 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/7d388199-9f34-415c-b83c-d5f54d17d063-ssh-key-openstack-edpm-ipam\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9q5kq\" (UID: \"7d388199-9f34-415c-b83c-d5f54d17d063\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9q5kq" Feb 02 09:47:32 crc kubenswrapper[4764]: I0202 09:47:32.087800 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/7d388199-9f34-415c-b83c-d5f54d17d063-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9q5kq\" (UID: \"7d388199-9f34-415c-b83c-d5f54d17d063\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9q5kq" Feb 02 09:47:32 crc kubenswrapper[4764]: I0202 09:47:32.087828 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d388199-9f34-415c-b83c-d5f54d17d063-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9q5kq\" (UID: \"7d388199-9f34-415c-b83c-d5f54d17d063\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9q5kq" Feb 02 09:47:32 crc kubenswrapper[4764]: I0202 09:47:32.189535 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/7d388199-9f34-415c-b83c-d5f54d17d063-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9q5kq\" (UID: \"7d388199-9f34-415c-b83c-d5f54d17d063\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9q5kq" Feb 02 09:47:32 crc kubenswrapper[4764]: I0202 09:47:32.189634 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/7d388199-9f34-415c-b83c-d5f54d17d063-ssh-key-openstack-edpm-ipam\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9q5kq\" (UID: \"7d388199-9f34-415c-b83c-d5f54d17d063\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9q5kq" Feb 02 09:47:32 crc kubenswrapper[4764]: I0202 09:47:32.189696 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/7d388199-9f34-415c-b83c-d5f54d17d063-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9q5kq\" (UID: \"7d388199-9f34-415c-b83c-d5f54d17d063\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9q5kq" Feb 02 09:47:32 crc kubenswrapper[4764]: I0202 09:47:32.189739 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d388199-9f34-415c-b83c-d5f54d17d063-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9q5kq\" (UID: \"7d388199-9f34-415c-b83c-d5f54d17d063\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9q5kq" Feb 02 09:47:32 crc kubenswrapper[4764]: I0202 09:47:32.189867 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f2rf7\" (UniqueName: \"kubernetes.io/projected/7d388199-9f34-415c-b83c-d5f54d17d063-kube-api-access-f2rf7\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9q5kq\" (UID: \"7d388199-9f34-415c-b83c-d5f54d17d063\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9q5kq" Feb 02 09:47:32 crc kubenswrapper[4764]: I0202 09:47:32.189943 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d388199-9f34-415c-b83c-d5f54d17d063-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9q5kq\" (UID: \"7d388199-9f34-415c-b83c-d5f54d17d063\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9q5kq" Feb 02 09:47:32 crc kubenswrapper[4764]: I0202 09:47:32.190506 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/7d388199-9f34-415c-b83c-d5f54d17d063-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9q5kq\" (UID: \"7d388199-9f34-415c-b83c-d5f54d17d063\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9q5kq" Feb 02 09:47:32 crc kubenswrapper[4764]: I0202 09:47:32.191346 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d388199-9f34-415c-b83c-d5f54d17d063-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9q5kq\" (UID: \"7d388199-9f34-415c-b83c-d5f54d17d063\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9q5kq" Feb 02 09:47:32 crc kubenswrapper[4764]: I0202 09:47:32.191495 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d388199-9f34-415c-b83c-d5f54d17d063-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9q5kq\" (UID: \"7d388199-9f34-415c-b83c-d5f54d17d063\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9q5kq" Feb 02 09:47:32 crc kubenswrapper[4764]: I0202 09:47:32.191584 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7d388199-9f34-415c-b83c-d5f54d17d063-ceph\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9q5kq\" (UID: \"7d388199-9f34-415c-b83c-d5f54d17d063\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9q5kq" Feb 02 09:47:32 crc kubenswrapper[4764]: I0202 09:47:32.191637 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d388199-9f34-415c-b83c-d5f54d17d063-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9q5kq\" (UID: \"7d388199-9f34-415c-b83c-d5f54d17d063\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9q5kq" Feb 02 09:47:32 crc kubenswrapper[4764]: I0202 09:47:32.191688 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7d388199-9f34-415c-b83c-d5f54d17d063-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9q5kq\" (UID: \"7d388199-9f34-415c-b83c-d5f54d17d063\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9q5kq" Feb 02 09:47:32 crc kubenswrapper[4764]: I0202 09:47:32.191749 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d388199-9f34-415c-b83c-d5f54d17d063-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9q5kq\" (UID: \"7d388199-9f34-415c-b83c-d5f54d17d063\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9q5kq" Feb 02 09:47:32 crc kubenswrapper[4764]: I0202 09:47:32.199039 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d388199-9f34-415c-b83c-d5f54d17d063-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9q5kq\" (UID: \"7d388199-9f34-415c-b83c-d5f54d17d063\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9q5kq" Feb 02 09:47:32 crc kubenswrapper[4764]: I0202 09:47:32.199321 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d388199-9f34-415c-b83c-d5f54d17d063-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9q5kq\" (UID: \"7d388199-9f34-415c-b83c-d5f54d17d063\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9q5kq" Feb 02 09:47:32 crc kubenswrapper[4764]: I0202 09:47:32.200013 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d388199-9f34-415c-b83c-d5f54d17d063-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9q5kq\" (UID: \"7d388199-9f34-415c-b83c-d5f54d17d063\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9q5kq" Feb 02 09:47:32 crc kubenswrapper[4764]: I0202 09:47:32.200661 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d388199-9f34-415c-b83c-d5f54d17d063-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9q5kq\" (UID: \"7d388199-9f34-415c-b83c-d5f54d17d063\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9q5kq" Feb 02 09:47:32 crc kubenswrapper[4764]: I0202 09:47:32.201734 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d388199-9f34-415c-b83c-d5f54d17d063-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9q5kq\" (UID: \"7d388199-9f34-415c-b83c-d5f54d17d063\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9q5kq" Feb 02 09:47:32 crc kubenswrapper[4764]: I0202 09:47:32.202550 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/7d388199-9f34-415c-b83c-d5f54d17d063-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9q5kq\" (UID: \"7d388199-9f34-415c-b83c-d5f54d17d063\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9q5kq" Feb 02 09:47:32 crc kubenswrapper[4764]: I0202 09:47:32.203271 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/7d388199-9f34-415c-b83c-d5f54d17d063-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9q5kq\" (UID: \"7d388199-9f34-415c-b83c-d5f54d17d063\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9q5kq" Feb 02 09:47:32 crc kubenswrapper[4764]: I0202 09:47:32.203600 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d388199-9f34-415c-b83c-d5f54d17d063-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9q5kq\" (UID: \"7d388199-9f34-415c-b83c-d5f54d17d063\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9q5kq" Feb 02 09:47:32 crc kubenswrapper[4764]: I0202 09:47:32.206628 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7d388199-9f34-415c-b83c-d5f54d17d063-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9q5kq\" (UID: \"7d388199-9f34-415c-b83c-d5f54d17d063\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9q5kq" Feb 02 09:47:32 crc kubenswrapper[4764]: I0202 09:47:32.206889 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/7d388199-9f34-415c-b83c-d5f54d17d063-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9q5kq\" (UID: \"7d388199-9f34-415c-b83c-d5f54d17d063\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9q5kq" Feb 02 09:47:32 crc kubenswrapper[4764]: I0202 09:47:32.209089 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/7d388199-9f34-415c-b83c-d5f54d17d063-ssh-key-openstack-edpm-ipam\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9q5kq\" (UID: \"7d388199-9f34-415c-b83c-d5f54d17d063\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9q5kq" Feb 02 09:47:32 crc kubenswrapper[4764]: I0202 09:47:32.212116 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f2rf7\" (UniqueName: \"kubernetes.io/projected/7d388199-9f34-415c-b83c-d5f54d17d063-kube-api-access-f2rf7\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9q5kq\" (UID: \"7d388199-9f34-415c-b83c-d5f54d17d063\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9q5kq" Feb 02 09:47:32 crc kubenswrapper[4764]: I0202 09:47:32.214537 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7d388199-9f34-415c-b83c-d5f54d17d063-ceph\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9q5kq\" (UID: \"7d388199-9f34-415c-b83c-d5f54d17d063\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9q5kq" Feb 02 09:47:32 crc kubenswrapper[4764]: I0202 09:47:32.319326 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9q5kq" Feb 02 09:47:32 crc kubenswrapper[4764]: I0202 09:47:32.818135 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9q5kq"] Feb 02 09:47:33 crc kubenswrapper[4764]: I0202 09:47:33.751385 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9q5kq" event={"ID":"7d388199-9f34-415c-b83c-d5f54d17d063","Type":"ContainerStarted","Data":"5d8deaa77f362a4223405716d8f784e4ad4dbcfd90f7f1ac8e51e0bbdecf3356"} Feb 02 09:47:33 crc kubenswrapper[4764]: I0202 09:47:33.751827 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9q5kq" event={"ID":"7d388199-9f34-415c-b83c-d5f54d17d063","Type":"ContainerStarted","Data":"64b363b92d0415f73b15905bc0f0c99d2b8fe2488c7d07349648b9ba8c6b00e1"} Feb 02 09:47:33 crc kubenswrapper[4764]: I0202 09:47:33.784309 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9q5kq" podStartSLOduration=2.382069951 podStartE2EDuration="2.78428149s" podCreationTimestamp="2026-02-02 09:47:31 +0000 UTC" firstStartedPulling="2026-02-02 09:47:32.828025058 +0000 UTC m=+2415.761749146" lastFinishedPulling="2026-02-02 09:47:33.230236597 +0000 UTC m=+2416.163960685" observedRunningTime="2026-02-02 09:47:33.780819928 +0000 UTC m=+2416.714544016" watchObservedRunningTime="2026-02-02 09:47:33.78428149 +0000 UTC m=+2416.718005618" Feb 02 09:47:34 crc kubenswrapper[4764]: I0202 09:47:34.826225 4764 scope.go:117] "RemoveContainer" containerID="be8fa5800ec42d1c141765dcfc6111f3ef96b05b6e8d5a9feaf2baa219897bcf" Feb 02 09:47:34 crc kubenswrapper[4764]: E0202 09:47:34.826851 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 09:47:46 crc kubenswrapper[4764]: I0202 09:47:46.826508 4764 scope.go:117] "RemoveContainer" containerID="be8fa5800ec42d1c141765dcfc6111f3ef96b05b6e8d5a9feaf2baa219897bcf" Feb 02 09:47:46 crc kubenswrapper[4764]: E0202 09:47:46.827479 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 09:47:59 crc kubenswrapper[4764]: I0202 09:47:59.826762 4764 scope.go:117] "RemoveContainer" containerID="be8fa5800ec42d1c141765dcfc6111f3ef96b05b6e8d5a9feaf2baa219897bcf" Feb 02 09:47:59 crc kubenswrapper[4764]: E0202 09:47:59.827853 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 09:48:06 crc kubenswrapper[4764]: I0202 09:48:06.061513 4764 generic.go:334] "Generic (PLEG): container finished" podID="7d388199-9f34-415c-b83c-d5f54d17d063" containerID="5d8deaa77f362a4223405716d8f784e4ad4dbcfd90f7f1ac8e51e0bbdecf3356" exitCode=0 Feb 02 09:48:06 crc kubenswrapper[4764]: I0202 09:48:06.061556 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9q5kq" event={"ID":"7d388199-9f34-415c-b83c-d5f54d17d063","Type":"ContainerDied","Data":"5d8deaa77f362a4223405716d8f784e4ad4dbcfd90f7f1ac8e51e0bbdecf3356"} Feb 02 09:48:07 crc kubenswrapper[4764]: I0202 09:48:07.572043 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9q5kq" Feb 02 09:48:07 crc kubenswrapper[4764]: I0202 09:48:07.673356 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7d388199-9f34-415c-b83c-d5f54d17d063-inventory\") pod \"7d388199-9f34-415c-b83c-d5f54d17d063\" (UID: \"7d388199-9f34-415c-b83c-d5f54d17d063\") " Feb 02 09:48:07 crc kubenswrapper[4764]: I0202 09:48:07.673448 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d388199-9f34-415c-b83c-d5f54d17d063-nova-combined-ca-bundle\") pod \"7d388199-9f34-415c-b83c-d5f54d17d063\" (UID: \"7d388199-9f34-415c-b83c-d5f54d17d063\") " Feb 02 09:48:07 crc kubenswrapper[4764]: I0202 09:48:07.673505 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d388199-9f34-415c-b83c-d5f54d17d063-ovn-combined-ca-bundle\") pod \"7d388199-9f34-415c-b83c-d5f54d17d063\" (UID: \"7d388199-9f34-415c-b83c-d5f54d17d063\") " Feb 02 09:48:07 crc kubenswrapper[4764]: I0202 09:48:07.673529 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d388199-9f34-415c-b83c-d5f54d17d063-libvirt-combined-ca-bundle\") pod \"7d388199-9f34-415c-b83c-d5f54d17d063\" (UID: \"7d388199-9f34-415c-b83c-d5f54d17d063\") " Feb 02 09:48:07 crc kubenswrapper[4764]: I0202 09:48:07.673616 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d388199-9f34-415c-b83c-d5f54d17d063-neutron-metadata-combined-ca-bundle\") pod \"7d388199-9f34-415c-b83c-d5f54d17d063\" (UID: \"7d388199-9f34-415c-b83c-d5f54d17d063\") " Feb 02 09:48:07 crc kubenswrapper[4764]: I0202 09:48:07.673647 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d388199-9f34-415c-b83c-d5f54d17d063-repo-setup-combined-ca-bundle\") pod \"7d388199-9f34-415c-b83c-d5f54d17d063\" (UID: \"7d388199-9f34-415c-b83c-d5f54d17d063\") " Feb 02 09:48:07 crc kubenswrapper[4764]: I0202 09:48:07.673672 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/7d388199-9f34-415c-b83c-d5f54d17d063-openstack-edpm-ipam-ovn-default-certs-0\") pod \"7d388199-9f34-415c-b83c-d5f54d17d063\" (UID: \"7d388199-9f34-415c-b83c-d5f54d17d063\") " Feb 02 09:48:07 crc kubenswrapper[4764]: I0202 09:48:07.673700 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/7d388199-9f34-415c-b83c-d5f54d17d063-ssh-key-openstack-edpm-ipam\") pod \"7d388199-9f34-415c-b83c-d5f54d17d063\" (UID: \"7d388199-9f34-415c-b83c-d5f54d17d063\") " Feb 02 09:48:07 crc kubenswrapper[4764]: I0202 09:48:07.673722 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/7d388199-9f34-415c-b83c-d5f54d17d063-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"7d388199-9f34-415c-b83c-d5f54d17d063\" (UID: \"7d388199-9f34-415c-b83c-d5f54d17d063\") " Feb 02 09:48:07 crc kubenswrapper[4764]: I0202 09:48:07.673748 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/7d388199-9f34-415c-b83c-d5f54d17d063-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"7d388199-9f34-415c-b83c-d5f54d17d063\" (UID: \"7d388199-9f34-415c-b83c-d5f54d17d063\") " Feb 02 09:48:07 crc kubenswrapper[4764]: I0202 09:48:07.674631 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f2rf7\" (UniqueName: \"kubernetes.io/projected/7d388199-9f34-415c-b83c-d5f54d17d063-kube-api-access-f2rf7\") pod \"7d388199-9f34-415c-b83c-d5f54d17d063\" (UID: \"7d388199-9f34-415c-b83c-d5f54d17d063\") " Feb 02 09:48:07 crc kubenswrapper[4764]: I0202 09:48:07.674689 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d388199-9f34-415c-b83c-d5f54d17d063-bootstrap-combined-ca-bundle\") pod \"7d388199-9f34-415c-b83c-d5f54d17d063\" (UID: \"7d388199-9f34-415c-b83c-d5f54d17d063\") " Feb 02 09:48:07 crc kubenswrapper[4764]: I0202 09:48:07.674708 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7d388199-9f34-415c-b83c-d5f54d17d063-ceph\") pod \"7d388199-9f34-415c-b83c-d5f54d17d063\" (UID: \"7d388199-9f34-415c-b83c-d5f54d17d063\") " Feb 02 09:48:07 crc kubenswrapper[4764]: I0202 09:48:07.685531 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d388199-9f34-415c-b83c-d5f54d17d063-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "7d388199-9f34-415c-b83c-d5f54d17d063" (UID: "7d388199-9f34-415c-b83c-d5f54d17d063"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:48:07 crc kubenswrapper[4764]: I0202 09:48:07.687351 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7d388199-9f34-415c-b83c-d5f54d17d063-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "7d388199-9f34-415c-b83c-d5f54d17d063" (UID: "7d388199-9f34-415c-b83c-d5f54d17d063"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:48:07 crc kubenswrapper[4764]: I0202 09:48:07.687561 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d388199-9f34-415c-b83c-d5f54d17d063-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "7d388199-9f34-415c-b83c-d5f54d17d063" (UID: "7d388199-9f34-415c-b83c-d5f54d17d063"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:48:07 crc kubenswrapper[4764]: I0202 09:48:07.687832 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d388199-9f34-415c-b83c-d5f54d17d063-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "7d388199-9f34-415c-b83c-d5f54d17d063" (UID: "7d388199-9f34-415c-b83c-d5f54d17d063"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:48:07 crc kubenswrapper[4764]: I0202 09:48:07.687913 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d388199-9f34-415c-b83c-d5f54d17d063-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "7d388199-9f34-415c-b83c-d5f54d17d063" (UID: "7d388199-9f34-415c-b83c-d5f54d17d063"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:48:07 crc kubenswrapper[4764]: I0202 09:48:07.688085 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7d388199-9f34-415c-b83c-d5f54d17d063-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "7d388199-9f34-415c-b83c-d5f54d17d063" (UID: "7d388199-9f34-415c-b83c-d5f54d17d063"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:48:07 crc kubenswrapper[4764]: I0202 09:48:07.688101 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d388199-9f34-415c-b83c-d5f54d17d063-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "7d388199-9f34-415c-b83c-d5f54d17d063" (UID: "7d388199-9f34-415c-b83c-d5f54d17d063"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:48:07 crc kubenswrapper[4764]: I0202 09:48:07.688190 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7d388199-9f34-415c-b83c-d5f54d17d063-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "7d388199-9f34-415c-b83c-d5f54d17d063" (UID: "7d388199-9f34-415c-b83c-d5f54d17d063"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:48:07 crc kubenswrapper[4764]: I0202 09:48:07.689017 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d388199-9f34-415c-b83c-d5f54d17d063-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "7d388199-9f34-415c-b83c-d5f54d17d063" (UID: "7d388199-9f34-415c-b83c-d5f54d17d063"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:48:07 crc kubenswrapper[4764]: I0202 09:48:07.689657 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d388199-9f34-415c-b83c-d5f54d17d063-ceph" (OuterVolumeSpecName: "ceph") pod "7d388199-9f34-415c-b83c-d5f54d17d063" (UID: "7d388199-9f34-415c-b83c-d5f54d17d063"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:48:07 crc kubenswrapper[4764]: I0202 09:48:07.698167 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7d388199-9f34-415c-b83c-d5f54d17d063-kube-api-access-f2rf7" (OuterVolumeSpecName: "kube-api-access-f2rf7") pod "7d388199-9f34-415c-b83c-d5f54d17d063" (UID: "7d388199-9f34-415c-b83c-d5f54d17d063"). InnerVolumeSpecName "kube-api-access-f2rf7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:48:07 crc kubenswrapper[4764]: I0202 09:48:07.713463 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d388199-9f34-415c-b83c-d5f54d17d063-inventory" (OuterVolumeSpecName: "inventory") pod "7d388199-9f34-415c-b83c-d5f54d17d063" (UID: "7d388199-9f34-415c-b83c-d5f54d17d063"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:48:07 crc kubenswrapper[4764]: I0202 09:48:07.728661 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d388199-9f34-415c-b83c-d5f54d17d063-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "7d388199-9f34-415c-b83c-d5f54d17d063" (UID: "7d388199-9f34-415c-b83c-d5f54d17d063"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:48:07 crc kubenswrapper[4764]: I0202 09:48:07.777693 4764 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d388199-9f34-415c-b83c-d5f54d17d063-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 09:48:07 crc kubenswrapper[4764]: I0202 09:48:07.777734 4764 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d388199-9f34-415c-b83c-d5f54d17d063-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 09:48:07 crc kubenswrapper[4764]: I0202 09:48:07.777747 4764 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d388199-9f34-415c-b83c-d5f54d17d063-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 09:48:07 crc kubenswrapper[4764]: I0202 09:48:07.777760 4764 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d388199-9f34-415c-b83c-d5f54d17d063-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 09:48:07 crc kubenswrapper[4764]: I0202 09:48:07.777775 4764 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d388199-9f34-415c-b83c-d5f54d17d063-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 09:48:07 crc kubenswrapper[4764]: I0202 09:48:07.777788 4764 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/7d388199-9f34-415c-b83c-d5f54d17d063-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Feb 02 09:48:07 crc kubenswrapper[4764]: I0202 09:48:07.777799 4764 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/7d388199-9f34-415c-b83c-d5f54d17d063-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 02 09:48:07 crc kubenswrapper[4764]: I0202 09:48:07.777811 4764 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/7d388199-9f34-415c-b83c-d5f54d17d063-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Feb 02 09:48:07 crc kubenswrapper[4764]: I0202 09:48:07.777825 4764 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/7d388199-9f34-415c-b83c-d5f54d17d063-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Feb 02 09:48:07 crc kubenswrapper[4764]: I0202 09:48:07.777838 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f2rf7\" (UniqueName: \"kubernetes.io/projected/7d388199-9f34-415c-b83c-d5f54d17d063-kube-api-access-f2rf7\") on node \"crc\" DevicePath \"\"" Feb 02 09:48:07 crc kubenswrapper[4764]: I0202 09:48:07.777851 4764 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d388199-9f34-415c-b83c-d5f54d17d063-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 09:48:07 crc kubenswrapper[4764]: I0202 09:48:07.777862 4764 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7d388199-9f34-415c-b83c-d5f54d17d063-ceph\") on node \"crc\" DevicePath \"\"" Feb 02 09:48:07 crc kubenswrapper[4764]: I0202 09:48:07.777875 4764 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7d388199-9f34-415c-b83c-d5f54d17d063-inventory\") on node \"crc\" DevicePath \"\"" Feb 02 09:48:08 crc kubenswrapper[4764]: I0202 09:48:08.103858 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9q5kq" Feb 02 09:48:08 crc kubenswrapper[4764]: I0202 09:48:08.104272 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9q5kq" event={"ID":"7d388199-9f34-415c-b83c-d5f54d17d063","Type":"ContainerDied","Data":"64b363b92d0415f73b15905bc0f0c99d2b8fe2488c7d07349648b9ba8c6b00e1"} Feb 02 09:48:08 crc kubenswrapper[4764]: I0202 09:48:08.104373 4764 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="64b363b92d0415f73b15905bc0f0c99d2b8fe2488c7d07349648b9ba8c6b00e1" Feb 02 09:48:08 crc kubenswrapper[4764]: I0202 09:48:08.224398 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-bs4pz"] Feb 02 09:48:08 crc kubenswrapper[4764]: E0202 09:48:08.224796 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d388199-9f34-415c-b83c-d5f54d17d063" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Feb 02 09:48:08 crc kubenswrapper[4764]: I0202 09:48:08.224812 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d388199-9f34-415c-b83c-d5f54d17d063" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Feb 02 09:48:08 crc kubenswrapper[4764]: I0202 09:48:08.225031 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d388199-9f34-415c-b83c-d5f54d17d063" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Feb 02 09:48:08 crc kubenswrapper[4764]: I0202 09:48:08.225663 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-bs4pz" Feb 02 09:48:08 crc kubenswrapper[4764]: I0202 09:48:08.228525 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Feb 02 09:48:08 crc kubenswrapper[4764]: I0202 09:48:08.228731 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 02 09:48:08 crc kubenswrapper[4764]: I0202 09:48:08.230035 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 02 09:48:08 crc kubenswrapper[4764]: I0202 09:48:08.230336 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 02 09:48:08 crc kubenswrapper[4764]: I0202 09:48:08.231200 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-xksq5" Feb 02 09:48:08 crc kubenswrapper[4764]: I0202 09:48:08.259833 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-bs4pz"] Feb 02 09:48:08 crc kubenswrapper[4764]: I0202 09:48:08.392920 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3858fd68-f1b8-4e73-a0c9-5fb26dcbedd6-inventory\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-bs4pz\" (UID: \"3858fd68-f1b8-4e73-a0c9-5fb26dcbedd6\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-bs4pz" Feb 02 09:48:08 crc kubenswrapper[4764]: I0202 09:48:08.393013 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3858fd68-f1b8-4e73-a0c9-5fb26dcbedd6-ceph\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-bs4pz\" (UID: \"3858fd68-f1b8-4e73-a0c9-5fb26dcbedd6\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-bs4pz" Feb 02 09:48:08 crc kubenswrapper[4764]: I0202 09:48:08.393054 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3858fd68-f1b8-4e73-a0c9-5fb26dcbedd6-ssh-key-openstack-edpm-ipam\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-bs4pz\" (UID: \"3858fd68-f1b8-4e73-a0c9-5fb26dcbedd6\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-bs4pz" Feb 02 09:48:08 crc kubenswrapper[4764]: I0202 09:48:08.393253 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pz7wd\" (UniqueName: \"kubernetes.io/projected/3858fd68-f1b8-4e73-a0c9-5fb26dcbedd6-kube-api-access-pz7wd\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-bs4pz\" (UID: \"3858fd68-f1b8-4e73-a0c9-5fb26dcbedd6\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-bs4pz" Feb 02 09:48:08 crc kubenswrapper[4764]: I0202 09:48:08.494304 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3858fd68-f1b8-4e73-a0c9-5fb26dcbedd6-inventory\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-bs4pz\" (UID: \"3858fd68-f1b8-4e73-a0c9-5fb26dcbedd6\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-bs4pz" Feb 02 09:48:08 crc kubenswrapper[4764]: I0202 09:48:08.494733 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3858fd68-f1b8-4e73-a0c9-5fb26dcbedd6-ceph\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-bs4pz\" (UID: \"3858fd68-f1b8-4e73-a0c9-5fb26dcbedd6\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-bs4pz" Feb 02 09:48:08 crc kubenswrapper[4764]: I0202 09:48:08.494762 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3858fd68-f1b8-4e73-a0c9-5fb26dcbedd6-ssh-key-openstack-edpm-ipam\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-bs4pz\" (UID: \"3858fd68-f1b8-4e73-a0c9-5fb26dcbedd6\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-bs4pz" Feb 02 09:48:08 crc kubenswrapper[4764]: I0202 09:48:08.494862 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pz7wd\" (UniqueName: \"kubernetes.io/projected/3858fd68-f1b8-4e73-a0c9-5fb26dcbedd6-kube-api-access-pz7wd\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-bs4pz\" (UID: \"3858fd68-f1b8-4e73-a0c9-5fb26dcbedd6\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-bs4pz" Feb 02 09:48:08 crc kubenswrapper[4764]: I0202 09:48:08.514067 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3858fd68-f1b8-4e73-a0c9-5fb26dcbedd6-inventory\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-bs4pz\" (UID: \"3858fd68-f1b8-4e73-a0c9-5fb26dcbedd6\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-bs4pz" Feb 02 09:48:08 crc kubenswrapper[4764]: I0202 09:48:08.520053 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3858fd68-f1b8-4e73-a0c9-5fb26dcbedd6-ssh-key-openstack-edpm-ipam\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-bs4pz\" (UID: \"3858fd68-f1b8-4e73-a0c9-5fb26dcbedd6\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-bs4pz" Feb 02 09:48:08 crc kubenswrapper[4764]: I0202 09:48:08.528679 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pz7wd\" (UniqueName: \"kubernetes.io/projected/3858fd68-f1b8-4e73-a0c9-5fb26dcbedd6-kube-api-access-pz7wd\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-bs4pz\" (UID: \"3858fd68-f1b8-4e73-a0c9-5fb26dcbedd6\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-bs4pz" Feb 02 09:48:08 crc kubenswrapper[4764]: I0202 09:48:08.529435 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3858fd68-f1b8-4e73-a0c9-5fb26dcbedd6-ceph\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-bs4pz\" (UID: \"3858fd68-f1b8-4e73-a0c9-5fb26dcbedd6\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-bs4pz" Feb 02 09:48:08 crc kubenswrapper[4764]: I0202 09:48:08.544431 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-bs4pz" Feb 02 09:48:09 crc kubenswrapper[4764]: I0202 09:48:09.063126 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-bs4pz"] Feb 02 09:48:09 crc kubenswrapper[4764]: I0202 09:48:09.138252 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-bs4pz" event={"ID":"3858fd68-f1b8-4e73-a0c9-5fb26dcbedd6","Type":"ContainerStarted","Data":"94076b49d6866397cde5e4abba83cd8c899fd0065158d298918607af4057e390"} Feb 02 09:48:10 crc kubenswrapper[4764]: I0202 09:48:10.150149 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-bs4pz" event={"ID":"3858fd68-f1b8-4e73-a0c9-5fb26dcbedd6","Type":"ContainerStarted","Data":"51d0b163cbad6d8d467f46ad7490642db7111e99a921e852c7a63e93784c9262"} Feb 02 09:48:10 crc kubenswrapper[4764]: I0202 09:48:10.185999 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-bs4pz" podStartSLOduration=1.733380972 podStartE2EDuration="2.185973056s" podCreationTimestamp="2026-02-02 09:48:08 +0000 UTC" firstStartedPulling="2026-02-02 09:48:09.073888604 +0000 UTC m=+2452.007612682" lastFinishedPulling="2026-02-02 09:48:09.526480668 +0000 UTC m=+2452.460204766" observedRunningTime="2026-02-02 09:48:10.173613137 +0000 UTC m=+2453.107337235" watchObservedRunningTime="2026-02-02 09:48:10.185973056 +0000 UTC m=+2453.119697134" Feb 02 09:48:12 crc kubenswrapper[4764]: I0202 09:48:12.826743 4764 scope.go:117] "RemoveContainer" containerID="be8fa5800ec42d1c141765dcfc6111f3ef96b05b6e8d5a9feaf2baa219897bcf" Feb 02 09:48:12 crc kubenswrapper[4764]: E0202 09:48:12.827548 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 09:48:15 crc kubenswrapper[4764]: I0202 09:48:15.210134 4764 generic.go:334] "Generic (PLEG): container finished" podID="3858fd68-f1b8-4e73-a0c9-5fb26dcbedd6" containerID="51d0b163cbad6d8d467f46ad7490642db7111e99a921e852c7a63e93784c9262" exitCode=0 Feb 02 09:48:15 crc kubenswrapper[4764]: I0202 09:48:15.210310 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-bs4pz" event={"ID":"3858fd68-f1b8-4e73-a0c9-5fb26dcbedd6","Type":"ContainerDied","Data":"51d0b163cbad6d8d467f46ad7490642db7111e99a921e852c7a63e93784c9262"} Feb 02 09:48:16 crc kubenswrapper[4764]: I0202 09:48:16.685668 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-bs4pz" Feb 02 09:48:16 crc kubenswrapper[4764]: I0202 09:48:16.886301 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3858fd68-f1b8-4e73-a0c9-5fb26dcbedd6-ssh-key-openstack-edpm-ipam\") pod \"3858fd68-f1b8-4e73-a0c9-5fb26dcbedd6\" (UID: \"3858fd68-f1b8-4e73-a0c9-5fb26dcbedd6\") " Feb 02 09:48:16 crc kubenswrapper[4764]: I0202 09:48:16.886711 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3858fd68-f1b8-4e73-a0c9-5fb26dcbedd6-inventory\") pod \"3858fd68-f1b8-4e73-a0c9-5fb26dcbedd6\" (UID: \"3858fd68-f1b8-4e73-a0c9-5fb26dcbedd6\") " Feb 02 09:48:16 crc kubenswrapper[4764]: I0202 09:48:16.887035 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3858fd68-f1b8-4e73-a0c9-5fb26dcbedd6-ceph\") pod \"3858fd68-f1b8-4e73-a0c9-5fb26dcbedd6\" (UID: \"3858fd68-f1b8-4e73-a0c9-5fb26dcbedd6\") " Feb 02 09:48:16 crc kubenswrapper[4764]: I0202 09:48:16.887188 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pz7wd\" (UniqueName: \"kubernetes.io/projected/3858fd68-f1b8-4e73-a0c9-5fb26dcbedd6-kube-api-access-pz7wd\") pod \"3858fd68-f1b8-4e73-a0c9-5fb26dcbedd6\" (UID: \"3858fd68-f1b8-4e73-a0c9-5fb26dcbedd6\") " Feb 02 09:48:16 crc kubenswrapper[4764]: I0202 09:48:16.892393 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3858fd68-f1b8-4e73-a0c9-5fb26dcbedd6-ceph" (OuterVolumeSpecName: "ceph") pod "3858fd68-f1b8-4e73-a0c9-5fb26dcbedd6" (UID: "3858fd68-f1b8-4e73-a0c9-5fb26dcbedd6"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:48:16 crc kubenswrapper[4764]: I0202 09:48:16.896157 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3858fd68-f1b8-4e73-a0c9-5fb26dcbedd6-kube-api-access-pz7wd" (OuterVolumeSpecName: "kube-api-access-pz7wd") pod "3858fd68-f1b8-4e73-a0c9-5fb26dcbedd6" (UID: "3858fd68-f1b8-4e73-a0c9-5fb26dcbedd6"). InnerVolumeSpecName "kube-api-access-pz7wd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:48:16 crc kubenswrapper[4764]: I0202 09:48:16.911430 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3858fd68-f1b8-4e73-a0c9-5fb26dcbedd6-inventory" (OuterVolumeSpecName: "inventory") pod "3858fd68-f1b8-4e73-a0c9-5fb26dcbedd6" (UID: "3858fd68-f1b8-4e73-a0c9-5fb26dcbedd6"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:48:16 crc kubenswrapper[4764]: I0202 09:48:16.918168 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3858fd68-f1b8-4e73-a0c9-5fb26dcbedd6-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "3858fd68-f1b8-4e73-a0c9-5fb26dcbedd6" (UID: "3858fd68-f1b8-4e73-a0c9-5fb26dcbedd6"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:48:16 crc kubenswrapper[4764]: I0202 09:48:16.990153 4764 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3858fd68-f1b8-4e73-a0c9-5fb26dcbedd6-ceph\") on node \"crc\" DevicePath \"\"" Feb 02 09:48:16 crc kubenswrapper[4764]: I0202 09:48:16.990214 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pz7wd\" (UniqueName: \"kubernetes.io/projected/3858fd68-f1b8-4e73-a0c9-5fb26dcbedd6-kube-api-access-pz7wd\") on node \"crc\" DevicePath \"\"" Feb 02 09:48:16 crc kubenswrapper[4764]: I0202 09:48:16.990235 4764 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3858fd68-f1b8-4e73-a0c9-5fb26dcbedd6-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 02 09:48:16 crc kubenswrapper[4764]: I0202 09:48:16.990254 4764 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3858fd68-f1b8-4e73-a0c9-5fb26dcbedd6-inventory\") on node \"crc\" DevicePath \"\"" Feb 02 09:48:17 crc kubenswrapper[4764]: I0202 09:48:17.237745 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-bs4pz" event={"ID":"3858fd68-f1b8-4e73-a0c9-5fb26dcbedd6","Type":"ContainerDied","Data":"94076b49d6866397cde5e4abba83cd8c899fd0065158d298918607af4057e390"} Feb 02 09:48:17 crc kubenswrapper[4764]: I0202 09:48:17.237789 4764 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="94076b49d6866397cde5e4abba83cd8c899fd0065158d298918607af4057e390" Feb 02 09:48:17 crc kubenswrapper[4764]: I0202 09:48:17.238118 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-bs4pz" Feb 02 09:48:17 crc kubenswrapper[4764]: I0202 09:48:17.337807 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-2qbnb"] Feb 02 09:48:17 crc kubenswrapper[4764]: E0202 09:48:17.338165 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3858fd68-f1b8-4e73-a0c9-5fb26dcbedd6" containerName="ceph-client-edpm-deployment-openstack-edpm-ipam" Feb 02 09:48:17 crc kubenswrapper[4764]: I0202 09:48:17.338181 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="3858fd68-f1b8-4e73-a0c9-5fb26dcbedd6" containerName="ceph-client-edpm-deployment-openstack-edpm-ipam" Feb 02 09:48:17 crc kubenswrapper[4764]: I0202 09:48:17.338355 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="3858fd68-f1b8-4e73-a0c9-5fb26dcbedd6" containerName="ceph-client-edpm-deployment-openstack-edpm-ipam" Feb 02 09:48:17 crc kubenswrapper[4764]: I0202 09:48:17.338963 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-2qbnb" Feb 02 09:48:17 crc kubenswrapper[4764]: I0202 09:48:17.342447 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 02 09:48:17 crc kubenswrapper[4764]: I0202 09:48:17.342726 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 02 09:48:17 crc kubenswrapper[4764]: I0202 09:48:17.344377 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-xksq5" Feb 02 09:48:17 crc kubenswrapper[4764]: I0202 09:48:17.345208 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Feb 02 09:48:17 crc kubenswrapper[4764]: I0202 09:48:17.345408 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Feb 02 09:48:17 crc kubenswrapper[4764]: I0202 09:48:17.345600 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 02 09:48:17 crc kubenswrapper[4764]: I0202 09:48:17.349370 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-2qbnb"] Feb 02 09:48:17 crc kubenswrapper[4764]: I0202 09:48:17.396894 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/e6da450f-20b5-41ef-b9bb-a0df89b3d4f3-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-2qbnb\" (UID: \"e6da450f-20b5-41ef-b9bb-a0df89b3d4f3\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-2qbnb" Feb 02 09:48:17 crc kubenswrapper[4764]: I0202 09:48:17.396977 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e6da450f-20b5-41ef-b9bb-a0df89b3d4f3-ssh-key-openstack-edpm-ipam\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-2qbnb\" (UID: \"e6da450f-20b5-41ef-b9bb-a0df89b3d4f3\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-2qbnb" Feb 02 09:48:17 crc kubenswrapper[4764]: I0202 09:48:17.397068 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e6da450f-20b5-41ef-b9bb-a0df89b3d4f3-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-2qbnb\" (UID: \"e6da450f-20b5-41ef-b9bb-a0df89b3d4f3\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-2qbnb" Feb 02 09:48:17 crc kubenswrapper[4764]: I0202 09:48:17.397132 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e6da450f-20b5-41ef-b9bb-a0df89b3d4f3-ceph\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-2qbnb\" (UID: \"e6da450f-20b5-41ef-b9bb-a0df89b3d4f3\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-2qbnb" Feb 02 09:48:17 crc kubenswrapper[4764]: I0202 09:48:17.397155 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bw74x\" (UniqueName: \"kubernetes.io/projected/e6da450f-20b5-41ef-b9bb-a0df89b3d4f3-kube-api-access-bw74x\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-2qbnb\" (UID: \"e6da450f-20b5-41ef-b9bb-a0df89b3d4f3\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-2qbnb" Feb 02 09:48:17 crc kubenswrapper[4764]: I0202 09:48:17.397407 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6da450f-20b5-41ef-b9bb-a0df89b3d4f3-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-2qbnb\" (UID: \"e6da450f-20b5-41ef-b9bb-a0df89b3d4f3\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-2qbnb" Feb 02 09:48:17 crc kubenswrapper[4764]: I0202 09:48:17.500759 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e6da450f-20b5-41ef-b9bb-a0df89b3d4f3-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-2qbnb\" (UID: \"e6da450f-20b5-41ef-b9bb-a0df89b3d4f3\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-2qbnb" Feb 02 09:48:17 crc kubenswrapper[4764]: I0202 09:48:17.500844 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e6da450f-20b5-41ef-b9bb-a0df89b3d4f3-ceph\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-2qbnb\" (UID: \"e6da450f-20b5-41ef-b9bb-a0df89b3d4f3\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-2qbnb" Feb 02 09:48:17 crc kubenswrapper[4764]: I0202 09:48:17.500875 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bw74x\" (UniqueName: \"kubernetes.io/projected/e6da450f-20b5-41ef-b9bb-a0df89b3d4f3-kube-api-access-bw74x\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-2qbnb\" (UID: \"e6da450f-20b5-41ef-b9bb-a0df89b3d4f3\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-2qbnb" Feb 02 09:48:17 crc kubenswrapper[4764]: I0202 09:48:17.500974 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6da450f-20b5-41ef-b9bb-a0df89b3d4f3-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-2qbnb\" (UID: \"e6da450f-20b5-41ef-b9bb-a0df89b3d4f3\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-2qbnb" Feb 02 09:48:17 crc kubenswrapper[4764]: I0202 09:48:17.501063 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/e6da450f-20b5-41ef-b9bb-a0df89b3d4f3-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-2qbnb\" (UID: \"e6da450f-20b5-41ef-b9bb-a0df89b3d4f3\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-2qbnb" Feb 02 09:48:17 crc kubenswrapper[4764]: I0202 09:48:17.501097 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e6da450f-20b5-41ef-b9bb-a0df89b3d4f3-ssh-key-openstack-edpm-ipam\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-2qbnb\" (UID: \"e6da450f-20b5-41ef-b9bb-a0df89b3d4f3\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-2qbnb" Feb 02 09:48:17 crc kubenswrapper[4764]: I0202 09:48:17.503162 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/e6da450f-20b5-41ef-b9bb-a0df89b3d4f3-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-2qbnb\" (UID: \"e6da450f-20b5-41ef-b9bb-a0df89b3d4f3\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-2qbnb" Feb 02 09:48:17 crc kubenswrapper[4764]: I0202 09:48:17.506333 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e6da450f-20b5-41ef-b9bb-a0df89b3d4f3-ssh-key-openstack-edpm-ipam\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-2qbnb\" (UID: \"e6da450f-20b5-41ef-b9bb-a0df89b3d4f3\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-2qbnb" Feb 02 09:48:17 crc kubenswrapper[4764]: I0202 09:48:17.518251 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e6da450f-20b5-41ef-b9bb-a0df89b3d4f3-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-2qbnb\" (UID: \"e6da450f-20b5-41ef-b9bb-a0df89b3d4f3\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-2qbnb" Feb 02 09:48:17 crc kubenswrapper[4764]: I0202 09:48:17.518771 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e6da450f-20b5-41ef-b9bb-a0df89b3d4f3-ceph\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-2qbnb\" (UID: \"e6da450f-20b5-41ef-b9bb-a0df89b3d4f3\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-2qbnb" Feb 02 09:48:17 crc kubenswrapper[4764]: I0202 09:48:17.521036 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bw74x\" (UniqueName: \"kubernetes.io/projected/e6da450f-20b5-41ef-b9bb-a0df89b3d4f3-kube-api-access-bw74x\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-2qbnb\" (UID: \"e6da450f-20b5-41ef-b9bb-a0df89b3d4f3\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-2qbnb" Feb 02 09:48:17 crc kubenswrapper[4764]: I0202 09:48:17.529313 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6da450f-20b5-41ef-b9bb-a0df89b3d4f3-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-2qbnb\" (UID: \"e6da450f-20b5-41ef-b9bb-a0df89b3d4f3\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-2qbnb" Feb 02 09:48:17 crc kubenswrapper[4764]: I0202 09:48:17.663269 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-2qbnb" Feb 02 09:48:18 crc kubenswrapper[4764]: I0202 09:48:18.196132 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-2qbnb"] Feb 02 09:48:18 crc kubenswrapper[4764]: W0202 09:48:18.202156 4764 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode6da450f_20b5_41ef_b9bb_a0df89b3d4f3.slice/crio-f755f2eea6c4c661a8c1c27b75764179c75f8656126f737339850c6800376bf0 WatchSource:0}: Error finding container f755f2eea6c4c661a8c1c27b75764179c75f8656126f737339850c6800376bf0: Status 404 returned error can't find the container with id f755f2eea6c4c661a8c1c27b75764179c75f8656126f737339850c6800376bf0 Feb 02 09:48:18 crc kubenswrapper[4764]: I0202 09:48:18.249353 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-2qbnb" event={"ID":"e6da450f-20b5-41ef-b9bb-a0df89b3d4f3","Type":"ContainerStarted","Data":"f755f2eea6c4c661a8c1c27b75764179c75f8656126f737339850c6800376bf0"} Feb 02 09:48:18 crc kubenswrapper[4764]: I0202 09:48:18.656448 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 02 09:48:19 crc kubenswrapper[4764]: I0202 09:48:19.263153 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-2qbnb" event={"ID":"e6da450f-20b5-41ef-b9bb-a0df89b3d4f3","Type":"ContainerStarted","Data":"0d0d732782d05467a6bcadcf36f2365de8849ba98ab1b7a7ed4f72f2e5bc6c40"} Feb 02 09:48:19 crc kubenswrapper[4764]: I0202 09:48:19.297351 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-2qbnb" podStartSLOduration=1.849011619 podStartE2EDuration="2.29733043s" podCreationTimestamp="2026-02-02 09:48:17 +0000 UTC" firstStartedPulling="2026-02-02 09:48:18.204150012 +0000 UTC m=+2461.137874110" lastFinishedPulling="2026-02-02 09:48:18.652468833 +0000 UTC m=+2461.586192921" observedRunningTime="2026-02-02 09:48:19.288237917 +0000 UTC m=+2462.221962015" watchObservedRunningTime="2026-02-02 09:48:19.29733043 +0000 UTC m=+2462.231054528" Feb 02 09:48:23 crc kubenswrapper[4764]: I0202 09:48:23.826671 4764 scope.go:117] "RemoveContainer" containerID="be8fa5800ec42d1c141765dcfc6111f3ef96b05b6e8d5a9feaf2baa219897bcf" Feb 02 09:48:23 crc kubenswrapper[4764]: E0202 09:48:23.827903 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 09:48:36 crc kubenswrapper[4764]: I0202 09:48:36.825509 4764 scope.go:117] "RemoveContainer" containerID="be8fa5800ec42d1c141765dcfc6111f3ef96b05b6e8d5a9feaf2baa219897bcf" Feb 02 09:48:36 crc kubenswrapper[4764]: E0202 09:48:36.826293 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 09:48:51 crc kubenswrapper[4764]: I0202 09:48:51.826850 4764 scope.go:117] "RemoveContainer" containerID="be8fa5800ec42d1c141765dcfc6111f3ef96b05b6e8d5a9feaf2baa219897bcf" Feb 02 09:48:51 crc kubenswrapper[4764]: E0202 09:48:51.828096 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 09:49:05 crc kubenswrapper[4764]: I0202 09:49:05.826436 4764 scope.go:117] "RemoveContainer" containerID="be8fa5800ec42d1c141765dcfc6111f3ef96b05b6e8d5a9feaf2baa219897bcf" Feb 02 09:49:05 crc kubenswrapper[4764]: E0202 09:49:05.827384 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 09:49:16 crc kubenswrapper[4764]: I0202 09:49:16.826241 4764 scope.go:117] "RemoveContainer" containerID="be8fa5800ec42d1c141765dcfc6111f3ef96b05b6e8d5a9feaf2baa219897bcf" Feb 02 09:49:16 crc kubenswrapper[4764]: E0202 09:49:16.826802 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 09:49:30 crc kubenswrapper[4764]: I0202 09:49:30.830262 4764 scope.go:117] "RemoveContainer" containerID="be8fa5800ec42d1c141765dcfc6111f3ef96b05b6e8d5a9feaf2baa219897bcf" Feb 02 09:49:30 crc kubenswrapper[4764]: E0202 09:49:30.834821 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 09:49:35 crc kubenswrapper[4764]: I0202 09:49:35.037854 4764 generic.go:334] "Generic (PLEG): container finished" podID="e6da450f-20b5-41ef-b9bb-a0df89b3d4f3" containerID="0d0d732782d05467a6bcadcf36f2365de8849ba98ab1b7a7ed4f72f2e5bc6c40" exitCode=0 Feb 02 09:49:35 crc kubenswrapper[4764]: I0202 09:49:35.037995 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-2qbnb" event={"ID":"e6da450f-20b5-41ef-b9bb-a0df89b3d4f3","Type":"ContainerDied","Data":"0d0d732782d05467a6bcadcf36f2365de8849ba98ab1b7a7ed4f72f2e5bc6c40"} Feb 02 09:49:36 crc kubenswrapper[4764]: I0202 09:49:36.527054 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-2qbnb" Feb 02 09:49:36 crc kubenswrapper[4764]: I0202 09:49:36.662460 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/e6da450f-20b5-41ef-b9bb-a0df89b3d4f3-ovncontroller-config-0\") pod \"e6da450f-20b5-41ef-b9bb-a0df89b3d4f3\" (UID: \"e6da450f-20b5-41ef-b9bb-a0df89b3d4f3\") " Feb 02 09:49:36 crc kubenswrapper[4764]: I0202 09:49:36.662552 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6da450f-20b5-41ef-b9bb-a0df89b3d4f3-ovn-combined-ca-bundle\") pod \"e6da450f-20b5-41ef-b9bb-a0df89b3d4f3\" (UID: \"e6da450f-20b5-41ef-b9bb-a0df89b3d4f3\") " Feb 02 09:49:36 crc kubenswrapper[4764]: I0202 09:49:36.663349 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bw74x\" (UniqueName: \"kubernetes.io/projected/e6da450f-20b5-41ef-b9bb-a0df89b3d4f3-kube-api-access-bw74x\") pod \"e6da450f-20b5-41ef-b9bb-a0df89b3d4f3\" (UID: \"e6da450f-20b5-41ef-b9bb-a0df89b3d4f3\") " Feb 02 09:49:36 crc kubenswrapper[4764]: I0202 09:49:36.663800 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e6da450f-20b5-41ef-b9bb-a0df89b3d4f3-ssh-key-openstack-edpm-ipam\") pod \"e6da450f-20b5-41ef-b9bb-a0df89b3d4f3\" (UID: \"e6da450f-20b5-41ef-b9bb-a0df89b3d4f3\") " Feb 02 09:49:36 crc kubenswrapper[4764]: I0202 09:49:36.663866 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e6da450f-20b5-41ef-b9bb-a0df89b3d4f3-inventory\") pod \"e6da450f-20b5-41ef-b9bb-a0df89b3d4f3\" (UID: \"e6da450f-20b5-41ef-b9bb-a0df89b3d4f3\") " Feb 02 09:49:36 crc kubenswrapper[4764]: I0202 09:49:36.663891 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e6da450f-20b5-41ef-b9bb-a0df89b3d4f3-ceph\") pod \"e6da450f-20b5-41ef-b9bb-a0df89b3d4f3\" (UID: \"e6da450f-20b5-41ef-b9bb-a0df89b3d4f3\") " Feb 02 09:49:36 crc kubenswrapper[4764]: I0202 09:49:36.671482 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e6da450f-20b5-41ef-b9bb-a0df89b3d4f3-ceph" (OuterVolumeSpecName: "ceph") pod "e6da450f-20b5-41ef-b9bb-a0df89b3d4f3" (UID: "e6da450f-20b5-41ef-b9bb-a0df89b3d4f3"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:49:36 crc kubenswrapper[4764]: I0202 09:49:36.672028 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e6da450f-20b5-41ef-b9bb-a0df89b3d4f3-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "e6da450f-20b5-41ef-b9bb-a0df89b3d4f3" (UID: "e6da450f-20b5-41ef-b9bb-a0df89b3d4f3"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:49:36 crc kubenswrapper[4764]: I0202 09:49:36.673218 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e6da450f-20b5-41ef-b9bb-a0df89b3d4f3-kube-api-access-bw74x" (OuterVolumeSpecName: "kube-api-access-bw74x") pod "e6da450f-20b5-41ef-b9bb-a0df89b3d4f3" (UID: "e6da450f-20b5-41ef-b9bb-a0df89b3d4f3"). InnerVolumeSpecName "kube-api-access-bw74x". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:49:36 crc kubenswrapper[4764]: I0202 09:49:36.690232 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e6da450f-20b5-41ef-b9bb-a0df89b3d4f3-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "e6da450f-20b5-41ef-b9bb-a0df89b3d4f3" (UID: "e6da450f-20b5-41ef-b9bb-a0df89b3d4f3"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:49:36 crc kubenswrapper[4764]: I0202 09:49:36.704836 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e6da450f-20b5-41ef-b9bb-a0df89b3d4f3-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "e6da450f-20b5-41ef-b9bb-a0df89b3d4f3" (UID: "e6da450f-20b5-41ef-b9bb-a0df89b3d4f3"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:49:36 crc kubenswrapper[4764]: I0202 09:49:36.722272 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e6da450f-20b5-41ef-b9bb-a0df89b3d4f3-inventory" (OuterVolumeSpecName: "inventory") pod "e6da450f-20b5-41ef-b9bb-a0df89b3d4f3" (UID: "e6da450f-20b5-41ef-b9bb-a0df89b3d4f3"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:49:36 crc kubenswrapper[4764]: I0202 09:49:36.766489 4764 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e6da450f-20b5-41ef-b9bb-a0df89b3d4f3-inventory\") on node \"crc\" DevicePath \"\"" Feb 02 09:49:36 crc kubenswrapper[4764]: I0202 09:49:36.766546 4764 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e6da450f-20b5-41ef-b9bb-a0df89b3d4f3-ceph\") on node \"crc\" DevicePath \"\"" Feb 02 09:49:36 crc kubenswrapper[4764]: I0202 09:49:36.766567 4764 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/e6da450f-20b5-41ef-b9bb-a0df89b3d4f3-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Feb 02 09:49:36 crc kubenswrapper[4764]: I0202 09:49:36.766595 4764 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6da450f-20b5-41ef-b9bb-a0df89b3d4f3-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 09:49:36 crc kubenswrapper[4764]: I0202 09:49:36.766613 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bw74x\" (UniqueName: \"kubernetes.io/projected/e6da450f-20b5-41ef-b9bb-a0df89b3d4f3-kube-api-access-bw74x\") on node \"crc\" DevicePath \"\"" Feb 02 09:49:36 crc kubenswrapper[4764]: I0202 09:49:36.766639 4764 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e6da450f-20b5-41ef-b9bb-a0df89b3d4f3-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 02 09:49:37 crc kubenswrapper[4764]: I0202 09:49:37.067073 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-2qbnb" event={"ID":"e6da450f-20b5-41ef-b9bb-a0df89b3d4f3","Type":"ContainerDied","Data":"f755f2eea6c4c661a8c1c27b75764179c75f8656126f737339850c6800376bf0"} Feb 02 09:49:37 crc kubenswrapper[4764]: I0202 09:49:37.067149 4764 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f755f2eea6c4c661a8c1c27b75764179c75f8656126f737339850c6800376bf0" Feb 02 09:49:37 crc kubenswrapper[4764]: I0202 09:49:37.067266 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-2qbnb" Feb 02 09:49:37 crc kubenswrapper[4764]: I0202 09:49:37.207855 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6pxwj"] Feb 02 09:49:37 crc kubenswrapper[4764]: E0202 09:49:37.208455 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6da450f-20b5-41ef-b9bb-a0df89b3d4f3" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Feb 02 09:49:37 crc kubenswrapper[4764]: I0202 09:49:37.208479 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6da450f-20b5-41ef-b9bb-a0df89b3d4f3" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Feb 02 09:49:37 crc kubenswrapper[4764]: I0202 09:49:37.208729 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="e6da450f-20b5-41ef-b9bb-a0df89b3d4f3" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Feb 02 09:49:37 crc kubenswrapper[4764]: I0202 09:49:37.209644 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6pxwj" Feb 02 09:49:37 crc kubenswrapper[4764]: I0202 09:49:37.212969 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 02 09:49:37 crc kubenswrapper[4764]: I0202 09:49:37.213407 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Feb 02 09:49:37 crc kubenswrapper[4764]: I0202 09:49:37.213868 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-xksq5" Feb 02 09:49:37 crc kubenswrapper[4764]: I0202 09:49:37.214197 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Feb 02 09:49:37 crc kubenswrapper[4764]: I0202 09:49:37.214753 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Feb 02 09:49:37 crc kubenswrapper[4764]: I0202 09:49:37.215256 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 02 09:49:37 crc kubenswrapper[4764]: I0202 09:49:37.217068 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 02 09:49:37 crc kubenswrapper[4764]: I0202 09:49:37.221139 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6pxwj"] Feb 02 09:49:37 crc kubenswrapper[4764]: I0202 09:49:37.278190 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/a74b118e-ee59-4e87-967d-c8332a898a4c-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-6pxwj\" (UID: \"a74b118e-ee59-4e87-967d-c8332a898a4c\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6pxwj" Feb 02 09:49:37 crc kubenswrapper[4764]: I0202 09:49:37.278588 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/a74b118e-ee59-4e87-967d-c8332a898a4c-ceph\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-6pxwj\" (UID: \"a74b118e-ee59-4e87-967d-c8332a898a4c\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6pxwj" Feb 02 09:49:37 crc kubenswrapper[4764]: I0202 09:49:37.278736 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/a74b118e-ee59-4e87-967d-c8332a898a4c-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-6pxwj\" (UID: \"a74b118e-ee59-4e87-967d-c8332a898a4c\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6pxwj" Feb 02 09:49:37 crc kubenswrapper[4764]: I0202 09:49:37.278861 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a74b118e-ee59-4e87-967d-c8332a898a4c-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-6pxwj\" (UID: \"a74b118e-ee59-4e87-967d-c8332a898a4c\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6pxwj" Feb 02 09:49:37 crc kubenswrapper[4764]: I0202 09:49:37.279410 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bncqt\" (UniqueName: \"kubernetes.io/projected/a74b118e-ee59-4e87-967d-c8332a898a4c-kube-api-access-bncqt\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-6pxwj\" (UID: \"a74b118e-ee59-4e87-967d-c8332a898a4c\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6pxwj" Feb 02 09:49:37 crc kubenswrapper[4764]: I0202 09:49:37.279570 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a74b118e-ee59-4e87-967d-c8332a898a4c-ssh-key-openstack-edpm-ipam\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-6pxwj\" (UID: \"a74b118e-ee59-4e87-967d-c8332a898a4c\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6pxwj" Feb 02 09:49:37 crc kubenswrapper[4764]: I0202 09:49:37.279639 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a74b118e-ee59-4e87-967d-c8332a898a4c-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-6pxwj\" (UID: \"a74b118e-ee59-4e87-967d-c8332a898a4c\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6pxwj" Feb 02 09:49:37 crc kubenswrapper[4764]: I0202 09:49:37.380301 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a74b118e-ee59-4e87-967d-c8332a898a4c-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-6pxwj\" (UID: \"a74b118e-ee59-4e87-967d-c8332a898a4c\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6pxwj" Feb 02 09:49:37 crc kubenswrapper[4764]: I0202 09:49:37.380389 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bncqt\" (UniqueName: \"kubernetes.io/projected/a74b118e-ee59-4e87-967d-c8332a898a4c-kube-api-access-bncqt\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-6pxwj\" (UID: \"a74b118e-ee59-4e87-967d-c8332a898a4c\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6pxwj" Feb 02 09:49:37 crc kubenswrapper[4764]: I0202 09:49:37.380431 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a74b118e-ee59-4e87-967d-c8332a898a4c-ssh-key-openstack-edpm-ipam\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-6pxwj\" (UID: \"a74b118e-ee59-4e87-967d-c8332a898a4c\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6pxwj" Feb 02 09:49:37 crc kubenswrapper[4764]: I0202 09:49:37.380463 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a74b118e-ee59-4e87-967d-c8332a898a4c-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-6pxwj\" (UID: \"a74b118e-ee59-4e87-967d-c8332a898a4c\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6pxwj" Feb 02 09:49:37 crc kubenswrapper[4764]: I0202 09:49:37.380497 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/a74b118e-ee59-4e87-967d-c8332a898a4c-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-6pxwj\" (UID: \"a74b118e-ee59-4e87-967d-c8332a898a4c\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6pxwj" Feb 02 09:49:37 crc kubenswrapper[4764]: I0202 09:49:37.380525 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/a74b118e-ee59-4e87-967d-c8332a898a4c-ceph\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-6pxwj\" (UID: \"a74b118e-ee59-4e87-967d-c8332a898a4c\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6pxwj" Feb 02 09:49:37 crc kubenswrapper[4764]: I0202 09:49:37.380540 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/a74b118e-ee59-4e87-967d-c8332a898a4c-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-6pxwj\" (UID: \"a74b118e-ee59-4e87-967d-c8332a898a4c\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6pxwj" Feb 02 09:49:37 crc kubenswrapper[4764]: I0202 09:49:37.385980 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a74b118e-ee59-4e87-967d-c8332a898a4c-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-6pxwj\" (UID: \"a74b118e-ee59-4e87-967d-c8332a898a4c\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6pxwj" Feb 02 09:49:37 crc kubenswrapper[4764]: I0202 09:49:37.386700 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a74b118e-ee59-4e87-967d-c8332a898a4c-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-6pxwj\" (UID: \"a74b118e-ee59-4e87-967d-c8332a898a4c\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6pxwj" Feb 02 09:49:37 crc kubenswrapper[4764]: I0202 09:49:37.387242 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/a74b118e-ee59-4e87-967d-c8332a898a4c-ceph\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-6pxwj\" (UID: \"a74b118e-ee59-4e87-967d-c8332a898a4c\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6pxwj" Feb 02 09:49:37 crc kubenswrapper[4764]: I0202 09:49:37.390248 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a74b118e-ee59-4e87-967d-c8332a898a4c-ssh-key-openstack-edpm-ipam\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-6pxwj\" (UID: \"a74b118e-ee59-4e87-967d-c8332a898a4c\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6pxwj" Feb 02 09:49:37 crc kubenswrapper[4764]: I0202 09:49:37.391332 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/a74b118e-ee59-4e87-967d-c8332a898a4c-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-6pxwj\" (UID: \"a74b118e-ee59-4e87-967d-c8332a898a4c\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6pxwj" Feb 02 09:49:37 crc kubenswrapper[4764]: I0202 09:49:37.394455 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/a74b118e-ee59-4e87-967d-c8332a898a4c-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-6pxwj\" (UID: \"a74b118e-ee59-4e87-967d-c8332a898a4c\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6pxwj" Feb 02 09:49:37 crc kubenswrapper[4764]: I0202 09:49:37.397839 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bncqt\" (UniqueName: \"kubernetes.io/projected/a74b118e-ee59-4e87-967d-c8332a898a4c-kube-api-access-bncqt\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-6pxwj\" (UID: \"a74b118e-ee59-4e87-967d-c8332a898a4c\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6pxwj" Feb 02 09:49:37 crc kubenswrapper[4764]: I0202 09:49:37.535674 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6pxwj" Feb 02 09:49:38 crc kubenswrapper[4764]: I0202 09:49:38.184813 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6pxwj"] Feb 02 09:49:39 crc kubenswrapper[4764]: I0202 09:49:39.100225 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6pxwj" event={"ID":"a74b118e-ee59-4e87-967d-c8332a898a4c","Type":"ContainerStarted","Data":"d523a5da7d25a72f9663a8818f5b159525a881c4de70be43a75111533b1ac694"} Feb 02 09:49:40 crc kubenswrapper[4764]: I0202 09:49:40.122778 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6pxwj" event={"ID":"a74b118e-ee59-4e87-967d-c8332a898a4c","Type":"ContainerStarted","Data":"6eaee8b00f43dfcf61a0002d2e9feb30ee4443245b56df0da14b61915aa649f0"} Feb 02 09:49:40 crc kubenswrapper[4764]: I0202 09:49:40.154969 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6pxwj" podStartSLOduration=2.276328891 podStartE2EDuration="3.154928074s" podCreationTimestamp="2026-02-02 09:49:37 +0000 UTC" firstStartedPulling="2026-02-02 09:49:38.204007891 +0000 UTC m=+2541.137731979" lastFinishedPulling="2026-02-02 09:49:39.082607064 +0000 UTC m=+2542.016331162" observedRunningTime="2026-02-02 09:49:40.151501794 +0000 UTC m=+2543.085225892" watchObservedRunningTime="2026-02-02 09:49:40.154928074 +0000 UTC m=+2543.088652172" Feb 02 09:49:41 crc kubenswrapper[4764]: I0202 09:49:41.825672 4764 scope.go:117] "RemoveContainer" containerID="be8fa5800ec42d1c141765dcfc6111f3ef96b05b6e8d5a9feaf2baa219897bcf" Feb 02 09:49:41 crc kubenswrapper[4764]: E0202 09:49:41.826341 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 09:49:56 crc kubenswrapper[4764]: I0202 09:49:56.825993 4764 scope.go:117] "RemoveContainer" containerID="be8fa5800ec42d1c141765dcfc6111f3ef96b05b6e8d5a9feaf2baa219897bcf" Feb 02 09:49:57 crc kubenswrapper[4764]: I0202 09:49:57.299676 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" event={"ID":"d192f670-9f9d-4539-9641-e4bed73acdd4","Type":"ContainerStarted","Data":"8055d380e91f36c3b5458616d2925cea740cbdb1b01f92fbfc5d98ef1fb69aaa"} Feb 02 09:50:43 crc kubenswrapper[4764]: I0202 09:50:43.216289 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-h8cvm"] Feb 02 09:50:43 crc kubenswrapper[4764]: I0202 09:50:43.220687 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-h8cvm" Feb 02 09:50:43 crc kubenswrapper[4764]: I0202 09:50:43.240968 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-h8cvm"] Feb 02 09:50:43 crc kubenswrapper[4764]: I0202 09:50:43.277270 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6c92a40a-b0f3-4a95-9a82-a302a837db10-catalog-content\") pod \"community-operators-h8cvm\" (UID: \"6c92a40a-b0f3-4a95-9a82-a302a837db10\") " pod="openshift-marketplace/community-operators-h8cvm" Feb 02 09:50:43 crc kubenswrapper[4764]: I0202 09:50:43.277334 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6c92a40a-b0f3-4a95-9a82-a302a837db10-utilities\") pod \"community-operators-h8cvm\" (UID: \"6c92a40a-b0f3-4a95-9a82-a302a837db10\") " pod="openshift-marketplace/community-operators-h8cvm" Feb 02 09:50:43 crc kubenswrapper[4764]: I0202 09:50:43.277366 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4pqm4\" (UniqueName: \"kubernetes.io/projected/6c92a40a-b0f3-4a95-9a82-a302a837db10-kube-api-access-4pqm4\") pod \"community-operators-h8cvm\" (UID: \"6c92a40a-b0f3-4a95-9a82-a302a837db10\") " pod="openshift-marketplace/community-operators-h8cvm" Feb 02 09:50:43 crc kubenswrapper[4764]: I0202 09:50:43.379941 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6c92a40a-b0f3-4a95-9a82-a302a837db10-catalog-content\") pod \"community-operators-h8cvm\" (UID: \"6c92a40a-b0f3-4a95-9a82-a302a837db10\") " pod="openshift-marketplace/community-operators-h8cvm" Feb 02 09:50:43 crc kubenswrapper[4764]: I0202 09:50:43.380220 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6c92a40a-b0f3-4a95-9a82-a302a837db10-utilities\") pod \"community-operators-h8cvm\" (UID: \"6c92a40a-b0f3-4a95-9a82-a302a837db10\") " pod="openshift-marketplace/community-operators-h8cvm" Feb 02 09:50:43 crc kubenswrapper[4764]: I0202 09:50:43.380307 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4pqm4\" (UniqueName: \"kubernetes.io/projected/6c92a40a-b0f3-4a95-9a82-a302a837db10-kube-api-access-4pqm4\") pod \"community-operators-h8cvm\" (UID: \"6c92a40a-b0f3-4a95-9a82-a302a837db10\") " pod="openshift-marketplace/community-operators-h8cvm" Feb 02 09:50:43 crc kubenswrapper[4764]: I0202 09:50:43.380889 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6c92a40a-b0f3-4a95-9a82-a302a837db10-catalog-content\") pod \"community-operators-h8cvm\" (UID: \"6c92a40a-b0f3-4a95-9a82-a302a837db10\") " pod="openshift-marketplace/community-operators-h8cvm" Feb 02 09:50:43 crc kubenswrapper[4764]: I0202 09:50:43.381116 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6c92a40a-b0f3-4a95-9a82-a302a837db10-utilities\") pod \"community-operators-h8cvm\" (UID: \"6c92a40a-b0f3-4a95-9a82-a302a837db10\") " pod="openshift-marketplace/community-operators-h8cvm" Feb 02 09:50:43 crc kubenswrapper[4764]: I0202 09:50:43.411644 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4pqm4\" (UniqueName: \"kubernetes.io/projected/6c92a40a-b0f3-4a95-9a82-a302a837db10-kube-api-access-4pqm4\") pod \"community-operators-h8cvm\" (UID: \"6c92a40a-b0f3-4a95-9a82-a302a837db10\") " pod="openshift-marketplace/community-operators-h8cvm" Feb 02 09:50:43 crc kubenswrapper[4764]: I0202 09:50:43.551528 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-h8cvm" Feb 02 09:50:44 crc kubenswrapper[4764]: I0202 09:50:44.072589 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-h8cvm"] Feb 02 09:50:44 crc kubenswrapper[4764]: I0202 09:50:44.779361 4764 generic.go:334] "Generic (PLEG): container finished" podID="6c92a40a-b0f3-4a95-9a82-a302a837db10" containerID="43407fb07a9a42b4248bb97c27c50f10c0149ff5fca520a27a7c2b3a2b1c1960" exitCode=0 Feb 02 09:50:44 crc kubenswrapper[4764]: I0202 09:50:44.779441 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h8cvm" event={"ID":"6c92a40a-b0f3-4a95-9a82-a302a837db10","Type":"ContainerDied","Data":"43407fb07a9a42b4248bb97c27c50f10c0149ff5fca520a27a7c2b3a2b1c1960"} Feb 02 09:50:44 crc kubenswrapper[4764]: I0202 09:50:44.779508 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h8cvm" event={"ID":"6c92a40a-b0f3-4a95-9a82-a302a837db10","Type":"ContainerStarted","Data":"db6e53616465335e5e53be6d1794f035ea98f1094cf163f93552035f360647d2"} Feb 02 09:50:44 crc kubenswrapper[4764]: I0202 09:50:44.783352 4764 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 02 09:50:46 crc kubenswrapper[4764]: I0202 09:50:46.794954 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h8cvm" event={"ID":"6c92a40a-b0f3-4a95-9a82-a302a837db10","Type":"ContainerStarted","Data":"1ffb4677bc25d5eb13ef2392072805a49fefd390eb682c99e53cc544dc60908f"} Feb 02 09:50:47 crc kubenswrapper[4764]: I0202 09:50:47.803292 4764 generic.go:334] "Generic (PLEG): container finished" podID="a74b118e-ee59-4e87-967d-c8332a898a4c" containerID="6eaee8b00f43dfcf61a0002d2e9feb30ee4443245b56df0da14b61915aa649f0" exitCode=0 Feb 02 09:50:47 crc kubenswrapper[4764]: I0202 09:50:47.803350 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6pxwj" event={"ID":"a74b118e-ee59-4e87-967d-c8332a898a4c","Type":"ContainerDied","Data":"6eaee8b00f43dfcf61a0002d2e9feb30ee4443245b56df0da14b61915aa649f0"} Feb 02 09:50:47 crc kubenswrapper[4764]: I0202 09:50:47.805104 4764 generic.go:334] "Generic (PLEG): container finished" podID="6c92a40a-b0f3-4a95-9a82-a302a837db10" containerID="1ffb4677bc25d5eb13ef2392072805a49fefd390eb682c99e53cc544dc60908f" exitCode=0 Feb 02 09:50:47 crc kubenswrapper[4764]: I0202 09:50:47.805129 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h8cvm" event={"ID":"6c92a40a-b0f3-4a95-9a82-a302a837db10","Type":"ContainerDied","Data":"1ffb4677bc25d5eb13ef2392072805a49fefd390eb682c99e53cc544dc60908f"} Feb 02 09:50:48 crc kubenswrapper[4764]: I0202 09:50:48.816215 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h8cvm" event={"ID":"6c92a40a-b0f3-4a95-9a82-a302a837db10","Type":"ContainerStarted","Data":"29e2e326a0b89ff609e1aca8841bb040e485d020877d2d68c3404307346946bf"} Feb 02 09:50:48 crc kubenswrapper[4764]: I0202 09:50:48.835518 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-h8cvm" podStartSLOduration=2.126247409 podStartE2EDuration="5.835499922s" podCreationTimestamp="2026-02-02 09:50:43 +0000 UTC" firstStartedPulling="2026-02-02 09:50:44.783076494 +0000 UTC m=+2607.716800592" lastFinishedPulling="2026-02-02 09:50:48.492329017 +0000 UTC m=+2611.426053105" observedRunningTime="2026-02-02 09:50:48.8315485 +0000 UTC m=+2611.765272588" watchObservedRunningTime="2026-02-02 09:50:48.835499922 +0000 UTC m=+2611.769224010" Feb 02 09:50:49 crc kubenswrapper[4764]: I0202 09:50:49.139287 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6pxwj" Feb 02 09:50:49 crc kubenswrapper[4764]: I0202 09:50:49.291224 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a74b118e-ee59-4e87-967d-c8332a898a4c-neutron-metadata-combined-ca-bundle\") pod \"a74b118e-ee59-4e87-967d-c8332a898a4c\" (UID: \"a74b118e-ee59-4e87-967d-c8332a898a4c\") " Feb 02 09:50:49 crc kubenswrapper[4764]: I0202 09:50:49.291309 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a74b118e-ee59-4e87-967d-c8332a898a4c-ssh-key-openstack-edpm-ipam\") pod \"a74b118e-ee59-4e87-967d-c8332a898a4c\" (UID: \"a74b118e-ee59-4e87-967d-c8332a898a4c\") " Feb 02 09:50:49 crc kubenswrapper[4764]: I0202 09:50:49.291340 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bncqt\" (UniqueName: \"kubernetes.io/projected/a74b118e-ee59-4e87-967d-c8332a898a4c-kube-api-access-bncqt\") pod \"a74b118e-ee59-4e87-967d-c8332a898a4c\" (UID: \"a74b118e-ee59-4e87-967d-c8332a898a4c\") " Feb 02 09:50:49 crc kubenswrapper[4764]: I0202 09:50:49.291408 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/a74b118e-ee59-4e87-967d-c8332a898a4c-neutron-ovn-metadata-agent-neutron-config-0\") pod \"a74b118e-ee59-4e87-967d-c8332a898a4c\" (UID: \"a74b118e-ee59-4e87-967d-c8332a898a4c\") " Feb 02 09:50:49 crc kubenswrapper[4764]: I0202 09:50:49.291436 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/a74b118e-ee59-4e87-967d-c8332a898a4c-nova-metadata-neutron-config-0\") pod \"a74b118e-ee59-4e87-967d-c8332a898a4c\" (UID: \"a74b118e-ee59-4e87-967d-c8332a898a4c\") " Feb 02 09:50:49 crc kubenswrapper[4764]: I0202 09:50:49.291471 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a74b118e-ee59-4e87-967d-c8332a898a4c-inventory\") pod \"a74b118e-ee59-4e87-967d-c8332a898a4c\" (UID: \"a74b118e-ee59-4e87-967d-c8332a898a4c\") " Feb 02 09:50:49 crc kubenswrapper[4764]: I0202 09:50:49.291525 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/a74b118e-ee59-4e87-967d-c8332a898a4c-ceph\") pod \"a74b118e-ee59-4e87-967d-c8332a898a4c\" (UID: \"a74b118e-ee59-4e87-967d-c8332a898a4c\") " Feb 02 09:50:49 crc kubenswrapper[4764]: I0202 09:50:49.311002 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a74b118e-ee59-4e87-967d-c8332a898a4c-kube-api-access-bncqt" (OuterVolumeSpecName: "kube-api-access-bncqt") pod "a74b118e-ee59-4e87-967d-c8332a898a4c" (UID: "a74b118e-ee59-4e87-967d-c8332a898a4c"). InnerVolumeSpecName "kube-api-access-bncqt". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:50:49 crc kubenswrapper[4764]: I0202 09:50:49.312018 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a74b118e-ee59-4e87-967d-c8332a898a4c-ceph" (OuterVolumeSpecName: "ceph") pod "a74b118e-ee59-4e87-967d-c8332a898a4c" (UID: "a74b118e-ee59-4e87-967d-c8332a898a4c"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:50:49 crc kubenswrapper[4764]: I0202 09:50:49.312061 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a74b118e-ee59-4e87-967d-c8332a898a4c-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "a74b118e-ee59-4e87-967d-c8332a898a4c" (UID: "a74b118e-ee59-4e87-967d-c8332a898a4c"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:50:49 crc kubenswrapper[4764]: I0202 09:50:49.322018 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a74b118e-ee59-4e87-967d-c8332a898a4c-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "a74b118e-ee59-4e87-967d-c8332a898a4c" (UID: "a74b118e-ee59-4e87-967d-c8332a898a4c"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:50:49 crc kubenswrapper[4764]: I0202 09:50:49.322638 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a74b118e-ee59-4e87-967d-c8332a898a4c-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "a74b118e-ee59-4e87-967d-c8332a898a4c" (UID: "a74b118e-ee59-4e87-967d-c8332a898a4c"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:50:49 crc kubenswrapper[4764]: I0202 09:50:49.338593 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a74b118e-ee59-4e87-967d-c8332a898a4c-inventory" (OuterVolumeSpecName: "inventory") pod "a74b118e-ee59-4e87-967d-c8332a898a4c" (UID: "a74b118e-ee59-4e87-967d-c8332a898a4c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:50:49 crc kubenswrapper[4764]: I0202 09:50:49.341047 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a74b118e-ee59-4e87-967d-c8332a898a4c-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "a74b118e-ee59-4e87-967d-c8332a898a4c" (UID: "a74b118e-ee59-4e87-967d-c8332a898a4c"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:50:49 crc kubenswrapper[4764]: I0202 09:50:49.393754 4764 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a74b118e-ee59-4e87-967d-c8332a898a4c-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 09:50:49 crc kubenswrapper[4764]: I0202 09:50:49.393780 4764 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a74b118e-ee59-4e87-967d-c8332a898a4c-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 02 09:50:49 crc kubenswrapper[4764]: I0202 09:50:49.393790 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bncqt\" (UniqueName: \"kubernetes.io/projected/a74b118e-ee59-4e87-967d-c8332a898a4c-kube-api-access-bncqt\") on node \"crc\" DevicePath \"\"" Feb 02 09:50:49 crc kubenswrapper[4764]: I0202 09:50:49.393799 4764 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/a74b118e-ee59-4e87-967d-c8332a898a4c-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Feb 02 09:50:49 crc kubenswrapper[4764]: I0202 09:50:49.393809 4764 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/a74b118e-ee59-4e87-967d-c8332a898a4c-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Feb 02 09:50:49 crc kubenswrapper[4764]: I0202 09:50:49.393818 4764 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a74b118e-ee59-4e87-967d-c8332a898a4c-inventory\") on node \"crc\" DevicePath \"\"" Feb 02 09:50:49 crc kubenswrapper[4764]: I0202 09:50:49.393827 4764 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/a74b118e-ee59-4e87-967d-c8332a898a4c-ceph\") on node \"crc\" DevicePath \"\"" Feb 02 09:50:49 crc kubenswrapper[4764]: I0202 09:50:49.842910 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6pxwj" Feb 02 09:50:49 crc kubenswrapper[4764]: I0202 09:50:49.888674 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6pxwj" event={"ID":"a74b118e-ee59-4e87-967d-c8332a898a4c","Type":"ContainerDied","Data":"d523a5da7d25a72f9663a8818f5b159525a881c4de70be43a75111533b1ac694"} Feb 02 09:50:49 crc kubenswrapper[4764]: I0202 09:50:49.888721 4764 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d523a5da7d25a72f9663a8818f5b159525a881c4de70be43a75111533b1ac694" Feb 02 09:50:50 crc kubenswrapper[4764]: E0202 09:50:50.030647 4764 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda74b118e_ee59_4e87_967d_c8332a898a4c.slice\": RecentStats: unable to find data in memory cache]" Feb 02 09:50:50 crc kubenswrapper[4764]: I0202 09:50:50.061000 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ccmw8"] Feb 02 09:50:50 crc kubenswrapper[4764]: E0202 09:50:50.062014 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a74b118e-ee59-4e87-967d-c8332a898a4c" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Feb 02 09:50:50 crc kubenswrapper[4764]: I0202 09:50:50.062105 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="a74b118e-ee59-4e87-967d-c8332a898a4c" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Feb 02 09:50:50 crc kubenswrapper[4764]: I0202 09:50:50.062451 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="a74b118e-ee59-4e87-967d-c8332a898a4c" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Feb 02 09:50:50 crc kubenswrapper[4764]: I0202 09:50:50.064199 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ccmw8"] Feb 02 09:50:50 crc kubenswrapper[4764]: I0202 09:50:50.064407 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ccmw8" Feb 02 09:50:50 crc kubenswrapper[4764]: I0202 09:50:50.073926 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 02 09:50:50 crc kubenswrapper[4764]: I0202 09:50:50.074181 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 02 09:50:50 crc kubenswrapper[4764]: I0202 09:50:50.074288 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Feb 02 09:50:50 crc kubenswrapper[4764]: I0202 09:50:50.075075 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-xksq5" Feb 02 09:50:50 crc kubenswrapper[4764]: I0202 09:50:50.075244 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Feb 02 09:50:50 crc kubenswrapper[4764]: I0202 09:50:50.079058 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 02 09:50:50 crc kubenswrapper[4764]: I0202 09:50:50.109998 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zskz8\" (UniqueName: \"kubernetes.io/projected/f232691b-63e5-4e8a-9659-a6591c7dbe61-kube-api-access-zskz8\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-ccmw8\" (UID: \"f232691b-63e5-4e8a-9659-a6591c7dbe61\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ccmw8" Feb 02 09:50:50 crc kubenswrapper[4764]: I0202 09:50:50.110381 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f232691b-63e5-4e8a-9659-a6591c7dbe61-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-ccmw8\" (UID: \"f232691b-63e5-4e8a-9659-a6591c7dbe61\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ccmw8" Feb 02 09:50:50 crc kubenswrapper[4764]: I0202 09:50:50.110522 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f232691b-63e5-4e8a-9659-a6591c7dbe61-ceph\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-ccmw8\" (UID: \"f232691b-63e5-4e8a-9659-a6591c7dbe61\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ccmw8" Feb 02 09:50:50 crc kubenswrapper[4764]: I0202 09:50:50.110637 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f232691b-63e5-4e8a-9659-a6591c7dbe61-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-ccmw8\" (UID: \"f232691b-63e5-4e8a-9659-a6591c7dbe61\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ccmw8" Feb 02 09:50:50 crc kubenswrapper[4764]: I0202 09:50:50.110740 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/f232691b-63e5-4e8a-9659-a6591c7dbe61-ssh-key-openstack-edpm-ipam\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-ccmw8\" (UID: \"f232691b-63e5-4e8a-9659-a6591c7dbe61\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ccmw8" Feb 02 09:50:50 crc kubenswrapper[4764]: I0202 09:50:50.110855 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/f232691b-63e5-4e8a-9659-a6591c7dbe61-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-ccmw8\" (UID: \"f232691b-63e5-4e8a-9659-a6591c7dbe61\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ccmw8" Feb 02 09:50:50 crc kubenswrapper[4764]: I0202 09:50:50.213288 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f232691b-63e5-4e8a-9659-a6591c7dbe61-ceph\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-ccmw8\" (UID: \"f232691b-63e5-4e8a-9659-a6591c7dbe61\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ccmw8" Feb 02 09:50:50 crc kubenswrapper[4764]: I0202 09:50:50.213675 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f232691b-63e5-4e8a-9659-a6591c7dbe61-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-ccmw8\" (UID: \"f232691b-63e5-4e8a-9659-a6591c7dbe61\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ccmw8" Feb 02 09:50:50 crc kubenswrapper[4764]: I0202 09:50:50.213729 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/f232691b-63e5-4e8a-9659-a6591c7dbe61-ssh-key-openstack-edpm-ipam\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-ccmw8\" (UID: \"f232691b-63e5-4e8a-9659-a6591c7dbe61\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ccmw8" Feb 02 09:50:50 crc kubenswrapper[4764]: I0202 09:50:50.213815 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/f232691b-63e5-4e8a-9659-a6591c7dbe61-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-ccmw8\" (UID: \"f232691b-63e5-4e8a-9659-a6591c7dbe61\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ccmw8" Feb 02 09:50:50 crc kubenswrapper[4764]: I0202 09:50:50.214005 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zskz8\" (UniqueName: \"kubernetes.io/projected/f232691b-63e5-4e8a-9659-a6591c7dbe61-kube-api-access-zskz8\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-ccmw8\" (UID: \"f232691b-63e5-4e8a-9659-a6591c7dbe61\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ccmw8" Feb 02 09:50:50 crc kubenswrapper[4764]: I0202 09:50:50.214100 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f232691b-63e5-4e8a-9659-a6591c7dbe61-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-ccmw8\" (UID: \"f232691b-63e5-4e8a-9659-a6591c7dbe61\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ccmw8" Feb 02 09:50:50 crc kubenswrapper[4764]: I0202 09:50:50.217987 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/f232691b-63e5-4e8a-9659-a6591c7dbe61-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-ccmw8\" (UID: \"f232691b-63e5-4e8a-9659-a6591c7dbe61\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ccmw8" Feb 02 09:50:50 crc kubenswrapper[4764]: I0202 09:50:50.218229 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f232691b-63e5-4e8a-9659-a6591c7dbe61-ceph\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-ccmw8\" (UID: \"f232691b-63e5-4e8a-9659-a6591c7dbe61\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ccmw8" Feb 02 09:50:50 crc kubenswrapper[4764]: I0202 09:50:50.220052 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/f232691b-63e5-4e8a-9659-a6591c7dbe61-ssh-key-openstack-edpm-ipam\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-ccmw8\" (UID: \"f232691b-63e5-4e8a-9659-a6591c7dbe61\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ccmw8" Feb 02 09:50:50 crc kubenswrapper[4764]: I0202 09:50:50.220531 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f232691b-63e5-4e8a-9659-a6591c7dbe61-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-ccmw8\" (UID: \"f232691b-63e5-4e8a-9659-a6591c7dbe61\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ccmw8" Feb 02 09:50:50 crc kubenswrapper[4764]: I0202 09:50:50.227399 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f232691b-63e5-4e8a-9659-a6591c7dbe61-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-ccmw8\" (UID: \"f232691b-63e5-4e8a-9659-a6591c7dbe61\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ccmw8" Feb 02 09:50:50 crc kubenswrapper[4764]: I0202 09:50:50.231172 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zskz8\" (UniqueName: \"kubernetes.io/projected/f232691b-63e5-4e8a-9659-a6591c7dbe61-kube-api-access-zskz8\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-ccmw8\" (UID: \"f232691b-63e5-4e8a-9659-a6591c7dbe61\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ccmw8" Feb 02 09:50:50 crc kubenswrapper[4764]: I0202 09:50:50.386919 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ccmw8" Feb 02 09:50:50 crc kubenswrapper[4764]: I0202 09:50:50.963652 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ccmw8"] Feb 02 09:50:51 crc kubenswrapper[4764]: I0202 09:50:51.870602 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ccmw8" event={"ID":"f232691b-63e5-4e8a-9659-a6591c7dbe61","Type":"ContainerStarted","Data":"12a84df6b158299a6d6b2bb5516dee742153afdbd44c739484dd554d2e9c4a25"} Feb 02 09:50:52 crc kubenswrapper[4764]: I0202 09:50:52.884484 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ccmw8" event={"ID":"f232691b-63e5-4e8a-9659-a6591c7dbe61","Type":"ContainerStarted","Data":"2cb018e3c0e98b8f27e7bcfcb6b833c345f54b2bf1c530fa08de2b47d270c7e3"} Feb 02 09:50:52 crc kubenswrapper[4764]: I0202 09:50:52.913012 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ccmw8" podStartSLOduration=3.389928502 podStartE2EDuration="3.912994432s" podCreationTimestamp="2026-02-02 09:50:49 +0000 UTC" firstStartedPulling="2026-02-02 09:50:50.975979504 +0000 UTC m=+2613.909703592" lastFinishedPulling="2026-02-02 09:50:51.499045424 +0000 UTC m=+2614.432769522" observedRunningTime="2026-02-02 09:50:52.910650777 +0000 UTC m=+2615.844374865" watchObservedRunningTime="2026-02-02 09:50:52.912994432 +0000 UTC m=+2615.846718520" Feb 02 09:50:53 crc kubenswrapper[4764]: I0202 09:50:53.552390 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-h8cvm" Feb 02 09:50:53 crc kubenswrapper[4764]: I0202 09:50:53.553391 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-h8cvm" Feb 02 09:50:53 crc kubenswrapper[4764]: I0202 09:50:53.605631 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-h8cvm" Feb 02 09:50:53 crc kubenswrapper[4764]: I0202 09:50:53.956606 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-h8cvm" Feb 02 09:50:54 crc kubenswrapper[4764]: I0202 09:50:54.011911 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-h8cvm"] Feb 02 09:50:55 crc kubenswrapper[4764]: I0202 09:50:55.910337 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-h8cvm" podUID="6c92a40a-b0f3-4a95-9a82-a302a837db10" containerName="registry-server" containerID="cri-o://29e2e326a0b89ff609e1aca8841bb040e485d020877d2d68c3404307346946bf" gracePeriod=2 Feb 02 09:50:56 crc kubenswrapper[4764]: I0202 09:50:56.528626 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-h8cvm" Feb 02 09:50:56 crc kubenswrapper[4764]: I0202 09:50:56.641296 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4pqm4\" (UniqueName: \"kubernetes.io/projected/6c92a40a-b0f3-4a95-9a82-a302a837db10-kube-api-access-4pqm4\") pod \"6c92a40a-b0f3-4a95-9a82-a302a837db10\" (UID: \"6c92a40a-b0f3-4a95-9a82-a302a837db10\") " Feb 02 09:50:56 crc kubenswrapper[4764]: I0202 09:50:56.641509 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6c92a40a-b0f3-4a95-9a82-a302a837db10-catalog-content\") pod \"6c92a40a-b0f3-4a95-9a82-a302a837db10\" (UID: \"6c92a40a-b0f3-4a95-9a82-a302a837db10\") " Feb 02 09:50:56 crc kubenswrapper[4764]: I0202 09:50:56.641534 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6c92a40a-b0f3-4a95-9a82-a302a837db10-utilities\") pod \"6c92a40a-b0f3-4a95-9a82-a302a837db10\" (UID: \"6c92a40a-b0f3-4a95-9a82-a302a837db10\") " Feb 02 09:50:56 crc kubenswrapper[4764]: I0202 09:50:56.642212 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6c92a40a-b0f3-4a95-9a82-a302a837db10-utilities" (OuterVolumeSpecName: "utilities") pod "6c92a40a-b0f3-4a95-9a82-a302a837db10" (UID: "6c92a40a-b0f3-4a95-9a82-a302a837db10"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 09:50:56 crc kubenswrapper[4764]: I0202 09:50:56.657165 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6c92a40a-b0f3-4a95-9a82-a302a837db10-kube-api-access-4pqm4" (OuterVolumeSpecName: "kube-api-access-4pqm4") pod "6c92a40a-b0f3-4a95-9a82-a302a837db10" (UID: "6c92a40a-b0f3-4a95-9a82-a302a837db10"). InnerVolumeSpecName "kube-api-access-4pqm4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:50:56 crc kubenswrapper[4764]: I0202 09:50:56.688387 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6c92a40a-b0f3-4a95-9a82-a302a837db10-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6c92a40a-b0f3-4a95-9a82-a302a837db10" (UID: "6c92a40a-b0f3-4a95-9a82-a302a837db10"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 09:50:56 crc kubenswrapper[4764]: I0202 09:50:56.743398 4764 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6c92a40a-b0f3-4a95-9a82-a302a837db10-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 09:50:56 crc kubenswrapper[4764]: I0202 09:50:56.743430 4764 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6c92a40a-b0f3-4a95-9a82-a302a837db10-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 09:50:56 crc kubenswrapper[4764]: I0202 09:50:56.743440 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4pqm4\" (UniqueName: \"kubernetes.io/projected/6c92a40a-b0f3-4a95-9a82-a302a837db10-kube-api-access-4pqm4\") on node \"crc\" DevicePath \"\"" Feb 02 09:50:56 crc kubenswrapper[4764]: I0202 09:50:56.931377 4764 generic.go:334] "Generic (PLEG): container finished" podID="6c92a40a-b0f3-4a95-9a82-a302a837db10" containerID="29e2e326a0b89ff609e1aca8841bb040e485d020877d2d68c3404307346946bf" exitCode=0 Feb 02 09:50:56 crc kubenswrapper[4764]: I0202 09:50:56.931457 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-h8cvm" Feb 02 09:50:56 crc kubenswrapper[4764]: I0202 09:50:56.931452 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h8cvm" event={"ID":"6c92a40a-b0f3-4a95-9a82-a302a837db10","Type":"ContainerDied","Data":"29e2e326a0b89ff609e1aca8841bb040e485d020877d2d68c3404307346946bf"} Feb 02 09:50:56 crc kubenswrapper[4764]: I0202 09:50:56.931747 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h8cvm" event={"ID":"6c92a40a-b0f3-4a95-9a82-a302a837db10","Type":"ContainerDied","Data":"db6e53616465335e5e53be6d1794f035ea98f1094cf163f93552035f360647d2"} Feb 02 09:50:56 crc kubenswrapper[4764]: I0202 09:50:56.931773 4764 scope.go:117] "RemoveContainer" containerID="29e2e326a0b89ff609e1aca8841bb040e485d020877d2d68c3404307346946bf" Feb 02 09:50:56 crc kubenswrapper[4764]: I0202 09:50:56.971094 4764 scope.go:117] "RemoveContainer" containerID="1ffb4677bc25d5eb13ef2392072805a49fefd390eb682c99e53cc544dc60908f" Feb 02 09:50:57 crc kubenswrapper[4764]: I0202 09:50:57.012856 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-h8cvm"] Feb 02 09:50:57 crc kubenswrapper[4764]: I0202 09:50:57.036801 4764 scope.go:117] "RemoveContainer" containerID="43407fb07a9a42b4248bb97c27c50f10c0149ff5fca520a27a7c2b3a2b1c1960" Feb 02 09:50:57 crc kubenswrapper[4764]: I0202 09:50:57.050299 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-h8cvm"] Feb 02 09:50:57 crc kubenswrapper[4764]: I0202 09:50:57.061512 4764 scope.go:117] "RemoveContainer" containerID="29e2e326a0b89ff609e1aca8841bb040e485d020877d2d68c3404307346946bf" Feb 02 09:50:57 crc kubenswrapper[4764]: E0202 09:50:57.062097 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"29e2e326a0b89ff609e1aca8841bb040e485d020877d2d68c3404307346946bf\": container with ID starting with 29e2e326a0b89ff609e1aca8841bb040e485d020877d2d68c3404307346946bf not found: ID does not exist" containerID="29e2e326a0b89ff609e1aca8841bb040e485d020877d2d68c3404307346946bf" Feb 02 09:50:57 crc kubenswrapper[4764]: I0202 09:50:57.062158 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"29e2e326a0b89ff609e1aca8841bb040e485d020877d2d68c3404307346946bf"} err="failed to get container status \"29e2e326a0b89ff609e1aca8841bb040e485d020877d2d68c3404307346946bf\": rpc error: code = NotFound desc = could not find container \"29e2e326a0b89ff609e1aca8841bb040e485d020877d2d68c3404307346946bf\": container with ID starting with 29e2e326a0b89ff609e1aca8841bb040e485d020877d2d68c3404307346946bf not found: ID does not exist" Feb 02 09:50:57 crc kubenswrapper[4764]: I0202 09:50:57.062189 4764 scope.go:117] "RemoveContainer" containerID="1ffb4677bc25d5eb13ef2392072805a49fefd390eb682c99e53cc544dc60908f" Feb 02 09:50:57 crc kubenswrapper[4764]: E0202 09:50:57.066095 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1ffb4677bc25d5eb13ef2392072805a49fefd390eb682c99e53cc544dc60908f\": container with ID starting with 1ffb4677bc25d5eb13ef2392072805a49fefd390eb682c99e53cc544dc60908f not found: ID does not exist" containerID="1ffb4677bc25d5eb13ef2392072805a49fefd390eb682c99e53cc544dc60908f" Feb 02 09:50:57 crc kubenswrapper[4764]: I0202 09:50:57.066152 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ffb4677bc25d5eb13ef2392072805a49fefd390eb682c99e53cc544dc60908f"} err="failed to get container status \"1ffb4677bc25d5eb13ef2392072805a49fefd390eb682c99e53cc544dc60908f\": rpc error: code = NotFound desc = could not find container \"1ffb4677bc25d5eb13ef2392072805a49fefd390eb682c99e53cc544dc60908f\": container with ID starting with 1ffb4677bc25d5eb13ef2392072805a49fefd390eb682c99e53cc544dc60908f not found: ID does not exist" Feb 02 09:50:57 crc kubenswrapper[4764]: I0202 09:50:57.066179 4764 scope.go:117] "RemoveContainer" containerID="43407fb07a9a42b4248bb97c27c50f10c0149ff5fca520a27a7c2b3a2b1c1960" Feb 02 09:50:57 crc kubenswrapper[4764]: E0202 09:50:57.066508 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"43407fb07a9a42b4248bb97c27c50f10c0149ff5fca520a27a7c2b3a2b1c1960\": container with ID starting with 43407fb07a9a42b4248bb97c27c50f10c0149ff5fca520a27a7c2b3a2b1c1960 not found: ID does not exist" containerID="43407fb07a9a42b4248bb97c27c50f10c0149ff5fca520a27a7c2b3a2b1c1960" Feb 02 09:50:57 crc kubenswrapper[4764]: I0202 09:50:57.066564 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"43407fb07a9a42b4248bb97c27c50f10c0149ff5fca520a27a7c2b3a2b1c1960"} err="failed to get container status \"43407fb07a9a42b4248bb97c27c50f10c0149ff5fca520a27a7c2b3a2b1c1960\": rpc error: code = NotFound desc = could not find container \"43407fb07a9a42b4248bb97c27c50f10c0149ff5fca520a27a7c2b3a2b1c1960\": container with ID starting with 43407fb07a9a42b4248bb97c27c50f10c0149ff5fca520a27a7c2b3a2b1c1960 not found: ID does not exist" Feb 02 09:50:57 crc kubenswrapper[4764]: I0202 09:50:57.839081 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6c92a40a-b0f3-4a95-9a82-a302a837db10" path="/var/lib/kubelet/pods/6c92a40a-b0f3-4a95-9a82-a302a837db10/volumes" Feb 02 09:51:43 crc kubenswrapper[4764]: I0202 09:51:43.024423 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-7fdcx"] Feb 02 09:51:43 crc kubenswrapper[4764]: E0202 09:51:43.025357 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c92a40a-b0f3-4a95-9a82-a302a837db10" containerName="extract-utilities" Feb 02 09:51:43 crc kubenswrapper[4764]: I0202 09:51:43.025373 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c92a40a-b0f3-4a95-9a82-a302a837db10" containerName="extract-utilities" Feb 02 09:51:43 crc kubenswrapper[4764]: E0202 09:51:43.025389 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c92a40a-b0f3-4a95-9a82-a302a837db10" containerName="registry-server" Feb 02 09:51:43 crc kubenswrapper[4764]: I0202 09:51:43.025397 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c92a40a-b0f3-4a95-9a82-a302a837db10" containerName="registry-server" Feb 02 09:51:43 crc kubenswrapper[4764]: E0202 09:51:43.025418 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c92a40a-b0f3-4a95-9a82-a302a837db10" containerName="extract-content" Feb 02 09:51:43 crc kubenswrapper[4764]: I0202 09:51:43.025424 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c92a40a-b0f3-4a95-9a82-a302a837db10" containerName="extract-content" Feb 02 09:51:43 crc kubenswrapper[4764]: I0202 09:51:43.025578 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="6c92a40a-b0f3-4a95-9a82-a302a837db10" containerName="registry-server" Feb 02 09:51:43 crc kubenswrapper[4764]: I0202 09:51:43.026815 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7fdcx" Feb 02 09:51:43 crc kubenswrapper[4764]: I0202 09:51:43.050095 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7fdcx"] Feb 02 09:51:43 crc kubenswrapper[4764]: I0202 09:51:43.102991 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9e2c6c0a-a8b6-4a12-9a41-b54621595d15-catalog-content\") pod \"redhat-operators-7fdcx\" (UID: \"9e2c6c0a-a8b6-4a12-9a41-b54621595d15\") " pod="openshift-marketplace/redhat-operators-7fdcx" Feb 02 09:51:43 crc kubenswrapper[4764]: I0202 09:51:43.103092 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fdr6t\" (UniqueName: \"kubernetes.io/projected/9e2c6c0a-a8b6-4a12-9a41-b54621595d15-kube-api-access-fdr6t\") pod \"redhat-operators-7fdcx\" (UID: \"9e2c6c0a-a8b6-4a12-9a41-b54621595d15\") " pod="openshift-marketplace/redhat-operators-7fdcx" Feb 02 09:51:43 crc kubenswrapper[4764]: I0202 09:51:43.103127 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9e2c6c0a-a8b6-4a12-9a41-b54621595d15-utilities\") pod \"redhat-operators-7fdcx\" (UID: \"9e2c6c0a-a8b6-4a12-9a41-b54621595d15\") " pod="openshift-marketplace/redhat-operators-7fdcx" Feb 02 09:51:43 crc kubenswrapper[4764]: I0202 09:51:43.204491 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fdr6t\" (UniqueName: \"kubernetes.io/projected/9e2c6c0a-a8b6-4a12-9a41-b54621595d15-kube-api-access-fdr6t\") pod \"redhat-operators-7fdcx\" (UID: \"9e2c6c0a-a8b6-4a12-9a41-b54621595d15\") " pod="openshift-marketplace/redhat-operators-7fdcx" Feb 02 09:51:43 crc kubenswrapper[4764]: I0202 09:51:43.204559 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9e2c6c0a-a8b6-4a12-9a41-b54621595d15-utilities\") pod \"redhat-operators-7fdcx\" (UID: \"9e2c6c0a-a8b6-4a12-9a41-b54621595d15\") " pod="openshift-marketplace/redhat-operators-7fdcx" Feb 02 09:51:43 crc kubenswrapper[4764]: I0202 09:51:43.204625 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9e2c6c0a-a8b6-4a12-9a41-b54621595d15-catalog-content\") pod \"redhat-operators-7fdcx\" (UID: \"9e2c6c0a-a8b6-4a12-9a41-b54621595d15\") " pod="openshift-marketplace/redhat-operators-7fdcx" Feb 02 09:51:43 crc kubenswrapper[4764]: I0202 09:51:43.205037 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9e2c6c0a-a8b6-4a12-9a41-b54621595d15-utilities\") pod \"redhat-operators-7fdcx\" (UID: \"9e2c6c0a-a8b6-4a12-9a41-b54621595d15\") " pod="openshift-marketplace/redhat-operators-7fdcx" Feb 02 09:51:43 crc kubenswrapper[4764]: I0202 09:51:43.205054 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9e2c6c0a-a8b6-4a12-9a41-b54621595d15-catalog-content\") pod \"redhat-operators-7fdcx\" (UID: \"9e2c6c0a-a8b6-4a12-9a41-b54621595d15\") " pod="openshift-marketplace/redhat-operators-7fdcx" Feb 02 09:51:43 crc kubenswrapper[4764]: I0202 09:51:43.230870 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fdr6t\" (UniqueName: \"kubernetes.io/projected/9e2c6c0a-a8b6-4a12-9a41-b54621595d15-kube-api-access-fdr6t\") pod \"redhat-operators-7fdcx\" (UID: \"9e2c6c0a-a8b6-4a12-9a41-b54621595d15\") " pod="openshift-marketplace/redhat-operators-7fdcx" Feb 02 09:51:43 crc kubenswrapper[4764]: I0202 09:51:43.355673 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7fdcx" Feb 02 09:51:43 crc kubenswrapper[4764]: I0202 09:51:43.628140 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7fdcx"] Feb 02 09:51:44 crc kubenswrapper[4764]: I0202 09:51:44.362262 4764 generic.go:334] "Generic (PLEG): container finished" podID="9e2c6c0a-a8b6-4a12-9a41-b54621595d15" containerID="d7fe69040cadb898ad91df346c5d186f7d2ba501aac66dd9cdd7693c1c6a37d6" exitCode=0 Feb 02 09:51:44 crc kubenswrapper[4764]: I0202 09:51:44.362300 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7fdcx" event={"ID":"9e2c6c0a-a8b6-4a12-9a41-b54621595d15","Type":"ContainerDied","Data":"d7fe69040cadb898ad91df346c5d186f7d2ba501aac66dd9cdd7693c1c6a37d6"} Feb 02 09:51:44 crc kubenswrapper[4764]: I0202 09:51:44.362325 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7fdcx" event={"ID":"9e2c6c0a-a8b6-4a12-9a41-b54621595d15","Type":"ContainerStarted","Data":"a77fa32a36ef6a781cf3c86ac367ab75c245cb1577f79e3fb3567913944dd53a"} Feb 02 09:51:45 crc kubenswrapper[4764]: I0202 09:51:45.372467 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7fdcx" event={"ID":"9e2c6c0a-a8b6-4a12-9a41-b54621595d15","Type":"ContainerStarted","Data":"bab989d32993290867760b867fc20026b0a27f9aeaa428f87362d34407c4e064"} Feb 02 09:51:50 crc kubenswrapper[4764]: I0202 09:51:50.455605 4764 generic.go:334] "Generic (PLEG): container finished" podID="9e2c6c0a-a8b6-4a12-9a41-b54621595d15" containerID="bab989d32993290867760b867fc20026b0a27f9aeaa428f87362d34407c4e064" exitCode=0 Feb 02 09:51:50 crc kubenswrapper[4764]: I0202 09:51:50.455734 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7fdcx" event={"ID":"9e2c6c0a-a8b6-4a12-9a41-b54621595d15","Type":"ContainerDied","Data":"bab989d32993290867760b867fc20026b0a27f9aeaa428f87362d34407c4e064"} Feb 02 09:51:51 crc kubenswrapper[4764]: I0202 09:51:51.466661 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7fdcx" event={"ID":"9e2c6c0a-a8b6-4a12-9a41-b54621595d15","Type":"ContainerStarted","Data":"92b212fb5750f25580330f3e39c1945df57ec2b9bd215093f389aa66fc4c5ecd"} Feb 02 09:51:53 crc kubenswrapper[4764]: I0202 09:51:53.356492 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-7fdcx" Feb 02 09:51:53 crc kubenswrapper[4764]: I0202 09:51:53.356746 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-7fdcx" Feb 02 09:51:54 crc kubenswrapper[4764]: I0202 09:51:54.405041 4764 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-7fdcx" podUID="9e2c6c0a-a8b6-4a12-9a41-b54621595d15" containerName="registry-server" probeResult="failure" output=< Feb 02 09:51:54 crc kubenswrapper[4764]: timeout: failed to connect service ":50051" within 1s Feb 02 09:51:54 crc kubenswrapper[4764]: > Feb 02 09:52:03 crc kubenswrapper[4764]: I0202 09:52:03.402438 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-7fdcx" Feb 02 09:52:03 crc kubenswrapper[4764]: I0202 09:52:03.433881 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-7fdcx" podStartSLOduration=14.904658899 podStartE2EDuration="21.433856533s" podCreationTimestamp="2026-02-02 09:51:42 +0000 UTC" firstStartedPulling="2026-02-02 09:51:44.363686471 +0000 UTC m=+2667.297410559" lastFinishedPulling="2026-02-02 09:51:50.892884075 +0000 UTC m=+2673.826608193" observedRunningTime="2026-02-02 09:51:51.495038346 +0000 UTC m=+2674.428762444" watchObservedRunningTime="2026-02-02 09:52:03.433856533 +0000 UTC m=+2686.367580641" Feb 02 09:52:03 crc kubenswrapper[4764]: I0202 09:52:03.470774 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-7fdcx" Feb 02 09:52:03 crc kubenswrapper[4764]: I0202 09:52:03.645277 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7fdcx"] Feb 02 09:52:04 crc kubenswrapper[4764]: I0202 09:52:04.589268 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-7fdcx" podUID="9e2c6c0a-a8b6-4a12-9a41-b54621595d15" containerName="registry-server" containerID="cri-o://92b212fb5750f25580330f3e39c1945df57ec2b9bd215093f389aa66fc4c5ecd" gracePeriod=2 Feb 02 09:52:05 crc kubenswrapper[4764]: I0202 09:52:05.022842 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7fdcx" Feb 02 09:52:05 crc kubenswrapper[4764]: I0202 09:52:05.184236 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9e2c6c0a-a8b6-4a12-9a41-b54621595d15-utilities\") pod \"9e2c6c0a-a8b6-4a12-9a41-b54621595d15\" (UID: \"9e2c6c0a-a8b6-4a12-9a41-b54621595d15\") " Feb 02 09:52:05 crc kubenswrapper[4764]: I0202 09:52:05.184333 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fdr6t\" (UniqueName: \"kubernetes.io/projected/9e2c6c0a-a8b6-4a12-9a41-b54621595d15-kube-api-access-fdr6t\") pod \"9e2c6c0a-a8b6-4a12-9a41-b54621595d15\" (UID: \"9e2c6c0a-a8b6-4a12-9a41-b54621595d15\") " Feb 02 09:52:05 crc kubenswrapper[4764]: I0202 09:52:05.184700 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9e2c6c0a-a8b6-4a12-9a41-b54621595d15-catalog-content\") pod \"9e2c6c0a-a8b6-4a12-9a41-b54621595d15\" (UID: \"9e2c6c0a-a8b6-4a12-9a41-b54621595d15\") " Feb 02 09:52:05 crc kubenswrapper[4764]: I0202 09:52:05.185556 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9e2c6c0a-a8b6-4a12-9a41-b54621595d15-utilities" (OuterVolumeSpecName: "utilities") pod "9e2c6c0a-a8b6-4a12-9a41-b54621595d15" (UID: "9e2c6c0a-a8b6-4a12-9a41-b54621595d15"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 09:52:05 crc kubenswrapper[4764]: I0202 09:52:05.186845 4764 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9e2c6c0a-a8b6-4a12-9a41-b54621595d15-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 09:52:05 crc kubenswrapper[4764]: I0202 09:52:05.192136 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e2c6c0a-a8b6-4a12-9a41-b54621595d15-kube-api-access-fdr6t" (OuterVolumeSpecName: "kube-api-access-fdr6t") pod "9e2c6c0a-a8b6-4a12-9a41-b54621595d15" (UID: "9e2c6c0a-a8b6-4a12-9a41-b54621595d15"). InnerVolumeSpecName "kube-api-access-fdr6t". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:52:05 crc kubenswrapper[4764]: I0202 09:52:05.298655 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fdr6t\" (UniqueName: \"kubernetes.io/projected/9e2c6c0a-a8b6-4a12-9a41-b54621595d15-kube-api-access-fdr6t\") on node \"crc\" DevicePath \"\"" Feb 02 09:52:05 crc kubenswrapper[4764]: I0202 09:52:05.350757 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9e2c6c0a-a8b6-4a12-9a41-b54621595d15-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9e2c6c0a-a8b6-4a12-9a41-b54621595d15" (UID: "9e2c6c0a-a8b6-4a12-9a41-b54621595d15"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 09:52:05 crc kubenswrapper[4764]: I0202 09:52:05.400721 4764 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9e2c6c0a-a8b6-4a12-9a41-b54621595d15-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 09:52:05 crc kubenswrapper[4764]: I0202 09:52:05.599674 4764 generic.go:334] "Generic (PLEG): container finished" podID="9e2c6c0a-a8b6-4a12-9a41-b54621595d15" containerID="92b212fb5750f25580330f3e39c1945df57ec2b9bd215093f389aa66fc4c5ecd" exitCode=0 Feb 02 09:52:05 crc kubenswrapper[4764]: I0202 09:52:05.599738 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7fdcx" event={"ID":"9e2c6c0a-a8b6-4a12-9a41-b54621595d15","Type":"ContainerDied","Data":"92b212fb5750f25580330f3e39c1945df57ec2b9bd215093f389aa66fc4c5ecd"} Feb 02 09:52:05 crc kubenswrapper[4764]: I0202 09:52:05.599779 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7fdcx" event={"ID":"9e2c6c0a-a8b6-4a12-9a41-b54621595d15","Type":"ContainerDied","Data":"a77fa32a36ef6a781cf3c86ac367ab75c245cb1577f79e3fb3567913944dd53a"} Feb 02 09:52:05 crc kubenswrapper[4764]: I0202 09:52:05.599810 4764 scope.go:117] "RemoveContainer" containerID="92b212fb5750f25580330f3e39c1945df57ec2b9bd215093f389aa66fc4c5ecd" Feb 02 09:52:05 crc kubenswrapper[4764]: I0202 09:52:05.600023 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7fdcx" Feb 02 09:52:05 crc kubenswrapper[4764]: I0202 09:52:05.638375 4764 scope.go:117] "RemoveContainer" containerID="bab989d32993290867760b867fc20026b0a27f9aeaa428f87362d34407c4e064" Feb 02 09:52:05 crc kubenswrapper[4764]: I0202 09:52:05.670302 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7fdcx"] Feb 02 09:52:05 crc kubenswrapper[4764]: I0202 09:52:05.673593 4764 scope.go:117] "RemoveContainer" containerID="d7fe69040cadb898ad91df346c5d186f7d2ba501aac66dd9cdd7693c1c6a37d6" Feb 02 09:52:05 crc kubenswrapper[4764]: I0202 09:52:05.682810 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-7fdcx"] Feb 02 09:52:05 crc kubenswrapper[4764]: I0202 09:52:05.726238 4764 scope.go:117] "RemoveContainer" containerID="92b212fb5750f25580330f3e39c1945df57ec2b9bd215093f389aa66fc4c5ecd" Feb 02 09:52:05 crc kubenswrapper[4764]: E0202 09:52:05.726859 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"92b212fb5750f25580330f3e39c1945df57ec2b9bd215093f389aa66fc4c5ecd\": container with ID starting with 92b212fb5750f25580330f3e39c1945df57ec2b9bd215093f389aa66fc4c5ecd not found: ID does not exist" containerID="92b212fb5750f25580330f3e39c1945df57ec2b9bd215093f389aa66fc4c5ecd" Feb 02 09:52:05 crc kubenswrapper[4764]: I0202 09:52:05.726911 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"92b212fb5750f25580330f3e39c1945df57ec2b9bd215093f389aa66fc4c5ecd"} err="failed to get container status \"92b212fb5750f25580330f3e39c1945df57ec2b9bd215093f389aa66fc4c5ecd\": rpc error: code = NotFound desc = could not find container \"92b212fb5750f25580330f3e39c1945df57ec2b9bd215093f389aa66fc4c5ecd\": container with ID starting with 92b212fb5750f25580330f3e39c1945df57ec2b9bd215093f389aa66fc4c5ecd not found: ID does not exist" Feb 02 09:52:05 crc kubenswrapper[4764]: I0202 09:52:05.727049 4764 scope.go:117] "RemoveContainer" containerID="bab989d32993290867760b867fc20026b0a27f9aeaa428f87362d34407c4e064" Feb 02 09:52:05 crc kubenswrapper[4764]: E0202 09:52:05.727876 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bab989d32993290867760b867fc20026b0a27f9aeaa428f87362d34407c4e064\": container with ID starting with bab989d32993290867760b867fc20026b0a27f9aeaa428f87362d34407c4e064 not found: ID does not exist" containerID="bab989d32993290867760b867fc20026b0a27f9aeaa428f87362d34407c4e064" Feb 02 09:52:05 crc kubenswrapper[4764]: I0202 09:52:05.728014 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bab989d32993290867760b867fc20026b0a27f9aeaa428f87362d34407c4e064"} err="failed to get container status \"bab989d32993290867760b867fc20026b0a27f9aeaa428f87362d34407c4e064\": rpc error: code = NotFound desc = could not find container \"bab989d32993290867760b867fc20026b0a27f9aeaa428f87362d34407c4e064\": container with ID starting with bab989d32993290867760b867fc20026b0a27f9aeaa428f87362d34407c4e064 not found: ID does not exist" Feb 02 09:52:05 crc kubenswrapper[4764]: I0202 09:52:05.728056 4764 scope.go:117] "RemoveContainer" containerID="d7fe69040cadb898ad91df346c5d186f7d2ba501aac66dd9cdd7693c1c6a37d6" Feb 02 09:52:05 crc kubenswrapper[4764]: E0202 09:52:05.728952 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d7fe69040cadb898ad91df346c5d186f7d2ba501aac66dd9cdd7693c1c6a37d6\": container with ID starting with d7fe69040cadb898ad91df346c5d186f7d2ba501aac66dd9cdd7693c1c6a37d6 not found: ID does not exist" containerID="d7fe69040cadb898ad91df346c5d186f7d2ba501aac66dd9cdd7693c1c6a37d6" Feb 02 09:52:05 crc kubenswrapper[4764]: I0202 09:52:05.729013 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d7fe69040cadb898ad91df346c5d186f7d2ba501aac66dd9cdd7693c1c6a37d6"} err="failed to get container status \"d7fe69040cadb898ad91df346c5d186f7d2ba501aac66dd9cdd7693c1c6a37d6\": rpc error: code = NotFound desc = could not find container \"d7fe69040cadb898ad91df346c5d186f7d2ba501aac66dd9cdd7693c1c6a37d6\": container with ID starting with d7fe69040cadb898ad91df346c5d186f7d2ba501aac66dd9cdd7693c1c6a37d6 not found: ID does not exist" Feb 02 09:52:05 crc kubenswrapper[4764]: I0202 09:52:05.843266 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9e2c6c0a-a8b6-4a12-9a41-b54621595d15" path="/var/lib/kubelet/pods/9e2c6c0a-a8b6-4a12-9a41-b54621595d15/volumes" Feb 02 09:52:13 crc kubenswrapper[4764]: I0202 09:52:13.523735 4764 patch_prober.go:28] interesting pod/machine-config-daemon-4ndm4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 09:52:13 crc kubenswrapper[4764]: I0202 09:52:13.524360 4764 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 09:52:43 crc kubenswrapper[4764]: I0202 09:52:43.523254 4764 patch_prober.go:28] interesting pod/machine-config-daemon-4ndm4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 09:52:43 crc kubenswrapper[4764]: I0202 09:52:43.524391 4764 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 09:53:13 crc kubenswrapper[4764]: I0202 09:53:13.522840 4764 patch_prober.go:28] interesting pod/machine-config-daemon-4ndm4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 09:53:13 crc kubenswrapper[4764]: I0202 09:53:13.523410 4764 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 09:53:13 crc kubenswrapper[4764]: I0202 09:53:13.523460 4764 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" Feb 02 09:53:13 crc kubenswrapper[4764]: I0202 09:53:13.524168 4764 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8055d380e91f36c3b5458616d2925cea740cbdb1b01f92fbfc5d98ef1fb69aaa"} pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 02 09:53:13 crc kubenswrapper[4764]: I0202 09:53:13.524215 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" containerName="machine-config-daemon" containerID="cri-o://8055d380e91f36c3b5458616d2925cea740cbdb1b01f92fbfc5d98ef1fb69aaa" gracePeriod=600 Feb 02 09:53:14 crc kubenswrapper[4764]: I0202 09:53:14.319078 4764 generic.go:334] "Generic (PLEG): container finished" podID="d192f670-9f9d-4539-9641-e4bed73acdd4" containerID="8055d380e91f36c3b5458616d2925cea740cbdb1b01f92fbfc5d98ef1fb69aaa" exitCode=0 Feb 02 09:53:14 crc kubenswrapper[4764]: I0202 09:53:14.319305 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" event={"ID":"d192f670-9f9d-4539-9641-e4bed73acdd4","Type":"ContainerDied","Data":"8055d380e91f36c3b5458616d2925cea740cbdb1b01f92fbfc5d98ef1fb69aaa"} Feb 02 09:53:14 crc kubenswrapper[4764]: I0202 09:53:14.319642 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" event={"ID":"d192f670-9f9d-4539-9641-e4bed73acdd4","Type":"ContainerStarted","Data":"db4ebd4a8dbee216b70cb175f017f1799f3ca01a887f602ad42b9f295964da5d"} Feb 02 09:53:14 crc kubenswrapper[4764]: I0202 09:53:14.319669 4764 scope.go:117] "RemoveContainer" containerID="be8fa5800ec42d1c141765dcfc6111f3ef96b05b6e8d5a9feaf2baa219897bcf" Feb 02 09:54:22 crc kubenswrapper[4764]: I0202 09:54:22.051265 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-smclx"] Feb 02 09:54:22 crc kubenswrapper[4764]: E0202 09:54:22.055012 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e2c6c0a-a8b6-4a12-9a41-b54621595d15" containerName="extract-content" Feb 02 09:54:22 crc kubenswrapper[4764]: I0202 09:54:22.055047 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e2c6c0a-a8b6-4a12-9a41-b54621595d15" containerName="extract-content" Feb 02 09:54:22 crc kubenswrapper[4764]: E0202 09:54:22.055080 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e2c6c0a-a8b6-4a12-9a41-b54621595d15" containerName="registry-server" Feb 02 09:54:22 crc kubenswrapper[4764]: I0202 09:54:22.055088 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e2c6c0a-a8b6-4a12-9a41-b54621595d15" containerName="registry-server" Feb 02 09:54:22 crc kubenswrapper[4764]: E0202 09:54:22.055120 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e2c6c0a-a8b6-4a12-9a41-b54621595d15" containerName="extract-utilities" Feb 02 09:54:22 crc kubenswrapper[4764]: I0202 09:54:22.055129 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e2c6c0a-a8b6-4a12-9a41-b54621595d15" containerName="extract-utilities" Feb 02 09:54:22 crc kubenswrapper[4764]: I0202 09:54:22.055344 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e2c6c0a-a8b6-4a12-9a41-b54621595d15" containerName="registry-server" Feb 02 09:54:22 crc kubenswrapper[4764]: I0202 09:54:22.056961 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-smclx" Feb 02 09:54:22 crc kubenswrapper[4764]: I0202 09:54:22.107297 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-smclx"] Feb 02 09:54:22 crc kubenswrapper[4764]: I0202 09:54:22.134725 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1506a594-0c9c-422e-ae6f-ae7a5111864b-catalog-content\") pod \"certified-operators-smclx\" (UID: \"1506a594-0c9c-422e-ae6f-ae7a5111864b\") " pod="openshift-marketplace/certified-operators-smclx" Feb 02 09:54:22 crc kubenswrapper[4764]: I0202 09:54:22.134831 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1506a594-0c9c-422e-ae6f-ae7a5111864b-utilities\") pod \"certified-operators-smclx\" (UID: \"1506a594-0c9c-422e-ae6f-ae7a5111864b\") " pod="openshift-marketplace/certified-operators-smclx" Feb 02 09:54:22 crc kubenswrapper[4764]: I0202 09:54:22.134889 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cg77z\" (UniqueName: \"kubernetes.io/projected/1506a594-0c9c-422e-ae6f-ae7a5111864b-kube-api-access-cg77z\") pod \"certified-operators-smclx\" (UID: \"1506a594-0c9c-422e-ae6f-ae7a5111864b\") " pod="openshift-marketplace/certified-operators-smclx" Feb 02 09:54:22 crc kubenswrapper[4764]: I0202 09:54:22.236761 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1506a594-0c9c-422e-ae6f-ae7a5111864b-catalog-content\") pod \"certified-operators-smclx\" (UID: \"1506a594-0c9c-422e-ae6f-ae7a5111864b\") " pod="openshift-marketplace/certified-operators-smclx" Feb 02 09:54:22 crc kubenswrapper[4764]: I0202 09:54:22.236879 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1506a594-0c9c-422e-ae6f-ae7a5111864b-utilities\") pod \"certified-operators-smclx\" (UID: \"1506a594-0c9c-422e-ae6f-ae7a5111864b\") " pod="openshift-marketplace/certified-operators-smclx" Feb 02 09:54:22 crc kubenswrapper[4764]: I0202 09:54:22.236912 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cg77z\" (UniqueName: \"kubernetes.io/projected/1506a594-0c9c-422e-ae6f-ae7a5111864b-kube-api-access-cg77z\") pod \"certified-operators-smclx\" (UID: \"1506a594-0c9c-422e-ae6f-ae7a5111864b\") " pod="openshift-marketplace/certified-operators-smclx" Feb 02 09:54:22 crc kubenswrapper[4764]: I0202 09:54:22.237299 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1506a594-0c9c-422e-ae6f-ae7a5111864b-catalog-content\") pod \"certified-operators-smclx\" (UID: \"1506a594-0c9c-422e-ae6f-ae7a5111864b\") " pod="openshift-marketplace/certified-operators-smclx" Feb 02 09:54:22 crc kubenswrapper[4764]: I0202 09:54:22.237421 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1506a594-0c9c-422e-ae6f-ae7a5111864b-utilities\") pod \"certified-operators-smclx\" (UID: \"1506a594-0c9c-422e-ae6f-ae7a5111864b\") " pod="openshift-marketplace/certified-operators-smclx" Feb 02 09:54:22 crc kubenswrapper[4764]: I0202 09:54:22.271041 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cg77z\" (UniqueName: \"kubernetes.io/projected/1506a594-0c9c-422e-ae6f-ae7a5111864b-kube-api-access-cg77z\") pod \"certified-operators-smclx\" (UID: \"1506a594-0c9c-422e-ae6f-ae7a5111864b\") " pod="openshift-marketplace/certified-operators-smclx" Feb 02 09:54:22 crc kubenswrapper[4764]: I0202 09:54:22.439366 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-smclx" Feb 02 09:54:22 crc kubenswrapper[4764]: I0202 09:54:22.963606 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-smclx"] Feb 02 09:54:22 crc kubenswrapper[4764]: I0202 09:54:22.999095 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-smclx" event={"ID":"1506a594-0c9c-422e-ae6f-ae7a5111864b","Type":"ContainerStarted","Data":"7ecf9ab61cb6ad7fef0952e7efa868fe2b89b0af7f214f535fe3c0589819e534"} Feb 02 09:54:24 crc kubenswrapper[4764]: I0202 09:54:24.010665 4764 generic.go:334] "Generic (PLEG): container finished" podID="1506a594-0c9c-422e-ae6f-ae7a5111864b" containerID="422e810bf5876c904aa377f5d29618b33c88d55a50c2351ade91a906b2ecd197" exitCode=0 Feb 02 09:54:24 crc kubenswrapper[4764]: I0202 09:54:24.010837 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-smclx" event={"ID":"1506a594-0c9c-422e-ae6f-ae7a5111864b","Type":"ContainerDied","Data":"422e810bf5876c904aa377f5d29618b33c88d55a50c2351ade91a906b2ecd197"} Feb 02 09:54:25 crc kubenswrapper[4764]: I0202 09:54:25.024900 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-smclx" event={"ID":"1506a594-0c9c-422e-ae6f-ae7a5111864b","Type":"ContainerStarted","Data":"9b4a19f842f5b24c053aa5b072f16485345e5ebf3a80544b144841ae0830961c"} Feb 02 09:54:28 crc kubenswrapper[4764]: I0202 09:54:28.059197 4764 generic.go:334] "Generic (PLEG): container finished" podID="1506a594-0c9c-422e-ae6f-ae7a5111864b" containerID="9b4a19f842f5b24c053aa5b072f16485345e5ebf3a80544b144841ae0830961c" exitCode=0 Feb 02 09:54:28 crc kubenswrapper[4764]: I0202 09:54:28.059286 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-smclx" event={"ID":"1506a594-0c9c-422e-ae6f-ae7a5111864b","Type":"ContainerDied","Data":"9b4a19f842f5b24c053aa5b072f16485345e5ebf3a80544b144841ae0830961c"} Feb 02 09:54:29 crc kubenswrapper[4764]: I0202 09:54:29.072677 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-smclx" event={"ID":"1506a594-0c9c-422e-ae6f-ae7a5111864b","Type":"ContainerStarted","Data":"a4ddc174a1358dd288b4de9f7fb1cdf212bed82d72a822d4f5e9c6145ad69643"} Feb 02 09:54:29 crc kubenswrapper[4764]: I0202 09:54:29.108754 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-smclx" podStartSLOduration=2.579265642 podStartE2EDuration="7.108729969s" podCreationTimestamp="2026-02-02 09:54:22 +0000 UTC" firstStartedPulling="2026-02-02 09:54:24.012966257 +0000 UTC m=+2826.946690345" lastFinishedPulling="2026-02-02 09:54:28.542430584 +0000 UTC m=+2831.476154672" observedRunningTime="2026-02-02 09:54:29.104858656 +0000 UTC m=+2832.038582794" watchObservedRunningTime="2026-02-02 09:54:29.108729969 +0000 UTC m=+2832.042454057" Feb 02 09:54:32 crc kubenswrapper[4764]: I0202 09:54:32.441267 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-smclx" Feb 02 09:54:32 crc kubenswrapper[4764]: I0202 09:54:32.442575 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-smclx" Feb 02 09:54:32 crc kubenswrapper[4764]: I0202 09:54:32.534215 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-smclx" Feb 02 09:54:33 crc kubenswrapper[4764]: I0202 09:54:33.194193 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-smclx" Feb 02 09:54:33 crc kubenswrapper[4764]: I0202 09:54:33.255118 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-smclx"] Feb 02 09:54:35 crc kubenswrapper[4764]: I0202 09:54:35.141777 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-smclx" podUID="1506a594-0c9c-422e-ae6f-ae7a5111864b" containerName="registry-server" containerID="cri-o://a4ddc174a1358dd288b4de9f7fb1cdf212bed82d72a822d4f5e9c6145ad69643" gracePeriod=2 Feb 02 09:54:36 crc kubenswrapper[4764]: I0202 09:54:36.090092 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-smclx" Feb 02 09:54:36 crc kubenswrapper[4764]: I0202 09:54:36.160373 4764 generic.go:334] "Generic (PLEG): container finished" podID="1506a594-0c9c-422e-ae6f-ae7a5111864b" containerID="a4ddc174a1358dd288b4de9f7fb1cdf212bed82d72a822d4f5e9c6145ad69643" exitCode=0 Feb 02 09:54:36 crc kubenswrapper[4764]: I0202 09:54:36.160412 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-smclx" event={"ID":"1506a594-0c9c-422e-ae6f-ae7a5111864b","Type":"ContainerDied","Data":"a4ddc174a1358dd288b4de9f7fb1cdf212bed82d72a822d4f5e9c6145ad69643"} Feb 02 09:54:36 crc kubenswrapper[4764]: I0202 09:54:36.160439 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-smclx" event={"ID":"1506a594-0c9c-422e-ae6f-ae7a5111864b","Type":"ContainerDied","Data":"7ecf9ab61cb6ad7fef0952e7efa868fe2b89b0af7f214f535fe3c0589819e534"} Feb 02 09:54:36 crc kubenswrapper[4764]: I0202 09:54:36.160456 4764 scope.go:117] "RemoveContainer" containerID="a4ddc174a1358dd288b4de9f7fb1cdf212bed82d72a822d4f5e9c6145ad69643" Feb 02 09:54:36 crc kubenswrapper[4764]: I0202 09:54:36.160588 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-smclx" Feb 02 09:54:36 crc kubenswrapper[4764]: I0202 09:54:36.183358 4764 scope.go:117] "RemoveContainer" containerID="9b4a19f842f5b24c053aa5b072f16485345e5ebf3a80544b144841ae0830961c" Feb 02 09:54:36 crc kubenswrapper[4764]: I0202 09:54:36.207801 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1506a594-0c9c-422e-ae6f-ae7a5111864b-catalog-content\") pod \"1506a594-0c9c-422e-ae6f-ae7a5111864b\" (UID: \"1506a594-0c9c-422e-ae6f-ae7a5111864b\") " Feb 02 09:54:36 crc kubenswrapper[4764]: I0202 09:54:36.207898 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cg77z\" (UniqueName: \"kubernetes.io/projected/1506a594-0c9c-422e-ae6f-ae7a5111864b-kube-api-access-cg77z\") pod \"1506a594-0c9c-422e-ae6f-ae7a5111864b\" (UID: \"1506a594-0c9c-422e-ae6f-ae7a5111864b\") " Feb 02 09:54:36 crc kubenswrapper[4764]: I0202 09:54:36.208089 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1506a594-0c9c-422e-ae6f-ae7a5111864b-utilities\") pod \"1506a594-0c9c-422e-ae6f-ae7a5111864b\" (UID: \"1506a594-0c9c-422e-ae6f-ae7a5111864b\") " Feb 02 09:54:36 crc kubenswrapper[4764]: I0202 09:54:36.208899 4764 scope.go:117] "RemoveContainer" containerID="422e810bf5876c904aa377f5d29618b33c88d55a50c2351ade91a906b2ecd197" Feb 02 09:54:36 crc kubenswrapper[4764]: I0202 09:54:36.215887 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1506a594-0c9c-422e-ae6f-ae7a5111864b-utilities" (OuterVolumeSpecName: "utilities") pod "1506a594-0c9c-422e-ae6f-ae7a5111864b" (UID: "1506a594-0c9c-422e-ae6f-ae7a5111864b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 09:54:36 crc kubenswrapper[4764]: I0202 09:54:36.217177 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1506a594-0c9c-422e-ae6f-ae7a5111864b-kube-api-access-cg77z" (OuterVolumeSpecName: "kube-api-access-cg77z") pod "1506a594-0c9c-422e-ae6f-ae7a5111864b" (UID: "1506a594-0c9c-422e-ae6f-ae7a5111864b"). InnerVolumeSpecName "kube-api-access-cg77z". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:54:36 crc kubenswrapper[4764]: I0202 09:54:36.255611 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1506a594-0c9c-422e-ae6f-ae7a5111864b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1506a594-0c9c-422e-ae6f-ae7a5111864b" (UID: "1506a594-0c9c-422e-ae6f-ae7a5111864b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 09:54:36 crc kubenswrapper[4764]: I0202 09:54:36.291191 4764 scope.go:117] "RemoveContainer" containerID="a4ddc174a1358dd288b4de9f7fb1cdf212bed82d72a822d4f5e9c6145ad69643" Feb 02 09:54:36 crc kubenswrapper[4764]: E0202 09:54:36.291739 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a4ddc174a1358dd288b4de9f7fb1cdf212bed82d72a822d4f5e9c6145ad69643\": container with ID starting with a4ddc174a1358dd288b4de9f7fb1cdf212bed82d72a822d4f5e9c6145ad69643 not found: ID does not exist" containerID="a4ddc174a1358dd288b4de9f7fb1cdf212bed82d72a822d4f5e9c6145ad69643" Feb 02 09:54:36 crc kubenswrapper[4764]: I0202 09:54:36.291790 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a4ddc174a1358dd288b4de9f7fb1cdf212bed82d72a822d4f5e9c6145ad69643"} err="failed to get container status \"a4ddc174a1358dd288b4de9f7fb1cdf212bed82d72a822d4f5e9c6145ad69643\": rpc error: code = NotFound desc = could not find container \"a4ddc174a1358dd288b4de9f7fb1cdf212bed82d72a822d4f5e9c6145ad69643\": container with ID starting with a4ddc174a1358dd288b4de9f7fb1cdf212bed82d72a822d4f5e9c6145ad69643 not found: ID does not exist" Feb 02 09:54:36 crc kubenswrapper[4764]: I0202 09:54:36.291825 4764 scope.go:117] "RemoveContainer" containerID="9b4a19f842f5b24c053aa5b072f16485345e5ebf3a80544b144841ae0830961c" Feb 02 09:54:36 crc kubenswrapper[4764]: E0202 09:54:36.292231 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9b4a19f842f5b24c053aa5b072f16485345e5ebf3a80544b144841ae0830961c\": container with ID starting with 9b4a19f842f5b24c053aa5b072f16485345e5ebf3a80544b144841ae0830961c not found: ID does not exist" containerID="9b4a19f842f5b24c053aa5b072f16485345e5ebf3a80544b144841ae0830961c" Feb 02 09:54:36 crc kubenswrapper[4764]: I0202 09:54:36.292278 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9b4a19f842f5b24c053aa5b072f16485345e5ebf3a80544b144841ae0830961c"} err="failed to get container status \"9b4a19f842f5b24c053aa5b072f16485345e5ebf3a80544b144841ae0830961c\": rpc error: code = NotFound desc = could not find container \"9b4a19f842f5b24c053aa5b072f16485345e5ebf3a80544b144841ae0830961c\": container with ID starting with 9b4a19f842f5b24c053aa5b072f16485345e5ebf3a80544b144841ae0830961c not found: ID does not exist" Feb 02 09:54:36 crc kubenswrapper[4764]: I0202 09:54:36.292300 4764 scope.go:117] "RemoveContainer" containerID="422e810bf5876c904aa377f5d29618b33c88d55a50c2351ade91a906b2ecd197" Feb 02 09:54:36 crc kubenswrapper[4764]: E0202 09:54:36.292494 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"422e810bf5876c904aa377f5d29618b33c88d55a50c2351ade91a906b2ecd197\": container with ID starting with 422e810bf5876c904aa377f5d29618b33c88d55a50c2351ade91a906b2ecd197 not found: ID does not exist" containerID="422e810bf5876c904aa377f5d29618b33c88d55a50c2351ade91a906b2ecd197" Feb 02 09:54:36 crc kubenswrapper[4764]: I0202 09:54:36.292520 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"422e810bf5876c904aa377f5d29618b33c88d55a50c2351ade91a906b2ecd197"} err="failed to get container status \"422e810bf5876c904aa377f5d29618b33c88d55a50c2351ade91a906b2ecd197\": rpc error: code = NotFound desc = could not find container \"422e810bf5876c904aa377f5d29618b33c88d55a50c2351ade91a906b2ecd197\": container with ID starting with 422e810bf5876c904aa377f5d29618b33c88d55a50c2351ade91a906b2ecd197 not found: ID does not exist" Feb 02 09:54:36 crc kubenswrapper[4764]: I0202 09:54:36.310202 4764 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1506a594-0c9c-422e-ae6f-ae7a5111864b-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 09:54:36 crc kubenswrapper[4764]: I0202 09:54:36.310413 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cg77z\" (UniqueName: \"kubernetes.io/projected/1506a594-0c9c-422e-ae6f-ae7a5111864b-kube-api-access-cg77z\") on node \"crc\" DevicePath \"\"" Feb 02 09:54:36 crc kubenswrapper[4764]: I0202 09:54:36.310482 4764 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1506a594-0c9c-422e-ae6f-ae7a5111864b-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 09:54:36 crc kubenswrapper[4764]: I0202 09:54:36.495434 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-smclx"] Feb 02 09:54:36 crc kubenswrapper[4764]: I0202 09:54:36.503834 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-smclx"] Feb 02 09:54:37 crc kubenswrapper[4764]: I0202 09:54:37.841656 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1506a594-0c9c-422e-ae6f-ae7a5111864b" path="/var/lib/kubelet/pods/1506a594-0c9c-422e-ae6f-ae7a5111864b/volumes" Feb 02 09:55:10 crc kubenswrapper[4764]: I0202 09:55:10.466110 4764 generic.go:334] "Generic (PLEG): container finished" podID="f232691b-63e5-4e8a-9659-a6591c7dbe61" containerID="2cb018e3c0e98b8f27e7bcfcb6b833c345f54b2bf1c530fa08de2b47d270c7e3" exitCode=0 Feb 02 09:55:10 crc kubenswrapper[4764]: I0202 09:55:10.466214 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ccmw8" event={"ID":"f232691b-63e5-4e8a-9659-a6591c7dbe61","Type":"ContainerDied","Data":"2cb018e3c0e98b8f27e7bcfcb6b833c345f54b2bf1c530fa08de2b47d270c7e3"} Feb 02 09:55:11 crc kubenswrapper[4764]: I0202 09:55:11.913543 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ccmw8" Feb 02 09:55:11 crc kubenswrapper[4764]: I0202 09:55:11.996517 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f232691b-63e5-4e8a-9659-a6591c7dbe61-inventory\") pod \"f232691b-63e5-4e8a-9659-a6591c7dbe61\" (UID: \"f232691b-63e5-4e8a-9659-a6591c7dbe61\") " Feb 02 09:55:11 crc kubenswrapper[4764]: I0202 09:55:11.996631 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/f232691b-63e5-4e8a-9659-a6591c7dbe61-libvirt-secret-0\") pod \"f232691b-63e5-4e8a-9659-a6591c7dbe61\" (UID: \"f232691b-63e5-4e8a-9659-a6591c7dbe61\") " Feb 02 09:55:11 crc kubenswrapper[4764]: I0202 09:55:11.996962 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f232691b-63e5-4e8a-9659-a6591c7dbe61-ceph\") pod \"f232691b-63e5-4e8a-9659-a6591c7dbe61\" (UID: \"f232691b-63e5-4e8a-9659-a6591c7dbe61\") " Feb 02 09:55:11 crc kubenswrapper[4764]: I0202 09:55:11.997010 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/f232691b-63e5-4e8a-9659-a6591c7dbe61-ssh-key-openstack-edpm-ipam\") pod \"f232691b-63e5-4e8a-9659-a6591c7dbe61\" (UID: \"f232691b-63e5-4e8a-9659-a6591c7dbe61\") " Feb 02 09:55:11 crc kubenswrapper[4764]: I0202 09:55:11.997089 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f232691b-63e5-4e8a-9659-a6591c7dbe61-libvirt-combined-ca-bundle\") pod \"f232691b-63e5-4e8a-9659-a6591c7dbe61\" (UID: \"f232691b-63e5-4e8a-9659-a6591c7dbe61\") " Feb 02 09:55:11 crc kubenswrapper[4764]: I0202 09:55:11.997178 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zskz8\" (UniqueName: \"kubernetes.io/projected/f232691b-63e5-4e8a-9659-a6591c7dbe61-kube-api-access-zskz8\") pod \"f232691b-63e5-4e8a-9659-a6591c7dbe61\" (UID: \"f232691b-63e5-4e8a-9659-a6591c7dbe61\") " Feb 02 09:55:12 crc kubenswrapper[4764]: I0202 09:55:12.002064 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f232691b-63e5-4e8a-9659-a6591c7dbe61-kube-api-access-zskz8" (OuterVolumeSpecName: "kube-api-access-zskz8") pod "f232691b-63e5-4e8a-9659-a6591c7dbe61" (UID: "f232691b-63e5-4e8a-9659-a6591c7dbe61"). InnerVolumeSpecName "kube-api-access-zskz8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:55:12 crc kubenswrapper[4764]: I0202 09:55:12.005917 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f232691b-63e5-4e8a-9659-a6591c7dbe61-ceph" (OuterVolumeSpecName: "ceph") pod "f232691b-63e5-4e8a-9659-a6591c7dbe61" (UID: "f232691b-63e5-4e8a-9659-a6591c7dbe61"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:55:12 crc kubenswrapper[4764]: I0202 09:55:12.007158 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f232691b-63e5-4e8a-9659-a6591c7dbe61-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "f232691b-63e5-4e8a-9659-a6591c7dbe61" (UID: "f232691b-63e5-4e8a-9659-a6591c7dbe61"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:55:12 crc kubenswrapper[4764]: I0202 09:55:12.021310 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f232691b-63e5-4e8a-9659-a6591c7dbe61-inventory" (OuterVolumeSpecName: "inventory") pod "f232691b-63e5-4e8a-9659-a6591c7dbe61" (UID: "f232691b-63e5-4e8a-9659-a6591c7dbe61"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:55:12 crc kubenswrapper[4764]: I0202 09:55:12.022120 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f232691b-63e5-4e8a-9659-a6591c7dbe61-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "f232691b-63e5-4e8a-9659-a6591c7dbe61" (UID: "f232691b-63e5-4e8a-9659-a6591c7dbe61"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:55:12 crc kubenswrapper[4764]: I0202 09:55:12.038807 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f232691b-63e5-4e8a-9659-a6591c7dbe61-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "f232691b-63e5-4e8a-9659-a6591c7dbe61" (UID: "f232691b-63e5-4e8a-9659-a6591c7dbe61"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:55:12 crc kubenswrapper[4764]: I0202 09:55:12.098563 4764 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f232691b-63e5-4e8a-9659-a6591c7dbe61-ceph\") on node \"crc\" DevicePath \"\"" Feb 02 09:55:12 crc kubenswrapper[4764]: I0202 09:55:12.098590 4764 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/f232691b-63e5-4e8a-9659-a6591c7dbe61-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 02 09:55:12 crc kubenswrapper[4764]: I0202 09:55:12.098603 4764 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f232691b-63e5-4e8a-9659-a6591c7dbe61-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 09:55:12 crc kubenswrapper[4764]: I0202 09:55:12.098612 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zskz8\" (UniqueName: \"kubernetes.io/projected/f232691b-63e5-4e8a-9659-a6591c7dbe61-kube-api-access-zskz8\") on node \"crc\" DevicePath \"\"" Feb 02 09:55:12 crc kubenswrapper[4764]: I0202 09:55:12.098621 4764 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f232691b-63e5-4e8a-9659-a6591c7dbe61-inventory\") on node \"crc\" DevicePath \"\"" Feb 02 09:55:12 crc kubenswrapper[4764]: I0202 09:55:12.098629 4764 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/f232691b-63e5-4e8a-9659-a6591c7dbe61-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Feb 02 09:55:12 crc kubenswrapper[4764]: I0202 09:55:12.486761 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ccmw8" event={"ID":"f232691b-63e5-4e8a-9659-a6591c7dbe61","Type":"ContainerDied","Data":"12a84df6b158299a6d6b2bb5516dee742153afdbd44c739484dd554d2e9c4a25"} Feb 02 09:55:12 crc kubenswrapper[4764]: I0202 09:55:12.486802 4764 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="12a84df6b158299a6d6b2bb5516dee742153afdbd44c739484dd554d2e9c4a25" Feb 02 09:55:12 crc kubenswrapper[4764]: I0202 09:55:12.486867 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ccmw8" Feb 02 09:55:12 crc kubenswrapper[4764]: I0202 09:55:12.600640 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-h4h49"] Feb 02 09:55:12 crc kubenswrapper[4764]: E0202 09:55:12.601078 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1506a594-0c9c-422e-ae6f-ae7a5111864b" containerName="extract-content" Feb 02 09:55:12 crc kubenswrapper[4764]: I0202 09:55:12.601101 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="1506a594-0c9c-422e-ae6f-ae7a5111864b" containerName="extract-content" Feb 02 09:55:12 crc kubenswrapper[4764]: E0202 09:55:12.601134 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1506a594-0c9c-422e-ae6f-ae7a5111864b" containerName="registry-server" Feb 02 09:55:12 crc kubenswrapper[4764]: I0202 09:55:12.601143 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="1506a594-0c9c-422e-ae6f-ae7a5111864b" containerName="registry-server" Feb 02 09:55:12 crc kubenswrapper[4764]: E0202 09:55:12.601158 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f232691b-63e5-4e8a-9659-a6591c7dbe61" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Feb 02 09:55:12 crc kubenswrapper[4764]: I0202 09:55:12.601167 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="f232691b-63e5-4e8a-9659-a6591c7dbe61" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Feb 02 09:55:12 crc kubenswrapper[4764]: E0202 09:55:12.601190 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1506a594-0c9c-422e-ae6f-ae7a5111864b" containerName="extract-utilities" Feb 02 09:55:12 crc kubenswrapper[4764]: I0202 09:55:12.601198 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="1506a594-0c9c-422e-ae6f-ae7a5111864b" containerName="extract-utilities" Feb 02 09:55:12 crc kubenswrapper[4764]: I0202 09:55:12.601419 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="1506a594-0c9c-422e-ae6f-ae7a5111864b" containerName="registry-server" Feb 02 09:55:12 crc kubenswrapper[4764]: I0202 09:55:12.601439 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="f232691b-63e5-4e8a-9659-a6591c7dbe61" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Feb 02 09:55:12 crc kubenswrapper[4764]: I0202 09:55:12.602140 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-h4h49" Feb 02 09:55:12 crc kubenswrapper[4764]: I0202 09:55:12.610324 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ceph-nova" Feb 02 09:55:12 crc kubenswrapper[4764]: I0202 09:55:12.610534 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Feb 02 09:55:12 crc kubenswrapper[4764]: I0202 09:55:12.610565 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Feb 02 09:55:12 crc kubenswrapper[4764]: I0202 09:55:12.610562 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 02 09:55:12 crc kubenswrapper[4764]: I0202 09:55:12.610871 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-xksq5" Feb 02 09:55:12 crc kubenswrapper[4764]: I0202 09:55:12.611082 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Feb 02 09:55:12 crc kubenswrapper[4764]: I0202 09:55:12.611327 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Feb 02 09:55:12 crc kubenswrapper[4764]: I0202 09:55:12.614047 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 02 09:55:12 crc kubenswrapper[4764]: I0202 09:55:12.617741 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-h4h49"] Feb 02 09:55:12 crc kubenswrapper[4764]: I0202 09:55:12.624393 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 02 09:55:12 crc kubenswrapper[4764]: I0202 09:55:12.654623 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/b5b2ecb9-0adf-489b-8a98-42cad0682d09-nova-migration-ssh-key-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-h4h49\" (UID: \"b5b2ecb9-0adf-489b-8a98-42cad0682d09\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-h4h49" Feb 02 09:55:12 crc kubenswrapper[4764]: I0202 09:55:12.654866 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/b5b2ecb9-0adf-489b-8a98-42cad0682d09-nova-cell1-compute-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-h4h49\" (UID: \"b5b2ecb9-0adf-489b-8a98-42cad0682d09\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-h4h49" Feb 02 09:55:12 crc kubenswrapper[4764]: I0202 09:55:12.654975 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b5b2ecb9-0adf-489b-8a98-42cad0682d09-ceph\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-h4h49\" (UID: \"b5b2ecb9-0adf-489b-8a98-42cad0682d09\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-h4h49" Feb 02 09:55:12 crc kubenswrapper[4764]: I0202 09:55:12.655080 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b5b2ecb9-0adf-489b-8a98-42cad0682d09-ssh-key-openstack-edpm-ipam\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-h4h49\" (UID: \"b5b2ecb9-0adf-489b-8a98-42cad0682d09\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-h4h49" Feb 02 09:55:12 crc kubenswrapper[4764]: I0202 09:55:12.655174 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/b5b2ecb9-0adf-489b-8a98-42cad0682d09-nova-cell1-compute-config-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-h4h49\" (UID: \"b5b2ecb9-0adf-489b-8a98-42cad0682d09\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-h4h49" Feb 02 09:55:12 crc kubenswrapper[4764]: I0202 09:55:12.655296 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b5b2ecb9-0adf-489b-8a98-42cad0682d09-inventory\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-h4h49\" (UID: \"b5b2ecb9-0adf-489b-8a98-42cad0682d09\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-h4h49" Feb 02 09:55:12 crc kubenswrapper[4764]: I0202 09:55:12.655390 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/b5b2ecb9-0adf-489b-8a98-42cad0682d09-nova-extra-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-h4h49\" (UID: \"b5b2ecb9-0adf-489b-8a98-42cad0682d09\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-h4h49" Feb 02 09:55:12 crc kubenswrapper[4764]: I0202 09:55:12.655622 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jrfmv\" (UniqueName: \"kubernetes.io/projected/b5b2ecb9-0adf-489b-8a98-42cad0682d09-kube-api-access-jrfmv\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-h4h49\" (UID: \"b5b2ecb9-0adf-489b-8a98-42cad0682d09\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-h4h49" Feb 02 09:55:12 crc kubenswrapper[4764]: I0202 09:55:12.655743 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph-nova-0\" (UniqueName: \"kubernetes.io/configmap/b5b2ecb9-0adf-489b-8a98-42cad0682d09-ceph-nova-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-h4h49\" (UID: \"b5b2ecb9-0adf-489b-8a98-42cad0682d09\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-h4h49" Feb 02 09:55:12 crc kubenswrapper[4764]: I0202 09:55:12.656027 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/b5b2ecb9-0adf-489b-8a98-42cad0682d09-nova-migration-ssh-key-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-h4h49\" (UID: \"b5b2ecb9-0adf-489b-8a98-42cad0682d09\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-h4h49" Feb 02 09:55:12 crc kubenswrapper[4764]: I0202 09:55:12.656078 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-custom-ceph-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5b2ecb9-0adf-489b-8a98-42cad0682d09-nova-custom-ceph-combined-ca-bundle\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-h4h49\" (UID: \"b5b2ecb9-0adf-489b-8a98-42cad0682d09\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-h4h49" Feb 02 09:55:12 crc kubenswrapper[4764]: I0202 09:55:12.758003 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/b5b2ecb9-0adf-489b-8a98-42cad0682d09-nova-extra-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-h4h49\" (UID: \"b5b2ecb9-0adf-489b-8a98-42cad0682d09\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-h4h49" Feb 02 09:55:12 crc kubenswrapper[4764]: I0202 09:55:12.758281 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jrfmv\" (UniqueName: \"kubernetes.io/projected/b5b2ecb9-0adf-489b-8a98-42cad0682d09-kube-api-access-jrfmv\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-h4h49\" (UID: \"b5b2ecb9-0adf-489b-8a98-42cad0682d09\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-h4h49" Feb 02 09:55:12 crc kubenswrapper[4764]: I0202 09:55:12.758327 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph-nova-0\" (UniqueName: \"kubernetes.io/configmap/b5b2ecb9-0adf-489b-8a98-42cad0682d09-ceph-nova-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-h4h49\" (UID: \"b5b2ecb9-0adf-489b-8a98-42cad0682d09\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-h4h49" Feb 02 09:55:12 crc kubenswrapper[4764]: I0202 09:55:12.758372 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/b5b2ecb9-0adf-489b-8a98-42cad0682d09-nova-migration-ssh-key-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-h4h49\" (UID: \"b5b2ecb9-0adf-489b-8a98-42cad0682d09\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-h4h49" Feb 02 09:55:12 crc kubenswrapper[4764]: I0202 09:55:12.758389 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-custom-ceph-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5b2ecb9-0adf-489b-8a98-42cad0682d09-nova-custom-ceph-combined-ca-bundle\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-h4h49\" (UID: \"b5b2ecb9-0adf-489b-8a98-42cad0682d09\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-h4h49" Feb 02 09:55:12 crc kubenswrapper[4764]: I0202 09:55:12.758421 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/b5b2ecb9-0adf-489b-8a98-42cad0682d09-nova-migration-ssh-key-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-h4h49\" (UID: \"b5b2ecb9-0adf-489b-8a98-42cad0682d09\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-h4h49" Feb 02 09:55:12 crc kubenswrapper[4764]: I0202 09:55:12.759130 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/b5b2ecb9-0adf-489b-8a98-42cad0682d09-nova-extra-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-h4h49\" (UID: \"b5b2ecb9-0adf-489b-8a98-42cad0682d09\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-h4h49" Feb 02 09:55:12 crc kubenswrapper[4764]: I0202 09:55:12.759305 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph-nova-0\" (UniqueName: \"kubernetes.io/configmap/b5b2ecb9-0adf-489b-8a98-42cad0682d09-ceph-nova-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-h4h49\" (UID: \"b5b2ecb9-0adf-489b-8a98-42cad0682d09\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-h4h49" Feb 02 09:55:12 crc kubenswrapper[4764]: I0202 09:55:12.760133 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/b5b2ecb9-0adf-489b-8a98-42cad0682d09-nova-cell1-compute-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-h4h49\" (UID: \"b5b2ecb9-0adf-489b-8a98-42cad0682d09\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-h4h49" Feb 02 09:55:12 crc kubenswrapper[4764]: I0202 09:55:12.760193 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b5b2ecb9-0adf-489b-8a98-42cad0682d09-ceph\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-h4h49\" (UID: \"b5b2ecb9-0adf-489b-8a98-42cad0682d09\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-h4h49" Feb 02 09:55:12 crc kubenswrapper[4764]: I0202 09:55:12.760236 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b5b2ecb9-0adf-489b-8a98-42cad0682d09-ssh-key-openstack-edpm-ipam\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-h4h49\" (UID: \"b5b2ecb9-0adf-489b-8a98-42cad0682d09\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-h4h49" Feb 02 09:55:12 crc kubenswrapper[4764]: I0202 09:55:12.760276 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/b5b2ecb9-0adf-489b-8a98-42cad0682d09-nova-cell1-compute-config-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-h4h49\" (UID: \"b5b2ecb9-0adf-489b-8a98-42cad0682d09\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-h4h49" Feb 02 09:55:12 crc kubenswrapper[4764]: I0202 09:55:12.760365 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b5b2ecb9-0adf-489b-8a98-42cad0682d09-inventory\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-h4h49\" (UID: \"b5b2ecb9-0adf-489b-8a98-42cad0682d09\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-h4h49" Feb 02 09:55:12 crc kubenswrapper[4764]: I0202 09:55:12.762872 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/b5b2ecb9-0adf-489b-8a98-42cad0682d09-nova-migration-ssh-key-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-h4h49\" (UID: \"b5b2ecb9-0adf-489b-8a98-42cad0682d09\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-h4h49" Feb 02 09:55:12 crc kubenswrapper[4764]: I0202 09:55:12.763289 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/b5b2ecb9-0adf-489b-8a98-42cad0682d09-nova-cell1-compute-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-h4h49\" (UID: \"b5b2ecb9-0adf-489b-8a98-42cad0682d09\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-h4h49" Feb 02 09:55:12 crc kubenswrapper[4764]: I0202 09:55:12.764067 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-custom-ceph-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5b2ecb9-0adf-489b-8a98-42cad0682d09-nova-custom-ceph-combined-ca-bundle\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-h4h49\" (UID: \"b5b2ecb9-0adf-489b-8a98-42cad0682d09\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-h4h49" Feb 02 09:55:12 crc kubenswrapper[4764]: I0202 09:55:12.765785 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b5b2ecb9-0adf-489b-8a98-42cad0682d09-ssh-key-openstack-edpm-ipam\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-h4h49\" (UID: \"b5b2ecb9-0adf-489b-8a98-42cad0682d09\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-h4h49" Feb 02 09:55:12 crc kubenswrapper[4764]: I0202 09:55:12.766703 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b5b2ecb9-0adf-489b-8a98-42cad0682d09-inventory\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-h4h49\" (UID: \"b5b2ecb9-0adf-489b-8a98-42cad0682d09\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-h4h49" Feb 02 09:55:12 crc kubenswrapper[4764]: I0202 09:55:12.767483 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/b5b2ecb9-0adf-489b-8a98-42cad0682d09-nova-cell1-compute-config-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-h4h49\" (UID: \"b5b2ecb9-0adf-489b-8a98-42cad0682d09\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-h4h49" Feb 02 09:55:12 crc kubenswrapper[4764]: I0202 09:55:12.768253 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b5b2ecb9-0adf-489b-8a98-42cad0682d09-ceph\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-h4h49\" (UID: \"b5b2ecb9-0adf-489b-8a98-42cad0682d09\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-h4h49" Feb 02 09:55:12 crc kubenswrapper[4764]: I0202 09:55:12.770321 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/b5b2ecb9-0adf-489b-8a98-42cad0682d09-nova-migration-ssh-key-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-h4h49\" (UID: \"b5b2ecb9-0adf-489b-8a98-42cad0682d09\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-h4h49" Feb 02 09:55:12 crc kubenswrapper[4764]: I0202 09:55:12.775138 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jrfmv\" (UniqueName: \"kubernetes.io/projected/b5b2ecb9-0adf-489b-8a98-42cad0682d09-kube-api-access-jrfmv\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-h4h49\" (UID: \"b5b2ecb9-0adf-489b-8a98-42cad0682d09\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-h4h49" Feb 02 09:55:12 crc kubenswrapper[4764]: I0202 09:55:12.924994 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-h4h49" Feb 02 09:55:13 crc kubenswrapper[4764]: I0202 09:55:13.552821 4764 patch_prober.go:28] interesting pod/machine-config-daemon-4ndm4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 09:55:13 crc kubenswrapper[4764]: I0202 09:55:13.553198 4764 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 09:55:13 crc kubenswrapper[4764]: I0202 09:55:13.565712 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-h4h49"] Feb 02 09:55:13 crc kubenswrapper[4764]: W0202 09:55:13.572746 4764 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb5b2ecb9_0adf_489b_8a98_42cad0682d09.slice/crio-3fae7a6f3747f6c1a04a48a95afb88cfcf4bc3a6264ff417a878f4de13e18f24 WatchSource:0}: Error finding container 3fae7a6f3747f6c1a04a48a95afb88cfcf4bc3a6264ff417a878f4de13e18f24: Status 404 returned error can't find the container with id 3fae7a6f3747f6c1a04a48a95afb88cfcf4bc3a6264ff417a878f4de13e18f24 Feb 02 09:55:14 crc kubenswrapper[4764]: I0202 09:55:14.507723 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-h4h49" event={"ID":"b5b2ecb9-0adf-489b-8a98-42cad0682d09","Type":"ContainerStarted","Data":"07e9aaec2f3424ba14737b634dfa12abab4530d230a736ff0e62feb7647c00e4"} Feb 02 09:55:14 crc kubenswrapper[4764]: I0202 09:55:14.508012 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-h4h49" event={"ID":"b5b2ecb9-0adf-489b-8a98-42cad0682d09","Type":"ContainerStarted","Data":"3fae7a6f3747f6c1a04a48a95afb88cfcf4bc3a6264ff417a878f4de13e18f24"} Feb 02 09:55:14 crc kubenswrapper[4764]: I0202 09:55:14.544399 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-h4h49" podStartSLOduration=2.039149742 podStartE2EDuration="2.54438002s" podCreationTimestamp="2026-02-02 09:55:12 +0000 UTC" firstStartedPulling="2026-02-02 09:55:13.575977375 +0000 UTC m=+2876.509701463" lastFinishedPulling="2026-02-02 09:55:14.081207613 +0000 UTC m=+2877.014931741" observedRunningTime="2026-02-02 09:55:14.535028261 +0000 UTC m=+2877.468752379" watchObservedRunningTime="2026-02-02 09:55:14.54438002 +0000 UTC m=+2877.478104108" Feb 02 09:55:43 crc kubenswrapper[4764]: I0202 09:55:43.523540 4764 patch_prober.go:28] interesting pod/machine-config-daemon-4ndm4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 09:55:43 crc kubenswrapper[4764]: I0202 09:55:43.524166 4764 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 09:56:03 crc kubenswrapper[4764]: I0202 09:56:03.065205 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-5s927"] Feb 02 09:56:03 crc kubenswrapper[4764]: I0202 09:56:03.069038 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5s927" Feb 02 09:56:03 crc kubenswrapper[4764]: I0202 09:56:03.089145 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5s927"] Feb 02 09:56:03 crc kubenswrapper[4764]: I0202 09:56:03.211406 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n6h94\" (UniqueName: \"kubernetes.io/projected/ff1fe6eb-a979-4a77-9d81-c60b464aa5b9-kube-api-access-n6h94\") pod \"redhat-marketplace-5s927\" (UID: \"ff1fe6eb-a979-4a77-9d81-c60b464aa5b9\") " pod="openshift-marketplace/redhat-marketplace-5s927" Feb 02 09:56:03 crc kubenswrapper[4764]: I0202 09:56:03.211586 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff1fe6eb-a979-4a77-9d81-c60b464aa5b9-catalog-content\") pod \"redhat-marketplace-5s927\" (UID: \"ff1fe6eb-a979-4a77-9d81-c60b464aa5b9\") " pod="openshift-marketplace/redhat-marketplace-5s927" Feb 02 09:56:03 crc kubenswrapper[4764]: I0202 09:56:03.211619 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff1fe6eb-a979-4a77-9d81-c60b464aa5b9-utilities\") pod \"redhat-marketplace-5s927\" (UID: \"ff1fe6eb-a979-4a77-9d81-c60b464aa5b9\") " pod="openshift-marketplace/redhat-marketplace-5s927" Feb 02 09:56:03 crc kubenswrapper[4764]: I0202 09:56:03.313545 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff1fe6eb-a979-4a77-9d81-c60b464aa5b9-catalog-content\") pod \"redhat-marketplace-5s927\" (UID: \"ff1fe6eb-a979-4a77-9d81-c60b464aa5b9\") " pod="openshift-marketplace/redhat-marketplace-5s927" Feb 02 09:56:03 crc kubenswrapper[4764]: I0202 09:56:03.313598 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff1fe6eb-a979-4a77-9d81-c60b464aa5b9-utilities\") pod \"redhat-marketplace-5s927\" (UID: \"ff1fe6eb-a979-4a77-9d81-c60b464aa5b9\") " pod="openshift-marketplace/redhat-marketplace-5s927" Feb 02 09:56:03 crc kubenswrapper[4764]: I0202 09:56:03.313763 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n6h94\" (UniqueName: \"kubernetes.io/projected/ff1fe6eb-a979-4a77-9d81-c60b464aa5b9-kube-api-access-n6h94\") pod \"redhat-marketplace-5s927\" (UID: \"ff1fe6eb-a979-4a77-9d81-c60b464aa5b9\") " pod="openshift-marketplace/redhat-marketplace-5s927" Feb 02 09:56:03 crc kubenswrapper[4764]: I0202 09:56:03.314145 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff1fe6eb-a979-4a77-9d81-c60b464aa5b9-catalog-content\") pod \"redhat-marketplace-5s927\" (UID: \"ff1fe6eb-a979-4a77-9d81-c60b464aa5b9\") " pod="openshift-marketplace/redhat-marketplace-5s927" Feb 02 09:56:03 crc kubenswrapper[4764]: I0202 09:56:03.314265 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff1fe6eb-a979-4a77-9d81-c60b464aa5b9-utilities\") pod \"redhat-marketplace-5s927\" (UID: \"ff1fe6eb-a979-4a77-9d81-c60b464aa5b9\") " pod="openshift-marketplace/redhat-marketplace-5s927" Feb 02 09:56:03 crc kubenswrapper[4764]: I0202 09:56:03.342499 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n6h94\" (UniqueName: \"kubernetes.io/projected/ff1fe6eb-a979-4a77-9d81-c60b464aa5b9-kube-api-access-n6h94\") pod \"redhat-marketplace-5s927\" (UID: \"ff1fe6eb-a979-4a77-9d81-c60b464aa5b9\") " pod="openshift-marketplace/redhat-marketplace-5s927" Feb 02 09:56:03 crc kubenswrapper[4764]: I0202 09:56:03.409636 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5s927" Feb 02 09:56:03 crc kubenswrapper[4764]: I0202 09:56:03.919643 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5s927"] Feb 02 09:56:04 crc kubenswrapper[4764]: I0202 09:56:04.024028 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5s927" event={"ID":"ff1fe6eb-a979-4a77-9d81-c60b464aa5b9","Type":"ContainerStarted","Data":"b3661f2f87fff2c3bcf6b7be2adcdff8e57c348e84b35f5c44ec09301dd4e850"} Feb 02 09:56:05 crc kubenswrapper[4764]: I0202 09:56:05.037451 4764 generic.go:334] "Generic (PLEG): container finished" podID="ff1fe6eb-a979-4a77-9d81-c60b464aa5b9" containerID="2de6df2457affad0acb063c9a3a48ce8ae2b252fcef8bc4c8e881d3430e6cf07" exitCode=0 Feb 02 09:56:05 crc kubenswrapper[4764]: I0202 09:56:05.037543 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5s927" event={"ID":"ff1fe6eb-a979-4a77-9d81-c60b464aa5b9","Type":"ContainerDied","Data":"2de6df2457affad0acb063c9a3a48ce8ae2b252fcef8bc4c8e881d3430e6cf07"} Feb 02 09:56:05 crc kubenswrapper[4764]: I0202 09:56:05.041764 4764 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 02 09:56:06 crc kubenswrapper[4764]: I0202 09:56:06.051130 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5s927" event={"ID":"ff1fe6eb-a979-4a77-9d81-c60b464aa5b9","Type":"ContainerStarted","Data":"271680b0520c122c913341eedfab000a5f45b71beab2dc0b9c39e063719b7dda"} Feb 02 09:56:07 crc kubenswrapper[4764]: I0202 09:56:07.064171 4764 generic.go:334] "Generic (PLEG): container finished" podID="ff1fe6eb-a979-4a77-9d81-c60b464aa5b9" containerID="271680b0520c122c913341eedfab000a5f45b71beab2dc0b9c39e063719b7dda" exitCode=0 Feb 02 09:56:07 crc kubenswrapper[4764]: I0202 09:56:07.064228 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5s927" event={"ID":"ff1fe6eb-a979-4a77-9d81-c60b464aa5b9","Type":"ContainerDied","Data":"271680b0520c122c913341eedfab000a5f45b71beab2dc0b9c39e063719b7dda"} Feb 02 09:56:08 crc kubenswrapper[4764]: I0202 09:56:08.077845 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5s927" event={"ID":"ff1fe6eb-a979-4a77-9d81-c60b464aa5b9","Type":"ContainerStarted","Data":"c74f9f3b0b86c9abda218e3fdddc387bd81607a34a8bd2846f7530f4816f4f45"} Feb 02 09:56:08 crc kubenswrapper[4764]: I0202 09:56:08.116901 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-5s927" podStartSLOduration=2.683744645 podStartE2EDuration="5.116873079s" podCreationTimestamp="2026-02-02 09:56:03 +0000 UTC" firstStartedPulling="2026-02-02 09:56:05.041356101 +0000 UTC m=+2927.975080189" lastFinishedPulling="2026-02-02 09:56:07.474484535 +0000 UTC m=+2930.408208623" observedRunningTime="2026-02-02 09:56:08.110132019 +0000 UTC m=+2931.043856137" watchObservedRunningTime="2026-02-02 09:56:08.116873079 +0000 UTC m=+2931.050597167" Feb 02 09:56:13 crc kubenswrapper[4764]: I0202 09:56:13.409809 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-5s927" Feb 02 09:56:13 crc kubenswrapper[4764]: I0202 09:56:13.410828 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-5s927" Feb 02 09:56:13 crc kubenswrapper[4764]: I0202 09:56:13.486264 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-5s927" Feb 02 09:56:13 crc kubenswrapper[4764]: I0202 09:56:13.523295 4764 patch_prober.go:28] interesting pod/machine-config-daemon-4ndm4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 09:56:13 crc kubenswrapper[4764]: I0202 09:56:13.523384 4764 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 09:56:13 crc kubenswrapper[4764]: I0202 09:56:13.523459 4764 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" Feb 02 09:56:13 crc kubenswrapper[4764]: I0202 09:56:13.524883 4764 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"db4ebd4a8dbee216b70cb175f017f1799f3ca01a887f602ad42b9f295964da5d"} pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 02 09:56:13 crc kubenswrapper[4764]: I0202 09:56:13.524998 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" containerName="machine-config-daemon" containerID="cri-o://db4ebd4a8dbee216b70cb175f017f1799f3ca01a887f602ad42b9f295964da5d" gracePeriod=600 Feb 02 09:56:13 crc kubenswrapper[4764]: E0202 09:56:13.663235 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 09:56:14 crc kubenswrapper[4764]: I0202 09:56:14.151080 4764 generic.go:334] "Generic (PLEG): container finished" podID="d192f670-9f9d-4539-9641-e4bed73acdd4" containerID="db4ebd4a8dbee216b70cb175f017f1799f3ca01a887f602ad42b9f295964da5d" exitCode=0 Feb 02 09:56:14 crc kubenswrapper[4764]: I0202 09:56:14.152181 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" event={"ID":"d192f670-9f9d-4539-9641-e4bed73acdd4","Type":"ContainerDied","Data":"db4ebd4a8dbee216b70cb175f017f1799f3ca01a887f602ad42b9f295964da5d"} Feb 02 09:56:14 crc kubenswrapper[4764]: I0202 09:56:14.152327 4764 scope.go:117] "RemoveContainer" containerID="8055d380e91f36c3b5458616d2925cea740cbdb1b01f92fbfc5d98ef1fb69aaa" Feb 02 09:56:14 crc kubenswrapper[4764]: I0202 09:56:14.153292 4764 scope.go:117] "RemoveContainer" containerID="db4ebd4a8dbee216b70cb175f017f1799f3ca01a887f602ad42b9f295964da5d" Feb 02 09:56:14 crc kubenswrapper[4764]: E0202 09:56:14.153774 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 09:56:14 crc kubenswrapper[4764]: I0202 09:56:14.251311 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-5s927" Feb 02 09:56:14 crc kubenswrapper[4764]: I0202 09:56:14.314427 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5s927"] Feb 02 09:56:16 crc kubenswrapper[4764]: I0202 09:56:16.193684 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-5s927" podUID="ff1fe6eb-a979-4a77-9d81-c60b464aa5b9" containerName="registry-server" containerID="cri-o://c74f9f3b0b86c9abda218e3fdddc387bd81607a34a8bd2846f7530f4816f4f45" gracePeriod=2 Feb 02 09:56:16 crc kubenswrapper[4764]: I0202 09:56:16.676962 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5s927" Feb 02 09:56:16 crc kubenswrapper[4764]: I0202 09:56:16.866522 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n6h94\" (UniqueName: \"kubernetes.io/projected/ff1fe6eb-a979-4a77-9d81-c60b464aa5b9-kube-api-access-n6h94\") pod \"ff1fe6eb-a979-4a77-9d81-c60b464aa5b9\" (UID: \"ff1fe6eb-a979-4a77-9d81-c60b464aa5b9\") " Feb 02 09:56:16 crc kubenswrapper[4764]: I0202 09:56:16.866844 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff1fe6eb-a979-4a77-9d81-c60b464aa5b9-utilities\") pod \"ff1fe6eb-a979-4a77-9d81-c60b464aa5b9\" (UID: \"ff1fe6eb-a979-4a77-9d81-c60b464aa5b9\") " Feb 02 09:56:16 crc kubenswrapper[4764]: I0202 09:56:16.867118 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff1fe6eb-a979-4a77-9d81-c60b464aa5b9-catalog-content\") pod \"ff1fe6eb-a979-4a77-9d81-c60b464aa5b9\" (UID: \"ff1fe6eb-a979-4a77-9d81-c60b464aa5b9\") " Feb 02 09:56:16 crc kubenswrapper[4764]: I0202 09:56:16.867680 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ff1fe6eb-a979-4a77-9d81-c60b464aa5b9-utilities" (OuterVolumeSpecName: "utilities") pod "ff1fe6eb-a979-4a77-9d81-c60b464aa5b9" (UID: "ff1fe6eb-a979-4a77-9d81-c60b464aa5b9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 09:56:16 crc kubenswrapper[4764]: I0202 09:56:16.867905 4764 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff1fe6eb-a979-4a77-9d81-c60b464aa5b9-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 09:56:16 crc kubenswrapper[4764]: I0202 09:56:16.889557 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff1fe6eb-a979-4a77-9d81-c60b464aa5b9-kube-api-access-n6h94" (OuterVolumeSpecName: "kube-api-access-n6h94") pod "ff1fe6eb-a979-4a77-9d81-c60b464aa5b9" (UID: "ff1fe6eb-a979-4a77-9d81-c60b464aa5b9"). InnerVolumeSpecName "kube-api-access-n6h94". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:56:16 crc kubenswrapper[4764]: I0202 09:56:16.890306 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ff1fe6eb-a979-4a77-9d81-c60b464aa5b9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ff1fe6eb-a979-4a77-9d81-c60b464aa5b9" (UID: "ff1fe6eb-a979-4a77-9d81-c60b464aa5b9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 09:56:16 crc kubenswrapper[4764]: I0202 09:56:16.970190 4764 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff1fe6eb-a979-4a77-9d81-c60b464aa5b9-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 09:56:16 crc kubenswrapper[4764]: I0202 09:56:16.970217 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n6h94\" (UniqueName: \"kubernetes.io/projected/ff1fe6eb-a979-4a77-9d81-c60b464aa5b9-kube-api-access-n6h94\") on node \"crc\" DevicePath \"\"" Feb 02 09:56:17 crc kubenswrapper[4764]: I0202 09:56:17.204320 4764 generic.go:334] "Generic (PLEG): container finished" podID="ff1fe6eb-a979-4a77-9d81-c60b464aa5b9" containerID="c74f9f3b0b86c9abda218e3fdddc387bd81607a34a8bd2846f7530f4816f4f45" exitCode=0 Feb 02 09:56:17 crc kubenswrapper[4764]: I0202 09:56:17.204371 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5s927" event={"ID":"ff1fe6eb-a979-4a77-9d81-c60b464aa5b9","Type":"ContainerDied","Data":"c74f9f3b0b86c9abda218e3fdddc387bd81607a34a8bd2846f7530f4816f4f45"} Feb 02 09:56:17 crc kubenswrapper[4764]: I0202 09:56:17.204407 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5s927" event={"ID":"ff1fe6eb-a979-4a77-9d81-c60b464aa5b9","Type":"ContainerDied","Data":"b3661f2f87fff2c3bcf6b7be2adcdff8e57c348e84b35f5c44ec09301dd4e850"} Feb 02 09:56:17 crc kubenswrapper[4764]: I0202 09:56:17.204433 4764 scope.go:117] "RemoveContainer" containerID="c74f9f3b0b86c9abda218e3fdddc387bd81607a34a8bd2846f7530f4816f4f45" Feb 02 09:56:17 crc kubenswrapper[4764]: I0202 09:56:17.204578 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5s927" Feb 02 09:56:17 crc kubenswrapper[4764]: I0202 09:56:17.242678 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5s927"] Feb 02 09:56:17 crc kubenswrapper[4764]: I0202 09:56:17.249647 4764 scope.go:117] "RemoveContainer" containerID="271680b0520c122c913341eedfab000a5f45b71beab2dc0b9c39e063719b7dda" Feb 02 09:56:17 crc kubenswrapper[4764]: I0202 09:56:17.251581 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-5s927"] Feb 02 09:56:17 crc kubenswrapper[4764]: I0202 09:56:17.277648 4764 scope.go:117] "RemoveContainer" containerID="2de6df2457affad0acb063c9a3a48ce8ae2b252fcef8bc4c8e881d3430e6cf07" Feb 02 09:56:17 crc kubenswrapper[4764]: I0202 09:56:17.316382 4764 scope.go:117] "RemoveContainer" containerID="c74f9f3b0b86c9abda218e3fdddc387bd81607a34a8bd2846f7530f4816f4f45" Feb 02 09:56:17 crc kubenswrapper[4764]: E0202 09:56:17.317172 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c74f9f3b0b86c9abda218e3fdddc387bd81607a34a8bd2846f7530f4816f4f45\": container with ID starting with c74f9f3b0b86c9abda218e3fdddc387bd81607a34a8bd2846f7530f4816f4f45 not found: ID does not exist" containerID="c74f9f3b0b86c9abda218e3fdddc387bd81607a34a8bd2846f7530f4816f4f45" Feb 02 09:56:17 crc kubenswrapper[4764]: I0202 09:56:17.317238 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c74f9f3b0b86c9abda218e3fdddc387bd81607a34a8bd2846f7530f4816f4f45"} err="failed to get container status \"c74f9f3b0b86c9abda218e3fdddc387bd81607a34a8bd2846f7530f4816f4f45\": rpc error: code = NotFound desc = could not find container \"c74f9f3b0b86c9abda218e3fdddc387bd81607a34a8bd2846f7530f4816f4f45\": container with ID starting with c74f9f3b0b86c9abda218e3fdddc387bd81607a34a8bd2846f7530f4816f4f45 not found: ID does not exist" Feb 02 09:56:17 crc kubenswrapper[4764]: I0202 09:56:17.317280 4764 scope.go:117] "RemoveContainer" containerID="271680b0520c122c913341eedfab000a5f45b71beab2dc0b9c39e063719b7dda" Feb 02 09:56:17 crc kubenswrapper[4764]: E0202 09:56:17.317886 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"271680b0520c122c913341eedfab000a5f45b71beab2dc0b9c39e063719b7dda\": container with ID starting with 271680b0520c122c913341eedfab000a5f45b71beab2dc0b9c39e063719b7dda not found: ID does not exist" containerID="271680b0520c122c913341eedfab000a5f45b71beab2dc0b9c39e063719b7dda" Feb 02 09:56:17 crc kubenswrapper[4764]: I0202 09:56:17.317927 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"271680b0520c122c913341eedfab000a5f45b71beab2dc0b9c39e063719b7dda"} err="failed to get container status \"271680b0520c122c913341eedfab000a5f45b71beab2dc0b9c39e063719b7dda\": rpc error: code = NotFound desc = could not find container \"271680b0520c122c913341eedfab000a5f45b71beab2dc0b9c39e063719b7dda\": container with ID starting with 271680b0520c122c913341eedfab000a5f45b71beab2dc0b9c39e063719b7dda not found: ID does not exist" Feb 02 09:56:17 crc kubenswrapper[4764]: I0202 09:56:17.317971 4764 scope.go:117] "RemoveContainer" containerID="2de6df2457affad0acb063c9a3a48ce8ae2b252fcef8bc4c8e881d3430e6cf07" Feb 02 09:56:17 crc kubenswrapper[4764]: E0202 09:56:17.318330 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2de6df2457affad0acb063c9a3a48ce8ae2b252fcef8bc4c8e881d3430e6cf07\": container with ID starting with 2de6df2457affad0acb063c9a3a48ce8ae2b252fcef8bc4c8e881d3430e6cf07 not found: ID does not exist" containerID="2de6df2457affad0acb063c9a3a48ce8ae2b252fcef8bc4c8e881d3430e6cf07" Feb 02 09:56:17 crc kubenswrapper[4764]: I0202 09:56:17.318588 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2de6df2457affad0acb063c9a3a48ce8ae2b252fcef8bc4c8e881d3430e6cf07"} err="failed to get container status \"2de6df2457affad0acb063c9a3a48ce8ae2b252fcef8bc4c8e881d3430e6cf07\": rpc error: code = NotFound desc = could not find container \"2de6df2457affad0acb063c9a3a48ce8ae2b252fcef8bc4c8e881d3430e6cf07\": container with ID starting with 2de6df2457affad0acb063c9a3a48ce8ae2b252fcef8bc4c8e881d3430e6cf07 not found: ID does not exist" Feb 02 09:56:17 crc kubenswrapper[4764]: I0202 09:56:17.838808 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ff1fe6eb-a979-4a77-9d81-c60b464aa5b9" path="/var/lib/kubelet/pods/ff1fe6eb-a979-4a77-9d81-c60b464aa5b9/volumes" Feb 02 09:56:26 crc kubenswrapper[4764]: I0202 09:56:26.826607 4764 scope.go:117] "RemoveContainer" containerID="db4ebd4a8dbee216b70cb175f017f1799f3ca01a887f602ad42b9f295964da5d" Feb 02 09:56:26 crc kubenswrapper[4764]: E0202 09:56:26.827481 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 09:56:38 crc kubenswrapper[4764]: I0202 09:56:38.827002 4764 scope.go:117] "RemoveContainer" containerID="db4ebd4a8dbee216b70cb175f017f1799f3ca01a887f602ad42b9f295964da5d" Feb 02 09:56:38 crc kubenswrapper[4764]: E0202 09:56:38.827915 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 09:56:49 crc kubenswrapper[4764]: I0202 09:56:49.826265 4764 scope.go:117] "RemoveContainer" containerID="db4ebd4a8dbee216b70cb175f017f1799f3ca01a887f602ad42b9f295964da5d" Feb 02 09:56:49 crc kubenswrapper[4764]: E0202 09:56:49.827046 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 09:57:01 crc kubenswrapper[4764]: I0202 09:57:01.826575 4764 scope.go:117] "RemoveContainer" containerID="db4ebd4a8dbee216b70cb175f017f1799f3ca01a887f602ad42b9f295964da5d" Feb 02 09:57:01 crc kubenswrapper[4764]: E0202 09:57:01.827628 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 09:57:15 crc kubenswrapper[4764]: I0202 09:57:15.826500 4764 scope.go:117] "RemoveContainer" containerID="db4ebd4a8dbee216b70cb175f017f1799f3ca01a887f602ad42b9f295964da5d" Feb 02 09:57:15 crc kubenswrapper[4764]: E0202 09:57:15.827296 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 09:57:30 crc kubenswrapper[4764]: I0202 09:57:30.826392 4764 scope.go:117] "RemoveContainer" containerID="db4ebd4a8dbee216b70cb175f017f1799f3ca01a887f602ad42b9f295964da5d" Feb 02 09:57:30 crc kubenswrapper[4764]: E0202 09:57:30.828666 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 09:57:43 crc kubenswrapper[4764]: I0202 09:57:43.826409 4764 scope.go:117] "RemoveContainer" containerID="db4ebd4a8dbee216b70cb175f017f1799f3ca01a887f602ad42b9f295964da5d" Feb 02 09:57:43 crc kubenswrapper[4764]: E0202 09:57:43.827716 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 09:57:58 crc kubenswrapper[4764]: I0202 09:57:58.825866 4764 scope.go:117] "RemoveContainer" containerID="db4ebd4a8dbee216b70cb175f017f1799f3ca01a887f602ad42b9f295964da5d" Feb 02 09:57:58 crc kubenswrapper[4764]: E0202 09:57:58.826701 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 09:58:02 crc kubenswrapper[4764]: I0202 09:58:02.174176 4764 generic.go:334] "Generic (PLEG): container finished" podID="b5b2ecb9-0adf-489b-8a98-42cad0682d09" containerID="07e9aaec2f3424ba14737b634dfa12abab4530d230a736ff0e62feb7647c00e4" exitCode=0 Feb 02 09:58:02 crc kubenswrapper[4764]: I0202 09:58:02.174313 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-h4h49" event={"ID":"b5b2ecb9-0adf-489b-8a98-42cad0682d09","Type":"ContainerDied","Data":"07e9aaec2f3424ba14737b634dfa12abab4530d230a736ff0e62feb7647c00e4"} Feb 02 09:58:03 crc kubenswrapper[4764]: I0202 09:58:03.601290 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-h4h49" Feb 02 09:58:03 crc kubenswrapper[4764]: I0202 09:58:03.743276 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/b5b2ecb9-0adf-489b-8a98-42cad0682d09-nova-cell1-compute-config-0\") pod \"b5b2ecb9-0adf-489b-8a98-42cad0682d09\" (UID: \"b5b2ecb9-0adf-489b-8a98-42cad0682d09\") " Feb 02 09:58:03 crc kubenswrapper[4764]: I0202 09:58:03.743380 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/b5b2ecb9-0adf-489b-8a98-42cad0682d09-nova-migration-ssh-key-1\") pod \"b5b2ecb9-0adf-489b-8a98-42cad0682d09\" (UID: \"b5b2ecb9-0adf-489b-8a98-42cad0682d09\") " Feb 02 09:58:03 crc kubenswrapper[4764]: I0202 09:58:03.743592 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b5b2ecb9-0adf-489b-8a98-42cad0682d09-ssh-key-openstack-edpm-ipam\") pod \"b5b2ecb9-0adf-489b-8a98-42cad0682d09\" (UID: \"b5b2ecb9-0adf-489b-8a98-42cad0682d09\") " Feb 02 09:58:03 crc kubenswrapper[4764]: I0202 09:58:03.743802 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b5b2ecb9-0adf-489b-8a98-42cad0682d09-inventory\") pod \"b5b2ecb9-0adf-489b-8a98-42cad0682d09\" (UID: \"b5b2ecb9-0adf-489b-8a98-42cad0682d09\") " Feb 02 09:58:03 crc kubenswrapper[4764]: I0202 09:58:03.743839 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/b5b2ecb9-0adf-489b-8a98-42cad0682d09-nova-extra-config-0\") pod \"b5b2ecb9-0adf-489b-8a98-42cad0682d09\" (UID: \"b5b2ecb9-0adf-489b-8a98-42cad0682d09\") " Feb 02 09:58:03 crc kubenswrapper[4764]: I0202 09:58:03.743996 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/b5b2ecb9-0adf-489b-8a98-42cad0682d09-nova-cell1-compute-config-1\") pod \"b5b2ecb9-0adf-489b-8a98-42cad0682d09\" (UID: \"b5b2ecb9-0adf-489b-8a98-42cad0682d09\") " Feb 02 09:58:03 crc kubenswrapper[4764]: I0202 09:58:03.744022 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph-nova-0\" (UniqueName: \"kubernetes.io/configmap/b5b2ecb9-0adf-489b-8a98-42cad0682d09-ceph-nova-0\") pod \"b5b2ecb9-0adf-489b-8a98-42cad0682d09\" (UID: \"b5b2ecb9-0adf-489b-8a98-42cad0682d09\") " Feb 02 09:58:03 crc kubenswrapper[4764]: I0202 09:58:03.744414 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b5b2ecb9-0adf-489b-8a98-42cad0682d09-ceph\") pod \"b5b2ecb9-0adf-489b-8a98-42cad0682d09\" (UID: \"b5b2ecb9-0adf-489b-8a98-42cad0682d09\") " Feb 02 09:58:03 crc kubenswrapper[4764]: I0202 09:58:03.744440 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/b5b2ecb9-0adf-489b-8a98-42cad0682d09-nova-migration-ssh-key-0\") pod \"b5b2ecb9-0adf-489b-8a98-42cad0682d09\" (UID: \"b5b2ecb9-0adf-489b-8a98-42cad0682d09\") " Feb 02 09:58:03 crc kubenswrapper[4764]: I0202 09:58:03.744558 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jrfmv\" (UniqueName: \"kubernetes.io/projected/b5b2ecb9-0adf-489b-8a98-42cad0682d09-kube-api-access-jrfmv\") pod \"b5b2ecb9-0adf-489b-8a98-42cad0682d09\" (UID: \"b5b2ecb9-0adf-489b-8a98-42cad0682d09\") " Feb 02 09:58:03 crc kubenswrapper[4764]: I0202 09:58:03.744580 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-custom-ceph-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5b2ecb9-0adf-489b-8a98-42cad0682d09-nova-custom-ceph-combined-ca-bundle\") pod \"b5b2ecb9-0adf-489b-8a98-42cad0682d09\" (UID: \"b5b2ecb9-0adf-489b-8a98-42cad0682d09\") " Feb 02 09:58:03 crc kubenswrapper[4764]: I0202 09:58:03.756740 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5b2ecb9-0adf-489b-8a98-42cad0682d09-nova-custom-ceph-combined-ca-bundle" (OuterVolumeSpecName: "nova-custom-ceph-combined-ca-bundle") pod "b5b2ecb9-0adf-489b-8a98-42cad0682d09" (UID: "b5b2ecb9-0adf-489b-8a98-42cad0682d09"). InnerVolumeSpecName "nova-custom-ceph-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:58:03 crc kubenswrapper[4764]: I0202 09:58:03.767799 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5b2ecb9-0adf-489b-8a98-42cad0682d09-ceph" (OuterVolumeSpecName: "ceph") pod "b5b2ecb9-0adf-489b-8a98-42cad0682d09" (UID: "b5b2ecb9-0adf-489b-8a98-42cad0682d09"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:58:03 crc kubenswrapper[4764]: I0202 09:58:03.772195 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b5b2ecb9-0adf-489b-8a98-42cad0682d09-kube-api-access-jrfmv" (OuterVolumeSpecName: "kube-api-access-jrfmv") pod "b5b2ecb9-0adf-489b-8a98-42cad0682d09" (UID: "b5b2ecb9-0adf-489b-8a98-42cad0682d09"). InnerVolumeSpecName "kube-api-access-jrfmv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:58:03 crc kubenswrapper[4764]: I0202 09:58:03.792347 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5b2ecb9-0adf-489b-8a98-42cad0682d09-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "b5b2ecb9-0adf-489b-8a98-42cad0682d09" (UID: "b5b2ecb9-0adf-489b-8a98-42cad0682d09"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:58:03 crc kubenswrapper[4764]: I0202 09:58:03.794914 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5b2ecb9-0adf-489b-8a98-42cad0682d09-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "b5b2ecb9-0adf-489b-8a98-42cad0682d09" (UID: "b5b2ecb9-0adf-489b-8a98-42cad0682d09"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:58:03 crc kubenswrapper[4764]: I0202 09:58:03.803086 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b5b2ecb9-0adf-489b-8a98-42cad0682d09-ceph-nova-0" (OuterVolumeSpecName: "ceph-nova-0") pod "b5b2ecb9-0adf-489b-8a98-42cad0682d09" (UID: "b5b2ecb9-0adf-489b-8a98-42cad0682d09"). InnerVolumeSpecName "ceph-nova-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:58:03 crc kubenswrapper[4764]: I0202 09:58:03.811473 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5b2ecb9-0adf-489b-8a98-42cad0682d09-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "b5b2ecb9-0adf-489b-8a98-42cad0682d09" (UID: "b5b2ecb9-0adf-489b-8a98-42cad0682d09"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:58:03 crc kubenswrapper[4764]: I0202 09:58:03.817008 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5b2ecb9-0adf-489b-8a98-42cad0682d09-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "b5b2ecb9-0adf-489b-8a98-42cad0682d09" (UID: "b5b2ecb9-0adf-489b-8a98-42cad0682d09"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:58:03 crc kubenswrapper[4764]: I0202 09:58:03.822618 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b5b2ecb9-0adf-489b-8a98-42cad0682d09-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "b5b2ecb9-0adf-489b-8a98-42cad0682d09" (UID: "b5b2ecb9-0adf-489b-8a98-42cad0682d09"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:58:03 crc kubenswrapper[4764]: I0202 09:58:03.836322 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5b2ecb9-0adf-489b-8a98-42cad0682d09-inventory" (OuterVolumeSpecName: "inventory") pod "b5b2ecb9-0adf-489b-8a98-42cad0682d09" (UID: "b5b2ecb9-0adf-489b-8a98-42cad0682d09"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:58:03 crc kubenswrapper[4764]: I0202 09:58:03.841663 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5b2ecb9-0adf-489b-8a98-42cad0682d09-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "b5b2ecb9-0adf-489b-8a98-42cad0682d09" (UID: "b5b2ecb9-0adf-489b-8a98-42cad0682d09"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:58:03 crc kubenswrapper[4764]: I0202 09:58:03.846975 4764 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/b5b2ecb9-0adf-489b-8a98-42cad0682d09-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Feb 02 09:58:03 crc kubenswrapper[4764]: I0202 09:58:03.847000 4764 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/b5b2ecb9-0adf-489b-8a98-42cad0682d09-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Feb 02 09:58:03 crc kubenswrapper[4764]: I0202 09:58:03.847027 4764 reconciler_common.go:293] "Volume detached for volume \"ceph-nova-0\" (UniqueName: \"kubernetes.io/configmap/b5b2ecb9-0adf-489b-8a98-42cad0682d09-ceph-nova-0\") on node \"crc\" DevicePath \"\"" Feb 02 09:58:03 crc kubenswrapper[4764]: I0202 09:58:03.847040 4764 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b5b2ecb9-0adf-489b-8a98-42cad0682d09-ceph\") on node \"crc\" DevicePath \"\"" Feb 02 09:58:03 crc kubenswrapper[4764]: I0202 09:58:03.847050 4764 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/b5b2ecb9-0adf-489b-8a98-42cad0682d09-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Feb 02 09:58:03 crc kubenswrapper[4764]: I0202 09:58:03.847063 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jrfmv\" (UniqueName: \"kubernetes.io/projected/b5b2ecb9-0adf-489b-8a98-42cad0682d09-kube-api-access-jrfmv\") on node \"crc\" DevicePath \"\"" Feb 02 09:58:03 crc kubenswrapper[4764]: I0202 09:58:03.847073 4764 reconciler_common.go:293] "Volume detached for volume \"nova-custom-ceph-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5b2ecb9-0adf-489b-8a98-42cad0682d09-nova-custom-ceph-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 09:58:03 crc kubenswrapper[4764]: I0202 09:58:03.847082 4764 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/b5b2ecb9-0adf-489b-8a98-42cad0682d09-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Feb 02 09:58:03 crc kubenswrapper[4764]: I0202 09:58:03.847105 4764 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/b5b2ecb9-0adf-489b-8a98-42cad0682d09-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Feb 02 09:58:03 crc kubenswrapper[4764]: I0202 09:58:03.847114 4764 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b5b2ecb9-0adf-489b-8a98-42cad0682d09-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 02 09:58:03 crc kubenswrapper[4764]: I0202 09:58:03.847242 4764 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b5b2ecb9-0adf-489b-8a98-42cad0682d09-inventory\") on node \"crc\" DevicePath \"\"" Feb 02 09:58:04 crc kubenswrapper[4764]: I0202 09:58:04.193997 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-h4h49" event={"ID":"b5b2ecb9-0adf-489b-8a98-42cad0682d09","Type":"ContainerDied","Data":"3fae7a6f3747f6c1a04a48a95afb88cfcf4bc3a6264ff417a878f4de13e18f24"} Feb 02 09:58:04 crc kubenswrapper[4764]: I0202 09:58:04.194044 4764 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3fae7a6f3747f6c1a04a48a95afb88cfcf4bc3a6264ff417a878f4de13e18f24" Feb 02 09:58:04 crc kubenswrapper[4764]: I0202 09:58:04.194077 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-h4h49" Feb 02 09:58:13 crc kubenswrapper[4764]: I0202 09:58:13.826806 4764 scope.go:117] "RemoveContainer" containerID="db4ebd4a8dbee216b70cb175f017f1799f3ca01a887f602ad42b9f295964da5d" Feb 02 09:58:13 crc kubenswrapper[4764]: E0202 09:58:13.828418 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 09:58:19 crc kubenswrapper[4764]: I0202 09:58:19.217370 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-backup-0"] Feb 02 09:58:19 crc kubenswrapper[4764]: E0202 09:58:19.218222 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff1fe6eb-a979-4a77-9d81-c60b464aa5b9" containerName="extract-utilities" Feb 02 09:58:19 crc kubenswrapper[4764]: I0202 09:58:19.218235 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff1fe6eb-a979-4a77-9d81-c60b464aa5b9" containerName="extract-utilities" Feb 02 09:58:19 crc kubenswrapper[4764]: E0202 09:58:19.218262 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5b2ecb9-0adf-489b-8a98-42cad0682d09" containerName="nova-custom-ceph-edpm-deployment-openstack-edpm-ipam" Feb 02 09:58:19 crc kubenswrapper[4764]: I0202 09:58:19.218270 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5b2ecb9-0adf-489b-8a98-42cad0682d09" containerName="nova-custom-ceph-edpm-deployment-openstack-edpm-ipam" Feb 02 09:58:19 crc kubenswrapper[4764]: E0202 09:58:19.218277 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff1fe6eb-a979-4a77-9d81-c60b464aa5b9" containerName="registry-server" Feb 02 09:58:19 crc kubenswrapper[4764]: I0202 09:58:19.218283 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff1fe6eb-a979-4a77-9d81-c60b464aa5b9" containerName="registry-server" Feb 02 09:58:19 crc kubenswrapper[4764]: E0202 09:58:19.218300 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff1fe6eb-a979-4a77-9d81-c60b464aa5b9" containerName="extract-content" Feb 02 09:58:19 crc kubenswrapper[4764]: I0202 09:58:19.218306 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff1fe6eb-a979-4a77-9d81-c60b464aa5b9" containerName="extract-content" Feb 02 09:58:19 crc kubenswrapper[4764]: I0202 09:58:19.218448 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff1fe6eb-a979-4a77-9d81-c60b464aa5b9" containerName="registry-server" Feb 02 09:58:19 crc kubenswrapper[4764]: I0202 09:58:19.218472 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="b5b2ecb9-0adf-489b-8a98-42cad0682d09" containerName="nova-custom-ceph-edpm-deployment-openstack-edpm-ipam" Feb 02 09:58:19 crc kubenswrapper[4764]: I0202 09:58:19.219520 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-backup-0" Feb 02 09:58:19 crc kubenswrapper[4764]: I0202 09:58:19.224207 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Feb 02 09:58:19 crc kubenswrapper[4764]: I0202 09:58:19.224252 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-backup-config-data" Feb 02 09:58:19 crc kubenswrapper[4764]: I0202 09:58:19.233484 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-volume-volume1-0"] Feb 02 09:58:19 crc kubenswrapper[4764]: I0202 09:58:19.238103 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-volume-volume1-0" Feb 02 09:58:19 crc kubenswrapper[4764]: I0202 09:58:19.258106 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-volume-volume1-config-data" Feb 02 09:58:19 crc kubenswrapper[4764]: I0202 09:58:19.265823 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-volume-volume1-0"] Feb 02 09:58:19 crc kubenswrapper[4764]: I0202 09:58:19.270156 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/371fb9ce-9ba4-429b-a68c-62d2e12d688c-ceph\") pod \"cinder-backup-0\" (UID: \"371fb9ce-9ba4-429b-a68c-62d2e12d688c\") " pod="openstack/cinder-backup-0" Feb 02 09:58:19 crc kubenswrapper[4764]: I0202 09:58:19.270195 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/371fb9ce-9ba4-429b-a68c-62d2e12d688c-lib-modules\") pod \"cinder-backup-0\" (UID: \"371fb9ce-9ba4-429b-a68c-62d2e12d688c\") " pod="openstack/cinder-backup-0" Feb 02 09:58:19 crc kubenswrapper[4764]: I0202 09:58:19.270214 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/371fb9ce-9ba4-429b-a68c-62d2e12d688c-sys\") pod \"cinder-backup-0\" (UID: \"371fb9ce-9ba4-429b-a68c-62d2e12d688c\") " pod="openstack/cinder-backup-0" Feb 02 09:58:19 crc kubenswrapper[4764]: I0202 09:58:19.270281 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/1c67d8f7-21c3-456d-af99-c96d04090915-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"1c67d8f7-21c3-456d-af99-c96d04090915\") " pod="openstack/cinder-volume-volume1-0" Feb 02 09:58:19 crc kubenswrapper[4764]: I0202 09:58:19.270346 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1c67d8f7-21c3-456d-af99-c96d04090915-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"1c67d8f7-21c3-456d-af99-c96d04090915\") " pod="openstack/cinder-volume-volume1-0" Feb 02 09:58:19 crc kubenswrapper[4764]: I0202 09:58:19.270373 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/1c67d8f7-21c3-456d-af99-c96d04090915-run\") pod \"cinder-volume-volume1-0\" (UID: \"1c67d8f7-21c3-456d-af99-c96d04090915\") " pod="openstack/cinder-volume-volume1-0" Feb 02 09:58:19 crc kubenswrapper[4764]: I0202 09:58:19.270407 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/371fb9ce-9ba4-429b-a68c-62d2e12d688c-run\") pod \"cinder-backup-0\" (UID: \"371fb9ce-9ba4-429b-a68c-62d2e12d688c\") " pod="openstack/cinder-backup-0" Feb 02 09:58:19 crc kubenswrapper[4764]: I0202 09:58:19.270441 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/371fb9ce-9ba4-429b-a68c-62d2e12d688c-config-data\") pod \"cinder-backup-0\" (UID: \"371fb9ce-9ba4-429b-a68c-62d2e12d688c\") " pod="openstack/cinder-backup-0" Feb 02 09:58:19 crc kubenswrapper[4764]: I0202 09:58:19.270497 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/1c67d8f7-21c3-456d-af99-c96d04090915-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"1c67d8f7-21c3-456d-af99-c96d04090915\") " pod="openstack/cinder-volume-volume1-0" Feb 02 09:58:19 crc kubenswrapper[4764]: I0202 09:58:19.270533 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-66qr8\" (UniqueName: \"kubernetes.io/projected/1c67d8f7-21c3-456d-af99-c96d04090915-kube-api-access-66qr8\") pod \"cinder-volume-volume1-0\" (UID: \"1c67d8f7-21c3-456d-af99-c96d04090915\") " pod="openstack/cinder-volume-volume1-0" Feb 02 09:58:19 crc kubenswrapper[4764]: I0202 09:58:19.270582 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1c67d8f7-21c3-456d-af99-c96d04090915-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"1c67d8f7-21c3-456d-af99-c96d04090915\") " pod="openstack/cinder-volume-volume1-0" Feb 02 09:58:19 crc kubenswrapper[4764]: I0202 09:58:19.270605 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/1c67d8f7-21c3-456d-af99-c96d04090915-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"1c67d8f7-21c3-456d-af99-c96d04090915\") " pod="openstack/cinder-volume-volume1-0" Feb 02 09:58:19 crc kubenswrapper[4764]: I0202 09:58:19.270629 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1c67d8f7-21c3-456d-af99-c96d04090915-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"1c67d8f7-21c3-456d-af99-c96d04090915\") " pod="openstack/cinder-volume-volume1-0" Feb 02 09:58:19 crc kubenswrapper[4764]: I0202 09:58:19.270651 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/371fb9ce-9ba4-429b-a68c-62d2e12d688c-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"371fb9ce-9ba4-429b-a68c-62d2e12d688c\") " pod="openstack/cinder-backup-0" Feb 02 09:58:19 crc kubenswrapper[4764]: I0202 09:58:19.270671 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/371fb9ce-9ba4-429b-a68c-62d2e12d688c-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"371fb9ce-9ba4-429b-a68c-62d2e12d688c\") " pod="openstack/cinder-backup-0" Feb 02 09:58:19 crc kubenswrapper[4764]: I0202 09:58:19.270687 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/1c67d8f7-21c3-456d-af99-c96d04090915-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"1c67d8f7-21c3-456d-af99-c96d04090915\") " pod="openstack/cinder-volume-volume1-0" Feb 02 09:58:19 crc kubenswrapper[4764]: I0202 09:58:19.270767 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/371fb9ce-9ba4-429b-a68c-62d2e12d688c-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"371fb9ce-9ba4-429b-a68c-62d2e12d688c\") " pod="openstack/cinder-backup-0" Feb 02 09:58:19 crc kubenswrapper[4764]: I0202 09:58:19.270794 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nnhsg\" (UniqueName: \"kubernetes.io/projected/371fb9ce-9ba4-429b-a68c-62d2e12d688c-kube-api-access-nnhsg\") pod \"cinder-backup-0\" (UID: \"371fb9ce-9ba4-429b-a68c-62d2e12d688c\") " pod="openstack/cinder-backup-0" Feb 02 09:58:19 crc kubenswrapper[4764]: I0202 09:58:19.270829 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/371fb9ce-9ba4-429b-a68c-62d2e12d688c-scripts\") pod \"cinder-backup-0\" (UID: \"371fb9ce-9ba4-429b-a68c-62d2e12d688c\") " pod="openstack/cinder-backup-0" Feb 02 09:58:19 crc kubenswrapper[4764]: I0202 09:58:19.270870 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/1c67d8f7-21c3-456d-af99-c96d04090915-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"1c67d8f7-21c3-456d-af99-c96d04090915\") " pod="openstack/cinder-volume-volume1-0" Feb 02 09:58:19 crc kubenswrapper[4764]: I0202 09:58:19.270917 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/371fb9ce-9ba4-429b-a68c-62d2e12d688c-dev\") pod \"cinder-backup-0\" (UID: \"371fb9ce-9ba4-429b-a68c-62d2e12d688c\") " pod="openstack/cinder-backup-0" Feb 02 09:58:19 crc kubenswrapper[4764]: I0202 09:58:19.270947 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/371fb9ce-9ba4-429b-a68c-62d2e12d688c-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"371fb9ce-9ba4-429b-a68c-62d2e12d688c\") " pod="openstack/cinder-backup-0" Feb 02 09:58:19 crc kubenswrapper[4764]: I0202 09:58:19.270985 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1c67d8f7-21c3-456d-af99-c96d04090915-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"1c67d8f7-21c3-456d-af99-c96d04090915\") " pod="openstack/cinder-volume-volume1-0" Feb 02 09:58:19 crc kubenswrapper[4764]: I0202 09:58:19.271012 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/371fb9ce-9ba4-429b-a68c-62d2e12d688c-config-data-custom\") pod \"cinder-backup-0\" (UID: \"371fb9ce-9ba4-429b-a68c-62d2e12d688c\") " pod="openstack/cinder-backup-0" Feb 02 09:58:19 crc kubenswrapper[4764]: I0202 09:58:19.271058 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/371fb9ce-9ba4-429b-a68c-62d2e12d688c-etc-nvme\") pod \"cinder-backup-0\" (UID: \"371fb9ce-9ba4-429b-a68c-62d2e12d688c\") " pod="openstack/cinder-backup-0" Feb 02 09:58:19 crc kubenswrapper[4764]: I0202 09:58:19.271120 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/371fb9ce-9ba4-429b-a68c-62d2e12d688c-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"371fb9ce-9ba4-429b-a68c-62d2e12d688c\") " pod="openstack/cinder-backup-0" Feb 02 09:58:19 crc kubenswrapper[4764]: I0202 09:58:19.271152 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c67d8f7-21c3-456d-af99-c96d04090915-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"1c67d8f7-21c3-456d-af99-c96d04090915\") " pod="openstack/cinder-volume-volume1-0" Feb 02 09:58:19 crc kubenswrapper[4764]: I0202 09:58:19.271168 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/1c67d8f7-21c3-456d-af99-c96d04090915-dev\") pod \"cinder-volume-volume1-0\" (UID: \"1c67d8f7-21c3-456d-af99-c96d04090915\") " pod="openstack/cinder-volume-volume1-0" Feb 02 09:58:19 crc kubenswrapper[4764]: I0202 09:58:19.271278 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/1c67d8f7-21c3-456d-af99-c96d04090915-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"1c67d8f7-21c3-456d-af99-c96d04090915\") " pod="openstack/cinder-volume-volume1-0" Feb 02 09:58:19 crc kubenswrapper[4764]: I0202 09:58:19.271342 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/1c67d8f7-21c3-456d-af99-c96d04090915-sys\") pod \"cinder-volume-volume1-0\" (UID: \"1c67d8f7-21c3-456d-af99-c96d04090915\") " pod="openstack/cinder-volume-volume1-0" Feb 02 09:58:19 crc kubenswrapper[4764]: I0202 09:58:19.271361 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/371fb9ce-9ba4-429b-a68c-62d2e12d688c-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"371fb9ce-9ba4-429b-a68c-62d2e12d688c\") " pod="openstack/cinder-backup-0" Feb 02 09:58:19 crc kubenswrapper[4764]: I0202 09:58:19.271379 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/1c67d8f7-21c3-456d-af99-c96d04090915-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"1c67d8f7-21c3-456d-af99-c96d04090915\") " pod="openstack/cinder-volume-volume1-0" Feb 02 09:58:19 crc kubenswrapper[4764]: I0202 09:58:19.355975 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-backup-0"] Feb 02 09:58:19 crc kubenswrapper[4764]: I0202 09:58:19.372005 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/1c67d8f7-21c3-456d-af99-c96d04090915-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"1c67d8f7-21c3-456d-af99-c96d04090915\") " pod="openstack/cinder-volume-volume1-0" Feb 02 09:58:19 crc kubenswrapper[4764]: I0202 09:58:19.372058 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/371fb9ce-9ba4-429b-a68c-62d2e12d688c-dev\") pod \"cinder-backup-0\" (UID: \"371fb9ce-9ba4-429b-a68c-62d2e12d688c\") " pod="openstack/cinder-backup-0" Feb 02 09:58:19 crc kubenswrapper[4764]: I0202 09:58:19.372076 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/371fb9ce-9ba4-429b-a68c-62d2e12d688c-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"371fb9ce-9ba4-429b-a68c-62d2e12d688c\") " pod="openstack/cinder-backup-0" Feb 02 09:58:19 crc kubenswrapper[4764]: I0202 09:58:19.372092 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1c67d8f7-21c3-456d-af99-c96d04090915-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"1c67d8f7-21c3-456d-af99-c96d04090915\") " pod="openstack/cinder-volume-volume1-0" Feb 02 09:58:19 crc kubenswrapper[4764]: I0202 09:58:19.372112 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/371fb9ce-9ba4-429b-a68c-62d2e12d688c-config-data-custom\") pod \"cinder-backup-0\" (UID: \"371fb9ce-9ba4-429b-a68c-62d2e12d688c\") " pod="openstack/cinder-backup-0" Feb 02 09:58:19 crc kubenswrapper[4764]: I0202 09:58:19.372143 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/371fb9ce-9ba4-429b-a68c-62d2e12d688c-etc-nvme\") pod \"cinder-backup-0\" (UID: \"371fb9ce-9ba4-429b-a68c-62d2e12d688c\") " pod="openstack/cinder-backup-0" Feb 02 09:58:19 crc kubenswrapper[4764]: I0202 09:58:19.372161 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/371fb9ce-9ba4-429b-a68c-62d2e12d688c-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"371fb9ce-9ba4-429b-a68c-62d2e12d688c\") " pod="openstack/cinder-backup-0" Feb 02 09:58:19 crc kubenswrapper[4764]: I0202 09:58:19.372181 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c67d8f7-21c3-456d-af99-c96d04090915-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"1c67d8f7-21c3-456d-af99-c96d04090915\") " pod="openstack/cinder-volume-volume1-0" Feb 02 09:58:19 crc kubenswrapper[4764]: I0202 09:58:19.372196 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/1c67d8f7-21c3-456d-af99-c96d04090915-dev\") pod \"cinder-volume-volume1-0\" (UID: \"1c67d8f7-21c3-456d-af99-c96d04090915\") " pod="openstack/cinder-volume-volume1-0" Feb 02 09:58:19 crc kubenswrapper[4764]: I0202 09:58:19.372214 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/1c67d8f7-21c3-456d-af99-c96d04090915-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"1c67d8f7-21c3-456d-af99-c96d04090915\") " pod="openstack/cinder-volume-volume1-0" Feb 02 09:58:19 crc kubenswrapper[4764]: I0202 09:58:19.372231 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/1c67d8f7-21c3-456d-af99-c96d04090915-sys\") pod \"cinder-volume-volume1-0\" (UID: \"1c67d8f7-21c3-456d-af99-c96d04090915\") " pod="openstack/cinder-volume-volume1-0" Feb 02 09:58:19 crc kubenswrapper[4764]: I0202 09:58:19.372246 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/371fb9ce-9ba4-429b-a68c-62d2e12d688c-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"371fb9ce-9ba4-429b-a68c-62d2e12d688c\") " pod="openstack/cinder-backup-0" Feb 02 09:58:19 crc kubenswrapper[4764]: I0202 09:58:19.372262 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/1c67d8f7-21c3-456d-af99-c96d04090915-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"1c67d8f7-21c3-456d-af99-c96d04090915\") " pod="openstack/cinder-volume-volume1-0" Feb 02 09:58:19 crc kubenswrapper[4764]: I0202 09:58:19.372277 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/371fb9ce-9ba4-429b-a68c-62d2e12d688c-ceph\") pod \"cinder-backup-0\" (UID: \"371fb9ce-9ba4-429b-a68c-62d2e12d688c\") " pod="openstack/cinder-backup-0" Feb 02 09:58:19 crc kubenswrapper[4764]: I0202 09:58:19.372292 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/371fb9ce-9ba4-429b-a68c-62d2e12d688c-lib-modules\") pod \"cinder-backup-0\" (UID: \"371fb9ce-9ba4-429b-a68c-62d2e12d688c\") " pod="openstack/cinder-backup-0" Feb 02 09:58:19 crc kubenswrapper[4764]: I0202 09:58:19.372306 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/371fb9ce-9ba4-429b-a68c-62d2e12d688c-sys\") pod \"cinder-backup-0\" (UID: \"371fb9ce-9ba4-429b-a68c-62d2e12d688c\") " pod="openstack/cinder-backup-0" Feb 02 09:58:19 crc kubenswrapper[4764]: I0202 09:58:19.372326 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/1c67d8f7-21c3-456d-af99-c96d04090915-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"1c67d8f7-21c3-456d-af99-c96d04090915\") " pod="openstack/cinder-volume-volume1-0" Feb 02 09:58:19 crc kubenswrapper[4764]: I0202 09:58:19.372341 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1c67d8f7-21c3-456d-af99-c96d04090915-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"1c67d8f7-21c3-456d-af99-c96d04090915\") " pod="openstack/cinder-volume-volume1-0" Feb 02 09:58:19 crc kubenswrapper[4764]: I0202 09:58:19.372357 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/1c67d8f7-21c3-456d-af99-c96d04090915-run\") pod \"cinder-volume-volume1-0\" (UID: \"1c67d8f7-21c3-456d-af99-c96d04090915\") " pod="openstack/cinder-volume-volume1-0" Feb 02 09:58:19 crc kubenswrapper[4764]: I0202 09:58:19.372383 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/371fb9ce-9ba4-429b-a68c-62d2e12d688c-run\") pod \"cinder-backup-0\" (UID: \"371fb9ce-9ba4-429b-a68c-62d2e12d688c\") " pod="openstack/cinder-backup-0" Feb 02 09:58:19 crc kubenswrapper[4764]: I0202 09:58:19.372397 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/371fb9ce-9ba4-429b-a68c-62d2e12d688c-config-data\") pod \"cinder-backup-0\" (UID: \"371fb9ce-9ba4-429b-a68c-62d2e12d688c\") " pod="openstack/cinder-backup-0" Feb 02 09:58:19 crc kubenswrapper[4764]: I0202 09:58:19.372412 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/1c67d8f7-21c3-456d-af99-c96d04090915-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"1c67d8f7-21c3-456d-af99-c96d04090915\") " pod="openstack/cinder-volume-volume1-0" Feb 02 09:58:19 crc kubenswrapper[4764]: I0202 09:58:19.372426 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-66qr8\" (UniqueName: \"kubernetes.io/projected/1c67d8f7-21c3-456d-af99-c96d04090915-kube-api-access-66qr8\") pod \"cinder-volume-volume1-0\" (UID: \"1c67d8f7-21c3-456d-af99-c96d04090915\") " pod="openstack/cinder-volume-volume1-0" Feb 02 09:58:19 crc kubenswrapper[4764]: I0202 09:58:19.372444 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1c67d8f7-21c3-456d-af99-c96d04090915-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"1c67d8f7-21c3-456d-af99-c96d04090915\") " pod="openstack/cinder-volume-volume1-0" Feb 02 09:58:19 crc kubenswrapper[4764]: I0202 09:58:19.372458 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/1c67d8f7-21c3-456d-af99-c96d04090915-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"1c67d8f7-21c3-456d-af99-c96d04090915\") " pod="openstack/cinder-volume-volume1-0" Feb 02 09:58:19 crc kubenswrapper[4764]: I0202 09:58:19.372475 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1c67d8f7-21c3-456d-af99-c96d04090915-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"1c67d8f7-21c3-456d-af99-c96d04090915\") " pod="openstack/cinder-volume-volume1-0" Feb 02 09:58:19 crc kubenswrapper[4764]: I0202 09:58:19.372494 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/371fb9ce-9ba4-429b-a68c-62d2e12d688c-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"371fb9ce-9ba4-429b-a68c-62d2e12d688c\") " pod="openstack/cinder-backup-0" Feb 02 09:58:19 crc kubenswrapper[4764]: I0202 09:58:19.372510 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/371fb9ce-9ba4-429b-a68c-62d2e12d688c-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"371fb9ce-9ba4-429b-a68c-62d2e12d688c\") " pod="openstack/cinder-backup-0" Feb 02 09:58:19 crc kubenswrapper[4764]: I0202 09:58:19.372524 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/1c67d8f7-21c3-456d-af99-c96d04090915-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"1c67d8f7-21c3-456d-af99-c96d04090915\") " pod="openstack/cinder-volume-volume1-0" Feb 02 09:58:19 crc kubenswrapper[4764]: I0202 09:58:19.372553 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/371fb9ce-9ba4-429b-a68c-62d2e12d688c-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"371fb9ce-9ba4-429b-a68c-62d2e12d688c\") " pod="openstack/cinder-backup-0" Feb 02 09:58:19 crc kubenswrapper[4764]: I0202 09:58:19.372569 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nnhsg\" (UniqueName: \"kubernetes.io/projected/371fb9ce-9ba4-429b-a68c-62d2e12d688c-kube-api-access-nnhsg\") pod \"cinder-backup-0\" (UID: \"371fb9ce-9ba4-429b-a68c-62d2e12d688c\") " pod="openstack/cinder-backup-0" Feb 02 09:58:19 crc kubenswrapper[4764]: I0202 09:58:19.372584 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/371fb9ce-9ba4-429b-a68c-62d2e12d688c-scripts\") pod \"cinder-backup-0\" (UID: \"371fb9ce-9ba4-429b-a68c-62d2e12d688c\") " pod="openstack/cinder-backup-0" Feb 02 09:58:19 crc kubenswrapper[4764]: I0202 09:58:19.373479 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/371fb9ce-9ba4-429b-a68c-62d2e12d688c-sys\") pod \"cinder-backup-0\" (UID: \"371fb9ce-9ba4-429b-a68c-62d2e12d688c\") " pod="openstack/cinder-backup-0" Feb 02 09:58:19 crc kubenswrapper[4764]: I0202 09:58:19.373678 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/1c67d8f7-21c3-456d-af99-c96d04090915-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"1c67d8f7-21c3-456d-af99-c96d04090915\") " pod="openstack/cinder-volume-volume1-0" Feb 02 09:58:19 crc kubenswrapper[4764]: I0202 09:58:19.373711 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/371fb9ce-9ba4-429b-a68c-62d2e12d688c-dev\") pod \"cinder-backup-0\" (UID: \"371fb9ce-9ba4-429b-a68c-62d2e12d688c\") " pod="openstack/cinder-backup-0" Feb 02 09:58:19 crc kubenswrapper[4764]: I0202 09:58:19.373735 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/371fb9ce-9ba4-429b-a68c-62d2e12d688c-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"371fb9ce-9ba4-429b-a68c-62d2e12d688c\") " pod="openstack/cinder-backup-0" Feb 02 09:58:19 crc kubenswrapper[4764]: I0202 09:58:19.377117 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/1c67d8f7-21c3-456d-af99-c96d04090915-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"1c67d8f7-21c3-456d-af99-c96d04090915\") " pod="openstack/cinder-volume-volume1-0" Feb 02 09:58:19 crc kubenswrapper[4764]: I0202 09:58:19.377185 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/1c67d8f7-21c3-456d-af99-c96d04090915-sys\") pod \"cinder-volume-volume1-0\" (UID: \"1c67d8f7-21c3-456d-af99-c96d04090915\") " pod="openstack/cinder-volume-volume1-0" Feb 02 09:58:19 crc kubenswrapper[4764]: I0202 09:58:19.377225 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/371fb9ce-9ba4-429b-a68c-62d2e12d688c-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"371fb9ce-9ba4-429b-a68c-62d2e12d688c\") " pod="openstack/cinder-backup-0" Feb 02 09:58:19 crc kubenswrapper[4764]: I0202 09:58:19.377253 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/371fb9ce-9ba4-429b-a68c-62d2e12d688c-lib-modules\") pod \"cinder-backup-0\" (UID: \"371fb9ce-9ba4-429b-a68c-62d2e12d688c\") " pod="openstack/cinder-backup-0" Feb 02 09:58:19 crc kubenswrapper[4764]: I0202 09:58:19.377275 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/371fb9ce-9ba4-429b-a68c-62d2e12d688c-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"371fb9ce-9ba4-429b-a68c-62d2e12d688c\") " pod="openstack/cinder-backup-0" Feb 02 09:58:19 crc kubenswrapper[4764]: I0202 09:58:19.377299 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/1c67d8f7-21c3-456d-af99-c96d04090915-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"1c67d8f7-21c3-456d-af99-c96d04090915\") " pod="openstack/cinder-volume-volume1-0" Feb 02 09:58:19 crc kubenswrapper[4764]: I0202 09:58:19.377323 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1c67d8f7-21c3-456d-af99-c96d04090915-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"1c67d8f7-21c3-456d-af99-c96d04090915\") " pod="openstack/cinder-volume-volume1-0" Feb 02 09:58:19 crc kubenswrapper[4764]: I0202 09:58:19.377430 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/371fb9ce-9ba4-429b-a68c-62d2e12d688c-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"371fb9ce-9ba4-429b-a68c-62d2e12d688c\") " pod="openstack/cinder-backup-0" Feb 02 09:58:19 crc kubenswrapper[4764]: I0202 09:58:19.377610 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/371fb9ce-9ba4-429b-a68c-62d2e12d688c-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"371fb9ce-9ba4-429b-a68c-62d2e12d688c\") " pod="openstack/cinder-backup-0" Feb 02 09:58:19 crc kubenswrapper[4764]: I0202 09:58:19.380348 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/371fb9ce-9ba4-429b-a68c-62d2e12d688c-etc-nvme\") pod \"cinder-backup-0\" (UID: \"371fb9ce-9ba4-429b-a68c-62d2e12d688c\") " pod="openstack/cinder-backup-0" Feb 02 09:58:19 crc kubenswrapper[4764]: I0202 09:58:19.380469 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/371fb9ce-9ba4-429b-a68c-62d2e12d688c-run\") pod \"cinder-backup-0\" (UID: \"371fb9ce-9ba4-429b-a68c-62d2e12d688c\") " pod="openstack/cinder-backup-0" Feb 02 09:58:19 crc kubenswrapper[4764]: I0202 09:58:19.380495 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/1c67d8f7-21c3-456d-af99-c96d04090915-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"1c67d8f7-21c3-456d-af99-c96d04090915\") " pod="openstack/cinder-volume-volume1-0" Feb 02 09:58:19 crc kubenswrapper[4764]: I0202 09:58:19.383379 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/1c67d8f7-21c3-456d-af99-c96d04090915-dev\") pod \"cinder-volume-volume1-0\" (UID: \"1c67d8f7-21c3-456d-af99-c96d04090915\") " pod="openstack/cinder-volume-volume1-0" Feb 02 09:58:19 crc kubenswrapper[4764]: I0202 09:58:19.383447 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/1c67d8f7-21c3-456d-af99-c96d04090915-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"1c67d8f7-21c3-456d-af99-c96d04090915\") " pod="openstack/cinder-volume-volume1-0" Feb 02 09:58:19 crc kubenswrapper[4764]: I0202 09:58:19.384168 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/1c67d8f7-21c3-456d-af99-c96d04090915-run\") pod \"cinder-volume-volume1-0\" (UID: \"1c67d8f7-21c3-456d-af99-c96d04090915\") " pod="openstack/cinder-volume-volume1-0" Feb 02 09:58:19 crc kubenswrapper[4764]: I0202 09:58:19.384235 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/1c67d8f7-21c3-456d-af99-c96d04090915-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"1c67d8f7-21c3-456d-af99-c96d04090915\") " pod="openstack/cinder-volume-volume1-0" Feb 02 09:58:19 crc kubenswrapper[4764]: I0202 09:58:19.387507 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1c67d8f7-21c3-456d-af99-c96d04090915-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"1c67d8f7-21c3-456d-af99-c96d04090915\") " pod="openstack/cinder-volume-volume1-0" Feb 02 09:58:19 crc kubenswrapper[4764]: I0202 09:58:19.393323 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/1c67d8f7-21c3-456d-af99-c96d04090915-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"1c67d8f7-21c3-456d-af99-c96d04090915\") " pod="openstack/cinder-volume-volume1-0" Feb 02 09:58:19 crc kubenswrapper[4764]: I0202 09:58:19.394328 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/371fb9ce-9ba4-429b-a68c-62d2e12d688c-scripts\") pod \"cinder-backup-0\" (UID: \"371fb9ce-9ba4-429b-a68c-62d2e12d688c\") " pod="openstack/cinder-backup-0" Feb 02 09:58:19 crc kubenswrapper[4764]: I0202 09:58:19.394745 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1c67d8f7-21c3-456d-af99-c96d04090915-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"1c67d8f7-21c3-456d-af99-c96d04090915\") " pod="openstack/cinder-volume-volume1-0" Feb 02 09:58:19 crc kubenswrapper[4764]: I0202 09:58:19.396599 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/371fb9ce-9ba4-429b-a68c-62d2e12d688c-config-data-custom\") pod \"cinder-backup-0\" (UID: \"371fb9ce-9ba4-429b-a68c-62d2e12d688c\") " pod="openstack/cinder-backup-0" Feb 02 09:58:19 crc kubenswrapper[4764]: I0202 09:58:19.397324 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/371fb9ce-9ba4-429b-a68c-62d2e12d688c-ceph\") pod \"cinder-backup-0\" (UID: \"371fb9ce-9ba4-429b-a68c-62d2e12d688c\") " pod="openstack/cinder-backup-0" Feb 02 09:58:19 crc kubenswrapper[4764]: I0202 09:58:19.400431 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/371fb9ce-9ba4-429b-a68c-62d2e12d688c-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"371fb9ce-9ba4-429b-a68c-62d2e12d688c\") " pod="openstack/cinder-backup-0" Feb 02 09:58:19 crc kubenswrapper[4764]: I0202 09:58:19.414058 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1c67d8f7-21c3-456d-af99-c96d04090915-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"1c67d8f7-21c3-456d-af99-c96d04090915\") " pod="openstack/cinder-volume-volume1-0" Feb 02 09:58:19 crc kubenswrapper[4764]: I0202 09:58:19.414803 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c67d8f7-21c3-456d-af99-c96d04090915-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"1c67d8f7-21c3-456d-af99-c96d04090915\") " pod="openstack/cinder-volume-volume1-0" Feb 02 09:58:19 crc kubenswrapper[4764]: I0202 09:58:19.426390 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nnhsg\" (UniqueName: \"kubernetes.io/projected/371fb9ce-9ba4-429b-a68c-62d2e12d688c-kube-api-access-nnhsg\") pod \"cinder-backup-0\" (UID: \"371fb9ce-9ba4-429b-a68c-62d2e12d688c\") " pod="openstack/cinder-backup-0" Feb 02 09:58:19 crc kubenswrapper[4764]: I0202 09:58:19.435052 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/371fb9ce-9ba4-429b-a68c-62d2e12d688c-config-data\") pod \"cinder-backup-0\" (UID: \"371fb9ce-9ba4-429b-a68c-62d2e12d688c\") " pod="openstack/cinder-backup-0" Feb 02 09:58:19 crc kubenswrapper[4764]: I0202 09:58:19.440484 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-66qr8\" (UniqueName: \"kubernetes.io/projected/1c67d8f7-21c3-456d-af99-c96d04090915-kube-api-access-66qr8\") pod \"cinder-volume-volume1-0\" (UID: \"1c67d8f7-21c3-456d-af99-c96d04090915\") " pod="openstack/cinder-volume-volume1-0" Feb 02 09:58:19 crc kubenswrapper[4764]: I0202 09:58:19.543715 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-backup-0" Feb 02 09:58:19 crc kubenswrapper[4764]: I0202 09:58:19.551517 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-volume-volume1-0" Feb 02 09:58:20 crc kubenswrapper[4764]: I0202 09:58:20.056453 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Feb 02 09:58:20 crc kubenswrapper[4764]: I0202 09:58:20.057793 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 02 09:58:20 crc kubenswrapper[4764]: I0202 09:58:20.065079 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Feb 02 09:58:20 crc kubenswrapper[4764]: I0202 09:58:20.065246 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Feb 02 09:58:20 crc kubenswrapper[4764]: I0202 09:58:20.065363 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-snxv4" Feb 02 09:58:20 crc kubenswrapper[4764]: I0202 09:58:20.065513 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Feb 02 09:58:20 crc kubenswrapper[4764]: I0202 09:58:20.075010 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 02 09:58:20 crc kubenswrapper[4764]: I0202 09:58:20.153201 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 02 09:58:20 crc kubenswrapper[4764]: I0202 09:58:20.157209 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 02 09:58:20 crc kubenswrapper[4764]: I0202 09:58:20.167391 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Feb 02 09:58:20 crc kubenswrapper[4764]: I0202 09:58:20.167580 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Feb 02 09:58:20 crc kubenswrapper[4764]: I0202 09:58:20.199460 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 02 09:58:20 crc kubenswrapper[4764]: I0202 09:58:20.224159 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/74048046-0204-4a32-957c-2c420c36a3d3-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"74048046-0204-4a32-957c-2c420c36a3d3\") " pod="openstack/glance-default-external-api-0" Feb 02 09:58:20 crc kubenswrapper[4764]: I0202 09:58:20.224232 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/74048046-0204-4a32-957c-2c420c36a3d3-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"74048046-0204-4a32-957c-2c420c36a3d3\") " pod="openstack/glance-default-external-api-0" Feb 02 09:58:20 crc kubenswrapper[4764]: I0202 09:58:20.224256 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/74048046-0204-4a32-957c-2c420c36a3d3-scripts\") pod \"glance-default-external-api-0\" (UID: \"74048046-0204-4a32-957c-2c420c36a3d3\") " pod="openstack/glance-default-external-api-0" Feb 02 09:58:20 crc kubenswrapper[4764]: I0202 09:58:20.224305 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/74048046-0204-4a32-957c-2c420c36a3d3-ceph\") pod \"glance-default-external-api-0\" (UID: \"74048046-0204-4a32-957c-2c420c36a3d3\") " pod="openstack/glance-default-external-api-0" Feb 02 09:58:20 crc kubenswrapper[4764]: I0202 09:58:20.224346 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74048046-0204-4a32-957c-2c420c36a3d3-config-data\") pod \"glance-default-external-api-0\" (UID: \"74048046-0204-4a32-957c-2c420c36a3d3\") " pod="openstack/glance-default-external-api-0" Feb 02 09:58:20 crc kubenswrapper[4764]: I0202 09:58:20.224380 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74048046-0204-4a32-957c-2c420c36a3d3-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"74048046-0204-4a32-957c-2c420c36a3d3\") " pod="openstack/glance-default-external-api-0" Feb 02 09:58:20 crc kubenswrapper[4764]: I0202 09:58:20.224403 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p6l7l\" (UniqueName: \"kubernetes.io/projected/74048046-0204-4a32-957c-2c420c36a3d3-kube-api-access-p6l7l\") pod \"glance-default-external-api-0\" (UID: \"74048046-0204-4a32-957c-2c420c36a3d3\") " pod="openstack/glance-default-external-api-0" Feb 02 09:58:20 crc kubenswrapper[4764]: I0202 09:58:20.224426 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"74048046-0204-4a32-957c-2c420c36a3d3\") " pod="openstack/glance-default-external-api-0" Feb 02 09:58:20 crc kubenswrapper[4764]: I0202 09:58:20.224450 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/74048046-0204-4a32-957c-2c420c36a3d3-logs\") pod \"glance-default-external-api-0\" (UID: \"74048046-0204-4a32-957c-2c420c36a3d3\") " pod="openstack/glance-default-external-api-0" Feb 02 09:58:20 crc kubenswrapper[4764]: I0202 09:58:20.245490 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-db-create-v6n5l"] Feb 02 09:58:20 crc kubenswrapper[4764]: I0202 09:58:20.247736 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-v6n5l" Feb 02 09:58:20 crc kubenswrapper[4764]: I0202 09:58:20.268146 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-create-v6n5l"] Feb 02 09:58:20 crc kubenswrapper[4764]: I0202 09:58:20.327343 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74048046-0204-4a32-957c-2c420c36a3d3-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"74048046-0204-4a32-957c-2c420c36a3d3\") " pod="openstack/glance-default-external-api-0" Feb 02 09:58:20 crc kubenswrapper[4764]: I0202 09:58:20.327379 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p6l7l\" (UniqueName: \"kubernetes.io/projected/74048046-0204-4a32-957c-2c420c36a3d3-kube-api-access-p6l7l\") pod \"glance-default-external-api-0\" (UID: \"74048046-0204-4a32-957c-2c420c36a3d3\") " pod="openstack/glance-default-external-api-0" Feb 02 09:58:20 crc kubenswrapper[4764]: I0202 09:58:20.327406 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"74048046-0204-4a32-957c-2c420c36a3d3\") " pod="openstack/glance-default-external-api-0" Feb 02 09:58:20 crc kubenswrapper[4764]: I0202 09:58:20.327435 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/74048046-0204-4a32-957c-2c420c36a3d3-logs\") pod \"glance-default-external-api-0\" (UID: \"74048046-0204-4a32-957c-2c420c36a3d3\") " pod="openstack/glance-default-external-api-0" Feb 02 09:58:20 crc kubenswrapper[4764]: I0202 09:58:20.327457 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a2b94631-9510-4324-b049-b4224b0b5825-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"a2b94631-9510-4324-b049-b4224b0b5825\") " pod="openstack/glance-default-internal-api-0" Feb 02 09:58:20 crc kubenswrapper[4764]: I0202 09:58:20.327483 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a2b94631-9510-4324-b049-b4224b0b5825-config-data\") pod \"glance-default-internal-api-0\" (UID: \"a2b94631-9510-4324-b049-b4224b0b5825\") " pod="openstack/glance-default-internal-api-0" Feb 02 09:58:20 crc kubenswrapper[4764]: I0202 09:58:20.327518 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/74048046-0204-4a32-957c-2c420c36a3d3-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"74048046-0204-4a32-957c-2c420c36a3d3\") " pod="openstack/glance-default-external-api-0" Feb 02 09:58:20 crc kubenswrapper[4764]: I0202 09:58:20.327575 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/a2b94631-9510-4324-b049-b4224b0b5825-ceph\") pod \"glance-default-internal-api-0\" (UID: \"a2b94631-9510-4324-b049-b4224b0b5825\") " pod="openstack/glance-default-internal-api-0" Feb 02 09:58:20 crc kubenswrapper[4764]: I0202 09:58:20.327598 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a2b94631-9510-4324-b049-b4224b0b5825-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"a2b94631-9510-4324-b049-b4224b0b5825\") " pod="openstack/glance-default-internal-api-0" Feb 02 09:58:20 crc kubenswrapper[4764]: I0202 09:58:20.327616 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/74048046-0204-4a32-957c-2c420c36a3d3-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"74048046-0204-4a32-957c-2c420c36a3d3\") " pod="openstack/glance-default-external-api-0" Feb 02 09:58:20 crc kubenswrapper[4764]: I0202 09:58:20.327633 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/74048046-0204-4a32-957c-2c420c36a3d3-scripts\") pod \"glance-default-external-api-0\" (UID: \"74048046-0204-4a32-957c-2c420c36a3d3\") " pod="openstack/glance-default-external-api-0" Feb 02 09:58:20 crc kubenswrapper[4764]: I0202 09:58:20.327659 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a2b94631-9510-4324-b049-b4224b0b5825-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"a2b94631-9510-4324-b049-b4224b0b5825\") " pod="openstack/glance-default-internal-api-0" Feb 02 09:58:20 crc kubenswrapper[4764]: I0202 09:58:20.327681 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6xvpd\" (UniqueName: \"kubernetes.io/projected/a2b94631-9510-4324-b049-b4224b0b5825-kube-api-access-6xvpd\") pod \"glance-default-internal-api-0\" (UID: \"a2b94631-9510-4324-b049-b4224b0b5825\") " pod="openstack/glance-default-internal-api-0" Feb 02 09:58:20 crc kubenswrapper[4764]: I0202 09:58:20.327709 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/74048046-0204-4a32-957c-2c420c36a3d3-ceph\") pod \"glance-default-external-api-0\" (UID: \"74048046-0204-4a32-957c-2c420c36a3d3\") " pod="openstack/glance-default-external-api-0" Feb 02 09:58:20 crc kubenswrapper[4764]: I0202 09:58:20.327731 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a2b94631-9510-4324-b049-b4224b0b5825-logs\") pod \"glance-default-internal-api-0\" (UID: \"a2b94631-9510-4324-b049-b4224b0b5825\") " pod="openstack/glance-default-internal-api-0" Feb 02 09:58:20 crc kubenswrapper[4764]: I0202 09:58:20.327751 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a2b94631-9510-4324-b049-b4224b0b5825-scripts\") pod \"glance-default-internal-api-0\" (UID: \"a2b94631-9510-4324-b049-b4224b0b5825\") " pod="openstack/glance-default-internal-api-0" Feb 02 09:58:20 crc kubenswrapper[4764]: I0202 09:58:20.327767 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"a2b94631-9510-4324-b049-b4224b0b5825\") " pod="openstack/glance-default-internal-api-0" Feb 02 09:58:20 crc kubenswrapper[4764]: I0202 09:58:20.327789 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74048046-0204-4a32-957c-2c420c36a3d3-config-data\") pod \"glance-default-external-api-0\" (UID: \"74048046-0204-4a32-957c-2c420c36a3d3\") " pod="openstack/glance-default-external-api-0" Feb 02 09:58:20 crc kubenswrapper[4764]: I0202 09:58:20.328498 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/74048046-0204-4a32-957c-2c420c36a3d3-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"74048046-0204-4a32-957c-2c420c36a3d3\") " pod="openstack/glance-default-external-api-0" Feb 02 09:58:20 crc kubenswrapper[4764]: I0202 09:58:20.328581 4764 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"74048046-0204-4a32-957c-2c420c36a3d3\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/glance-default-external-api-0" Feb 02 09:58:20 crc kubenswrapper[4764]: I0202 09:58:20.329666 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/74048046-0204-4a32-957c-2c420c36a3d3-logs\") pod \"glance-default-external-api-0\" (UID: \"74048046-0204-4a32-957c-2c420c36a3d3\") " pod="openstack/glance-default-external-api-0" Feb 02 09:58:20 crc kubenswrapper[4764]: I0202 09:58:20.347808 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/74048046-0204-4a32-957c-2c420c36a3d3-ceph\") pod \"glance-default-external-api-0\" (UID: \"74048046-0204-4a32-957c-2c420c36a3d3\") " pod="openstack/glance-default-external-api-0" Feb 02 09:58:20 crc kubenswrapper[4764]: I0202 09:58:20.348754 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74048046-0204-4a32-957c-2c420c36a3d3-config-data\") pod \"glance-default-external-api-0\" (UID: \"74048046-0204-4a32-957c-2c420c36a3d3\") " pod="openstack/glance-default-external-api-0" Feb 02 09:58:20 crc kubenswrapper[4764]: I0202 09:58:20.356054 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p6l7l\" (UniqueName: \"kubernetes.io/projected/74048046-0204-4a32-957c-2c420c36a3d3-kube-api-access-p6l7l\") pod \"glance-default-external-api-0\" (UID: \"74048046-0204-4a32-957c-2c420c36a3d3\") " pod="openstack/glance-default-external-api-0" Feb 02 09:58:20 crc kubenswrapper[4764]: I0202 09:58:20.360566 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/74048046-0204-4a32-957c-2c420c36a3d3-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"74048046-0204-4a32-957c-2c420c36a3d3\") " pod="openstack/glance-default-external-api-0" Feb 02 09:58:20 crc kubenswrapper[4764]: I0202 09:58:20.361780 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74048046-0204-4a32-957c-2c420c36a3d3-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"74048046-0204-4a32-957c-2c420c36a3d3\") " pod="openstack/glance-default-external-api-0" Feb 02 09:58:20 crc kubenswrapper[4764]: I0202 09:58:20.365526 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-5e6b-account-create-update-shmv8"] Feb 02 09:58:20 crc kubenswrapper[4764]: I0202 09:58:20.366006 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/74048046-0204-4a32-957c-2c420c36a3d3-scripts\") pod \"glance-default-external-api-0\" (UID: \"74048046-0204-4a32-957c-2c420c36a3d3\") " pod="openstack/glance-default-external-api-0" Feb 02 09:58:20 crc kubenswrapper[4764]: I0202 09:58:20.366631 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-5e6b-account-create-update-shmv8" Feb 02 09:58:20 crc kubenswrapper[4764]: I0202 09:58:20.397615 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-5e6b-account-create-update-shmv8"] Feb 02 09:58:20 crc kubenswrapper[4764]: I0202 09:58:20.402354 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-db-secret" Feb 02 09:58:20 crc kubenswrapper[4764]: I0202 09:58:20.431988 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/a2b94631-9510-4324-b049-b4224b0b5825-ceph\") pod \"glance-default-internal-api-0\" (UID: \"a2b94631-9510-4324-b049-b4224b0b5825\") " pod="openstack/glance-default-internal-api-0" Feb 02 09:58:20 crc kubenswrapper[4764]: I0202 09:58:20.432030 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a2b94631-9510-4324-b049-b4224b0b5825-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"a2b94631-9510-4324-b049-b4224b0b5825\") " pod="openstack/glance-default-internal-api-0" Feb 02 09:58:20 crc kubenswrapper[4764]: I0202 09:58:20.432083 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a2b94631-9510-4324-b049-b4224b0b5825-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"a2b94631-9510-4324-b049-b4224b0b5825\") " pod="openstack/glance-default-internal-api-0" Feb 02 09:58:20 crc kubenswrapper[4764]: I0202 09:58:20.432104 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6xvpd\" (UniqueName: \"kubernetes.io/projected/a2b94631-9510-4324-b049-b4224b0b5825-kube-api-access-6xvpd\") pod \"glance-default-internal-api-0\" (UID: \"a2b94631-9510-4324-b049-b4224b0b5825\") " pod="openstack/glance-default-internal-api-0" Feb 02 09:58:20 crc kubenswrapper[4764]: I0202 09:58:20.432144 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a2b94631-9510-4324-b049-b4224b0b5825-logs\") pod \"glance-default-internal-api-0\" (UID: \"a2b94631-9510-4324-b049-b4224b0b5825\") " pod="openstack/glance-default-internal-api-0" Feb 02 09:58:20 crc kubenswrapper[4764]: I0202 09:58:20.432166 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a2b94631-9510-4324-b049-b4224b0b5825-scripts\") pod \"glance-default-internal-api-0\" (UID: \"a2b94631-9510-4324-b049-b4224b0b5825\") " pod="openstack/glance-default-internal-api-0" Feb 02 09:58:20 crc kubenswrapper[4764]: I0202 09:58:20.432183 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"a2b94631-9510-4324-b049-b4224b0b5825\") " pod="openstack/glance-default-internal-api-0" Feb 02 09:58:20 crc kubenswrapper[4764]: I0202 09:58:20.432253 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a2b94631-9510-4324-b049-b4224b0b5825-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"a2b94631-9510-4324-b049-b4224b0b5825\") " pod="openstack/glance-default-internal-api-0" Feb 02 09:58:20 crc kubenswrapper[4764]: I0202 09:58:20.432277 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a2b94631-9510-4324-b049-b4224b0b5825-config-data\") pod \"glance-default-internal-api-0\" (UID: \"a2b94631-9510-4324-b049-b4224b0b5825\") " pod="openstack/glance-default-internal-api-0" Feb 02 09:58:20 crc kubenswrapper[4764]: I0202 09:58:20.443409 4764 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"a2b94631-9510-4324-b049-b4224b0b5825\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/glance-default-internal-api-0" Feb 02 09:58:20 crc kubenswrapper[4764]: I0202 09:58:20.443662 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a2b94631-9510-4324-b049-b4224b0b5825-logs\") pod \"glance-default-internal-api-0\" (UID: \"a2b94631-9510-4324-b049-b4224b0b5825\") " pod="openstack/glance-default-internal-api-0" Feb 02 09:58:20 crc kubenswrapper[4764]: I0202 09:58:20.443895 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a2b94631-9510-4324-b049-b4224b0b5825-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"a2b94631-9510-4324-b049-b4224b0b5825\") " pod="openstack/glance-default-internal-api-0" Feb 02 09:58:20 crc kubenswrapper[4764]: I0202 09:58:20.462973 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a2b94631-9510-4324-b049-b4224b0b5825-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"a2b94631-9510-4324-b049-b4224b0b5825\") " pod="openstack/glance-default-internal-api-0" Feb 02 09:58:20 crc kubenswrapper[4764]: I0202 09:58:20.467166 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a2b94631-9510-4324-b049-b4224b0b5825-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"a2b94631-9510-4324-b049-b4224b0b5825\") " pod="openstack/glance-default-internal-api-0" Feb 02 09:58:20 crc kubenswrapper[4764]: I0202 09:58:20.472447 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a2b94631-9510-4324-b049-b4224b0b5825-scripts\") pod \"glance-default-internal-api-0\" (UID: \"a2b94631-9510-4324-b049-b4224b0b5825\") " pod="openstack/glance-default-internal-api-0" Feb 02 09:58:20 crc kubenswrapper[4764]: I0202 09:58:20.477194 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"74048046-0204-4a32-957c-2c420c36a3d3\") " pod="openstack/glance-default-external-api-0" Feb 02 09:58:20 crc kubenswrapper[4764]: I0202 09:58:20.477879 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/a2b94631-9510-4324-b049-b4224b0b5825-ceph\") pod \"glance-default-internal-api-0\" (UID: \"a2b94631-9510-4324-b049-b4224b0b5825\") " pod="openstack/glance-default-internal-api-0" Feb 02 09:58:20 crc kubenswrapper[4764]: I0202 09:58:20.482110 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6xvpd\" (UniqueName: \"kubernetes.io/projected/a2b94631-9510-4324-b049-b4224b0b5825-kube-api-access-6xvpd\") pod \"glance-default-internal-api-0\" (UID: \"a2b94631-9510-4324-b049-b4224b0b5825\") " pod="openstack/glance-default-internal-api-0" Feb 02 09:58:20 crc kubenswrapper[4764]: I0202 09:58:20.495566 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a2b94631-9510-4324-b049-b4224b0b5825-config-data\") pod \"glance-default-internal-api-0\" (UID: \"a2b94631-9510-4324-b049-b4224b0b5825\") " pod="openstack/glance-default-internal-api-0" Feb 02 09:58:20 crc kubenswrapper[4764]: I0202 09:58:20.497832 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"a2b94631-9510-4324-b049-b4224b0b5825\") " pod="openstack/glance-default-internal-api-0" Feb 02 09:58:20 crc kubenswrapper[4764]: I0202 09:58:20.504655 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-volume-volume1-0"] Feb 02 09:58:20 crc kubenswrapper[4764]: I0202 09:58:20.528781 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-6d4f8f8755-dp7xg"] Feb 02 09:58:20 crc kubenswrapper[4764]: I0202 09:58:20.531879 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 02 09:58:20 crc kubenswrapper[4764]: I0202 09:58:20.534225 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f001efec-ca89-4c17-be01-3f42fd33e5c0-operator-scripts\") pod \"manila-db-create-v6n5l\" (UID: \"f001efec-ca89-4c17-be01-3f42fd33e5c0\") " pod="openstack/manila-db-create-v6n5l" Feb 02 09:58:20 crc kubenswrapper[4764]: I0202 09:58:20.534308 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ca0252ea-2836-433b-b26f-bee25fd3adb2-operator-scripts\") pod \"manila-5e6b-account-create-update-shmv8\" (UID: \"ca0252ea-2836-433b-b26f-bee25fd3adb2\") " pod="openstack/manila-5e6b-account-create-update-shmv8" Feb 02 09:58:20 crc kubenswrapper[4764]: I0202 09:58:20.534332 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h4jdb\" (UniqueName: \"kubernetes.io/projected/f001efec-ca89-4c17-be01-3f42fd33e5c0-kube-api-access-h4jdb\") pod \"manila-db-create-v6n5l\" (UID: \"f001efec-ca89-4c17-be01-3f42fd33e5c0\") " pod="openstack/manila-db-create-v6n5l" Feb 02 09:58:20 crc kubenswrapper[4764]: I0202 09:58:20.534368 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-znspp\" (UniqueName: \"kubernetes.io/projected/ca0252ea-2836-433b-b26f-bee25fd3adb2-kube-api-access-znspp\") pod \"manila-5e6b-account-create-update-shmv8\" (UID: \"ca0252ea-2836-433b-b26f-bee25fd3adb2\") " pod="openstack/manila-5e6b-account-create-update-shmv8" Feb 02 09:58:20 crc kubenswrapper[4764]: I0202 09:58:20.539223 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6d4f8f8755-dp7xg" Feb 02 09:58:20 crc kubenswrapper[4764]: I0202 09:58:20.560156 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Feb 02 09:58:20 crc kubenswrapper[4764]: I0202 09:58:20.560345 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Feb 02 09:58:20 crc kubenswrapper[4764]: I0202 09:58:20.560445 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-69nzg" Feb 02 09:58:20 crc kubenswrapper[4764]: I0202 09:58:20.560556 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Feb 02 09:58:20 crc kubenswrapper[4764]: I0202 09:58:20.568191 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6d4f8f8755-dp7xg"] Feb 02 09:58:20 crc kubenswrapper[4764]: I0202 09:58:20.637806 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ca0252ea-2836-433b-b26f-bee25fd3adb2-operator-scripts\") pod \"manila-5e6b-account-create-update-shmv8\" (UID: \"ca0252ea-2836-433b-b26f-bee25fd3adb2\") " pod="openstack/manila-5e6b-account-create-update-shmv8" Feb 02 09:58:20 crc kubenswrapper[4764]: I0202 09:58:20.637848 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h4jdb\" (UniqueName: \"kubernetes.io/projected/f001efec-ca89-4c17-be01-3f42fd33e5c0-kube-api-access-h4jdb\") pod \"manila-db-create-v6n5l\" (UID: \"f001efec-ca89-4c17-be01-3f42fd33e5c0\") " pod="openstack/manila-db-create-v6n5l" Feb 02 09:58:20 crc kubenswrapper[4764]: I0202 09:58:20.637888 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-znspp\" (UniqueName: \"kubernetes.io/projected/ca0252ea-2836-433b-b26f-bee25fd3adb2-kube-api-access-znspp\") pod \"manila-5e6b-account-create-update-shmv8\" (UID: \"ca0252ea-2836-433b-b26f-bee25fd3adb2\") " pod="openstack/manila-5e6b-account-create-update-shmv8" Feb 02 09:58:20 crc kubenswrapper[4764]: I0202 09:58:20.637979 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f001efec-ca89-4c17-be01-3f42fd33e5c0-operator-scripts\") pod \"manila-db-create-v6n5l\" (UID: \"f001efec-ca89-4c17-be01-3f42fd33e5c0\") " pod="openstack/manila-db-create-v6n5l" Feb 02 09:58:20 crc kubenswrapper[4764]: I0202 09:58:20.663979 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f001efec-ca89-4c17-be01-3f42fd33e5c0-operator-scripts\") pod \"manila-db-create-v6n5l\" (UID: \"f001efec-ca89-4c17-be01-3f42fd33e5c0\") " pod="openstack/manila-db-create-v6n5l" Feb 02 09:58:20 crc kubenswrapper[4764]: I0202 09:58:20.666839 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ca0252ea-2836-433b-b26f-bee25fd3adb2-operator-scripts\") pod \"manila-5e6b-account-create-update-shmv8\" (UID: \"ca0252ea-2836-433b-b26f-bee25fd3adb2\") " pod="openstack/manila-5e6b-account-create-update-shmv8" Feb 02 09:58:20 crc kubenswrapper[4764]: I0202 09:58:20.680836 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-znspp\" (UniqueName: \"kubernetes.io/projected/ca0252ea-2836-433b-b26f-bee25fd3adb2-kube-api-access-znspp\") pod \"manila-5e6b-account-create-update-shmv8\" (UID: \"ca0252ea-2836-433b-b26f-bee25fd3adb2\") " pod="openstack/manila-5e6b-account-create-update-shmv8" Feb 02 09:58:20 crc kubenswrapper[4764]: I0202 09:58:20.688873 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h4jdb\" (UniqueName: \"kubernetes.io/projected/f001efec-ca89-4c17-be01-3f42fd33e5c0-kube-api-access-h4jdb\") pod \"manila-db-create-v6n5l\" (UID: \"f001efec-ca89-4c17-be01-3f42fd33e5c0\") " pod="openstack/manila-db-create-v6n5l" Feb 02 09:58:20 crc kubenswrapper[4764]: I0202 09:58:20.699290 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 02 09:58:20 crc kubenswrapper[4764]: I0202 09:58:20.699451 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 02 09:58:20 crc kubenswrapper[4764]: I0202 09:58:20.742296 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b79a9079-9d96-4edc-97fc-b9bdf122e4d8-logs\") pod \"horizon-6d4f8f8755-dp7xg\" (UID: \"b79a9079-9d96-4edc-97fc-b9bdf122e4d8\") " pod="openstack/horizon-6d4f8f8755-dp7xg" Feb 02 09:58:20 crc kubenswrapper[4764]: I0202 09:58:20.742377 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nct92\" (UniqueName: \"kubernetes.io/projected/b79a9079-9d96-4edc-97fc-b9bdf122e4d8-kube-api-access-nct92\") pod \"horizon-6d4f8f8755-dp7xg\" (UID: \"b79a9079-9d96-4edc-97fc-b9bdf122e4d8\") " pod="openstack/horizon-6d4f8f8755-dp7xg" Feb 02 09:58:20 crc kubenswrapper[4764]: I0202 09:58:20.742408 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b79a9079-9d96-4edc-97fc-b9bdf122e4d8-scripts\") pod \"horizon-6d4f8f8755-dp7xg\" (UID: \"b79a9079-9d96-4edc-97fc-b9bdf122e4d8\") " pod="openstack/horizon-6d4f8f8755-dp7xg" Feb 02 09:58:20 crc kubenswrapper[4764]: I0202 09:58:20.742466 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/b79a9079-9d96-4edc-97fc-b9bdf122e4d8-horizon-secret-key\") pod \"horizon-6d4f8f8755-dp7xg\" (UID: \"b79a9079-9d96-4edc-97fc-b9bdf122e4d8\") " pod="openstack/horizon-6d4f8f8755-dp7xg" Feb 02 09:58:20 crc kubenswrapper[4764]: I0202 09:58:20.742570 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b79a9079-9d96-4edc-97fc-b9bdf122e4d8-config-data\") pod \"horizon-6d4f8f8755-dp7xg\" (UID: \"b79a9079-9d96-4edc-97fc-b9bdf122e4d8\") " pod="openstack/horizon-6d4f8f8755-dp7xg" Feb 02 09:58:20 crc kubenswrapper[4764]: I0202 09:58:20.745547 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-backup-0"] Feb 02 09:58:20 crc kubenswrapper[4764]: I0202 09:58:20.782792 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 02 09:58:20 crc kubenswrapper[4764]: I0202 09:58:20.852826 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b79a9079-9d96-4edc-97fc-b9bdf122e4d8-config-data\") pod \"horizon-6d4f8f8755-dp7xg\" (UID: \"b79a9079-9d96-4edc-97fc-b9bdf122e4d8\") " pod="openstack/horizon-6d4f8f8755-dp7xg" Feb 02 09:58:20 crc kubenswrapper[4764]: I0202 09:58:20.853255 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b79a9079-9d96-4edc-97fc-b9bdf122e4d8-logs\") pod \"horizon-6d4f8f8755-dp7xg\" (UID: \"b79a9079-9d96-4edc-97fc-b9bdf122e4d8\") " pod="openstack/horizon-6d4f8f8755-dp7xg" Feb 02 09:58:20 crc kubenswrapper[4764]: I0202 09:58:20.853287 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nct92\" (UniqueName: \"kubernetes.io/projected/b79a9079-9d96-4edc-97fc-b9bdf122e4d8-kube-api-access-nct92\") pod \"horizon-6d4f8f8755-dp7xg\" (UID: \"b79a9079-9d96-4edc-97fc-b9bdf122e4d8\") " pod="openstack/horizon-6d4f8f8755-dp7xg" Feb 02 09:58:20 crc kubenswrapper[4764]: I0202 09:58:20.853310 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b79a9079-9d96-4edc-97fc-b9bdf122e4d8-scripts\") pod \"horizon-6d4f8f8755-dp7xg\" (UID: \"b79a9079-9d96-4edc-97fc-b9bdf122e4d8\") " pod="openstack/horizon-6d4f8f8755-dp7xg" Feb 02 09:58:20 crc kubenswrapper[4764]: I0202 09:58:20.853353 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/b79a9079-9d96-4edc-97fc-b9bdf122e4d8-horizon-secret-key\") pod \"horizon-6d4f8f8755-dp7xg\" (UID: \"b79a9079-9d96-4edc-97fc-b9bdf122e4d8\") " pod="openstack/horizon-6d4f8f8755-dp7xg" Feb 02 09:58:20 crc kubenswrapper[4764]: I0202 09:58:20.855424 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b79a9079-9d96-4edc-97fc-b9bdf122e4d8-logs\") pod \"horizon-6d4f8f8755-dp7xg\" (UID: \"b79a9079-9d96-4edc-97fc-b9bdf122e4d8\") " pod="openstack/horizon-6d4f8f8755-dp7xg" Feb 02 09:58:20 crc kubenswrapper[4764]: I0202 09:58:20.856367 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b79a9079-9d96-4edc-97fc-b9bdf122e4d8-config-data\") pod \"horizon-6d4f8f8755-dp7xg\" (UID: \"b79a9079-9d96-4edc-97fc-b9bdf122e4d8\") " pod="openstack/horizon-6d4f8f8755-dp7xg" Feb 02 09:58:20 crc kubenswrapper[4764]: I0202 09:58:20.857129 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b79a9079-9d96-4edc-97fc-b9bdf122e4d8-scripts\") pod \"horizon-6d4f8f8755-dp7xg\" (UID: \"b79a9079-9d96-4edc-97fc-b9bdf122e4d8\") " pod="openstack/horizon-6d4f8f8755-dp7xg" Feb 02 09:58:20 crc kubenswrapper[4764]: I0202 09:58:20.870572 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/b79a9079-9d96-4edc-97fc-b9bdf122e4d8-horizon-secret-key\") pod \"horizon-6d4f8f8755-dp7xg\" (UID: \"b79a9079-9d96-4edc-97fc-b9bdf122e4d8\") " pod="openstack/horizon-6d4f8f8755-dp7xg" Feb 02 09:58:20 crc kubenswrapper[4764]: I0202 09:58:20.871155 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-5e6b-account-create-update-shmv8" Feb 02 09:58:20 crc kubenswrapper[4764]: I0202 09:58:20.892978 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-7f4b5b9665-qqgjw"] Feb 02 09:58:20 crc kubenswrapper[4764]: I0202 09:58:20.894430 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7f4b5b9665-qqgjw" Feb 02 09:58:20 crc kubenswrapper[4764]: I0202 09:58:20.895526 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-v6n5l" Feb 02 09:58:20 crc kubenswrapper[4764]: I0202 09:58:20.906185 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nct92\" (UniqueName: \"kubernetes.io/projected/b79a9079-9d96-4edc-97fc-b9bdf122e4d8-kube-api-access-nct92\") pod \"horizon-6d4f8f8755-dp7xg\" (UID: \"b79a9079-9d96-4edc-97fc-b9bdf122e4d8\") " pod="openstack/horizon-6d4f8f8755-dp7xg" Feb 02 09:58:20 crc kubenswrapper[4764]: I0202 09:58:20.934863 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7f4b5b9665-qqgjw"] Feb 02 09:58:20 crc kubenswrapper[4764]: I0202 09:58:20.989779 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6d4f8f8755-dp7xg" Feb 02 09:58:21 crc kubenswrapper[4764]: I0202 09:58:21.065632 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/165479bf-43a7-47fc-9083-c87cb1ee0f2c-config-data\") pod \"horizon-7f4b5b9665-qqgjw\" (UID: \"165479bf-43a7-47fc-9083-c87cb1ee0f2c\") " pod="openstack/horizon-7f4b5b9665-qqgjw" Feb 02 09:58:21 crc kubenswrapper[4764]: I0202 09:58:21.065676 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/165479bf-43a7-47fc-9083-c87cb1ee0f2c-horizon-secret-key\") pod \"horizon-7f4b5b9665-qqgjw\" (UID: \"165479bf-43a7-47fc-9083-c87cb1ee0f2c\") " pod="openstack/horizon-7f4b5b9665-qqgjw" Feb 02 09:58:21 crc kubenswrapper[4764]: I0202 09:58:21.065730 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/165479bf-43a7-47fc-9083-c87cb1ee0f2c-logs\") pod \"horizon-7f4b5b9665-qqgjw\" (UID: \"165479bf-43a7-47fc-9083-c87cb1ee0f2c\") " pod="openstack/horizon-7f4b5b9665-qqgjw" Feb 02 09:58:21 crc kubenswrapper[4764]: I0202 09:58:21.065751 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/165479bf-43a7-47fc-9083-c87cb1ee0f2c-scripts\") pod \"horizon-7f4b5b9665-qqgjw\" (UID: \"165479bf-43a7-47fc-9083-c87cb1ee0f2c\") " pod="openstack/horizon-7f4b5b9665-qqgjw" Feb 02 09:58:21 crc kubenswrapper[4764]: I0202 09:58:21.065778 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ftgmp\" (UniqueName: \"kubernetes.io/projected/165479bf-43a7-47fc-9083-c87cb1ee0f2c-kube-api-access-ftgmp\") pod \"horizon-7f4b5b9665-qqgjw\" (UID: \"165479bf-43a7-47fc-9083-c87cb1ee0f2c\") " pod="openstack/horizon-7f4b5b9665-qqgjw" Feb 02 09:58:21 crc kubenswrapper[4764]: I0202 09:58:21.168141 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/165479bf-43a7-47fc-9083-c87cb1ee0f2c-scripts\") pod \"horizon-7f4b5b9665-qqgjw\" (UID: \"165479bf-43a7-47fc-9083-c87cb1ee0f2c\") " pod="openstack/horizon-7f4b5b9665-qqgjw" Feb 02 09:58:21 crc kubenswrapper[4764]: I0202 09:58:21.168204 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ftgmp\" (UniqueName: \"kubernetes.io/projected/165479bf-43a7-47fc-9083-c87cb1ee0f2c-kube-api-access-ftgmp\") pod \"horizon-7f4b5b9665-qqgjw\" (UID: \"165479bf-43a7-47fc-9083-c87cb1ee0f2c\") " pod="openstack/horizon-7f4b5b9665-qqgjw" Feb 02 09:58:21 crc kubenswrapper[4764]: I0202 09:58:21.168317 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/165479bf-43a7-47fc-9083-c87cb1ee0f2c-config-data\") pod \"horizon-7f4b5b9665-qqgjw\" (UID: \"165479bf-43a7-47fc-9083-c87cb1ee0f2c\") " pod="openstack/horizon-7f4b5b9665-qqgjw" Feb 02 09:58:21 crc kubenswrapper[4764]: I0202 09:58:21.168357 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/165479bf-43a7-47fc-9083-c87cb1ee0f2c-horizon-secret-key\") pod \"horizon-7f4b5b9665-qqgjw\" (UID: \"165479bf-43a7-47fc-9083-c87cb1ee0f2c\") " pod="openstack/horizon-7f4b5b9665-qqgjw" Feb 02 09:58:21 crc kubenswrapper[4764]: I0202 09:58:21.168427 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/165479bf-43a7-47fc-9083-c87cb1ee0f2c-logs\") pod \"horizon-7f4b5b9665-qqgjw\" (UID: \"165479bf-43a7-47fc-9083-c87cb1ee0f2c\") " pod="openstack/horizon-7f4b5b9665-qqgjw" Feb 02 09:58:21 crc kubenswrapper[4764]: I0202 09:58:21.168962 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/165479bf-43a7-47fc-9083-c87cb1ee0f2c-logs\") pod \"horizon-7f4b5b9665-qqgjw\" (UID: \"165479bf-43a7-47fc-9083-c87cb1ee0f2c\") " pod="openstack/horizon-7f4b5b9665-qqgjw" Feb 02 09:58:21 crc kubenswrapper[4764]: I0202 09:58:21.169648 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/165479bf-43a7-47fc-9083-c87cb1ee0f2c-scripts\") pod \"horizon-7f4b5b9665-qqgjw\" (UID: \"165479bf-43a7-47fc-9083-c87cb1ee0f2c\") " pod="openstack/horizon-7f4b5b9665-qqgjw" Feb 02 09:58:21 crc kubenswrapper[4764]: I0202 09:58:21.171586 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/165479bf-43a7-47fc-9083-c87cb1ee0f2c-config-data\") pod \"horizon-7f4b5b9665-qqgjw\" (UID: \"165479bf-43a7-47fc-9083-c87cb1ee0f2c\") " pod="openstack/horizon-7f4b5b9665-qqgjw" Feb 02 09:58:21 crc kubenswrapper[4764]: I0202 09:58:21.187556 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/165479bf-43a7-47fc-9083-c87cb1ee0f2c-horizon-secret-key\") pod \"horizon-7f4b5b9665-qqgjw\" (UID: \"165479bf-43a7-47fc-9083-c87cb1ee0f2c\") " pod="openstack/horizon-7f4b5b9665-qqgjw" Feb 02 09:58:21 crc kubenswrapper[4764]: I0202 09:58:21.189637 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ftgmp\" (UniqueName: \"kubernetes.io/projected/165479bf-43a7-47fc-9083-c87cb1ee0f2c-kube-api-access-ftgmp\") pod \"horizon-7f4b5b9665-qqgjw\" (UID: \"165479bf-43a7-47fc-9083-c87cb1ee0f2c\") " pod="openstack/horizon-7f4b5b9665-qqgjw" Feb 02 09:58:21 crc kubenswrapper[4764]: I0202 09:58:21.371889 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"371fb9ce-9ba4-429b-a68c-62d2e12d688c","Type":"ContainerStarted","Data":"a089789dda7559f4a0b5cd6e67f3e48c64b54e42b0bb7bf949e9dafdf4a8450a"} Feb 02 09:58:21 crc kubenswrapper[4764]: I0202 09:58:21.372669 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7f4b5b9665-qqgjw" Feb 02 09:58:21 crc kubenswrapper[4764]: I0202 09:58:21.375338 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"1c67d8f7-21c3-456d-af99-c96d04090915","Type":"ContainerStarted","Data":"aeb56125830537a7446dab0eb0125a3f636c27a7d273404576d4b4b7cfd5ecfb"} Feb 02 09:58:21 crc kubenswrapper[4764]: I0202 09:58:21.678825 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 02 09:58:21 crc kubenswrapper[4764]: W0202 09:58:21.717777 4764 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod74048046_0204_4a32_957c_2c420c36a3d3.slice/crio-511775cc124826bc8d7abf3fda9538d110b35dbf169243066b5286751c33aac7 WatchSource:0}: Error finding container 511775cc124826bc8d7abf3fda9538d110b35dbf169243066b5286751c33aac7: Status 404 returned error can't find the container with id 511775cc124826bc8d7abf3fda9538d110b35dbf169243066b5286751c33aac7 Feb 02 09:58:21 crc kubenswrapper[4764]: I0202 09:58:21.946415 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-5e6b-account-create-update-shmv8"] Feb 02 09:58:22 crc kubenswrapper[4764]: I0202 09:58:21.976522 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-create-v6n5l"] Feb 02 09:58:22 crc kubenswrapper[4764]: I0202 09:58:21.985089 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6d4f8f8755-dp7xg"] Feb 02 09:58:22 crc kubenswrapper[4764]: W0202 09:58:22.040580 4764 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podca0252ea_2836_433b_b26f_bee25fd3adb2.slice/crio-7fa8b629dc27c26f001bd0ada4ba085624a794c9ec5288294c5c6cbadb204ea6 WatchSource:0}: Error finding container 7fa8b629dc27c26f001bd0ada4ba085624a794c9ec5288294c5c6cbadb204ea6: Status 404 returned error can't find the container with id 7fa8b629dc27c26f001bd0ada4ba085624a794c9ec5288294c5c6cbadb204ea6 Feb 02 09:58:22 crc kubenswrapper[4764]: I0202 09:58:22.313124 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7f4b5b9665-qqgjw"] Feb 02 09:58:22 crc kubenswrapper[4764]: I0202 09:58:22.413040 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 02 09:58:22 crc kubenswrapper[4764]: I0202 09:58:22.471327 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-5e6b-account-create-update-shmv8" event={"ID":"ca0252ea-2836-433b-b26f-bee25fd3adb2","Type":"ContainerStarted","Data":"7fa8b629dc27c26f001bd0ada4ba085624a794c9ec5288294c5c6cbadb204ea6"} Feb 02 09:58:22 crc kubenswrapper[4764]: I0202 09:58:22.523534 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"371fb9ce-9ba4-429b-a68c-62d2e12d688c","Type":"ContainerStarted","Data":"cdeb0f8f808cc5eea5b9e9a512f69aab102aca76bee4b83dec6662efd513e45d"} Feb 02 09:58:22 crc kubenswrapper[4764]: W0202 09:58:22.545858 4764 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda2b94631_9510_4324_b049_b4224b0b5825.slice/crio-442b679c36cc28925886866596e056c002ad9e6b8f5d57a3e2dea6e2a0122ea0 WatchSource:0}: Error finding container 442b679c36cc28925886866596e056c002ad9e6b8f5d57a3e2dea6e2a0122ea0: Status 404 returned error can't find the container with id 442b679c36cc28925886866596e056c002ad9e6b8f5d57a3e2dea6e2a0122ea0 Feb 02 09:58:22 crc kubenswrapper[4764]: W0202 09:58:22.557295 4764 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod165479bf_43a7_47fc_9083_c87cb1ee0f2c.slice/crio-5dc251dad9db660ac1714da21d2ad61cb71204b205a96fadba2e570712a989aa WatchSource:0}: Error finding container 5dc251dad9db660ac1714da21d2ad61cb71204b205a96fadba2e570712a989aa: Status 404 returned error can't find the container with id 5dc251dad9db660ac1714da21d2ad61cb71204b205a96fadba2e570712a989aa Feb 02 09:58:22 crc kubenswrapper[4764]: I0202 09:58:22.557537 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"74048046-0204-4a32-957c-2c420c36a3d3","Type":"ContainerStarted","Data":"511775cc124826bc8d7abf3fda9538d110b35dbf169243066b5286751c33aac7"} Feb 02 09:58:22 crc kubenswrapper[4764]: I0202 09:58:22.561464 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-v6n5l" event={"ID":"f001efec-ca89-4c17-be01-3f42fd33e5c0","Type":"ContainerStarted","Data":"f5c1b7838458bf480c48d4f8862d78d8d2ad95451f151514fc66ce939fa6a89b"} Feb 02 09:58:22 crc kubenswrapper[4764]: I0202 09:58:22.561490 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-v6n5l" event={"ID":"f001efec-ca89-4c17-be01-3f42fd33e5c0","Type":"ContainerStarted","Data":"5ee443372866aa9b442f46d12218a261db1bd22763a6e01df1591d5bfe936d70"} Feb 02 09:58:22 crc kubenswrapper[4764]: I0202 09:58:22.583892 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6d4f8f8755-dp7xg" event={"ID":"b79a9079-9d96-4edc-97fc-b9bdf122e4d8","Type":"ContainerStarted","Data":"d7b5b153040634feda48c21ad1fc5dd45a2e9e376c8028d9de2d75e11fe2d349"} Feb 02 09:58:22 crc kubenswrapper[4764]: I0202 09:58:22.600215 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"1c67d8f7-21c3-456d-af99-c96d04090915","Type":"ContainerStarted","Data":"c2a9c05e309552f6b5bb65c3a4319ac17e1c9455d41a70abb385198a832bc04e"} Feb 02 09:58:23 crc kubenswrapper[4764]: I0202 09:58:23.248786 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-db-create-v6n5l" podStartSLOduration=3.248770595 podStartE2EDuration="3.248770595s" podCreationTimestamp="2026-02-02 09:58:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:58:22.612210856 +0000 UTC m=+3065.545934944" watchObservedRunningTime="2026-02-02 09:58:23.248770595 +0000 UTC m=+3066.182494683" Feb 02 09:58:23 crc kubenswrapper[4764]: I0202 09:58:23.262852 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-6d4f8f8755-dp7xg"] Feb 02 09:58:23 crc kubenswrapper[4764]: I0202 09:58:23.270401 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-6c8bdc8fbb-gkt26"] Feb 02 09:58:23 crc kubenswrapper[4764]: I0202 09:58:23.271726 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6c8bdc8fbb-gkt26" Feb 02 09:58:23 crc kubenswrapper[4764]: I0202 09:58:23.275380 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-horizon-svc" Feb 02 09:58:23 crc kubenswrapper[4764]: I0202 09:58:23.310423 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6c8bdc8fbb-gkt26"] Feb 02 09:58:23 crc kubenswrapper[4764]: I0202 09:58:23.347533 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/bd8ea4a3-c3ab-4f61-8257-40adfbbaf5df-horizon-tls-certs\") pod \"horizon-6c8bdc8fbb-gkt26\" (UID: \"bd8ea4a3-c3ab-4f61-8257-40adfbbaf5df\") " pod="openstack/horizon-6c8bdc8fbb-gkt26" Feb 02 09:58:23 crc kubenswrapper[4764]: I0202 09:58:23.347575 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mhfvj\" (UniqueName: \"kubernetes.io/projected/bd8ea4a3-c3ab-4f61-8257-40adfbbaf5df-kube-api-access-mhfvj\") pod \"horizon-6c8bdc8fbb-gkt26\" (UID: \"bd8ea4a3-c3ab-4f61-8257-40adfbbaf5df\") " pod="openstack/horizon-6c8bdc8fbb-gkt26" Feb 02 09:58:23 crc kubenswrapper[4764]: I0202 09:58:23.347649 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd8ea4a3-c3ab-4f61-8257-40adfbbaf5df-combined-ca-bundle\") pod \"horizon-6c8bdc8fbb-gkt26\" (UID: \"bd8ea4a3-c3ab-4f61-8257-40adfbbaf5df\") " pod="openstack/horizon-6c8bdc8fbb-gkt26" Feb 02 09:58:23 crc kubenswrapper[4764]: I0202 09:58:23.347679 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bd8ea4a3-c3ab-4f61-8257-40adfbbaf5df-config-data\") pod \"horizon-6c8bdc8fbb-gkt26\" (UID: \"bd8ea4a3-c3ab-4f61-8257-40adfbbaf5df\") " pod="openstack/horizon-6c8bdc8fbb-gkt26" Feb 02 09:58:23 crc kubenswrapper[4764]: I0202 09:58:23.347699 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/bd8ea4a3-c3ab-4f61-8257-40adfbbaf5df-horizon-secret-key\") pod \"horizon-6c8bdc8fbb-gkt26\" (UID: \"bd8ea4a3-c3ab-4f61-8257-40adfbbaf5df\") " pod="openstack/horizon-6c8bdc8fbb-gkt26" Feb 02 09:58:23 crc kubenswrapper[4764]: I0202 09:58:23.347743 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bd8ea4a3-c3ab-4f61-8257-40adfbbaf5df-scripts\") pod \"horizon-6c8bdc8fbb-gkt26\" (UID: \"bd8ea4a3-c3ab-4f61-8257-40adfbbaf5df\") " pod="openstack/horizon-6c8bdc8fbb-gkt26" Feb 02 09:58:23 crc kubenswrapper[4764]: I0202 09:58:23.347768 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bd8ea4a3-c3ab-4f61-8257-40adfbbaf5df-logs\") pod \"horizon-6c8bdc8fbb-gkt26\" (UID: \"bd8ea4a3-c3ab-4f61-8257-40adfbbaf5df\") " pod="openstack/horizon-6c8bdc8fbb-gkt26" Feb 02 09:58:23 crc kubenswrapper[4764]: I0202 09:58:23.373553 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7f4b5b9665-qqgjw"] Feb 02 09:58:23 crc kubenswrapper[4764]: I0202 09:58:23.445131 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-698ff6d98d-6h7sk"] Feb 02 09:58:23 crc kubenswrapper[4764]: I0202 09:58:23.446550 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-698ff6d98d-6h7sk" Feb 02 09:58:23 crc kubenswrapper[4764]: I0202 09:58:23.461502 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/174bcb22-4370-4c3f-9141-9c1cee0d2542-horizon-secret-key\") pod \"horizon-698ff6d98d-6h7sk\" (UID: \"174bcb22-4370-4c3f-9141-9c1cee0d2542\") " pod="openstack/horizon-698ff6d98d-6h7sk" Feb 02 09:58:23 crc kubenswrapper[4764]: I0202 09:58:23.461605 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n7pgf\" (UniqueName: \"kubernetes.io/projected/174bcb22-4370-4c3f-9141-9c1cee0d2542-kube-api-access-n7pgf\") pod \"horizon-698ff6d98d-6h7sk\" (UID: \"174bcb22-4370-4c3f-9141-9c1cee0d2542\") " pod="openstack/horizon-698ff6d98d-6h7sk" Feb 02 09:58:23 crc kubenswrapper[4764]: I0202 09:58:23.461632 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/174bcb22-4370-4c3f-9141-9c1cee0d2542-combined-ca-bundle\") pod \"horizon-698ff6d98d-6h7sk\" (UID: \"174bcb22-4370-4c3f-9141-9c1cee0d2542\") " pod="openstack/horizon-698ff6d98d-6h7sk" Feb 02 09:58:23 crc kubenswrapper[4764]: I0202 09:58:23.461723 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd8ea4a3-c3ab-4f61-8257-40adfbbaf5df-combined-ca-bundle\") pod \"horizon-6c8bdc8fbb-gkt26\" (UID: \"bd8ea4a3-c3ab-4f61-8257-40adfbbaf5df\") " pod="openstack/horizon-6c8bdc8fbb-gkt26" Feb 02 09:58:23 crc kubenswrapper[4764]: I0202 09:58:23.461770 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/174bcb22-4370-4c3f-9141-9c1cee0d2542-logs\") pod \"horizon-698ff6d98d-6h7sk\" (UID: \"174bcb22-4370-4c3f-9141-9c1cee0d2542\") " pod="openstack/horizon-698ff6d98d-6h7sk" Feb 02 09:58:23 crc kubenswrapper[4764]: I0202 09:58:23.461801 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bd8ea4a3-c3ab-4f61-8257-40adfbbaf5df-config-data\") pod \"horizon-6c8bdc8fbb-gkt26\" (UID: \"bd8ea4a3-c3ab-4f61-8257-40adfbbaf5df\") " pod="openstack/horizon-6c8bdc8fbb-gkt26" Feb 02 09:58:23 crc kubenswrapper[4764]: I0202 09:58:23.461836 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/bd8ea4a3-c3ab-4f61-8257-40adfbbaf5df-horizon-secret-key\") pod \"horizon-6c8bdc8fbb-gkt26\" (UID: \"bd8ea4a3-c3ab-4f61-8257-40adfbbaf5df\") " pod="openstack/horizon-6c8bdc8fbb-gkt26" Feb 02 09:58:23 crc kubenswrapper[4764]: I0202 09:58:23.462042 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bd8ea4a3-c3ab-4f61-8257-40adfbbaf5df-scripts\") pod \"horizon-6c8bdc8fbb-gkt26\" (UID: \"bd8ea4a3-c3ab-4f61-8257-40adfbbaf5df\") " pod="openstack/horizon-6c8bdc8fbb-gkt26" Feb 02 09:58:23 crc kubenswrapper[4764]: I0202 09:58:23.462106 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/174bcb22-4370-4c3f-9141-9c1cee0d2542-horizon-tls-certs\") pod \"horizon-698ff6d98d-6h7sk\" (UID: \"174bcb22-4370-4c3f-9141-9c1cee0d2542\") " pod="openstack/horizon-698ff6d98d-6h7sk" Feb 02 09:58:23 crc kubenswrapper[4764]: I0202 09:58:23.462144 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bd8ea4a3-c3ab-4f61-8257-40adfbbaf5df-logs\") pod \"horizon-6c8bdc8fbb-gkt26\" (UID: \"bd8ea4a3-c3ab-4f61-8257-40adfbbaf5df\") " pod="openstack/horizon-6c8bdc8fbb-gkt26" Feb 02 09:58:23 crc kubenswrapper[4764]: I0202 09:58:23.462204 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/174bcb22-4370-4c3f-9141-9c1cee0d2542-scripts\") pod \"horizon-698ff6d98d-6h7sk\" (UID: \"174bcb22-4370-4c3f-9141-9c1cee0d2542\") " pod="openstack/horizon-698ff6d98d-6h7sk" Feb 02 09:58:23 crc kubenswrapper[4764]: I0202 09:58:23.462228 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/bd8ea4a3-c3ab-4f61-8257-40adfbbaf5df-horizon-tls-certs\") pod \"horizon-6c8bdc8fbb-gkt26\" (UID: \"bd8ea4a3-c3ab-4f61-8257-40adfbbaf5df\") " pod="openstack/horizon-6c8bdc8fbb-gkt26" Feb 02 09:58:23 crc kubenswrapper[4764]: I0202 09:58:23.462254 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mhfvj\" (UniqueName: \"kubernetes.io/projected/bd8ea4a3-c3ab-4f61-8257-40adfbbaf5df-kube-api-access-mhfvj\") pod \"horizon-6c8bdc8fbb-gkt26\" (UID: \"bd8ea4a3-c3ab-4f61-8257-40adfbbaf5df\") " pod="openstack/horizon-6c8bdc8fbb-gkt26" Feb 02 09:58:23 crc kubenswrapper[4764]: I0202 09:58:23.462298 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/174bcb22-4370-4c3f-9141-9c1cee0d2542-config-data\") pod \"horizon-698ff6d98d-6h7sk\" (UID: \"174bcb22-4370-4c3f-9141-9c1cee0d2542\") " pod="openstack/horizon-698ff6d98d-6h7sk" Feb 02 09:58:23 crc kubenswrapper[4764]: I0202 09:58:23.463040 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bd8ea4a3-c3ab-4f61-8257-40adfbbaf5df-scripts\") pod \"horizon-6c8bdc8fbb-gkt26\" (UID: \"bd8ea4a3-c3ab-4f61-8257-40adfbbaf5df\") " pod="openstack/horizon-6c8bdc8fbb-gkt26" Feb 02 09:58:23 crc kubenswrapper[4764]: I0202 09:58:23.464229 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bd8ea4a3-c3ab-4f61-8257-40adfbbaf5df-logs\") pod \"horizon-6c8bdc8fbb-gkt26\" (UID: \"bd8ea4a3-c3ab-4f61-8257-40adfbbaf5df\") " pod="openstack/horizon-6c8bdc8fbb-gkt26" Feb 02 09:58:23 crc kubenswrapper[4764]: I0202 09:58:23.470377 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bd8ea4a3-c3ab-4f61-8257-40adfbbaf5df-config-data\") pod \"horizon-6c8bdc8fbb-gkt26\" (UID: \"bd8ea4a3-c3ab-4f61-8257-40adfbbaf5df\") " pod="openstack/horizon-6c8bdc8fbb-gkt26" Feb 02 09:58:23 crc kubenswrapper[4764]: I0202 09:58:23.496923 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd8ea4a3-c3ab-4f61-8257-40adfbbaf5df-combined-ca-bundle\") pod \"horizon-6c8bdc8fbb-gkt26\" (UID: \"bd8ea4a3-c3ab-4f61-8257-40adfbbaf5df\") " pod="openstack/horizon-6c8bdc8fbb-gkt26" Feb 02 09:58:23 crc kubenswrapper[4764]: I0202 09:58:23.500893 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/bd8ea4a3-c3ab-4f61-8257-40adfbbaf5df-horizon-tls-certs\") pod \"horizon-6c8bdc8fbb-gkt26\" (UID: \"bd8ea4a3-c3ab-4f61-8257-40adfbbaf5df\") " pod="openstack/horizon-6c8bdc8fbb-gkt26" Feb 02 09:58:23 crc kubenswrapper[4764]: I0202 09:58:23.532696 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/bd8ea4a3-c3ab-4f61-8257-40adfbbaf5df-horizon-secret-key\") pod \"horizon-6c8bdc8fbb-gkt26\" (UID: \"bd8ea4a3-c3ab-4f61-8257-40adfbbaf5df\") " pod="openstack/horizon-6c8bdc8fbb-gkt26" Feb 02 09:58:23 crc kubenswrapper[4764]: I0202 09:58:23.533097 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mhfvj\" (UniqueName: \"kubernetes.io/projected/bd8ea4a3-c3ab-4f61-8257-40adfbbaf5df-kube-api-access-mhfvj\") pod \"horizon-6c8bdc8fbb-gkt26\" (UID: \"bd8ea4a3-c3ab-4f61-8257-40adfbbaf5df\") " pod="openstack/horizon-6c8bdc8fbb-gkt26" Feb 02 09:58:23 crc kubenswrapper[4764]: I0202 09:58:23.559761 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-698ff6d98d-6h7sk"] Feb 02 09:58:23 crc kubenswrapper[4764]: I0202 09:58:23.563770 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/174bcb22-4370-4c3f-9141-9c1cee0d2542-horizon-tls-certs\") pod \"horizon-698ff6d98d-6h7sk\" (UID: \"174bcb22-4370-4c3f-9141-9c1cee0d2542\") " pod="openstack/horizon-698ff6d98d-6h7sk" Feb 02 09:58:23 crc kubenswrapper[4764]: I0202 09:58:23.563830 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/174bcb22-4370-4c3f-9141-9c1cee0d2542-scripts\") pod \"horizon-698ff6d98d-6h7sk\" (UID: \"174bcb22-4370-4c3f-9141-9c1cee0d2542\") " pod="openstack/horizon-698ff6d98d-6h7sk" Feb 02 09:58:23 crc kubenswrapper[4764]: I0202 09:58:23.563875 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/174bcb22-4370-4c3f-9141-9c1cee0d2542-config-data\") pod \"horizon-698ff6d98d-6h7sk\" (UID: \"174bcb22-4370-4c3f-9141-9c1cee0d2542\") " pod="openstack/horizon-698ff6d98d-6h7sk" Feb 02 09:58:23 crc kubenswrapper[4764]: I0202 09:58:23.563896 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/174bcb22-4370-4c3f-9141-9c1cee0d2542-horizon-secret-key\") pod \"horizon-698ff6d98d-6h7sk\" (UID: \"174bcb22-4370-4c3f-9141-9c1cee0d2542\") " pod="openstack/horizon-698ff6d98d-6h7sk" Feb 02 09:58:23 crc kubenswrapper[4764]: I0202 09:58:23.565673 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/174bcb22-4370-4c3f-9141-9c1cee0d2542-scripts\") pod \"horizon-698ff6d98d-6h7sk\" (UID: \"174bcb22-4370-4c3f-9141-9c1cee0d2542\") " pod="openstack/horizon-698ff6d98d-6h7sk" Feb 02 09:58:23 crc kubenswrapper[4764]: I0202 09:58:23.569455 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n7pgf\" (UniqueName: \"kubernetes.io/projected/174bcb22-4370-4c3f-9141-9c1cee0d2542-kube-api-access-n7pgf\") pod \"horizon-698ff6d98d-6h7sk\" (UID: \"174bcb22-4370-4c3f-9141-9c1cee0d2542\") " pod="openstack/horizon-698ff6d98d-6h7sk" Feb 02 09:58:23 crc kubenswrapper[4764]: I0202 09:58:23.569488 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/174bcb22-4370-4c3f-9141-9c1cee0d2542-combined-ca-bundle\") pod \"horizon-698ff6d98d-6h7sk\" (UID: \"174bcb22-4370-4c3f-9141-9c1cee0d2542\") " pod="openstack/horizon-698ff6d98d-6h7sk" Feb 02 09:58:23 crc kubenswrapper[4764]: I0202 09:58:23.569551 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/174bcb22-4370-4c3f-9141-9c1cee0d2542-logs\") pod \"horizon-698ff6d98d-6h7sk\" (UID: \"174bcb22-4370-4c3f-9141-9c1cee0d2542\") " pod="openstack/horizon-698ff6d98d-6h7sk" Feb 02 09:58:23 crc kubenswrapper[4764]: I0202 09:58:23.569995 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/174bcb22-4370-4c3f-9141-9c1cee0d2542-logs\") pod \"horizon-698ff6d98d-6h7sk\" (UID: \"174bcb22-4370-4c3f-9141-9c1cee0d2542\") " pod="openstack/horizon-698ff6d98d-6h7sk" Feb 02 09:58:23 crc kubenswrapper[4764]: I0202 09:58:23.577017 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/174bcb22-4370-4c3f-9141-9c1cee0d2542-config-data\") pod \"horizon-698ff6d98d-6h7sk\" (UID: \"174bcb22-4370-4c3f-9141-9c1cee0d2542\") " pod="openstack/horizon-698ff6d98d-6h7sk" Feb 02 09:58:23 crc kubenswrapper[4764]: I0202 09:58:23.586316 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/174bcb22-4370-4c3f-9141-9c1cee0d2542-horizon-secret-key\") pod \"horizon-698ff6d98d-6h7sk\" (UID: \"174bcb22-4370-4c3f-9141-9c1cee0d2542\") " pod="openstack/horizon-698ff6d98d-6h7sk" Feb 02 09:58:23 crc kubenswrapper[4764]: I0202 09:58:23.592819 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/174bcb22-4370-4c3f-9141-9c1cee0d2542-horizon-tls-certs\") pod \"horizon-698ff6d98d-6h7sk\" (UID: \"174bcb22-4370-4c3f-9141-9c1cee0d2542\") " pod="openstack/horizon-698ff6d98d-6h7sk" Feb 02 09:58:23 crc kubenswrapper[4764]: I0202 09:58:23.601374 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/174bcb22-4370-4c3f-9141-9c1cee0d2542-combined-ca-bundle\") pod \"horizon-698ff6d98d-6h7sk\" (UID: \"174bcb22-4370-4c3f-9141-9c1cee0d2542\") " pod="openstack/horizon-698ff6d98d-6h7sk" Feb 02 09:58:23 crc kubenswrapper[4764]: I0202 09:58:23.622292 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n7pgf\" (UniqueName: \"kubernetes.io/projected/174bcb22-4370-4c3f-9141-9c1cee0d2542-kube-api-access-n7pgf\") pod \"horizon-698ff6d98d-6h7sk\" (UID: \"174bcb22-4370-4c3f-9141-9c1cee0d2542\") " pod="openstack/horizon-698ff6d98d-6h7sk" Feb 02 09:58:23 crc kubenswrapper[4764]: I0202 09:58:23.637035 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6c8bdc8fbb-gkt26" Feb 02 09:58:23 crc kubenswrapper[4764]: I0202 09:58:23.654576 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"1c67d8f7-21c3-456d-af99-c96d04090915","Type":"ContainerStarted","Data":"b2f3c82713aeb5a3ef94461b6e06d7872c2c7920165eac8b5f8e004baeaa651c"} Feb 02 09:58:23 crc kubenswrapper[4764]: I0202 09:58:23.680648 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-5e6b-account-create-update-shmv8" event={"ID":"ca0252ea-2836-433b-b26f-bee25fd3adb2","Type":"ContainerStarted","Data":"73fe62e3ef90007d6d8d663dcf54065a81911cd5c1fb0755c45efd55d939ee35"} Feb 02 09:58:23 crc kubenswrapper[4764]: I0202 09:58:23.684807 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-698ff6d98d-6h7sk" Feb 02 09:58:23 crc kubenswrapper[4764]: I0202 09:58:23.691624 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"371fb9ce-9ba4-429b-a68c-62d2e12d688c","Type":"ContainerStarted","Data":"58e16d30111f1f15db02f8a54a45ce207e13f63b2a9cf9ac3b09b73dc7208d85"} Feb 02 09:58:23 crc kubenswrapper[4764]: I0202 09:58:23.720904 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-volume-volume1-0" podStartSLOduration=3.439395518 podStartE2EDuration="4.720889137s" podCreationTimestamp="2026-02-02 09:58:19 +0000 UTC" firstStartedPulling="2026-02-02 09:58:20.507428861 +0000 UTC m=+3063.441152949" lastFinishedPulling="2026-02-02 09:58:21.78892248 +0000 UTC m=+3064.722646568" observedRunningTime="2026-02-02 09:58:23.690627862 +0000 UTC m=+3066.624351950" watchObservedRunningTime="2026-02-02 09:58:23.720889137 +0000 UTC m=+3066.654613225" Feb 02 09:58:23 crc kubenswrapper[4764]: I0202 09:58:23.722159 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"74048046-0204-4a32-957c-2c420c36a3d3","Type":"ContainerStarted","Data":"bb1926fd4aedc103f94030f79cb71ab14a09afd33a83fe613a11fbd6b252ac47"} Feb 02 09:58:23 crc kubenswrapper[4764]: I0202 09:58:23.730095 4764 generic.go:334] "Generic (PLEG): container finished" podID="f001efec-ca89-4c17-be01-3f42fd33e5c0" containerID="f5c1b7838458bf480c48d4f8862d78d8d2ad95451f151514fc66ce939fa6a89b" exitCode=0 Feb 02 09:58:23 crc kubenswrapper[4764]: I0202 09:58:23.730336 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-v6n5l" event={"ID":"f001efec-ca89-4c17-be01-3f42fd33e5c0","Type":"ContainerDied","Data":"f5c1b7838458bf480c48d4f8862d78d8d2ad95451f151514fc66ce939fa6a89b"} Feb 02 09:58:23 crc kubenswrapper[4764]: I0202 09:58:23.731863 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"a2b94631-9510-4324-b049-b4224b0b5825","Type":"ContainerStarted","Data":"442b679c36cc28925886866596e056c002ad9e6b8f5d57a3e2dea6e2a0122ea0"} Feb 02 09:58:23 crc kubenswrapper[4764]: I0202 09:58:23.734224 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-5e6b-account-create-update-shmv8" podStartSLOduration=3.734207701 podStartE2EDuration="3.734207701s" podCreationTimestamp="2026-02-02 09:58:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:58:23.725092119 +0000 UTC m=+3066.658816207" watchObservedRunningTime="2026-02-02 09:58:23.734207701 +0000 UTC m=+3066.667931789" Feb 02 09:58:23 crc kubenswrapper[4764]: I0202 09:58:23.736708 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7f4b5b9665-qqgjw" event={"ID":"165479bf-43a7-47fc-9083-c87cb1ee0f2c","Type":"ContainerStarted","Data":"5dc251dad9db660ac1714da21d2ad61cb71204b205a96fadba2e570712a989aa"} Feb 02 09:58:23 crc kubenswrapper[4764]: I0202 09:58:23.763703 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-backup-0" podStartSLOduration=3.752846429 podStartE2EDuration="4.763684216s" podCreationTimestamp="2026-02-02 09:58:19 +0000 UTC" firstStartedPulling="2026-02-02 09:58:20.780427125 +0000 UTC m=+3063.714151203" lastFinishedPulling="2026-02-02 09:58:21.791264902 +0000 UTC m=+3064.724988990" observedRunningTime="2026-02-02 09:58:23.757115301 +0000 UTC m=+3066.690839389" watchObservedRunningTime="2026-02-02 09:58:23.763684216 +0000 UTC m=+3066.697408294" Feb 02 09:58:24 crc kubenswrapper[4764]: I0202 09:58:24.358191 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-698ff6d98d-6h7sk"] Feb 02 09:58:24 crc kubenswrapper[4764]: I0202 09:58:24.465443 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6c8bdc8fbb-gkt26"] Feb 02 09:58:24 crc kubenswrapper[4764]: I0202 09:58:24.544868 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-backup-0" Feb 02 09:58:24 crc kubenswrapper[4764]: I0202 09:58:24.551847 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-volume-volume1-0" Feb 02 09:58:24 crc kubenswrapper[4764]: I0202 09:58:24.757479 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"74048046-0204-4a32-957c-2c420c36a3d3","Type":"ContainerStarted","Data":"c5a9fb18e5b1532b1ca1b36467b61dfd261eae764d1fb9a9f1ee1d7d3ec5fb84"} Feb 02 09:58:24 crc kubenswrapper[4764]: I0202 09:58:24.757548 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="74048046-0204-4a32-957c-2c420c36a3d3" containerName="glance-log" containerID="cri-o://bb1926fd4aedc103f94030f79cb71ab14a09afd33a83fe613a11fbd6b252ac47" gracePeriod=30 Feb 02 09:58:24 crc kubenswrapper[4764]: I0202 09:58:24.757758 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="74048046-0204-4a32-957c-2c420c36a3d3" containerName="glance-httpd" containerID="cri-o://c5a9fb18e5b1532b1ca1b36467b61dfd261eae764d1fb9a9f1ee1d7d3ec5fb84" gracePeriod=30 Feb 02 09:58:24 crc kubenswrapper[4764]: I0202 09:58:24.765063 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6c8bdc8fbb-gkt26" event={"ID":"bd8ea4a3-c3ab-4f61-8257-40adfbbaf5df","Type":"ContainerStarted","Data":"da638134da8211a36a9d8683d288afbf1a33da4f89967b1503e15dab9a594e74"} Feb 02 09:58:24 crc kubenswrapper[4764]: I0202 09:58:24.768754 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"a2b94631-9510-4324-b049-b4224b0b5825","Type":"ContainerStarted","Data":"cf0870900571b3cad366704bded5f6fc811fd93b947d9852c1a27df445a57563"} Feb 02 09:58:24 crc kubenswrapper[4764]: I0202 09:58:24.774463 4764 generic.go:334] "Generic (PLEG): container finished" podID="ca0252ea-2836-433b-b26f-bee25fd3adb2" containerID="73fe62e3ef90007d6d8d663dcf54065a81911cd5c1fb0755c45efd55d939ee35" exitCode=0 Feb 02 09:58:24 crc kubenswrapper[4764]: I0202 09:58:24.774530 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-5e6b-account-create-update-shmv8" event={"ID":"ca0252ea-2836-433b-b26f-bee25fd3adb2","Type":"ContainerDied","Data":"73fe62e3ef90007d6d8d663dcf54065a81911cd5c1fb0755c45efd55d939ee35"} Feb 02 09:58:24 crc kubenswrapper[4764]: I0202 09:58:24.778532 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-698ff6d98d-6h7sk" event={"ID":"174bcb22-4370-4c3f-9141-9c1cee0d2542","Type":"ContainerStarted","Data":"b23a17f67487e81d88566ad125073cc0c6630e568c8f19aaa6dc6767cac24c26"} Feb 02 09:58:24 crc kubenswrapper[4764]: I0202 09:58:24.793006 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=5.792987053 podStartE2EDuration="5.792987053s" podCreationTimestamp="2026-02-02 09:58:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:58:24.791761821 +0000 UTC m=+3067.725485929" watchObservedRunningTime="2026-02-02 09:58:24.792987053 +0000 UTC m=+3067.726711141" Feb 02 09:58:25 crc kubenswrapper[4764]: I0202 09:58:25.457396 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-v6n5l" Feb 02 09:58:25 crc kubenswrapper[4764]: I0202 09:58:25.518656 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f001efec-ca89-4c17-be01-3f42fd33e5c0-operator-scripts\") pod \"f001efec-ca89-4c17-be01-3f42fd33e5c0\" (UID: \"f001efec-ca89-4c17-be01-3f42fd33e5c0\") " Feb 02 09:58:25 crc kubenswrapper[4764]: I0202 09:58:25.518810 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h4jdb\" (UniqueName: \"kubernetes.io/projected/f001efec-ca89-4c17-be01-3f42fd33e5c0-kube-api-access-h4jdb\") pod \"f001efec-ca89-4c17-be01-3f42fd33e5c0\" (UID: \"f001efec-ca89-4c17-be01-3f42fd33e5c0\") " Feb 02 09:58:25 crc kubenswrapper[4764]: I0202 09:58:25.520709 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f001efec-ca89-4c17-be01-3f42fd33e5c0-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f001efec-ca89-4c17-be01-3f42fd33e5c0" (UID: "f001efec-ca89-4c17-be01-3f42fd33e5c0"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:58:25 crc kubenswrapper[4764]: I0202 09:58:25.530625 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f001efec-ca89-4c17-be01-3f42fd33e5c0-kube-api-access-h4jdb" (OuterVolumeSpecName: "kube-api-access-h4jdb") pod "f001efec-ca89-4c17-be01-3f42fd33e5c0" (UID: "f001efec-ca89-4c17-be01-3f42fd33e5c0"). InnerVolumeSpecName "kube-api-access-h4jdb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:58:25 crc kubenswrapper[4764]: I0202 09:58:25.621176 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h4jdb\" (UniqueName: \"kubernetes.io/projected/f001efec-ca89-4c17-be01-3f42fd33e5c0-kube-api-access-h4jdb\") on node \"crc\" DevicePath \"\"" Feb 02 09:58:25 crc kubenswrapper[4764]: I0202 09:58:25.621207 4764 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f001efec-ca89-4c17-be01-3f42fd33e5c0-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 09:58:25 crc kubenswrapper[4764]: I0202 09:58:25.639212 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 02 09:58:25 crc kubenswrapper[4764]: I0202 09:58:25.722624 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"74048046-0204-4a32-957c-2c420c36a3d3\" (UID: \"74048046-0204-4a32-957c-2c420c36a3d3\") " Feb 02 09:58:25 crc kubenswrapper[4764]: I0202 09:58:25.722673 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/74048046-0204-4a32-957c-2c420c36a3d3-logs\") pod \"74048046-0204-4a32-957c-2c420c36a3d3\" (UID: \"74048046-0204-4a32-957c-2c420c36a3d3\") " Feb 02 09:58:25 crc kubenswrapper[4764]: I0202 09:58:25.722784 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/74048046-0204-4a32-957c-2c420c36a3d3-public-tls-certs\") pod \"74048046-0204-4a32-957c-2c420c36a3d3\" (UID: \"74048046-0204-4a32-957c-2c420c36a3d3\") " Feb 02 09:58:25 crc kubenswrapper[4764]: I0202 09:58:25.722837 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/74048046-0204-4a32-957c-2c420c36a3d3-scripts\") pod \"74048046-0204-4a32-957c-2c420c36a3d3\" (UID: \"74048046-0204-4a32-957c-2c420c36a3d3\") " Feb 02 09:58:25 crc kubenswrapper[4764]: I0202 09:58:25.722893 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74048046-0204-4a32-957c-2c420c36a3d3-combined-ca-bundle\") pod \"74048046-0204-4a32-957c-2c420c36a3d3\" (UID: \"74048046-0204-4a32-957c-2c420c36a3d3\") " Feb 02 09:58:25 crc kubenswrapper[4764]: I0202 09:58:25.722953 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p6l7l\" (UniqueName: \"kubernetes.io/projected/74048046-0204-4a32-957c-2c420c36a3d3-kube-api-access-p6l7l\") pod \"74048046-0204-4a32-957c-2c420c36a3d3\" (UID: \"74048046-0204-4a32-957c-2c420c36a3d3\") " Feb 02 09:58:25 crc kubenswrapper[4764]: I0202 09:58:25.722990 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/74048046-0204-4a32-957c-2c420c36a3d3-ceph\") pod \"74048046-0204-4a32-957c-2c420c36a3d3\" (UID: \"74048046-0204-4a32-957c-2c420c36a3d3\") " Feb 02 09:58:25 crc kubenswrapper[4764]: I0202 09:58:25.723027 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/74048046-0204-4a32-957c-2c420c36a3d3-httpd-run\") pod \"74048046-0204-4a32-957c-2c420c36a3d3\" (UID: \"74048046-0204-4a32-957c-2c420c36a3d3\") " Feb 02 09:58:25 crc kubenswrapper[4764]: I0202 09:58:25.723109 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74048046-0204-4a32-957c-2c420c36a3d3-config-data\") pod \"74048046-0204-4a32-957c-2c420c36a3d3\" (UID: \"74048046-0204-4a32-957c-2c420c36a3d3\") " Feb 02 09:58:25 crc kubenswrapper[4764]: I0202 09:58:25.723533 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/74048046-0204-4a32-957c-2c420c36a3d3-logs" (OuterVolumeSpecName: "logs") pod "74048046-0204-4a32-957c-2c420c36a3d3" (UID: "74048046-0204-4a32-957c-2c420c36a3d3"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 09:58:25 crc kubenswrapper[4764]: I0202 09:58:25.729252 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/74048046-0204-4a32-957c-2c420c36a3d3-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "74048046-0204-4a32-957c-2c420c36a3d3" (UID: "74048046-0204-4a32-957c-2c420c36a3d3"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 09:58:25 crc kubenswrapper[4764]: I0202 09:58:25.732076 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "glance") pod "74048046-0204-4a32-957c-2c420c36a3d3" (UID: "74048046-0204-4a32-957c-2c420c36a3d3"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 02 09:58:25 crc kubenswrapper[4764]: I0202 09:58:25.745231 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/74048046-0204-4a32-957c-2c420c36a3d3-kube-api-access-p6l7l" (OuterVolumeSpecName: "kube-api-access-p6l7l") pod "74048046-0204-4a32-957c-2c420c36a3d3" (UID: "74048046-0204-4a32-957c-2c420c36a3d3"). InnerVolumeSpecName "kube-api-access-p6l7l". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:58:25 crc kubenswrapper[4764]: I0202 09:58:25.759383 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74048046-0204-4a32-957c-2c420c36a3d3-scripts" (OuterVolumeSpecName: "scripts") pod "74048046-0204-4a32-957c-2c420c36a3d3" (UID: "74048046-0204-4a32-957c-2c420c36a3d3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:58:25 crc kubenswrapper[4764]: I0202 09:58:25.765105 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/74048046-0204-4a32-957c-2c420c36a3d3-ceph" (OuterVolumeSpecName: "ceph") pod "74048046-0204-4a32-957c-2c420c36a3d3" (UID: "74048046-0204-4a32-957c-2c420c36a3d3"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:58:25 crc kubenswrapper[4764]: I0202 09:58:25.814101 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74048046-0204-4a32-957c-2c420c36a3d3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "74048046-0204-4a32-957c-2c420c36a3d3" (UID: "74048046-0204-4a32-957c-2c420c36a3d3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:58:25 crc kubenswrapper[4764]: I0202 09:58:25.824951 4764 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/74048046-0204-4a32-957c-2c420c36a3d3-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 09:58:25 crc kubenswrapper[4764]: I0202 09:58:25.824976 4764 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74048046-0204-4a32-957c-2c420c36a3d3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 09:58:25 crc kubenswrapper[4764]: I0202 09:58:25.824988 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p6l7l\" (UniqueName: \"kubernetes.io/projected/74048046-0204-4a32-957c-2c420c36a3d3-kube-api-access-p6l7l\") on node \"crc\" DevicePath \"\"" Feb 02 09:58:25 crc kubenswrapper[4764]: I0202 09:58:25.824997 4764 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/74048046-0204-4a32-957c-2c420c36a3d3-ceph\") on node \"crc\" DevicePath \"\"" Feb 02 09:58:25 crc kubenswrapper[4764]: I0202 09:58:25.825005 4764 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/74048046-0204-4a32-957c-2c420c36a3d3-httpd-run\") on node \"crc\" DevicePath \"\"" Feb 02 09:58:25 crc kubenswrapper[4764]: I0202 09:58:25.825032 4764 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Feb 02 09:58:25 crc kubenswrapper[4764]: I0202 09:58:25.825042 4764 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/74048046-0204-4a32-957c-2c420c36a3d3-logs\") on node \"crc\" DevicePath \"\"" Feb 02 09:58:25 crc kubenswrapper[4764]: I0202 09:58:25.841395 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="a2b94631-9510-4324-b049-b4224b0b5825" containerName="glance-log" containerID="cri-o://cf0870900571b3cad366704bded5f6fc811fd93b947d9852c1a27df445a57563" gracePeriod=30 Feb 02 09:58:25 crc kubenswrapper[4764]: I0202 09:58:25.841769 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="a2b94631-9510-4324-b049-b4224b0b5825" containerName="glance-httpd" containerID="cri-o://27a8ab6dee79fd0c78f6c2ec13caaa965b920e450548fec308ec61d320e1de72" gracePeriod=30 Feb 02 09:58:25 crc kubenswrapper[4764]: I0202 09:58:25.864320 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"a2b94631-9510-4324-b049-b4224b0b5825","Type":"ContainerStarted","Data":"27a8ab6dee79fd0c78f6c2ec13caaa965b920e450548fec308ec61d320e1de72"} Feb 02 09:58:25 crc kubenswrapper[4764]: I0202 09:58:25.879878 4764 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Feb 02 09:58:25 crc kubenswrapper[4764]: I0202 09:58:25.893523 4764 generic.go:334] "Generic (PLEG): container finished" podID="74048046-0204-4a32-957c-2c420c36a3d3" containerID="c5a9fb18e5b1532b1ca1b36467b61dfd261eae764d1fb9a9f1ee1d7d3ec5fb84" exitCode=143 Feb 02 09:58:25 crc kubenswrapper[4764]: I0202 09:58:25.893554 4764 generic.go:334] "Generic (PLEG): container finished" podID="74048046-0204-4a32-957c-2c420c36a3d3" containerID="bb1926fd4aedc103f94030f79cb71ab14a09afd33a83fe613a11fbd6b252ac47" exitCode=143 Feb 02 09:58:25 crc kubenswrapper[4764]: I0202 09:58:25.893619 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"74048046-0204-4a32-957c-2c420c36a3d3","Type":"ContainerDied","Data":"c5a9fb18e5b1532b1ca1b36467b61dfd261eae764d1fb9a9f1ee1d7d3ec5fb84"} Feb 02 09:58:25 crc kubenswrapper[4764]: I0202 09:58:25.893648 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"74048046-0204-4a32-957c-2c420c36a3d3","Type":"ContainerDied","Data":"bb1926fd4aedc103f94030f79cb71ab14a09afd33a83fe613a11fbd6b252ac47"} Feb 02 09:58:25 crc kubenswrapper[4764]: I0202 09:58:25.893659 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"74048046-0204-4a32-957c-2c420c36a3d3","Type":"ContainerDied","Data":"511775cc124826bc8d7abf3fda9538d110b35dbf169243066b5286751c33aac7"} Feb 02 09:58:25 crc kubenswrapper[4764]: I0202 09:58:25.893674 4764 scope.go:117] "RemoveContainer" containerID="c5a9fb18e5b1532b1ca1b36467b61dfd261eae764d1fb9a9f1ee1d7d3ec5fb84" Feb 02 09:58:25 crc kubenswrapper[4764]: I0202 09:58:25.893805 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 02 09:58:25 crc kubenswrapper[4764]: I0202 09:58:25.902282 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74048046-0204-4a32-957c-2c420c36a3d3-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "74048046-0204-4a32-957c-2c420c36a3d3" (UID: "74048046-0204-4a32-957c-2c420c36a3d3"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:58:25 crc kubenswrapper[4764]: I0202 09:58:25.904980 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-v6n5l" Feb 02 09:58:25 crc kubenswrapper[4764]: I0202 09:58:25.905380 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-v6n5l" event={"ID":"f001efec-ca89-4c17-be01-3f42fd33e5c0","Type":"ContainerDied","Data":"5ee443372866aa9b442f46d12218a261db1bd22763a6e01df1591d5bfe936d70"} Feb 02 09:58:25 crc kubenswrapper[4764]: I0202 09:58:25.905402 4764 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5ee443372866aa9b442f46d12218a261db1bd22763a6e01df1591d5bfe936d70" Feb 02 09:58:25 crc kubenswrapper[4764]: I0202 09:58:25.928401 4764 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/74048046-0204-4a32-957c-2c420c36a3d3-public-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 02 09:58:25 crc kubenswrapper[4764]: I0202 09:58:25.928430 4764 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Feb 02 09:58:25 crc kubenswrapper[4764]: I0202 09:58:25.935284 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=6.935266988 podStartE2EDuration="6.935266988s" podCreationTimestamp="2026-02-02 09:58:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:58:25.917537936 +0000 UTC m=+3068.851262024" watchObservedRunningTime="2026-02-02 09:58:25.935266988 +0000 UTC m=+3068.868991076" Feb 02 09:58:25 crc kubenswrapper[4764]: I0202 09:58:25.954160 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74048046-0204-4a32-957c-2c420c36a3d3-config-data" (OuterVolumeSpecName: "config-data") pod "74048046-0204-4a32-957c-2c420c36a3d3" (UID: "74048046-0204-4a32-957c-2c420c36a3d3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:58:25 crc kubenswrapper[4764]: I0202 09:58:25.956665 4764 scope.go:117] "RemoveContainer" containerID="bb1926fd4aedc103f94030f79cb71ab14a09afd33a83fe613a11fbd6b252ac47" Feb 02 09:58:26 crc kubenswrapper[4764]: I0202 09:58:26.027347 4764 scope.go:117] "RemoveContainer" containerID="c5a9fb18e5b1532b1ca1b36467b61dfd261eae764d1fb9a9f1ee1d7d3ec5fb84" Feb 02 09:58:26 crc kubenswrapper[4764]: I0202 09:58:26.030260 4764 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74048046-0204-4a32-957c-2c420c36a3d3-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 09:58:26 crc kubenswrapper[4764]: E0202 09:58:26.035664 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c5a9fb18e5b1532b1ca1b36467b61dfd261eae764d1fb9a9f1ee1d7d3ec5fb84\": container with ID starting with c5a9fb18e5b1532b1ca1b36467b61dfd261eae764d1fb9a9f1ee1d7d3ec5fb84 not found: ID does not exist" containerID="c5a9fb18e5b1532b1ca1b36467b61dfd261eae764d1fb9a9f1ee1d7d3ec5fb84" Feb 02 09:58:26 crc kubenswrapper[4764]: I0202 09:58:26.035773 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c5a9fb18e5b1532b1ca1b36467b61dfd261eae764d1fb9a9f1ee1d7d3ec5fb84"} err="failed to get container status \"c5a9fb18e5b1532b1ca1b36467b61dfd261eae764d1fb9a9f1ee1d7d3ec5fb84\": rpc error: code = NotFound desc = could not find container \"c5a9fb18e5b1532b1ca1b36467b61dfd261eae764d1fb9a9f1ee1d7d3ec5fb84\": container with ID starting with c5a9fb18e5b1532b1ca1b36467b61dfd261eae764d1fb9a9f1ee1d7d3ec5fb84 not found: ID does not exist" Feb 02 09:58:26 crc kubenswrapper[4764]: I0202 09:58:26.035870 4764 scope.go:117] "RemoveContainer" containerID="bb1926fd4aedc103f94030f79cb71ab14a09afd33a83fe613a11fbd6b252ac47" Feb 02 09:58:26 crc kubenswrapper[4764]: E0202 09:58:26.037609 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bb1926fd4aedc103f94030f79cb71ab14a09afd33a83fe613a11fbd6b252ac47\": container with ID starting with bb1926fd4aedc103f94030f79cb71ab14a09afd33a83fe613a11fbd6b252ac47 not found: ID does not exist" containerID="bb1926fd4aedc103f94030f79cb71ab14a09afd33a83fe613a11fbd6b252ac47" Feb 02 09:58:26 crc kubenswrapper[4764]: I0202 09:58:26.037648 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bb1926fd4aedc103f94030f79cb71ab14a09afd33a83fe613a11fbd6b252ac47"} err="failed to get container status \"bb1926fd4aedc103f94030f79cb71ab14a09afd33a83fe613a11fbd6b252ac47\": rpc error: code = NotFound desc = could not find container \"bb1926fd4aedc103f94030f79cb71ab14a09afd33a83fe613a11fbd6b252ac47\": container with ID starting with bb1926fd4aedc103f94030f79cb71ab14a09afd33a83fe613a11fbd6b252ac47 not found: ID does not exist" Feb 02 09:58:26 crc kubenswrapper[4764]: I0202 09:58:26.037675 4764 scope.go:117] "RemoveContainer" containerID="c5a9fb18e5b1532b1ca1b36467b61dfd261eae764d1fb9a9f1ee1d7d3ec5fb84" Feb 02 09:58:26 crc kubenswrapper[4764]: I0202 09:58:26.037950 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c5a9fb18e5b1532b1ca1b36467b61dfd261eae764d1fb9a9f1ee1d7d3ec5fb84"} err="failed to get container status \"c5a9fb18e5b1532b1ca1b36467b61dfd261eae764d1fb9a9f1ee1d7d3ec5fb84\": rpc error: code = NotFound desc = could not find container \"c5a9fb18e5b1532b1ca1b36467b61dfd261eae764d1fb9a9f1ee1d7d3ec5fb84\": container with ID starting with c5a9fb18e5b1532b1ca1b36467b61dfd261eae764d1fb9a9f1ee1d7d3ec5fb84 not found: ID does not exist" Feb 02 09:58:26 crc kubenswrapper[4764]: I0202 09:58:26.037970 4764 scope.go:117] "RemoveContainer" containerID="bb1926fd4aedc103f94030f79cb71ab14a09afd33a83fe613a11fbd6b252ac47" Feb 02 09:58:26 crc kubenswrapper[4764]: I0202 09:58:26.038478 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bb1926fd4aedc103f94030f79cb71ab14a09afd33a83fe613a11fbd6b252ac47"} err="failed to get container status \"bb1926fd4aedc103f94030f79cb71ab14a09afd33a83fe613a11fbd6b252ac47\": rpc error: code = NotFound desc = could not find container \"bb1926fd4aedc103f94030f79cb71ab14a09afd33a83fe613a11fbd6b252ac47\": container with ID starting with bb1926fd4aedc103f94030f79cb71ab14a09afd33a83fe613a11fbd6b252ac47 not found: ID does not exist" Feb 02 09:58:26 crc kubenswrapper[4764]: I0202 09:58:26.263319 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 02 09:58:26 crc kubenswrapper[4764]: I0202 09:58:26.271963 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 02 09:58:26 crc kubenswrapper[4764]: I0202 09:58:26.312020 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Feb 02 09:58:26 crc kubenswrapper[4764]: E0202 09:58:26.312432 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74048046-0204-4a32-957c-2c420c36a3d3" containerName="glance-httpd" Feb 02 09:58:26 crc kubenswrapper[4764]: I0202 09:58:26.312448 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="74048046-0204-4a32-957c-2c420c36a3d3" containerName="glance-httpd" Feb 02 09:58:26 crc kubenswrapper[4764]: E0202 09:58:26.312469 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f001efec-ca89-4c17-be01-3f42fd33e5c0" containerName="mariadb-database-create" Feb 02 09:58:26 crc kubenswrapper[4764]: I0202 09:58:26.312477 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="f001efec-ca89-4c17-be01-3f42fd33e5c0" containerName="mariadb-database-create" Feb 02 09:58:26 crc kubenswrapper[4764]: E0202 09:58:26.312504 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74048046-0204-4a32-957c-2c420c36a3d3" containerName="glance-log" Feb 02 09:58:26 crc kubenswrapper[4764]: I0202 09:58:26.312510 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="74048046-0204-4a32-957c-2c420c36a3d3" containerName="glance-log" Feb 02 09:58:26 crc kubenswrapper[4764]: I0202 09:58:26.312666 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="74048046-0204-4a32-957c-2c420c36a3d3" containerName="glance-httpd" Feb 02 09:58:26 crc kubenswrapper[4764]: I0202 09:58:26.312689 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="74048046-0204-4a32-957c-2c420c36a3d3" containerName="glance-log" Feb 02 09:58:26 crc kubenswrapper[4764]: I0202 09:58:26.312707 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="f001efec-ca89-4c17-be01-3f42fd33e5c0" containerName="mariadb-database-create" Feb 02 09:58:26 crc kubenswrapper[4764]: I0202 09:58:26.313733 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 02 09:58:26 crc kubenswrapper[4764]: I0202 09:58:26.325392 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Feb 02 09:58:26 crc kubenswrapper[4764]: I0202 09:58:26.325489 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Feb 02 09:58:26 crc kubenswrapper[4764]: I0202 09:58:26.346593 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 02 09:58:26 crc kubenswrapper[4764]: I0202 09:58:26.453072 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9f924031-1e49-4009-aeba-a4e5f3476373-logs\") pod \"glance-default-external-api-0\" (UID: \"9f924031-1e49-4009-aeba-a4e5f3476373\") " pod="openstack/glance-default-external-api-0" Feb 02 09:58:26 crc kubenswrapper[4764]: I0202 09:58:26.453130 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9f924031-1e49-4009-aeba-a4e5f3476373-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"9f924031-1e49-4009-aeba-a4e5f3476373\") " pod="openstack/glance-default-external-api-0" Feb 02 09:58:26 crc kubenswrapper[4764]: I0202 09:58:26.453164 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9f924031-1e49-4009-aeba-a4e5f3476373-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"9f924031-1e49-4009-aeba-a4e5f3476373\") " pod="openstack/glance-default-external-api-0" Feb 02 09:58:26 crc kubenswrapper[4764]: I0202 09:58:26.453182 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jdfzk\" (UniqueName: \"kubernetes.io/projected/9f924031-1e49-4009-aeba-a4e5f3476373-kube-api-access-jdfzk\") pod \"glance-default-external-api-0\" (UID: \"9f924031-1e49-4009-aeba-a4e5f3476373\") " pod="openstack/glance-default-external-api-0" Feb 02 09:58:26 crc kubenswrapper[4764]: I0202 09:58:26.453225 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"9f924031-1e49-4009-aeba-a4e5f3476373\") " pod="openstack/glance-default-external-api-0" Feb 02 09:58:26 crc kubenswrapper[4764]: I0202 09:58:26.453258 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9f924031-1e49-4009-aeba-a4e5f3476373-scripts\") pod \"glance-default-external-api-0\" (UID: \"9f924031-1e49-4009-aeba-a4e5f3476373\") " pod="openstack/glance-default-external-api-0" Feb 02 09:58:26 crc kubenswrapper[4764]: I0202 09:58:26.453284 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f924031-1e49-4009-aeba-a4e5f3476373-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"9f924031-1e49-4009-aeba-a4e5f3476373\") " pod="openstack/glance-default-external-api-0" Feb 02 09:58:26 crc kubenswrapper[4764]: I0202 09:58:26.453306 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f924031-1e49-4009-aeba-a4e5f3476373-config-data\") pod \"glance-default-external-api-0\" (UID: \"9f924031-1e49-4009-aeba-a4e5f3476373\") " pod="openstack/glance-default-external-api-0" Feb 02 09:58:26 crc kubenswrapper[4764]: I0202 09:58:26.453322 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/9f924031-1e49-4009-aeba-a4e5f3476373-ceph\") pod \"glance-default-external-api-0\" (UID: \"9f924031-1e49-4009-aeba-a4e5f3476373\") " pod="openstack/glance-default-external-api-0" Feb 02 09:58:26 crc kubenswrapper[4764]: I0202 09:58:26.538032 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-5e6b-account-create-update-shmv8" Feb 02 09:58:26 crc kubenswrapper[4764]: I0202 09:58:26.554992 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9f924031-1e49-4009-aeba-a4e5f3476373-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"9f924031-1e49-4009-aeba-a4e5f3476373\") " pod="openstack/glance-default-external-api-0" Feb 02 09:58:26 crc kubenswrapper[4764]: I0202 09:58:26.555053 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9f924031-1e49-4009-aeba-a4e5f3476373-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"9f924031-1e49-4009-aeba-a4e5f3476373\") " pod="openstack/glance-default-external-api-0" Feb 02 09:58:26 crc kubenswrapper[4764]: I0202 09:58:26.555078 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jdfzk\" (UniqueName: \"kubernetes.io/projected/9f924031-1e49-4009-aeba-a4e5f3476373-kube-api-access-jdfzk\") pod \"glance-default-external-api-0\" (UID: \"9f924031-1e49-4009-aeba-a4e5f3476373\") " pod="openstack/glance-default-external-api-0" Feb 02 09:58:26 crc kubenswrapper[4764]: I0202 09:58:26.555126 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"9f924031-1e49-4009-aeba-a4e5f3476373\") " pod="openstack/glance-default-external-api-0" Feb 02 09:58:26 crc kubenswrapper[4764]: I0202 09:58:26.555158 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9f924031-1e49-4009-aeba-a4e5f3476373-scripts\") pod \"glance-default-external-api-0\" (UID: \"9f924031-1e49-4009-aeba-a4e5f3476373\") " pod="openstack/glance-default-external-api-0" Feb 02 09:58:26 crc kubenswrapper[4764]: I0202 09:58:26.560307 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f924031-1e49-4009-aeba-a4e5f3476373-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"9f924031-1e49-4009-aeba-a4e5f3476373\") " pod="openstack/glance-default-external-api-0" Feb 02 09:58:26 crc kubenswrapper[4764]: I0202 09:58:26.560377 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f924031-1e49-4009-aeba-a4e5f3476373-config-data\") pod \"glance-default-external-api-0\" (UID: \"9f924031-1e49-4009-aeba-a4e5f3476373\") " pod="openstack/glance-default-external-api-0" Feb 02 09:58:26 crc kubenswrapper[4764]: I0202 09:58:26.560403 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/9f924031-1e49-4009-aeba-a4e5f3476373-ceph\") pod \"glance-default-external-api-0\" (UID: \"9f924031-1e49-4009-aeba-a4e5f3476373\") " pod="openstack/glance-default-external-api-0" Feb 02 09:58:26 crc kubenswrapper[4764]: I0202 09:58:26.560546 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9f924031-1e49-4009-aeba-a4e5f3476373-logs\") pod \"glance-default-external-api-0\" (UID: \"9f924031-1e49-4009-aeba-a4e5f3476373\") " pod="openstack/glance-default-external-api-0" Feb 02 09:58:26 crc kubenswrapper[4764]: I0202 09:58:26.567612 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9f924031-1e49-4009-aeba-a4e5f3476373-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"9f924031-1e49-4009-aeba-a4e5f3476373\") " pod="openstack/glance-default-external-api-0" Feb 02 09:58:26 crc kubenswrapper[4764]: I0202 09:58:26.567854 4764 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"9f924031-1e49-4009-aeba-a4e5f3476373\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/glance-default-external-api-0" Feb 02 09:58:26 crc kubenswrapper[4764]: I0202 09:58:26.586277 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9f924031-1e49-4009-aeba-a4e5f3476373-scripts\") pod \"glance-default-external-api-0\" (UID: \"9f924031-1e49-4009-aeba-a4e5f3476373\") " pod="openstack/glance-default-external-api-0" Feb 02 09:58:26 crc kubenswrapper[4764]: I0202 09:58:26.617847 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9f924031-1e49-4009-aeba-a4e5f3476373-logs\") pod \"glance-default-external-api-0\" (UID: \"9f924031-1e49-4009-aeba-a4e5f3476373\") " pod="openstack/glance-default-external-api-0" Feb 02 09:58:26 crc kubenswrapper[4764]: I0202 09:58:26.619066 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f924031-1e49-4009-aeba-a4e5f3476373-config-data\") pod \"glance-default-external-api-0\" (UID: \"9f924031-1e49-4009-aeba-a4e5f3476373\") " pod="openstack/glance-default-external-api-0" Feb 02 09:58:26 crc kubenswrapper[4764]: I0202 09:58:26.619650 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/9f924031-1e49-4009-aeba-a4e5f3476373-ceph\") pod \"glance-default-external-api-0\" (UID: \"9f924031-1e49-4009-aeba-a4e5f3476373\") " pod="openstack/glance-default-external-api-0" Feb 02 09:58:26 crc kubenswrapper[4764]: I0202 09:58:26.658203 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9f924031-1e49-4009-aeba-a4e5f3476373-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"9f924031-1e49-4009-aeba-a4e5f3476373\") " pod="openstack/glance-default-external-api-0" Feb 02 09:58:26 crc kubenswrapper[4764]: I0202 09:58:26.658765 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f924031-1e49-4009-aeba-a4e5f3476373-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"9f924031-1e49-4009-aeba-a4e5f3476373\") " pod="openstack/glance-default-external-api-0" Feb 02 09:58:26 crc kubenswrapper[4764]: I0202 09:58:26.663282 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ca0252ea-2836-433b-b26f-bee25fd3adb2-operator-scripts\") pod \"ca0252ea-2836-433b-b26f-bee25fd3adb2\" (UID: \"ca0252ea-2836-433b-b26f-bee25fd3adb2\") " Feb 02 09:58:26 crc kubenswrapper[4764]: I0202 09:58:26.663390 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-znspp\" (UniqueName: \"kubernetes.io/projected/ca0252ea-2836-433b-b26f-bee25fd3adb2-kube-api-access-znspp\") pod \"ca0252ea-2836-433b-b26f-bee25fd3adb2\" (UID: \"ca0252ea-2836-433b-b26f-bee25fd3adb2\") " Feb 02 09:58:26 crc kubenswrapper[4764]: I0202 09:58:26.669896 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jdfzk\" (UniqueName: \"kubernetes.io/projected/9f924031-1e49-4009-aeba-a4e5f3476373-kube-api-access-jdfzk\") pod \"glance-default-external-api-0\" (UID: \"9f924031-1e49-4009-aeba-a4e5f3476373\") " pod="openstack/glance-default-external-api-0" Feb 02 09:58:26 crc kubenswrapper[4764]: I0202 09:58:26.671600 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ca0252ea-2836-433b-b26f-bee25fd3adb2-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ca0252ea-2836-433b-b26f-bee25fd3adb2" (UID: "ca0252ea-2836-433b-b26f-bee25fd3adb2"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:58:26 crc kubenswrapper[4764]: I0202 09:58:26.677060 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ca0252ea-2836-433b-b26f-bee25fd3adb2-kube-api-access-znspp" (OuterVolumeSpecName: "kube-api-access-znspp") pod "ca0252ea-2836-433b-b26f-bee25fd3adb2" (UID: "ca0252ea-2836-433b-b26f-bee25fd3adb2"). InnerVolumeSpecName "kube-api-access-znspp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:58:26 crc kubenswrapper[4764]: I0202 09:58:26.687039 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"9f924031-1e49-4009-aeba-a4e5f3476373\") " pod="openstack/glance-default-external-api-0" Feb 02 09:58:26 crc kubenswrapper[4764]: I0202 09:58:26.788210 4764 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ca0252ea-2836-433b-b26f-bee25fd3adb2-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 09:58:26 crc kubenswrapper[4764]: I0202 09:58:26.788242 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-znspp\" (UniqueName: \"kubernetes.io/projected/ca0252ea-2836-433b-b26f-bee25fd3adb2-kube-api-access-znspp\") on node \"crc\" DevicePath \"\"" Feb 02 09:58:26 crc kubenswrapper[4764]: I0202 09:58:26.826045 4764 scope.go:117] "RemoveContainer" containerID="db4ebd4a8dbee216b70cb175f017f1799f3ca01a887f602ad42b9f295964da5d" Feb 02 09:58:26 crc kubenswrapper[4764]: E0202 09:58:26.826365 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 09:58:26 crc kubenswrapper[4764]: I0202 09:58:26.903517 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 02 09:58:26 crc kubenswrapper[4764]: I0202 09:58:26.948416 4764 generic.go:334] "Generic (PLEG): container finished" podID="a2b94631-9510-4324-b049-b4224b0b5825" containerID="27a8ab6dee79fd0c78f6c2ec13caaa965b920e450548fec308ec61d320e1de72" exitCode=143 Feb 02 09:58:26 crc kubenswrapper[4764]: I0202 09:58:26.948446 4764 generic.go:334] "Generic (PLEG): container finished" podID="a2b94631-9510-4324-b049-b4224b0b5825" containerID="cf0870900571b3cad366704bded5f6fc811fd93b947d9852c1a27df445a57563" exitCode=143 Feb 02 09:58:26 crc kubenswrapper[4764]: I0202 09:58:26.948566 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 02 09:58:26 crc kubenswrapper[4764]: I0202 09:58:26.948747 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"a2b94631-9510-4324-b049-b4224b0b5825","Type":"ContainerDied","Data":"27a8ab6dee79fd0c78f6c2ec13caaa965b920e450548fec308ec61d320e1de72"} Feb 02 09:58:26 crc kubenswrapper[4764]: I0202 09:58:26.948801 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"a2b94631-9510-4324-b049-b4224b0b5825","Type":"ContainerDied","Data":"cf0870900571b3cad366704bded5f6fc811fd93b947d9852c1a27df445a57563"} Feb 02 09:58:26 crc kubenswrapper[4764]: I0202 09:58:26.948816 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"a2b94631-9510-4324-b049-b4224b0b5825","Type":"ContainerDied","Data":"442b679c36cc28925886866596e056c002ad9e6b8f5d57a3e2dea6e2a0122ea0"} Feb 02 09:58:26 crc kubenswrapper[4764]: I0202 09:58:26.948833 4764 scope.go:117] "RemoveContainer" containerID="27a8ab6dee79fd0c78f6c2ec13caaa965b920e450548fec308ec61d320e1de72" Feb 02 09:58:26 crc kubenswrapper[4764]: I0202 09:58:26.953709 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-5e6b-account-create-update-shmv8" event={"ID":"ca0252ea-2836-433b-b26f-bee25fd3adb2","Type":"ContainerDied","Data":"7fa8b629dc27c26f001bd0ada4ba085624a794c9ec5288294c5c6cbadb204ea6"} Feb 02 09:58:26 crc kubenswrapper[4764]: I0202 09:58:26.953763 4764 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7fa8b629dc27c26f001bd0ada4ba085624a794c9ec5288294c5c6cbadb204ea6" Feb 02 09:58:26 crc kubenswrapper[4764]: I0202 09:58:26.953869 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-5e6b-account-create-update-shmv8" Feb 02 09:58:26 crc kubenswrapper[4764]: I0202 09:58:26.968562 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 02 09:58:26 crc kubenswrapper[4764]: I0202 09:58:26.994817 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/a2b94631-9510-4324-b049-b4224b0b5825-ceph\") pod \"a2b94631-9510-4324-b049-b4224b0b5825\" (UID: \"a2b94631-9510-4324-b049-b4224b0b5825\") " Feb 02 09:58:26 crc kubenswrapper[4764]: I0202 09:58:26.994861 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a2b94631-9510-4324-b049-b4224b0b5825-logs\") pod \"a2b94631-9510-4324-b049-b4224b0b5825\" (UID: \"a2b94631-9510-4324-b049-b4224b0b5825\") " Feb 02 09:58:26 crc kubenswrapper[4764]: I0202 09:58:26.994920 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6xvpd\" (UniqueName: \"kubernetes.io/projected/a2b94631-9510-4324-b049-b4224b0b5825-kube-api-access-6xvpd\") pod \"a2b94631-9510-4324-b049-b4224b0b5825\" (UID: \"a2b94631-9510-4324-b049-b4224b0b5825\") " Feb 02 09:58:26 crc kubenswrapper[4764]: I0202 09:58:26.994956 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a2b94631-9510-4324-b049-b4224b0b5825-internal-tls-certs\") pod \"a2b94631-9510-4324-b049-b4224b0b5825\" (UID: \"a2b94631-9510-4324-b049-b4224b0b5825\") " Feb 02 09:58:26 crc kubenswrapper[4764]: I0202 09:58:26.995000 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"a2b94631-9510-4324-b049-b4224b0b5825\" (UID: \"a2b94631-9510-4324-b049-b4224b0b5825\") " Feb 02 09:58:26 crc kubenswrapper[4764]: I0202 09:58:26.995030 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a2b94631-9510-4324-b049-b4224b0b5825-combined-ca-bundle\") pod \"a2b94631-9510-4324-b049-b4224b0b5825\" (UID: \"a2b94631-9510-4324-b049-b4224b0b5825\") " Feb 02 09:58:26 crc kubenswrapper[4764]: I0202 09:58:26.995073 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a2b94631-9510-4324-b049-b4224b0b5825-config-data\") pod \"a2b94631-9510-4324-b049-b4224b0b5825\" (UID: \"a2b94631-9510-4324-b049-b4224b0b5825\") " Feb 02 09:58:26 crc kubenswrapper[4764]: I0202 09:58:26.995138 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a2b94631-9510-4324-b049-b4224b0b5825-scripts\") pod \"a2b94631-9510-4324-b049-b4224b0b5825\" (UID: \"a2b94631-9510-4324-b049-b4224b0b5825\") " Feb 02 09:58:26 crc kubenswrapper[4764]: I0202 09:58:26.995156 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a2b94631-9510-4324-b049-b4224b0b5825-httpd-run\") pod \"a2b94631-9510-4324-b049-b4224b0b5825\" (UID: \"a2b94631-9510-4324-b049-b4224b0b5825\") " Feb 02 09:58:26 crc kubenswrapper[4764]: I0202 09:58:26.996131 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a2b94631-9510-4324-b049-b4224b0b5825-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "a2b94631-9510-4324-b049-b4224b0b5825" (UID: "a2b94631-9510-4324-b049-b4224b0b5825"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 09:58:27 crc kubenswrapper[4764]: I0202 09:58:27.003913 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a2b94631-9510-4324-b049-b4224b0b5825-ceph" (OuterVolumeSpecName: "ceph") pod "a2b94631-9510-4324-b049-b4224b0b5825" (UID: "a2b94631-9510-4324-b049-b4224b0b5825"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:58:27 crc kubenswrapper[4764]: I0202 09:58:27.005234 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a2b94631-9510-4324-b049-b4224b0b5825-logs" (OuterVolumeSpecName: "logs") pod "a2b94631-9510-4324-b049-b4224b0b5825" (UID: "a2b94631-9510-4324-b049-b4224b0b5825"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 09:58:27 crc kubenswrapper[4764]: I0202 09:58:27.006449 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "glance") pod "a2b94631-9510-4324-b049-b4224b0b5825" (UID: "a2b94631-9510-4324-b049-b4224b0b5825"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 02 09:58:27 crc kubenswrapper[4764]: I0202 09:58:27.009119 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a2b94631-9510-4324-b049-b4224b0b5825-kube-api-access-6xvpd" (OuterVolumeSpecName: "kube-api-access-6xvpd") pod "a2b94631-9510-4324-b049-b4224b0b5825" (UID: "a2b94631-9510-4324-b049-b4224b0b5825"). InnerVolumeSpecName "kube-api-access-6xvpd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:58:27 crc kubenswrapper[4764]: I0202 09:58:27.012575 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a2b94631-9510-4324-b049-b4224b0b5825-scripts" (OuterVolumeSpecName: "scripts") pod "a2b94631-9510-4324-b049-b4224b0b5825" (UID: "a2b94631-9510-4324-b049-b4224b0b5825"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:58:27 crc kubenswrapper[4764]: I0202 09:58:27.063468 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a2b94631-9510-4324-b049-b4224b0b5825-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a2b94631-9510-4324-b049-b4224b0b5825" (UID: "a2b94631-9510-4324-b049-b4224b0b5825"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:58:27 crc kubenswrapper[4764]: I0202 09:58:27.108387 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6xvpd\" (UniqueName: \"kubernetes.io/projected/a2b94631-9510-4324-b049-b4224b0b5825-kube-api-access-6xvpd\") on node \"crc\" DevicePath \"\"" Feb 02 09:58:27 crc kubenswrapper[4764]: I0202 09:58:27.108437 4764 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Feb 02 09:58:27 crc kubenswrapper[4764]: I0202 09:58:27.108448 4764 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a2b94631-9510-4324-b049-b4224b0b5825-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 09:58:27 crc kubenswrapper[4764]: I0202 09:58:27.108457 4764 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a2b94631-9510-4324-b049-b4224b0b5825-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 09:58:27 crc kubenswrapper[4764]: I0202 09:58:27.108467 4764 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a2b94631-9510-4324-b049-b4224b0b5825-httpd-run\") on node \"crc\" DevicePath \"\"" Feb 02 09:58:27 crc kubenswrapper[4764]: I0202 09:58:27.108475 4764 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/a2b94631-9510-4324-b049-b4224b0b5825-ceph\") on node \"crc\" DevicePath \"\"" Feb 02 09:58:27 crc kubenswrapper[4764]: I0202 09:58:27.108483 4764 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a2b94631-9510-4324-b049-b4224b0b5825-logs\") on node \"crc\" DevicePath \"\"" Feb 02 09:58:27 crc kubenswrapper[4764]: I0202 09:58:27.109435 4764 scope.go:117] "RemoveContainer" containerID="cf0870900571b3cad366704bded5f6fc811fd93b947d9852c1a27df445a57563" Feb 02 09:58:27 crc kubenswrapper[4764]: I0202 09:58:27.118630 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a2b94631-9510-4324-b049-b4224b0b5825-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "a2b94631-9510-4324-b049-b4224b0b5825" (UID: "a2b94631-9510-4324-b049-b4224b0b5825"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:58:27 crc kubenswrapper[4764]: I0202 09:58:27.119416 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a2b94631-9510-4324-b049-b4224b0b5825-config-data" (OuterVolumeSpecName: "config-data") pod "a2b94631-9510-4324-b049-b4224b0b5825" (UID: "a2b94631-9510-4324-b049-b4224b0b5825"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:58:27 crc kubenswrapper[4764]: I0202 09:58:27.150738 4764 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Feb 02 09:58:27 crc kubenswrapper[4764]: I0202 09:58:27.178139 4764 scope.go:117] "RemoveContainer" containerID="27a8ab6dee79fd0c78f6c2ec13caaa965b920e450548fec308ec61d320e1de72" Feb 02 09:58:27 crc kubenswrapper[4764]: E0202 09:58:27.178822 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"27a8ab6dee79fd0c78f6c2ec13caaa965b920e450548fec308ec61d320e1de72\": container with ID starting with 27a8ab6dee79fd0c78f6c2ec13caaa965b920e450548fec308ec61d320e1de72 not found: ID does not exist" containerID="27a8ab6dee79fd0c78f6c2ec13caaa965b920e450548fec308ec61d320e1de72" Feb 02 09:58:27 crc kubenswrapper[4764]: I0202 09:58:27.178863 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"27a8ab6dee79fd0c78f6c2ec13caaa965b920e450548fec308ec61d320e1de72"} err="failed to get container status \"27a8ab6dee79fd0c78f6c2ec13caaa965b920e450548fec308ec61d320e1de72\": rpc error: code = NotFound desc = could not find container \"27a8ab6dee79fd0c78f6c2ec13caaa965b920e450548fec308ec61d320e1de72\": container with ID starting with 27a8ab6dee79fd0c78f6c2ec13caaa965b920e450548fec308ec61d320e1de72 not found: ID does not exist" Feb 02 09:58:27 crc kubenswrapper[4764]: I0202 09:58:27.178884 4764 scope.go:117] "RemoveContainer" containerID="cf0870900571b3cad366704bded5f6fc811fd93b947d9852c1a27df445a57563" Feb 02 09:58:27 crc kubenswrapper[4764]: E0202 09:58:27.179066 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cf0870900571b3cad366704bded5f6fc811fd93b947d9852c1a27df445a57563\": container with ID starting with cf0870900571b3cad366704bded5f6fc811fd93b947d9852c1a27df445a57563 not found: ID does not exist" containerID="cf0870900571b3cad366704bded5f6fc811fd93b947d9852c1a27df445a57563" Feb 02 09:58:27 crc kubenswrapper[4764]: I0202 09:58:27.179089 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cf0870900571b3cad366704bded5f6fc811fd93b947d9852c1a27df445a57563"} err="failed to get container status \"cf0870900571b3cad366704bded5f6fc811fd93b947d9852c1a27df445a57563\": rpc error: code = NotFound desc = could not find container \"cf0870900571b3cad366704bded5f6fc811fd93b947d9852c1a27df445a57563\": container with ID starting with cf0870900571b3cad366704bded5f6fc811fd93b947d9852c1a27df445a57563 not found: ID does not exist" Feb 02 09:58:27 crc kubenswrapper[4764]: I0202 09:58:27.179102 4764 scope.go:117] "RemoveContainer" containerID="27a8ab6dee79fd0c78f6c2ec13caaa965b920e450548fec308ec61d320e1de72" Feb 02 09:58:27 crc kubenswrapper[4764]: I0202 09:58:27.179271 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"27a8ab6dee79fd0c78f6c2ec13caaa965b920e450548fec308ec61d320e1de72"} err="failed to get container status \"27a8ab6dee79fd0c78f6c2ec13caaa965b920e450548fec308ec61d320e1de72\": rpc error: code = NotFound desc = could not find container \"27a8ab6dee79fd0c78f6c2ec13caaa965b920e450548fec308ec61d320e1de72\": container with ID starting with 27a8ab6dee79fd0c78f6c2ec13caaa965b920e450548fec308ec61d320e1de72 not found: ID does not exist" Feb 02 09:58:27 crc kubenswrapper[4764]: I0202 09:58:27.179288 4764 scope.go:117] "RemoveContainer" containerID="cf0870900571b3cad366704bded5f6fc811fd93b947d9852c1a27df445a57563" Feb 02 09:58:27 crc kubenswrapper[4764]: I0202 09:58:27.179529 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cf0870900571b3cad366704bded5f6fc811fd93b947d9852c1a27df445a57563"} err="failed to get container status \"cf0870900571b3cad366704bded5f6fc811fd93b947d9852c1a27df445a57563\": rpc error: code = NotFound desc = could not find container \"cf0870900571b3cad366704bded5f6fc811fd93b947d9852c1a27df445a57563\": container with ID starting with cf0870900571b3cad366704bded5f6fc811fd93b947d9852c1a27df445a57563 not found: ID does not exist" Feb 02 09:58:27 crc kubenswrapper[4764]: I0202 09:58:27.210011 4764 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a2b94631-9510-4324-b049-b4224b0b5825-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 02 09:58:27 crc kubenswrapper[4764]: I0202 09:58:27.210076 4764 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Feb 02 09:58:27 crc kubenswrapper[4764]: I0202 09:58:27.210100 4764 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a2b94631-9510-4324-b049-b4224b0b5825-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 09:58:27 crc kubenswrapper[4764]: I0202 09:58:27.323228 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 02 09:58:27 crc kubenswrapper[4764]: I0202 09:58:27.343868 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 02 09:58:27 crc kubenswrapper[4764]: I0202 09:58:27.358991 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 02 09:58:27 crc kubenswrapper[4764]: E0202 09:58:27.359379 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2b94631-9510-4324-b049-b4224b0b5825" containerName="glance-httpd" Feb 02 09:58:27 crc kubenswrapper[4764]: I0202 09:58:27.359391 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2b94631-9510-4324-b049-b4224b0b5825" containerName="glance-httpd" Feb 02 09:58:27 crc kubenswrapper[4764]: E0202 09:58:27.359400 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2b94631-9510-4324-b049-b4224b0b5825" containerName="glance-log" Feb 02 09:58:27 crc kubenswrapper[4764]: I0202 09:58:27.359406 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2b94631-9510-4324-b049-b4224b0b5825" containerName="glance-log" Feb 02 09:58:27 crc kubenswrapper[4764]: E0202 09:58:27.359426 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca0252ea-2836-433b-b26f-bee25fd3adb2" containerName="mariadb-account-create-update" Feb 02 09:58:27 crc kubenswrapper[4764]: I0202 09:58:27.359432 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca0252ea-2836-433b-b26f-bee25fd3adb2" containerName="mariadb-account-create-update" Feb 02 09:58:27 crc kubenswrapper[4764]: I0202 09:58:27.359619 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="a2b94631-9510-4324-b049-b4224b0b5825" containerName="glance-log" Feb 02 09:58:27 crc kubenswrapper[4764]: I0202 09:58:27.359628 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="a2b94631-9510-4324-b049-b4224b0b5825" containerName="glance-httpd" Feb 02 09:58:27 crc kubenswrapper[4764]: I0202 09:58:27.359641 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="ca0252ea-2836-433b-b26f-bee25fd3adb2" containerName="mariadb-account-create-update" Feb 02 09:58:27 crc kubenswrapper[4764]: I0202 09:58:27.360591 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 02 09:58:27 crc kubenswrapper[4764]: I0202 09:58:27.367266 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Feb 02 09:58:27 crc kubenswrapper[4764]: I0202 09:58:27.367441 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Feb 02 09:58:27 crc kubenswrapper[4764]: I0202 09:58:27.371697 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 02 09:58:27 crc kubenswrapper[4764]: I0202 09:58:27.516622 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"687693a0-eb78-4922-991f-6738de00f468\") " pod="openstack/glance-default-internal-api-0" Feb 02 09:58:27 crc kubenswrapper[4764]: I0202 09:58:27.516911 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/687693a0-eb78-4922-991f-6738de00f468-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"687693a0-eb78-4922-991f-6738de00f468\") " pod="openstack/glance-default-internal-api-0" Feb 02 09:58:27 crc kubenswrapper[4764]: I0202 09:58:27.517026 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/687693a0-eb78-4922-991f-6738de00f468-scripts\") pod \"glance-default-internal-api-0\" (UID: \"687693a0-eb78-4922-991f-6738de00f468\") " pod="openstack/glance-default-internal-api-0" Feb 02 09:58:27 crc kubenswrapper[4764]: I0202 09:58:27.517107 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/687693a0-eb78-4922-991f-6738de00f468-config-data\") pod \"glance-default-internal-api-0\" (UID: \"687693a0-eb78-4922-991f-6738de00f468\") " pod="openstack/glance-default-internal-api-0" Feb 02 09:58:27 crc kubenswrapper[4764]: I0202 09:58:27.517177 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/687693a0-eb78-4922-991f-6738de00f468-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"687693a0-eb78-4922-991f-6738de00f468\") " pod="openstack/glance-default-internal-api-0" Feb 02 09:58:27 crc kubenswrapper[4764]: I0202 09:58:27.517289 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/687693a0-eb78-4922-991f-6738de00f468-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"687693a0-eb78-4922-991f-6738de00f468\") " pod="openstack/glance-default-internal-api-0" Feb 02 09:58:27 crc kubenswrapper[4764]: I0202 09:58:27.517370 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/687693a0-eb78-4922-991f-6738de00f468-logs\") pod \"glance-default-internal-api-0\" (UID: \"687693a0-eb78-4922-991f-6738de00f468\") " pod="openstack/glance-default-internal-api-0" Feb 02 09:58:27 crc kubenswrapper[4764]: I0202 09:58:27.517478 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/687693a0-eb78-4922-991f-6738de00f468-ceph\") pod \"glance-default-internal-api-0\" (UID: \"687693a0-eb78-4922-991f-6738de00f468\") " pod="openstack/glance-default-internal-api-0" Feb 02 09:58:27 crc kubenswrapper[4764]: I0202 09:58:27.517602 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hwvqx\" (UniqueName: \"kubernetes.io/projected/687693a0-eb78-4922-991f-6738de00f468-kube-api-access-hwvqx\") pod \"glance-default-internal-api-0\" (UID: \"687693a0-eb78-4922-991f-6738de00f468\") " pod="openstack/glance-default-internal-api-0" Feb 02 09:58:27 crc kubenswrapper[4764]: I0202 09:58:27.619036 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/687693a0-eb78-4922-991f-6738de00f468-scripts\") pod \"glance-default-internal-api-0\" (UID: \"687693a0-eb78-4922-991f-6738de00f468\") " pod="openstack/glance-default-internal-api-0" Feb 02 09:58:27 crc kubenswrapper[4764]: I0202 09:58:27.619083 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/687693a0-eb78-4922-991f-6738de00f468-config-data\") pod \"glance-default-internal-api-0\" (UID: \"687693a0-eb78-4922-991f-6738de00f468\") " pod="openstack/glance-default-internal-api-0" Feb 02 09:58:27 crc kubenswrapper[4764]: I0202 09:58:27.619101 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/687693a0-eb78-4922-991f-6738de00f468-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"687693a0-eb78-4922-991f-6738de00f468\") " pod="openstack/glance-default-internal-api-0" Feb 02 09:58:27 crc kubenswrapper[4764]: I0202 09:58:27.619140 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/687693a0-eb78-4922-991f-6738de00f468-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"687693a0-eb78-4922-991f-6738de00f468\") " pod="openstack/glance-default-internal-api-0" Feb 02 09:58:27 crc kubenswrapper[4764]: I0202 09:58:27.619163 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/687693a0-eb78-4922-991f-6738de00f468-logs\") pod \"glance-default-internal-api-0\" (UID: \"687693a0-eb78-4922-991f-6738de00f468\") " pod="openstack/glance-default-internal-api-0" Feb 02 09:58:27 crc kubenswrapper[4764]: I0202 09:58:27.619223 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/687693a0-eb78-4922-991f-6738de00f468-ceph\") pod \"glance-default-internal-api-0\" (UID: \"687693a0-eb78-4922-991f-6738de00f468\") " pod="openstack/glance-default-internal-api-0" Feb 02 09:58:27 crc kubenswrapper[4764]: I0202 09:58:27.619318 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hwvqx\" (UniqueName: \"kubernetes.io/projected/687693a0-eb78-4922-991f-6738de00f468-kube-api-access-hwvqx\") pod \"glance-default-internal-api-0\" (UID: \"687693a0-eb78-4922-991f-6738de00f468\") " pod="openstack/glance-default-internal-api-0" Feb 02 09:58:27 crc kubenswrapper[4764]: I0202 09:58:27.619390 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"687693a0-eb78-4922-991f-6738de00f468\") " pod="openstack/glance-default-internal-api-0" Feb 02 09:58:27 crc kubenswrapper[4764]: I0202 09:58:27.619411 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/687693a0-eb78-4922-991f-6738de00f468-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"687693a0-eb78-4922-991f-6738de00f468\") " pod="openstack/glance-default-internal-api-0" Feb 02 09:58:27 crc kubenswrapper[4764]: I0202 09:58:27.619823 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/687693a0-eb78-4922-991f-6738de00f468-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"687693a0-eb78-4922-991f-6738de00f468\") " pod="openstack/glance-default-internal-api-0" Feb 02 09:58:27 crc kubenswrapper[4764]: I0202 09:58:27.620666 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/687693a0-eb78-4922-991f-6738de00f468-logs\") pod \"glance-default-internal-api-0\" (UID: \"687693a0-eb78-4922-991f-6738de00f468\") " pod="openstack/glance-default-internal-api-0" Feb 02 09:58:27 crc kubenswrapper[4764]: I0202 09:58:27.622891 4764 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"687693a0-eb78-4922-991f-6738de00f468\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/glance-default-internal-api-0" Feb 02 09:58:27 crc kubenswrapper[4764]: I0202 09:58:27.633569 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/687693a0-eb78-4922-991f-6738de00f468-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"687693a0-eb78-4922-991f-6738de00f468\") " pod="openstack/glance-default-internal-api-0" Feb 02 09:58:27 crc kubenswrapper[4764]: I0202 09:58:27.633895 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/687693a0-eb78-4922-991f-6738de00f468-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"687693a0-eb78-4922-991f-6738de00f468\") " pod="openstack/glance-default-internal-api-0" Feb 02 09:58:27 crc kubenswrapper[4764]: I0202 09:58:27.635391 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/687693a0-eb78-4922-991f-6738de00f468-config-data\") pod \"glance-default-internal-api-0\" (UID: \"687693a0-eb78-4922-991f-6738de00f468\") " pod="openstack/glance-default-internal-api-0" Feb 02 09:58:27 crc kubenswrapper[4764]: I0202 09:58:27.643761 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/687693a0-eb78-4922-991f-6738de00f468-scripts\") pod \"glance-default-internal-api-0\" (UID: \"687693a0-eb78-4922-991f-6738de00f468\") " pod="openstack/glance-default-internal-api-0" Feb 02 09:58:27 crc kubenswrapper[4764]: I0202 09:58:27.647482 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/687693a0-eb78-4922-991f-6738de00f468-ceph\") pod \"glance-default-internal-api-0\" (UID: \"687693a0-eb78-4922-991f-6738de00f468\") " pod="openstack/glance-default-internal-api-0" Feb 02 09:58:27 crc kubenswrapper[4764]: I0202 09:58:27.654279 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hwvqx\" (UniqueName: \"kubernetes.io/projected/687693a0-eb78-4922-991f-6738de00f468-kube-api-access-hwvqx\") pod \"glance-default-internal-api-0\" (UID: \"687693a0-eb78-4922-991f-6738de00f468\") " pod="openstack/glance-default-internal-api-0" Feb 02 09:58:27 crc kubenswrapper[4764]: I0202 09:58:27.691239 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"687693a0-eb78-4922-991f-6738de00f468\") " pod="openstack/glance-default-internal-api-0" Feb 02 09:58:27 crc kubenswrapper[4764]: I0202 09:58:27.796770 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 02 09:58:27 crc kubenswrapper[4764]: I0202 09:58:27.846078 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="74048046-0204-4a32-957c-2c420c36a3d3" path="/var/lib/kubelet/pods/74048046-0204-4a32-957c-2c420c36a3d3/volumes" Feb 02 09:58:27 crc kubenswrapper[4764]: I0202 09:58:27.846924 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a2b94631-9510-4324-b049-b4224b0b5825" path="/var/lib/kubelet/pods/a2b94631-9510-4324-b049-b4224b0b5825/volumes" Feb 02 09:58:27 crc kubenswrapper[4764]: I0202 09:58:27.981062 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 02 09:58:27 crc kubenswrapper[4764]: I0202 09:58:27.984077 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"9f924031-1e49-4009-aeba-a4e5f3476373","Type":"ContainerStarted","Data":"76838d2918197b5acfb9332b05755231b5310f27b1ff8958fe12e1c7cee1423a"} Feb 02 09:58:28 crc kubenswrapper[4764]: I0202 09:58:28.716958 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 02 09:58:28 crc kubenswrapper[4764]: I0202 09:58:28.999255 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"687693a0-eb78-4922-991f-6738de00f468","Type":"ContainerStarted","Data":"9417550c3763893a0dd25338364384b5d434a27686dde3af14820f85f226d939"} Feb 02 09:58:29 crc kubenswrapper[4764]: I0202 09:58:29.005870 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"9f924031-1e49-4009-aeba-a4e5f3476373","Type":"ContainerStarted","Data":"a774f78b8dedf828227ce850aee54177dc3ba75277f49062eef897dda98d4bfc"} Feb 02 09:58:29 crc kubenswrapper[4764]: I0202 09:58:29.786819 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-volume-volume1-0" Feb 02 09:58:29 crc kubenswrapper[4764]: I0202 09:58:29.925060 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-backup-0" Feb 02 09:58:30 crc kubenswrapper[4764]: I0202 09:58:30.035036 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"687693a0-eb78-4922-991f-6738de00f468","Type":"ContainerStarted","Data":"da3547545ecd927f987c2ab629b57d9709ef9bcad4bcdf29896e439ee8ec878a"} Feb 02 09:58:30 crc kubenswrapper[4764]: I0202 09:58:30.036922 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"9f924031-1e49-4009-aeba-a4e5f3476373","Type":"ContainerStarted","Data":"8b301087312b721ab72b16f0a6718d780d21d533a36c52e65a2e050cc00e29fa"} Feb 02 09:58:30 crc kubenswrapper[4764]: I0202 09:58:30.637048 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=4.637029416 podStartE2EDuration="4.637029416s" podCreationTimestamp="2026-02-02 09:58:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:58:30.061114571 +0000 UTC m=+3072.994838659" watchObservedRunningTime="2026-02-02 09:58:30.637029416 +0000 UTC m=+3073.570753494" Feb 02 09:58:30 crc kubenswrapper[4764]: I0202 09:58:30.646369 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-db-sync-xvkkk"] Feb 02 09:58:30 crc kubenswrapper[4764]: I0202 09:58:30.647632 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-xvkkk" Feb 02 09:58:30 crc kubenswrapper[4764]: I0202 09:58:30.653172 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-manila-dockercfg-8bjjz" Feb 02 09:58:30 crc kubenswrapper[4764]: I0202 09:58:30.654093 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-sync-xvkkk"] Feb 02 09:58:30 crc kubenswrapper[4764]: I0202 09:58:30.655074 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-config-data" Feb 02 09:58:30 crc kubenswrapper[4764]: I0202 09:58:30.680498 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/1bf7db82-4332-485f-acbf-3edbf4269f50-job-config-data\") pod \"manila-db-sync-xvkkk\" (UID: \"1bf7db82-4332-485f-acbf-3edbf4269f50\") " pod="openstack/manila-db-sync-xvkkk" Feb 02 09:58:30 crc kubenswrapper[4764]: I0202 09:58:30.680564 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1bf7db82-4332-485f-acbf-3edbf4269f50-config-data\") pod \"manila-db-sync-xvkkk\" (UID: \"1bf7db82-4332-485f-acbf-3edbf4269f50\") " pod="openstack/manila-db-sync-xvkkk" Feb 02 09:58:30 crc kubenswrapper[4764]: I0202 09:58:30.680616 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bf7db82-4332-485f-acbf-3edbf4269f50-combined-ca-bundle\") pod \"manila-db-sync-xvkkk\" (UID: \"1bf7db82-4332-485f-acbf-3edbf4269f50\") " pod="openstack/manila-db-sync-xvkkk" Feb 02 09:58:30 crc kubenswrapper[4764]: I0202 09:58:30.680671 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kn7cc\" (UniqueName: \"kubernetes.io/projected/1bf7db82-4332-485f-acbf-3edbf4269f50-kube-api-access-kn7cc\") pod \"manila-db-sync-xvkkk\" (UID: \"1bf7db82-4332-485f-acbf-3edbf4269f50\") " pod="openstack/manila-db-sync-xvkkk" Feb 02 09:58:30 crc kubenswrapper[4764]: I0202 09:58:30.782075 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1bf7db82-4332-485f-acbf-3edbf4269f50-config-data\") pod \"manila-db-sync-xvkkk\" (UID: \"1bf7db82-4332-485f-acbf-3edbf4269f50\") " pod="openstack/manila-db-sync-xvkkk" Feb 02 09:58:30 crc kubenswrapper[4764]: I0202 09:58:30.782504 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bf7db82-4332-485f-acbf-3edbf4269f50-combined-ca-bundle\") pod \"manila-db-sync-xvkkk\" (UID: \"1bf7db82-4332-485f-acbf-3edbf4269f50\") " pod="openstack/manila-db-sync-xvkkk" Feb 02 09:58:30 crc kubenswrapper[4764]: I0202 09:58:30.782573 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kn7cc\" (UniqueName: \"kubernetes.io/projected/1bf7db82-4332-485f-acbf-3edbf4269f50-kube-api-access-kn7cc\") pod \"manila-db-sync-xvkkk\" (UID: \"1bf7db82-4332-485f-acbf-3edbf4269f50\") " pod="openstack/manila-db-sync-xvkkk" Feb 02 09:58:30 crc kubenswrapper[4764]: I0202 09:58:30.782648 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/1bf7db82-4332-485f-acbf-3edbf4269f50-job-config-data\") pod \"manila-db-sync-xvkkk\" (UID: \"1bf7db82-4332-485f-acbf-3edbf4269f50\") " pod="openstack/manila-db-sync-xvkkk" Feb 02 09:58:30 crc kubenswrapper[4764]: I0202 09:58:30.793071 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bf7db82-4332-485f-acbf-3edbf4269f50-combined-ca-bundle\") pod \"manila-db-sync-xvkkk\" (UID: \"1bf7db82-4332-485f-acbf-3edbf4269f50\") " pod="openstack/manila-db-sync-xvkkk" Feb 02 09:58:30 crc kubenswrapper[4764]: I0202 09:58:30.808047 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/1bf7db82-4332-485f-acbf-3edbf4269f50-job-config-data\") pod \"manila-db-sync-xvkkk\" (UID: \"1bf7db82-4332-485f-acbf-3edbf4269f50\") " pod="openstack/manila-db-sync-xvkkk" Feb 02 09:58:30 crc kubenswrapper[4764]: I0202 09:58:30.810043 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1bf7db82-4332-485f-acbf-3edbf4269f50-config-data\") pod \"manila-db-sync-xvkkk\" (UID: \"1bf7db82-4332-485f-acbf-3edbf4269f50\") " pod="openstack/manila-db-sync-xvkkk" Feb 02 09:58:30 crc kubenswrapper[4764]: I0202 09:58:30.818746 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kn7cc\" (UniqueName: \"kubernetes.io/projected/1bf7db82-4332-485f-acbf-3edbf4269f50-kube-api-access-kn7cc\") pod \"manila-db-sync-xvkkk\" (UID: \"1bf7db82-4332-485f-acbf-3edbf4269f50\") " pod="openstack/manila-db-sync-xvkkk" Feb 02 09:58:30 crc kubenswrapper[4764]: I0202 09:58:30.966818 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-xvkkk" Feb 02 09:58:31 crc kubenswrapper[4764]: I0202 09:58:31.080981 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"687693a0-eb78-4922-991f-6738de00f468","Type":"ContainerStarted","Data":"f1c7d9b28d62f55d411456894c32e624fe4484a1b76229151bc8698970200716"} Feb 02 09:58:31 crc kubenswrapper[4764]: I0202 09:58:31.126427 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=4.126408806 podStartE2EDuration="4.126408806s" podCreationTimestamp="2026-02-02 09:58:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:58:31.108259234 +0000 UTC m=+3074.041983342" watchObservedRunningTime="2026-02-02 09:58:31.126408806 +0000 UTC m=+3074.060132894" Feb 02 09:58:31 crc kubenswrapper[4764]: I0202 09:58:31.609796 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-sync-xvkkk"] Feb 02 09:58:36 crc kubenswrapper[4764]: I0202 09:58:36.968739 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Feb 02 09:58:36 crc kubenswrapper[4764]: I0202 09:58:36.969376 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Feb 02 09:58:37 crc kubenswrapper[4764]: I0202 09:58:37.002275 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Feb 02 09:58:37 crc kubenswrapper[4764]: I0202 09:58:37.039728 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Feb 02 09:58:37 crc kubenswrapper[4764]: I0202 09:58:37.143560 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6d4f8f8755-dp7xg" event={"ID":"b79a9079-9d96-4edc-97fc-b9bdf122e4d8","Type":"ContainerStarted","Data":"71fc5c9189af4e03a6f6fc0a3c0dbb2e1e002c4c3ba7e050a1b56f82fac1fc29"} Feb 02 09:58:37 crc kubenswrapper[4764]: I0202 09:58:37.147482 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7f4b5b9665-qqgjw" event={"ID":"165479bf-43a7-47fc-9083-c87cb1ee0f2c","Type":"ContainerStarted","Data":"5c0ef8d37db32c9fb6af9fafb8fd85ea7920d910afd0c5afef3e02e43598b8de"} Feb 02 09:58:37 crc kubenswrapper[4764]: I0202 09:58:37.149512 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-698ff6d98d-6h7sk" event={"ID":"174bcb22-4370-4c3f-9141-9c1cee0d2542","Type":"ContainerStarted","Data":"f7bc1f8a0d77ef7aa34ff43a69c2507b000866ce4bec8aa13d4aa885677916b1"} Feb 02 09:58:37 crc kubenswrapper[4764]: I0202 09:58:37.151528 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-xvkkk" event={"ID":"1bf7db82-4332-485f-acbf-3edbf4269f50","Type":"ContainerStarted","Data":"6d1307c11c429a051bcbc5b8747d4862c15d9ef65642394ad1e370bfc1389e46"} Feb 02 09:58:37 crc kubenswrapper[4764]: I0202 09:58:37.154435 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6c8bdc8fbb-gkt26" event={"ID":"bd8ea4a3-c3ab-4f61-8257-40adfbbaf5df","Type":"ContainerStarted","Data":"1522fd5feafe9818e7378965fcd9371ca91da359f83c68850824668df3553089"} Feb 02 09:58:37 crc kubenswrapper[4764]: I0202 09:58:37.154872 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Feb 02 09:58:37 crc kubenswrapper[4764]: I0202 09:58:37.154900 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Feb 02 09:58:37 crc kubenswrapper[4764]: I0202 09:58:37.981693 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Feb 02 09:58:37 crc kubenswrapper[4764]: I0202 09:58:37.981741 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Feb 02 09:58:38 crc kubenswrapper[4764]: I0202 09:58:38.021747 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Feb 02 09:58:38 crc kubenswrapper[4764]: I0202 09:58:38.023841 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Feb 02 09:58:38 crc kubenswrapper[4764]: I0202 09:58:38.165349 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7f4b5b9665-qqgjw" event={"ID":"165479bf-43a7-47fc-9083-c87cb1ee0f2c","Type":"ContainerStarted","Data":"1bea9ae58a50124665a0eed96e3eb3490c806bc91198c0e7e5d31d67d80fcad6"} Feb 02 09:58:38 crc kubenswrapper[4764]: I0202 09:58:38.165496 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7f4b5b9665-qqgjw" podUID="165479bf-43a7-47fc-9083-c87cb1ee0f2c" containerName="horizon-log" containerID="cri-o://5c0ef8d37db32c9fb6af9fafb8fd85ea7920d910afd0c5afef3e02e43598b8de" gracePeriod=30 Feb 02 09:58:38 crc kubenswrapper[4764]: I0202 09:58:38.166084 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7f4b5b9665-qqgjw" podUID="165479bf-43a7-47fc-9083-c87cb1ee0f2c" containerName="horizon" containerID="cri-o://1bea9ae58a50124665a0eed96e3eb3490c806bc91198c0e7e5d31d67d80fcad6" gracePeriod=30 Feb 02 09:58:38 crc kubenswrapper[4764]: I0202 09:58:38.172016 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-698ff6d98d-6h7sk" event={"ID":"174bcb22-4370-4c3f-9141-9c1cee0d2542","Type":"ContainerStarted","Data":"f3036f1933880bb4d7462593dbdc569bcde73e6eb5564cbbea91bca016c162ed"} Feb 02 09:58:38 crc kubenswrapper[4764]: I0202 09:58:38.176918 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6c8bdc8fbb-gkt26" event={"ID":"bd8ea4a3-c3ab-4f61-8257-40adfbbaf5df","Type":"ContainerStarted","Data":"a9ba8889704978e8cdf3c04a2ed778ba56d8623240747403033dcaf0c56a278a"} Feb 02 09:58:38 crc kubenswrapper[4764]: I0202 09:58:38.184044 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6d4f8f8755-dp7xg" event={"ID":"b79a9079-9d96-4edc-97fc-b9bdf122e4d8","Type":"ContainerStarted","Data":"46b17bc1ee2d2cee2f865d09367965cf4628ec9996c321118f0fea7ca29a0810"} Feb 02 09:58:38 crc kubenswrapper[4764]: I0202 09:58:38.188758 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Feb 02 09:58:38 crc kubenswrapper[4764]: I0202 09:58:38.188810 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Feb 02 09:58:38 crc kubenswrapper[4764]: I0202 09:58:38.189186 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-6d4f8f8755-dp7xg" podUID="b79a9079-9d96-4edc-97fc-b9bdf122e4d8" containerName="horizon-log" containerID="cri-o://71fc5c9189af4e03a6f6fc0a3c0dbb2e1e002c4c3ba7e050a1b56f82fac1fc29" gracePeriod=30 Feb 02 09:58:38 crc kubenswrapper[4764]: I0202 09:58:38.189540 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-6d4f8f8755-dp7xg" podUID="b79a9079-9d96-4edc-97fc-b9bdf122e4d8" containerName="horizon" containerID="cri-o://46b17bc1ee2d2cee2f865d09367965cf4628ec9996c321118f0fea7ca29a0810" gracePeriod=30 Feb 02 09:58:38 crc kubenswrapper[4764]: I0202 09:58:38.265299 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-7f4b5b9665-qqgjw" podStartSLOduration=4.282477649 podStartE2EDuration="18.265279501s" podCreationTimestamp="2026-02-02 09:58:20 +0000 UTC" firstStartedPulling="2026-02-02 09:58:22.633468362 +0000 UTC m=+3065.567192450" lastFinishedPulling="2026-02-02 09:58:36.616270204 +0000 UTC m=+3079.549994302" observedRunningTime="2026-02-02 09:58:38.239144966 +0000 UTC m=+3081.172869054" watchObservedRunningTime="2026-02-02 09:58:38.265279501 +0000 UTC m=+3081.199003589" Feb 02 09:58:38 crc kubenswrapper[4764]: I0202 09:58:38.267956 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-6d4f8f8755-dp7xg" podStartSLOduration=3.794006731 podStartE2EDuration="18.267950502s" podCreationTimestamp="2026-02-02 09:58:20 +0000 UTC" firstStartedPulling="2026-02-02 09:58:22.060839185 +0000 UTC m=+3064.994563273" lastFinishedPulling="2026-02-02 09:58:36.534782946 +0000 UTC m=+3079.468507044" observedRunningTime="2026-02-02 09:58:38.256562869 +0000 UTC m=+3081.190286957" watchObservedRunningTime="2026-02-02 09:58:38.267950502 +0000 UTC m=+3081.201674580" Feb 02 09:58:38 crc kubenswrapper[4764]: I0202 09:58:38.305011 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-698ff6d98d-6h7sk" podStartSLOduration=3.220498107 podStartE2EDuration="15.304997948s" podCreationTimestamp="2026-02-02 09:58:23 +0000 UTC" firstStartedPulling="2026-02-02 09:58:24.478848405 +0000 UTC m=+3067.412572483" lastFinishedPulling="2026-02-02 09:58:36.563348226 +0000 UTC m=+3079.497072324" observedRunningTime="2026-02-02 09:58:38.28140687 +0000 UTC m=+3081.215130958" watchObservedRunningTime="2026-02-02 09:58:38.304997948 +0000 UTC m=+3081.238722036" Feb 02 09:58:38 crc kubenswrapper[4764]: I0202 09:58:38.316827 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-6c8bdc8fbb-gkt26" podStartSLOduration=3.240023197 podStartE2EDuration="15.316812033s" podCreationTimestamp="2026-02-02 09:58:23 +0000 UTC" firstStartedPulling="2026-02-02 09:58:24.511492283 +0000 UTC m=+3067.445216371" lastFinishedPulling="2026-02-02 09:58:36.588281109 +0000 UTC m=+3079.522005207" observedRunningTime="2026-02-02 09:58:38.301361411 +0000 UTC m=+3081.235085499" watchObservedRunningTime="2026-02-02 09:58:38.316812033 +0000 UTC m=+3081.250536121" Feb 02 09:58:40 crc kubenswrapper[4764]: I0202 09:58:40.213445 4764 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 02 09:58:40 crc kubenswrapper[4764]: I0202 09:58:40.213699 4764 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 02 09:58:40 crc kubenswrapper[4764]: I0202 09:58:40.991645 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-6d4f8f8755-dp7xg" Feb 02 09:58:41 crc kubenswrapper[4764]: I0202 09:58:41.373805 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7f4b5b9665-qqgjw" Feb 02 09:58:41 crc kubenswrapper[4764]: I0202 09:58:41.564604 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Feb 02 09:58:41 crc kubenswrapper[4764]: I0202 09:58:41.564723 4764 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 02 09:58:41 crc kubenswrapper[4764]: I0202 09:58:41.598625 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Feb 02 09:58:41 crc kubenswrapper[4764]: I0202 09:58:41.723836 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Feb 02 09:58:41 crc kubenswrapper[4764]: I0202 09:58:41.723885 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Feb 02 09:58:41 crc kubenswrapper[4764]: I0202 09:58:41.829596 4764 scope.go:117] "RemoveContainer" containerID="db4ebd4a8dbee216b70cb175f017f1799f3ca01a887f602ad42b9f295964da5d" Feb 02 09:58:41 crc kubenswrapper[4764]: E0202 09:58:41.829823 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 09:58:43 crc kubenswrapper[4764]: I0202 09:58:43.638981 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-6c8bdc8fbb-gkt26" Feb 02 09:58:43 crc kubenswrapper[4764]: I0202 09:58:43.639519 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-6c8bdc8fbb-gkt26" Feb 02 09:58:43 crc kubenswrapper[4764]: I0202 09:58:43.686278 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-698ff6d98d-6h7sk" Feb 02 09:58:43 crc kubenswrapper[4764]: I0202 09:58:43.686317 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-698ff6d98d-6h7sk" Feb 02 09:58:45 crc kubenswrapper[4764]: I0202 09:58:45.259504 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-xvkkk" event={"ID":"1bf7db82-4332-485f-acbf-3edbf4269f50","Type":"ContainerStarted","Data":"d8a10c4c9ac37c4ec6829a95cae27abf3c0ab4bbb1602a4daf3f9b9a087b5478"} Feb 02 09:58:45 crc kubenswrapper[4764]: I0202 09:58:45.273358 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-db-sync-xvkkk" podStartSLOduration=7.778321825 podStartE2EDuration="15.273337626s" podCreationTimestamp="2026-02-02 09:58:30 +0000 UTC" firstStartedPulling="2026-02-02 09:58:36.509118563 +0000 UTC m=+3079.442842661" lastFinishedPulling="2026-02-02 09:58:44.004134374 +0000 UTC m=+3086.937858462" observedRunningTime="2026-02-02 09:58:45.27121185 +0000 UTC m=+3088.204935948" watchObservedRunningTime="2026-02-02 09:58:45.273337626 +0000 UTC m=+3088.207061734" Feb 02 09:58:53 crc kubenswrapper[4764]: I0202 09:58:53.639788 4764 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-6c8bdc8fbb-gkt26" podUID="bd8ea4a3-c3ab-4f61-8257-40adfbbaf5df" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.245:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.245:8443: connect: connection refused" Feb 02 09:58:53 crc kubenswrapper[4764]: I0202 09:58:53.687383 4764 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-698ff6d98d-6h7sk" podUID="174bcb22-4370-4c3f-9141-9c1cee0d2542" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.246:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.246:8443: connect: connection refused" Feb 02 09:58:55 crc kubenswrapper[4764]: I0202 09:58:55.371636 4764 generic.go:334] "Generic (PLEG): container finished" podID="1bf7db82-4332-485f-acbf-3edbf4269f50" containerID="d8a10c4c9ac37c4ec6829a95cae27abf3c0ab4bbb1602a4daf3f9b9a087b5478" exitCode=0 Feb 02 09:58:55 crc kubenswrapper[4764]: I0202 09:58:55.371685 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-xvkkk" event={"ID":"1bf7db82-4332-485f-acbf-3edbf4269f50","Type":"ContainerDied","Data":"d8a10c4c9ac37c4ec6829a95cae27abf3c0ab4bbb1602a4daf3f9b9a087b5478"} Feb 02 09:58:56 crc kubenswrapper[4764]: I0202 09:58:56.826071 4764 scope.go:117] "RemoveContainer" containerID="db4ebd4a8dbee216b70cb175f017f1799f3ca01a887f602ad42b9f295964da5d" Feb 02 09:58:56 crc kubenswrapper[4764]: E0202 09:58:56.826376 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 09:58:57 crc kubenswrapper[4764]: I0202 09:58:57.181516 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-xvkkk" Feb 02 09:58:57 crc kubenswrapper[4764]: I0202 09:58:57.253383 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/1bf7db82-4332-485f-acbf-3edbf4269f50-job-config-data\") pod \"1bf7db82-4332-485f-acbf-3edbf4269f50\" (UID: \"1bf7db82-4332-485f-acbf-3edbf4269f50\") " Feb 02 09:58:57 crc kubenswrapper[4764]: I0202 09:58:57.253481 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kn7cc\" (UniqueName: \"kubernetes.io/projected/1bf7db82-4332-485f-acbf-3edbf4269f50-kube-api-access-kn7cc\") pod \"1bf7db82-4332-485f-acbf-3edbf4269f50\" (UID: \"1bf7db82-4332-485f-acbf-3edbf4269f50\") " Feb 02 09:58:57 crc kubenswrapper[4764]: I0202 09:58:57.253581 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1bf7db82-4332-485f-acbf-3edbf4269f50-config-data\") pod \"1bf7db82-4332-485f-acbf-3edbf4269f50\" (UID: \"1bf7db82-4332-485f-acbf-3edbf4269f50\") " Feb 02 09:58:57 crc kubenswrapper[4764]: I0202 09:58:57.253612 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bf7db82-4332-485f-acbf-3edbf4269f50-combined-ca-bundle\") pod \"1bf7db82-4332-485f-acbf-3edbf4269f50\" (UID: \"1bf7db82-4332-485f-acbf-3edbf4269f50\") " Feb 02 09:58:57 crc kubenswrapper[4764]: I0202 09:58:57.275049 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7db82-4332-485f-acbf-3edbf4269f50-job-config-data" (OuterVolumeSpecName: "job-config-data") pod "1bf7db82-4332-485f-acbf-3edbf4269f50" (UID: "1bf7db82-4332-485f-acbf-3edbf4269f50"). InnerVolumeSpecName "job-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:58:57 crc kubenswrapper[4764]: I0202 09:58:57.277706 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7db82-4332-485f-acbf-3edbf4269f50-kube-api-access-kn7cc" (OuterVolumeSpecName: "kube-api-access-kn7cc") pod "1bf7db82-4332-485f-acbf-3edbf4269f50" (UID: "1bf7db82-4332-485f-acbf-3edbf4269f50"). InnerVolumeSpecName "kube-api-access-kn7cc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:58:57 crc kubenswrapper[4764]: I0202 09:58:57.277734 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7db82-4332-485f-acbf-3edbf4269f50-config-data" (OuterVolumeSpecName: "config-data") pod "1bf7db82-4332-485f-acbf-3edbf4269f50" (UID: "1bf7db82-4332-485f-acbf-3edbf4269f50"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:58:57 crc kubenswrapper[4764]: I0202 09:58:57.291358 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7db82-4332-485f-acbf-3edbf4269f50-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1bf7db82-4332-485f-acbf-3edbf4269f50" (UID: "1bf7db82-4332-485f-acbf-3edbf4269f50"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:58:57 crc kubenswrapper[4764]: I0202 09:58:57.358514 4764 reconciler_common.go:293] "Volume detached for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/1bf7db82-4332-485f-acbf-3edbf4269f50-job-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 09:58:57 crc kubenswrapper[4764]: I0202 09:58:57.358542 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kn7cc\" (UniqueName: \"kubernetes.io/projected/1bf7db82-4332-485f-acbf-3edbf4269f50-kube-api-access-kn7cc\") on node \"crc\" DevicePath \"\"" Feb 02 09:58:57 crc kubenswrapper[4764]: I0202 09:58:57.358553 4764 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1bf7db82-4332-485f-acbf-3edbf4269f50-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 09:58:57 crc kubenswrapper[4764]: I0202 09:58:57.358562 4764 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bf7db82-4332-485f-acbf-3edbf4269f50-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 09:58:57 crc kubenswrapper[4764]: I0202 09:58:57.388740 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-xvkkk" event={"ID":"1bf7db82-4332-485f-acbf-3edbf4269f50","Type":"ContainerDied","Data":"6d1307c11c429a051bcbc5b8747d4862c15d9ef65642394ad1e370bfc1389e46"} Feb 02 09:58:57 crc kubenswrapper[4764]: I0202 09:58:57.388776 4764 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6d1307c11c429a051bcbc5b8747d4862c15d9ef65642394ad1e370bfc1389e46" Feb 02 09:58:57 crc kubenswrapper[4764]: I0202 09:58:57.388840 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-xvkkk" Feb 02 09:58:57 crc kubenswrapper[4764]: I0202 09:58:57.863771 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-share-share1-0"] Feb 02 09:58:57 crc kubenswrapper[4764]: E0202 09:58:57.864583 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bf7db82-4332-485f-acbf-3edbf4269f50" containerName="manila-db-sync" Feb 02 09:58:57 crc kubenswrapper[4764]: I0202 09:58:57.864598 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bf7db82-4332-485f-acbf-3edbf4269f50" containerName="manila-db-sync" Feb 02 09:58:57 crc kubenswrapper[4764]: I0202 09:58:57.864977 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="1bf7db82-4332-485f-acbf-3edbf4269f50" containerName="manila-db-sync" Feb 02 09:58:57 crc kubenswrapper[4764]: I0202 09:58:57.866344 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-scheduler-0"] Feb 02 09:58:57 crc kubenswrapper[4764]: I0202 09:58:57.866528 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Feb 02 09:58:57 crc kubenswrapper[4764]: I0202 09:58:57.872655 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Feb 02 09:58:57 crc kubenswrapper[4764]: I0202 09:58:57.872736 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Feb 02 09:58:57 crc kubenswrapper[4764]: I0202 09:58:57.873307 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-scripts" Feb 02 09:58:57 crc kubenswrapper[4764]: I0202 09:58:57.879337 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Feb 02 09:58:57 crc kubenswrapper[4764]: I0202 09:58:57.887517 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-manila-dockercfg-8bjjz" Feb 02 09:58:57 crc kubenswrapper[4764]: I0202 09:58:57.887835 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-config-data" Feb 02 09:58:57 crc kubenswrapper[4764]: I0202 09:58:57.888041 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-share-share1-config-data" Feb 02 09:58:57 crc kubenswrapper[4764]: I0202 09:58:57.888210 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-scheduler-config-data" Feb 02 09:58:57 crc kubenswrapper[4764]: I0202 09:58:57.935452 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6d94cddcb9-n4wc4"] Feb 02 09:58:57 crc kubenswrapper[4764]: I0202 09:58:57.937212 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d94cddcb9-n4wc4" Feb 02 09:58:57 crc kubenswrapper[4764]: I0202 09:58:57.945962 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6d94cddcb9-n4wc4"] Feb 02 09:58:57 crc kubenswrapper[4764]: I0202 09:58:57.976708 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c0ba1c62-6619-4988-867b-f4a9e4d426dc-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"c0ba1c62-6619-4988-867b-f4a9e4d426dc\") " pod="openstack/manila-share-share1-0" Feb 02 09:58:57 crc kubenswrapper[4764]: I0202 09:58:57.976759 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d61062ba-f497-4f61-9b51-27e62888797e-ovsdbserver-sb\") pod \"dnsmasq-dns-6d94cddcb9-n4wc4\" (UID: \"d61062ba-f497-4f61-9b51-27e62888797e\") " pod="openstack/dnsmasq-dns-6d94cddcb9-n4wc4" Feb 02 09:58:57 crc kubenswrapper[4764]: I0202 09:58:57.976791 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c0ba1c62-6619-4988-867b-f4a9e4d426dc-scripts\") pod \"manila-share-share1-0\" (UID: \"c0ba1c62-6619-4988-867b-f4a9e4d426dc\") " pod="openstack/manila-share-share1-0" Feb 02 09:58:57 crc kubenswrapper[4764]: I0202 09:58:57.976821 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-78gmk\" (UniqueName: \"kubernetes.io/projected/c0ba1c62-6619-4988-867b-f4a9e4d426dc-kube-api-access-78gmk\") pod \"manila-share-share1-0\" (UID: \"c0ba1c62-6619-4988-867b-f4a9e4d426dc\") " pod="openstack/manila-share-share1-0" Feb 02 09:58:57 crc kubenswrapper[4764]: I0202 09:58:57.976885 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d61062ba-f497-4f61-9b51-27e62888797e-config\") pod \"dnsmasq-dns-6d94cddcb9-n4wc4\" (UID: \"d61062ba-f497-4f61-9b51-27e62888797e\") " pod="openstack/dnsmasq-dns-6d94cddcb9-n4wc4" Feb 02 09:58:57 crc kubenswrapper[4764]: I0202 09:58:57.976907 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b78ef5a7-a2a0-494d-b687-18384876989b-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"b78ef5a7-a2a0-494d-b687-18384876989b\") " pod="openstack/manila-scheduler-0" Feb 02 09:58:57 crc kubenswrapper[4764]: I0202 09:58:57.976923 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/c0ba1c62-6619-4988-867b-f4a9e4d426dc-ceph\") pod \"manila-share-share1-0\" (UID: \"c0ba1c62-6619-4988-867b-f4a9e4d426dc\") " pod="openstack/manila-share-share1-0" Feb 02 09:58:57 crc kubenswrapper[4764]: I0202 09:58:57.977084 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b78ef5a7-a2a0-494d-b687-18384876989b-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"b78ef5a7-a2a0-494d-b687-18384876989b\") " pod="openstack/manila-scheduler-0" Feb 02 09:58:57 crc kubenswrapper[4764]: I0202 09:58:57.977148 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b78ef5a7-a2a0-494d-b687-18384876989b-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"b78ef5a7-a2a0-494d-b687-18384876989b\") " pod="openstack/manila-scheduler-0" Feb 02 09:58:57 crc kubenswrapper[4764]: I0202 09:58:57.977173 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5rl5f\" (UniqueName: \"kubernetes.io/projected/d61062ba-f497-4f61-9b51-27e62888797e-kube-api-access-5rl5f\") pod \"dnsmasq-dns-6d94cddcb9-n4wc4\" (UID: \"d61062ba-f497-4f61-9b51-27e62888797e\") " pod="openstack/dnsmasq-dns-6d94cddcb9-n4wc4" Feb 02 09:58:57 crc kubenswrapper[4764]: I0202 09:58:57.977200 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0ba1c62-6619-4988-867b-f4a9e4d426dc-config-data\") pod \"manila-share-share1-0\" (UID: \"c0ba1c62-6619-4988-867b-f4a9e4d426dc\") " pod="openstack/manila-share-share1-0" Feb 02 09:58:57 crc kubenswrapper[4764]: I0202 09:58:57.977225 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0ba1c62-6619-4988-867b-f4a9e4d426dc-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"c0ba1c62-6619-4988-867b-f4a9e4d426dc\") " pod="openstack/manila-share-share1-0" Feb 02 09:58:57 crc kubenswrapper[4764]: I0202 09:58:57.977268 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d61062ba-f497-4f61-9b51-27e62888797e-ovsdbserver-nb\") pod \"dnsmasq-dns-6d94cddcb9-n4wc4\" (UID: \"d61062ba-f497-4f61-9b51-27e62888797e\") " pod="openstack/dnsmasq-dns-6d94cddcb9-n4wc4" Feb 02 09:58:57 crc kubenswrapper[4764]: I0202 09:58:57.977340 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b78ef5a7-a2a0-494d-b687-18384876989b-scripts\") pod \"manila-scheduler-0\" (UID: \"b78ef5a7-a2a0-494d-b687-18384876989b\") " pod="openstack/manila-scheduler-0" Feb 02 09:58:57 crc kubenswrapper[4764]: I0202 09:58:57.977377 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b78ef5a7-a2a0-494d-b687-18384876989b-config-data\") pod \"manila-scheduler-0\" (UID: \"b78ef5a7-a2a0-494d-b687-18384876989b\") " pod="openstack/manila-scheduler-0" Feb 02 09:58:57 crc kubenswrapper[4764]: I0202 09:58:57.977403 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/d61062ba-f497-4f61-9b51-27e62888797e-openstack-edpm-ipam\") pod \"dnsmasq-dns-6d94cddcb9-n4wc4\" (UID: \"d61062ba-f497-4f61-9b51-27e62888797e\") " pod="openstack/dnsmasq-dns-6d94cddcb9-n4wc4" Feb 02 09:58:57 crc kubenswrapper[4764]: I0202 09:58:57.977431 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d61062ba-f497-4f61-9b51-27e62888797e-dns-svc\") pod \"dnsmasq-dns-6d94cddcb9-n4wc4\" (UID: \"d61062ba-f497-4f61-9b51-27e62888797e\") " pod="openstack/dnsmasq-dns-6d94cddcb9-n4wc4" Feb 02 09:58:57 crc kubenswrapper[4764]: I0202 09:58:57.977484 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/c0ba1c62-6619-4988-867b-f4a9e4d426dc-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"c0ba1c62-6619-4988-867b-f4a9e4d426dc\") " pod="openstack/manila-share-share1-0" Feb 02 09:58:57 crc kubenswrapper[4764]: I0202 09:58:57.977559 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c0ba1c62-6619-4988-867b-f4a9e4d426dc-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"c0ba1c62-6619-4988-867b-f4a9e4d426dc\") " pod="openstack/manila-share-share1-0" Feb 02 09:58:57 crc kubenswrapper[4764]: I0202 09:58:57.977592 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r5tbq\" (UniqueName: \"kubernetes.io/projected/b78ef5a7-a2a0-494d-b687-18384876989b-kube-api-access-r5tbq\") pod \"manila-scheduler-0\" (UID: \"b78ef5a7-a2a0-494d-b687-18384876989b\") " pod="openstack/manila-scheduler-0" Feb 02 09:58:58 crc kubenswrapper[4764]: I0202 09:58:58.079022 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b78ef5a7-a2a0-494d-b687-18384876989b-scripts\") pod \"manila-scheduler-0\" (UID: \"b78ef5a7-a2a0-494d-b687-18384876989b\") " pod="openstack/manila-scheduler-0" Feb 02 09:58:58 crc kubenswrapper[4764]: I0202 09:58:58.079075 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b78ef5a7-a2a0-494d-b687-18384876989b-config-data\") pod \"manila-scheduler-0\" (UID: \"b78ef5a7-a2a0-494d-b687-18384876989b\") " pod="openstack/manila-scheduler-0" Feb 02 09:58:58 crc kubenswrapper[4764]: I0202 09:58:58.079106 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/d61062ba-f497-4f61-9b51-27e62888797e-openstack-edpm-ipam\") pod \"dnsmasq-dns-6d94cddcb9-n4wc4\" (UID: \"d61062ba-f497-4f61-9b51-27e62888797e\") " pod="openstack/dnsmasq-dns-6d94cddcb9-n4wc4" Feb 02 09:58:58 crc kubenswrapper[4764]: I0202 09:58:58.079127 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d61062ba-f497-4f61-9b51-27e62888797e-dns-svc\") pod \"dnsmasq-dns-6d94cddcb9-n4wc4\" (UID: \"d61062ba-f497-4f61-9b51-27e62888797e\") " pod="openstack/dnsmasq-dns-6d94cddcb9-n4wc4" Feb 02 09:58:58 crc kubenswrapper[4764]: I0202 09:58:58.079163 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/c0ba1c62-6619-4988-867b-f4a9e4d426dc-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"c0ba1c62-6619-4988-867b-f4a9e4d426dc\") " pod="openstack/manila-share-share1-0" Feb 02 09:58:58 crc kubenswrapper[4764]: I0202 09:58:58.079200 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c0ba1c62-6619-4988-867b-f4a9e4d426dc-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"c0ba1c62-6619-4988-867b-f4a9e4d426dc\") " pod="openstack/manila-share-share1-0" Feb 02 09:58:58 crc kubenswrapper[4764]: I0202 09:58:58.079220 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r5tbq\" (UniqueName: \"kubernetes.io/projected/b78ef5a7-a2a0-494d-b687-18384876989b-kube-api-access-r5tbq\") pod \"manila-scheduler-0\" (UID: \"b78ef5a7-a2a0-494d-b687-18384876989b\") " pod="openstack/manila-scheduler-0" Feb 02 09:58:58 crc kubenswrapper[4764]: I0202 09:58:58.079252 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c0ba1c62-6619-4988-867b-f4a9e4d426dc-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"c0ba1c62-6619-4988-867b-f4a9e4d426dc\") " pod="openstack/manila-share-share1-0" Feb 02 09:58:58 crc kubenswrapper[4764]: I0202 09:58:58.079275 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d61062ba-f497-4f61-9b51-27e62888797e-ovsdbserver-sb\") pod \"dnsmasq-dns-6d94cddcb9-n4wc4\" (UID: \"d61062ba-f497-4f61-9b51-27e62888797e\") " pod="openstack/dnsmasq-dns-6d94cddcb9-n4wc4" Feb 02 09:58:58 crc kubenswrapper[4764]: I0202 09:58:58.079293 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c0ba1c62-6619-4988-867b-f4a9e4d426dc-scripts\") pod \"manila-share-share1-0\" (UID: \"c0ba1c62-6619-4988-867b-f4a9e4d426dc\") " pod="openstack/manila-share-share1-0" Feb 02 09:58:58 crc kubenswrapper[4764]: I0202 09:58:58.079319 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-78gmk\" (UniqueName: \"kubernetes.io/projected/c0ba1c62-6619-4988-867b-f4a9e4d426dc-kube-api-access-78gmk\") pod \"manila-share-share1-0\" (UID: \"c0ba1c62-6619-4988-867b-f4a9e4d426dc\") " pod="openstack/manila-share-share1-0" Feb 02 09:58:58 crc kubenswrapper[4764]: I0202 09:58:58.079344 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d61062ba-f497-4f61-9b51-27e62888797e-config\") pod \"dnsmasq-dns-6d94cddcb9-n4wc4\" (UID: \"d61062ba-f497-4f61-9b51-27e62888797e\") " pod="openstack/dnsmasq-dns-6d94cddcb9-n4wc4" Feb 02 09:58:58 crc kubenswrapper[4764]: I0202 09:58:58.079365 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b78ef5a7-a2a0-494d-b687-18384876989b-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"b78ef5a7-a2a0-494d-b687-18384876989b\") " pod="openstack/manila-scheduler-0" Feb 02 09:58:58 crc kubenswrapper[4764]: I0202 09:58:58.079380 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/c0ba1c62-6619-4988-867b-f4a9e4d426dc-ceph\") pod \"manila-share-share1-0\" (UID: \"c0ba1c62-6619-4988-867b-f4a9e4d426dc\") " pod="openstack/manila-share-share1-0" Feb 02 09:58:58 crc kubenswrapper[4764]: I0202 09:58:58.079415 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b78ef5a7-a2a0-494d-b687-18384876989b-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"b78ef5a7-a2a0-494d-b687-18384876989b\") " pod="openstack/manila-scheduler-0" Feb 02 09:58:58 crc kubenswrapper[4764]: I0202 09:58:58.079433 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b78ef5a7-a2a0-494d-b687-18384876989b-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"b78ef5a7-a2a0-494d-b687-18384876989b\") " pod="openstack/manila-scheduler-0" Feb 02 09:58:58 crc kubenswrapper[4764]: I0202 09:58:58.079448 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5rl5f\" (UniqueName: \"kubernetes.io/projected/d61062ba-f497-4f61-9b51-27e62888797e-kube-api-access-5rl5f\") pod \"dnsmasq-dns-6d94cddcb9-n4wc4\" (UID: \"d61062ba-f497-4f61-9b51-27e62888797e\") " pod="openstack/dnsmasq-dns-6d94cddcb9-n4wc4" Feb 02 09:58:58 crc kubenswrapper[4764]: I0202 09:58:58.079470 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0ba1c62-6619-4988-867b-f4a9e4d426dc-config-data\") pod \"manila-share-share1-0\" (UID: \"c0ba1c62-6619-4988-867b-f4a9e4d426dc\") " pod="openstack/manila-share-share1-0" Feb 02 09:58:58 crc kubenswrapper[4764]: I0202 09:58:58.079488 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0ba1c62-6619-4988-867b-f4a9e4d426dc-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"c0ba1c62-6619-4988-867b-f4a9e4d426dc\") " pod="openstack/manila-share-share1-0" Feb 02 09:58:58 crc kubenswrapper[4764]: I0202 09:58:58.079509 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d61062ba-f497-4f61-9b51-27e62888797e-ovsdbserver-nb\") pod \"dnsmasq-dns-6d94cddcb9-n4wc4\" (UID: \"d61062ba-f497-4f61-9b51-27e62888797e\") " pod="openstack/dnsmasq-dns-6d94cddcb9-n4wc4" Feb 02 09:58:58 crc kubenswrapper[4764]: I0202 09:58:58.081174 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/c0ba1c62-6619-4988-867b-f4a9e4d426dc-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"c0ba1c62-6619-4988-867b-f4a9e4d426dc\") " pod="openstack/manila-share-share1-0" Feb 02 09:58:58 crc kubenswrapper[4764]: I0202 09:58:58.081436 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d61062ba-f497-4f61-9b51-27e62888797e-ovsdbserver-nb\") pod \"dnsmasq-dns-6d94cddcb9-n4wc4\" (UID: \"d61062ba-f497-4f61-9b51-27e62888797e\") " pod="openstack/dnsmasq-dns-6d94cddcb9-n4wc4" Feb 02 09:58:58 crc kubenswrapper[4764]: I0202 09:58:58.082330 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d61062ba-f497-4f61-9b51-27e62888797e-config\") pod \"dnsmasq-dns-6d94cddcb9-n4wc4\" (UID: \"d61062ba-f497-4f61-9b51-27e62888797e\") " pod="openstack/dnsmasq-dns-6d94cddcb9-n4wc4" Feb 02 09:58:58 crc kubenswrapper[4764]: I0202 09:58:58.086537 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b78ef5a7-a2a0-494d-b687-18384876989b-scripts\") pod \"manila-scheduler-0\" (UID: \"b78ef5a7-a2a0-494d-b687-18384876989b\") " pod="openstack/manila-scheduler-0" Feb 02 09:58:58 crc kubenswrapper[4764]: I0202 09:58:58.087271 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c0ba1c62-6619-4988-867b-f4a9e4d426dc-scripts\") pod \"manila-share-share1-0\" (UID: \"c0ba1c62-6619-4988-867b-f4a9e4d426dc\") " pod="openstack/manila-share-share1-0" Feb 02 09:58:58 crc kubenswrapper[4764]: I0202 09:58:58.087168 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/d61062ba-f497-4f61-9b51-27e62888797e-openstack-edpm-ipam\") pod \"dnsmasq-dns-6d94cddcb9-n4wc4\" (UID: \"d61062ba-f497-4f61-9b51-27e62888797e\") " pod="openstack/dnsmasq-dns-6d94cddcb9-n4wc4" Feb 02 09:58:58 crc kubenswrapper[4764]: I0202 09:58:58.088522 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b78ef5a7-a2a0-494d-b687-18384876989b-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"b78ef5a7-a2a0-494d-b687-18384876989b\") " pod="openstack/manila-scheduler-0" Feb 02 09:58:58 crc kubenswrapper[4764]: I0202 09:58:58.088647 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c0ba1c62-6619-4988-867b-f4a9e4d426dc-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"c0ba1c62-6619-4988-867b-f4a9e4d426dc\") " pod="openstack/manila-share-share1-0" Feb 02 09:58:58 crc kubenswrapper[4764]: I0202 09:58:58.089584 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d61062ba-f497-4f61-9b51-27e62888797e-ovsdbserver-sb\") pod \"dnsmasq-dns-6d94cddcb9-n4wc4\" (UID: \"d61062ba-f497-4f61-9b51-27e62888797e\") " pod="openstack/dnsmasq-dns-6d94cddcb9-n4wc4" Feb 02 09:58:58 crc kubenswrapper[4764]: I0202 09:58:58.089596 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0ba1c62-6619-4988-867b-f4a9e4d426dc-config-data\") pod \"manila-share-share1-0\" (UID: \"c0ba1c62-6619-4988-867b-f4a9e4d426dc\") " pod="openstack/manila-share-share1-0" Feb 02 09:58:58 crc kubenswrapper[4764]: I0202 09:58:58.091111 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d61062ba-f497-4f61-9b51-27e62888797e-dns-svc\") pod \"dnsmasq-dns-6d94cddcb9-n4wc4\" (UID: \"d61062ba-f497-4f61-9b51-27e62888797e\") " pod="openstack/dnsmasq-dns-6d94cddcb9-n4wc4" Feb 02 09:58:58 crc kubenswrapper[4764]: I0202 09:58:58.094803 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0ba1c62-6619-4988-867b-f4a9e4d426dc-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"c0ba1c62-6619-4988-867b-f4a9e4d426dc\") " pod="openstack/manila-share-share1-0" Feb 02 09:58:58 crc kubenswrapper[4764]: I0202 09:58:58.094811 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b78ef5a7-a2a0-494d-b687-18384876989b-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"b78ef5a7-a2a0-494d-b687-18384876989b\") " pod="openstack/manila-scheduler-0" Feb 02 09:58:58 crc kubenswrapper[4764]: I0202 09:58:58.095100 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/c0ba1c62-6619-4988-867b-f4a9e4d426dc-ceph\") pod \"manila-share-share1-0\" (UID: \"c0ba1c62-6619-4988-867b-f4a9e4d426dc\") " pod="openstack/manila-share-share1-0" Feb 02 09:58:58 crc kubenswrapper[4764]: I0202 09:58:58.095675 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b78ef5a7-a2a0-494d-b687-18384876989b-config-data\") pod \"manila-scheduler-0\" (UID: \"b78ef5a7-a2a0-494d-b687-18384876989b\") " pod="openstack/manila-scheduler-0" Feb 02 09:58:58 crc kubenswrapper[4764]: I0202 09:58:58.095862 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b78ef5a7-a2a0-494d-b687-18384876989b-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"b78ef5a7-a2a0-494d-b687-18384876989b\") " pod="openstack/manila-scheduler-0" Feb 02 09:58:58 crc kubenswrapper[4764]: I0202 09:58:58.098791 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c0ba1c62-6619-4988-867b-f4a9e4d426dc-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"c0ba1c62-6619-4988-867b-f4a9e4d426dc\") " pod="openstack/manila-share-share1-0" Feb 02 09:58:58 crc kubenswrapper[4764]: I0202 09:58:58.112100 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5rl5f\" (UniqueName: \"kubernetes.io/projected/d61062ba-f497-4f61-9b51-27e62888797e-kube-api-access-5rl5f\") pod \"dnsmasq-dns-6d94cddcb9-n4wc4\" (UID: \"d61062ba-f497-4f61-9b51-27e62888797e\") " pod="openstack/dnsmasq-dns-6d94cddcb9-n4wc4" Feb 02 09:58:58 crc kubenswrapper[4764]: I0202 09:58:58.121400 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r5tbq\" (UniqueName: \"kubernetes.io/projected/b78ef5a7-a2a0-494d-b687-18384876989b-kube-api-access-r5tbq\") pod \"manila-scheduler-0\" (UID: \"b78ef5a7-a2a0-494d-b687-18384876989b\") " pod="openstack/manila-scheduler-0" Feb 02 09:58:58 crc kubenswrapper[4764]: I0202 09:58:58.132083 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-78gmk\" (UniqueName: \"kubernetes.io/projected/c0ba1c62-6619-4988-867b-f4a9e4d426dc-kube-api-access-78gmk\") pod \"manila-share-share1-0\" (UID: \"c0ba1c62-6619-4988-867b-f4a9e4d426dc\") " pod="openstack/manila-share-share1-0" Feb 02 09:58:58 crc kubenswrapper[4764]: I0202 09:58:58.189861 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-api-0"] Feb 02 09:58:58 crc kubenswrapper[4764]: I0202 09:58:58.194368 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Feb 02 09:58:58 crc kubenswrapper[4764]: I0202 09:58:58.200138 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-api-config-data" Feb 02 09:58:58 crc kubenswrapper[4764]: I0202 09:58:58.200889 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Feb 02 09:58:58 crc kubenswrapper[4764]: I0202 09:58:58.210242 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Feb 02 09:58:58 crc kubenswrapper[4764]: I0202 09:58:58.224415 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Feb 02 09:58:58 crc kubenswrapper[4764]: I0202 09:58:58.275435 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d94cddcb9-n4wc4" Feb 02 09:58:58 crc kubenswrapper[4764]: I0202 09:58:58.281757 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/073343c4-a6d4-494f-8c19-7c27c3434eb8-config-data\") pod \"manila-api-0\" (UID: \"073343c4-a6d4-494f-8c19-7c27c3434eb8\") " pod="openstack/manila-api-0" Feb 02 09:58:58 crc kubenswrapper[4764]: I0202 09:58:58.281829 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wmq97\" (UniqueName: \"kubernetes.io/projected/073343c4-a6d4-494f-8c19-7c27c3434eb8-kube-api-access-wmq97\") pod \"manila-api-0\" (UID: \"073343c4-a6d4-494f-8c19-7c27c3434eb8\") " pod="openstack/manila-api-0" Feb 02 09:58:58 crc kubenswrapper[4764]: I0202 09:58:58.281849 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/073343c4-a6d4-494f-8c19-7c27c3434eb8-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"073343c4-a6d4-494f-8c19-7c27c3434eb8\") " pod="openstack/manila-api-0" Feb 02 09:58:58 crc kubenswrapper[4764]: I0202 09:58:58.281908 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/073343c4-a6d4-494f-8c19-7c27c3434eb8-scripts\") pod \"manila-api-0\" (UID: \"073343c4-a6d4-494f-8c19-7c27c3434eb8\") " pod="openstack/manila-api-0" Feb 02 09:58:58 crc kubenswrapper[4764]: I0202 09:58:58.281956 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/073343c4-a6d4-494f-8c19-7c27c3434eb8-etc-machine-id\") pod \"manila-api-0\" (UID: \"073343c4-a6d4-494f-8c19-7c27c3434eb8\") " pod="openstack/manila-api-0" Feb 02 09:58:58 crc kubenswrapper[4764]: I0202 09:58:58.282076 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/073343c4-a6d4-494f-8c19-7c27c3434eb8-config-data-custom\") pod \"manila-api-0\" (UID: \"073343c4-a6d4-494f-8c19-7c27c3434eb8\") " pod="openstack/manila-api-0" Feb 02 09:58:58 crc kubenswrapper[4764]: I0202 09:58:58.282180 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/073343c4-a6d4-494f-8c19-7c27c3434eb8-logs\") pod \"manila-api-0\" (UID: \"073343c4-a6d4-494f-8c19-7c27c3434eb8\") " pod="openstack/manila-api-0" Feb 02 09:58:58 crc kubenswrapper[4764]: I0202 09:58:58.389171 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wmq97\" (UniqueName: \"kubernetes.io/projected/073343c4-a6d4-494f-8c19-7c27c3434eb8-kube-api-access-wmq97\") pod \"manila-api-0\" (UID: \"073343c4-a6d4-494f-8c19-7c27c3434eb8\") " pod="openstack/manila-api-0" Feb 02 09:58:58 crc kubenswrapper[4764]: I0202 09:58:58.389218 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/073343c4-a6d4-494f-8c19-7c27c3434eb8-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"073343c4-a6d4-494f-8c19-7c27c3434eb8\") " pod="openstack/manila-api-0" Feb 02 09:58:58 crc kubenswrapper[4764]: I0202 09:58:58.389289 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/073343c4-a6d4-494f-8c19-7c27c3434eb8-scripts\") pod \"manila-api-0\" (UID: \"073343c4-a6d4-494f-8c19-7c27c3434eb8\") " pod="openstack/manila-api-0" Feb 02 09:58:58 crc kubenswrapper[4764]: I0202 09:58:58.389313 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/073343c4-a6d4-494f-8c19-7c27c3434eb8-etc-machine-id\") pod \"manila-api-0\" (UID: \"073343c4-a6d4-494f-8c19-7c27c3434eb8\") " pod="openstack/manila-api-0" Feb 02 09:58:58 crc kubenswrapper[4764]: I0202 09:58:58.389338 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/073343c4-a6d4-494f-8c19-7c27c3434eb8-config-data-custom\") pod \"manila-api-0\" (UID: \"073343c4-a6d4-494f-8c19-7c27c3434eb8\") " pod="openstack/manila-api-0" Feb 02 09:58:58 crc kubenswrapper[4764]: I0202 09:58:58.389371 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/073343c4-a6d4-494f-8c19-7c27c3434eb8-logs\") pod \"manila-api-0\" (UID: \"073343c4-a6d4-494f-8c19-7c27c3434eb8\") " pod="openstack/manila-api-0" Feb 02 09:58:58 crc kubenswrapper[4764]: I0202 09:58:58.389429 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/073343c4-a6d4-494f-8c19-7c27c3434eb8-config-data\") pod \"manila-api-0\" (UID: \"073343c4-a6d4-494f-8c19-7c27c3434eb8\") " pod="openstack/manila-api-0" Feb 02 09:58:58 crc kubenswrapper[4764]: I0202 09:58:58.389589 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/073343c4-a6d4-494f-8c19-7c27c3434eb8-etc-machine-id\") pod \"manila-api-0\" (UID: \"073343c4-a6d4-494f-8c19-7c27c3434eb8\") " pod="openstack/manila-api-0" Feb 02 09:58:58 crc kubenswrapper[4764]: I0202 09:58:58.395441 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/073343c4-a6d4-494f-8c19-7c27c3434eb8-logs\") pod \"manila-api-0\" (UID: \"073343c4-a6d4-494f-8c19-7c27c3434eb8\") " pod="openstack/manila-api-0" Feb 02 09:58:58 crc kubenswrapper[4764]: I0202 09:58:58.396329 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/073343c4-a6d4-494f-8c19-7c27c3434eb8-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"073343c4-a6d4-494f-8c19-7c27c3434eb8\") " pod="openstack/manila-api-0" Feb 02 09:58:58 crc kubenswrapper[4764]: I0202 09:58:58.406874 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/073343c4-a6d4-494f-8c19-7c27c3434eb8-scripts\") pod \"manila-api-0\" (UID: \"073343c4-a6d4-494f-8c19-7c27c3434eb8\") " pod="openstack/manila-api-0" Feb 02 09:58:58 crc kubenswrapper[4764]: I0202 09:58:58.407740 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wmq97\" (UniqueName: \"kubernetes.io/projected/073343c4-a6d4-494f-8c19-7c27c3434eb8-kube-api-access-wmq97\") pod \"manila-api-0\" (UID: \"073343c4-a6d4-494f-8c19-7c27c3434eb8\") " pod="openstack/manila-api-0" Feb 02 09:58:58 crc kubenswrapper[4764]: I0202 09:58:58.409281 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/073343c4-a6d4-494f-8c19-7c27c3434eb8-config-data\") pod \"manila-api-0\" (UID: \"073343c4-a6d4-494f-8c19-7c27c3434eb8\") " pod="openstack/manila-api-0" Feb 02 09:58:58 crc kubenswrapper[4764]: I0202 09:58:58.411724 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/073343c4-a6d4-494f-8c19-7c27c3434eb8-config-data-custom\") pod \"manila-api-0\" (UID: \"073343c4-a6d4-494f-8c19-7c27c3434eb8\") " pod="openstack/manila-api-0" Feb 02 09:58:58 crc kubenswrapper[4764]: I0202 09:58:58.520906 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Feb 02 09:58:59 crc kubenswrapper[4764]: I0202 09:58:59.063400 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Feb 02 09:58:59 crc kubenswrapper[4764]: I0202 09:58:59.246047 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Feb 02 09:58:59 crc kubenswrapper[4764]: I0202 09:58:59.282485 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6d94cddcb9-n4wc4"] Feb 02 09:58:59 crc kubenswrapper[4764]: W0202 09:58:59.294424 4764 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd61062ba_f497_4f61_9b51_27e62888797e.slice/crio-52ee2bbc353c997dd37f193503a462b36e047a24d211b8d299a3067e608f5ac9 WatchSource:0}: Error finding container 52ee2bbc353c997dd37f193503a462b36e047a24d211b8d299a3067e608f5ac9: Status 404 returned error can't find the container with id 52ee2bbc353c997dd37f193503a462b36e047a24d211b8d299a3067e608f5ac9 Feb 02 09:58:59 crc kubenswrapper[4764]: W0202 09:58:59.359175 4764 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod073343c4_a6d4_494f_8c19_7c27c3434eb8.slice/crio-dab5d367fcf043a35a3286c2bf2e400947c6ce75552425084823809205dbd84c WatchSource:0}: Error finding container dab5d367fcf043a35a3286c2bf2e400947c6ce75552425084823809205dbd84c: Status 404 returned error can't find the container with id dab5d367fcf043a35a3286c2bf2e400947c6ce75552425084823809205dbd84c Feb 02 09:58:59 crc kubenswrapper[4764]: I0202 09:58:59.380280 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Feb 02 09:58:59 crc kubenswrapper[4764]: I0202 09:58:59.449042 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"c0ba1c62-6619-4988-867b-f4a9e4d426dc","Type":"ContainerStarted","Data":"bf8b0c387315154e7bc4f1001ea4a7383094f96c8feb20161a8568e6decec43b"} Feb 02 09:58:59 crc kubenswrapper[4764]: I0202 09:58:59.463243 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"073343c4-a6d4-494f-8c19-7c27c3434eb8","Type":"ContainerStarted","Data":"dab5d367fcf043a35a3286c2bf2e400947c6ce75552425084823809205dbd84c"} Feb 02 09:58:59 crc kubenswrapper[4764]: I0202 09:58:59.464986 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"b78ef5a7-a2a0-494d-b687-18384876989b","Type":"ContainerStarted","Data":"6a92e0901d5e2d0f76df49de40945cbc8df9429b1736cec50e4bc4ed45a221c5"} Feb 02 09:58:59 crc kubenswrapper[4764]: I0202 09:58:59.482245 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d94cddcb9-n4wc4" event={"ID":"d61062ba-f497-4f61-9b51-27e62888797e","Type":"ContainerStarted","Data":"52ee2bbc353c997dd37f193503a462b36e047a24d211b8d299a3067e608f5ac9"} Feb 02 09:59:00 crc kubenswrapper[4764]: I0202 09:59:00.502877 4764 generic.go:334] "Generic (PLEG): container finished" podID="d61062ba-f497-4f61-9b51-27e62888797e" containerID="ac664ffb9a93bf11e7a5c123b96daf472ffc63ca854d537fa8d925e101664151" exitCode=0 Feb 02 09:59:00 crc kubenswrapper[4764]: I0202 09:59:00.503385 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d94cddcb9-n4wc4" event={"ID":"d61062ba-f497-4f61-9b51-27e62888797e","Type":"ContainerDied","Data":"ac664ffb9a93bf11e7a5c123b96daf472ffc63ca854d537fa8d925e101664151"} Feb 02 09:59:00 crc kubenswrapper[4764]: I0202 09:59:00.508906 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"073343c4-a6d4-494f-8c19-7c27c3434eb8","Type":"ContainerStarted","Data":"8dea9591b536e0efa9b3e0fd177e7d748c71b200695f870037c5913cc12e2d29"} Feb 02 09:59:01 crc kubenswrapper[4764]: I0202 09:59:01.518584 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"073343c4-a6d4-494f-8c19-7c27c3434eb8","Type":"ContainerStarted","Data":"9bc7a8a1bbf2a2668b7502da6653fd67e265469d30eb768107b15e29db06aab3"} Feb 02 09:59:01 crc kubenswrapper[4764]: I0202 09:59:01.519211 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/manila-api-0" Feb 02 09:59:01 crc kubenswrapper[4764]: I0202 09:59:01.521174 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"b78ef5a7-a2a0-494d-b687-18384876989b","Type":"ContainerStarted","Data":"68750e3d5d783461dcd76ac064febdaec06ae407bc81f56036014b1305a00528"} Feb 02 09:59:01 crc kubenswrapper[4764]: I0202 09:59:01.524668 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d94cddcb9-n4wc4" event={"ID":"d61062ba-f497-4f61-9b51-27e62888797e","Type":"ContainerStarted","Data":"9abceda1f2e9073f0b8c7d3abda845d35ef2c5d32178ceb67137042553505070"} Feb 02 09:59:01 crc kubenswrapper[4764]: I0202 09:59:01.524858 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6d94cddcb9-n4wc4" Feb 02 09:59:01 crc kubenswrapper[4764]: I0202 09:59:01.531545 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-api-0"] Feb 02 09:59:01 crc kubenswrapper[4764]: I0202 09:59:01.547884 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-api-0" podStartSLOduration=3.547867738 podStartE2EDuration="3.547867738s" podCreationTimestamp="2026-02-02 09:58:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:59:01.547336184 +0000 UTC m=+3104.481060262" watchObservedRunningTime="2026-02-02 09:59:01.547867738 +0000 UTC m=+3104.481591826" Feb 02 09:59:01 crc kubenswrapper[4764]: I0202 09:59:01.583277 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6d94cddcb9-n4wc4" podStartSLOduration=4.58325852 podStartE2EDuration="4.58325852s" podCreationTimestamp="2026-02-02 09:58:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:59:01.573613163 +0000 UTC m=+3104.507337241" watchObservedRunningTime="2026-02-02 09:59:01.58325852 +0000 UTC m=+3104.516982608" Feb 02 09:59:02 crc kubenswrapper[4764]: I0202 09:59:02.587369 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"b78ef5a7-a2a0-494d-b687-18384876989b","Type":"ContainerStarted","Data":"c4978209530c86d62dededf2405f3a1ef19fe1ccb720a7d1b81bfc6027c1da2b"} Feb 02 09:59:02 crc kubenswrapper[4764]: I0202 09:59:02.624124 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-scheduler-0" podStartSLOduration=4.740881475 podStartE2EDuration="5.624108976s" podCreationTimestamp="2026-02-02 09:58:57 +0000 UTC" firstStartedPulling="2026-02-02 09:58:59.271250451 +0000 UTC m=+3102.204974539" lastFinishedPulling="2026-02-02 09:59:00.154477952 +0000 UTC m=+3103.088202040" observedRunningTime="2026-02-02 09:59:02.621212009 +0000 UTC m=+3105.554936107" watchObservedRunningTime="2026-02-02 09:59:02.624108976 +0000 UTC m=+3105.557833064" Feb 02 09:59:03 crc kubenswrapper[4764]: I0202 09:59:03.593504 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-api-0" podUID="073343c4-a6d4-494f-8c19-7c27c3434eb8" containerName="manila-api-log" containerID="cri-o://8dea9591b536e0efa9b3e0fd177e7d748c71b200695f870037c5913cc12e2d29" gracePeriod=30 Feb 02 09:59:03 crc kubenswrapper[4764]: I0202 09:59:03.593890 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-api-0" podUID="073343c4-a6d4-494f-8c19-7c27c3434eb8" containerName="manila-api" containerID="cri-o://9bc7a8a1bbf2a2668b7502da6653fd67e265469d30eb768107b15e29db06aab3" gracePeriod=30 Feb 02 09:59:03 crc kubenswrapper[4764]: I0202 09:59:03.639005 4764 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-6c8bdc8fbb-gkt26" podUID="bd8ea4a3-c3ab-4f61-8257-40adfbbaf5df" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.245:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.245:8443: connect: connection refused" Feb 02 09:59:03 crc kubenswrapper[4764]: I0202 09:59:03.685908 4764 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-698ff6d98d-6h7sk" podUID="174bcb22-4370-4c3f-9141-9c1cee0d2542" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.246:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.246:8443: connect: connection refused" Feb 02 09:59:04 crc kubenswrapper[4764]: I0202 09:59:04.619280 4764 generic.go:334] "Generic (PLEG): container finished" podID="073343c4-a6d4-494f-8c19-7c27c3434eb8" containerID="9bc7a8a1bbf2a2668b7502da6653fd67e265469d30eb768107b15e29db06aab3" exitCode=0 Feb 02 09:59:04 crc kubenswrapper[4764]: I0202 09:59:04.620632 4764 generic.go:334] "Generic (PLEG): container finished" podID="073343c4-a6d4-494f-8c19-7c27c3434eb8" containerID="8dea9591b536e0efa9b3e0fd177e7d748c71b200695f870037c5913cc12e2d29" exitCode=143 Feb 02 09:59:04 crc kubenswrapper[4764]: I0202 09:59:04.620726 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"073343c4-a6d4-494f-8c19-7c27c3434eb8","Type":"ContainerDied","Data":"9bc7a8a1bbf2a2668b7502da6653fd67e265469d30eb768107b15e29db06aab3"} Feb 02 09:59:04 crc kubenswrapper[4764]: I0202 09:59:04.620843 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"073343c4-a6d4-494f-8c19-7c27c3434eb8","Type":"ContainerDied","Data":"8dea9591b536e0efa9b3e0fd177e7d748c71b200695f870037c5913cc12e2d29"} Feb 02 09:59:04 crc kubenswrapper[4764]: I0202 09:59:04.726474 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Feb 02 09:59:04 crc kubenswrapper[4764]: I0202 09:59:04.885506 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/073343c4-a6d4-494f-8c19-7c27c3434eb8-config-data\") pod \"073343c4-a6d4-494f-8c19-7c27c3434eb8\" (UID: \"073343c4-a6d4-494f-8c19-7c27c3434eb8\") " Feb 02 09:59:04 crc kubenswrapper[4764]: I0202 09:59:04.885680 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/073343c4-a6d4-494f-8c19-7c27c3434eb8-logs\") pod \"073343c4-a6d4-494f-8c19-7c27c3434eb8\" (UID: \"073343c4-a6d4-494f-8c19-7c27c3434eb8\") " Feb 02 09:59:04 crc kubenswrapper[4764]: I0202 09:59:04.886175 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/073343c4-a6d4-494f-8c19-7c27c3434eb8-logs" (OuterVolumeSpecName: "logs") pod "073343c4-a6d4-494f-8c19-7c27c3434eb8" (UID: "073343c4-a6d4-494f-8c19-7c27c3434eb8"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 09:59:04 crc kubenswrapper[4764]: I0202 09:59:04.886278 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/073343c4-a6d4-494f-8c19-7c27c3434eb8-scripts\") pod \"073343c4-a6d4-494f-8c19-7c27c3434eb8\" (UID: \"073343c4-a6d4-494f-8c19-7c27c3434eb8\") " Feb 02 09:59:04 crc kubenswrapper[4764]: I0202 09:59:04.886304 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/073343c4-a6d4-494f-8c19-7c27c3434eb8-etc-machine-id\") pod \"073343c4-a6d4-494f-8c19-7c27c3434eb8\" (UID: \"073343c4-a6d4-494f-8c19-7c27c3434eb8\") " Feb 02 09:59:04 crc kubenswrapper[4764]: I0202 09:59:04.886638 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wmq97\" (UniqueName: \"kubernetes.io/projected/073343c4-a6d4-494f-8c19-7c27c3434eb8-kube-api-access-wmq97\") pod \"073343c4-a6d4-494f-8c19-7c27c3434eb8\" (UID: \"073343c4-a6d4-494f-8c19-7c27c3434eb8\") " Feb 02 09:59:04 crc kubenswrapper[4764]: I0202 09:59:04.886693 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/073343c4-a6d4-494f-8c19-7c27c3434eb8-config-data-custom\") pod \"073343c4-a6d4-494f-8c19-7c27c3434eb8\" (UID: \"073343c4-a6d4-494f-8c19-7c27c3434eb8\") " Feb 02 09:59:04 crc kubenswrapper[4764]: I0202 09:59:04.886776 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/073343c4-a6d4-494f-8c19-7c27c3434eb8-combined-ca-bundle\") pod \"073343c4-a6d4-494f-8c19-7c27c3434eb8\" (UID: \"073343c4-a6d4-494f-8c19-7c27c3434eb8\") " Feb 02 09:59:04 crc kubenswrapper[4764]: I0202 09:59:04.886949 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/073343c4-a6d4-494f-8c19-7c27c3434eb8-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "073343c4-a6d4-494f-8c19-7c27c3434eb8" (UID: "073343c4-a6d4-494f-8c19-7c27c3434eb8"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 09:59:04 crc kubenswrapper[4764]: I0202 09:59:04.887271 4764 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/073343c4-a6d4-494f-8c19-7c27c3434eb8-logs\") on node \"crc\" DevicePath \"\"" Feb 02 09:59:04 crc kubenswrapper[4764]: I0202 09:59:04.887291 4764 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/073343c4-a6d4-494f-8c19-7c27c3434eb8-etc-machine-id\") on node \"crc\" DevicePath \"\"" Feb 02 09:59:04 crc kubenswrapper[4764]: I0202 09:59:04.891781 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/073343c4-a6d4-494f-8c19-7c27c3434eb8-scripts" (OuterVolumeSpecName: "scripts") pod "073343c4-a6d4-494f-8c19-7c27c3434eb8" (UID: "073343c4-a6d4-494f-8c19-7c27c3434eb8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:59:04 crc kubenswrapper[4764]: I0202 09:59:04.898329 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/073343c4-a6d4-494f-8c19-7c27c3434eb8-kube-api-access-wmq97" (OuterVolumeSpecName: "kube-api-access-wmq97") pod "073343c4-a6d4-494f-8c19-7c27c3434eb8" (UID: "073343c4-a6d4-494f-8c19-7c27c3434eb8"). InnerVolumeSpecName "kube-api-access-wmq97". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:59:04 crc kubenswrapper[4764]: I0202 09:59:04.899862 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/073343c4-a6d4-494f-8c19-7c27c3434eb8-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "073343c4-a6d4-494f-8c19-7c27c3434eb8" (UID: "073343c4-a6d4-494f-8c19-7c27c3434eb8"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:59:04 crc kubenswrapper[4764]: I0202 09:59:04.932023 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/073343c4-a6d4-494f-8c19-7c27c3434eb8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "073343c4-a6d4-494f-8c19-7c27c3434eb8" (UID: "073343c4-a6d4-494f-8c19-7c27c3434eb8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:59:04 crc kubenswrapper[4764]: I0202 09:59:04.948111 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/073343c4-a6d4-494f-8c19-7c27c3434eb8-config-data" (OuterVolumeSpecName: "config-data") pod "073343c4-a6d4-494f-8c19-7c27c3434eb8" (UID: "073343c4-a6d4-494f-8c19-7c27c3434eb8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:59:04 crc kubenswrapper[4764]: I0202 09:59:04.989304 4764 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/073343c4-a6d4-494f-8c19-7c27c3434eb8-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 09:59:04 crc kubenswrapper[4764]: I0202 09:59:04.989921 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wmq97\" (UniqueName: \"kubernetes.io/projected/073343c4-a6d4-494f-8c19-7c27c3434eb8-kube-api-access-wmq97\") on node \"crc\" DevicePath \"\"" Feb 02 09:59:04 crc kubenswrapper[4764]: I0202 09:59:04.990053 4764 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/073343c4-a6d4-494f-8c19-7c27c3434eb8-config-data-custom\") on node \"crc\" DevicePath \"\"" Feb 02 09:59:04 crc kubenswrapper[4764]: I0202 09:59:04.990130 4764 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/073343c4-a6d4-494f-8c19-7c27c3434eb8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 09:59:04 crc kubenswrapper[4764]: I0202 09:59:04.990200 4764 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/073343c4-a6d4-494f-8c19-7c27c3434eb8-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 09:59:05 crc kubenswrapper[4764]: I0202 09:59:05.632569 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"073343c4-a6d4-494f-8c19-7c27c3434eb8","Type":"ContainerDied","Data":"dab5d367fcf043a35a3286c2bf2e400947c6ce75552425084823809205dbd84c"} Feb 02 09:59:05 crc kubenswrapper[4764]: I0202 09:59:05.632638 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Feb 02 09:59:05 crc kubenswrapper[4764]: I0202 09:59:05.632853 4764 scope.go:117] "RemoveContainer" containerID="9bc7a8a1bbf2a2668b7502da6653fd67e265469d30eb768107b15e29db06aab3" Feb 02 09:59:05 crc kubenswrapper[4764]: I0202 09:59:05.673265 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-api-0"] Feb 02 09:59:05 crc kubenswrapper[4764]: I0202 09:59:05.680371 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-api-0"] Feb 02 09:59:05 crc kubenswrapper[4764]: I0202 09:59:05.696600 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-api-0"] Feb 02 09:59:05 crc kubenswrapper[4764]: E0202 09:59:05.696951 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="073343c4-a6d4-494f-8c19-7c27c3434eb8" containerName="manila-api" Feb 02 09:59:05 crc kubenswrapper[4764]: I0202 09:59:05.696965 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="073343c4-a6d4-494f-8c19-7c27c3434eb8" containerName="manila-api" Feb 02 09:59:05 crc kubenswrapper[4764]: E0202 09:59:05.696976 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="073343c4-a6d4-494f-8c19-7c27c3434eb8" containerName="manila-api-log" Feb 02 09:59:05 crc kubenswrapper[4764]: I0202 09:59:05.696982 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="073343c4-a6d4-494f-8c19-7c27c3434eb8" containerName="manila-api-log" Feb 02 09:59:05 crc kubenswrapper[4764]: I0202 09:59:05.697167 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="073343c4-a6d4-494f-8c19-7c27c3434eb8" containerName="manila-api" Feb 02 09:59:05 crc kubenswrapper[4764]: I0202 09:59:05.697192 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="073343c4-a6d4-494f-8c19-7c27c3434eb8" containerName="manila-api-log" Feb 02 09:59:05 crc kubenswrapper[4764]: I0202 09:59:05.698110 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Feb 02 09:59:05 crc kubenswrapper[4764]: I0202 09:59:05.698999 4764 scope.go:117] "RemoveContainer" containerID="8dea9591b536e0efa9b3e0fd177e7d748c71b200695f870037c5913cc12e2d29" Feb 02 09:59:05 crc kubenswrapper[4764]: I0202 09:59:05.703411 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-manila-internal-svc" Feb 02 09:59:05 crc kubenswrapper[4764]: I0202 09:59:05.703562 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-manila-public-svc" Feb 02 09:59:05 crc kubenswrapper[4764]: I0202 09:59:05.703711 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-api-config-data" Feb 02 09:59:05 crc kubenswrapper[4764]: I0202 09:59:05.726739 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Feb 02 09:59:05 crc kubenswrapper[4764]: I0202 09:59:05.807809 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eed25ff8-c8ee-41d6-bb33-89e4c5f30724-logs\") pod \"manila-api-0\" (UID: \"eed25ff8-c8ee-41d6-bb33-89e4c5f30724\") " pod="openstack/manila-api-0" Feb 02 09:59:05 crc kubenswrapper[4764]: I0202 09:59:05.807897 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/eed25ff8-c8ee-41d6-bb33-89e4c5f30724-public-tls-certs\") pod \"manila-api-0\" (UID: \"eed25ff8-c8ee-41d6-bb33-89e4c5f30724\") " pod="openstack/manila-api-0" Feb 02 09:59:05 crc kubenswrapper[4764]: I0202 09:59:05.807951 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eed25ff8-c8ee-41d6-bb33-89e4c5f30724-config-data\") pod \"manila-api-0\" (UID: \"eed25ff8-c8ee-41d6-bb33-89e4c5f30724\") " pod="openstack/manila-api-0" Feb 02 09:59:05 crc kubenswrapper[4764]: I0202 09:59:05.808045 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eed25ff8-c8ee-41d6-bb33-89e4c5f30724-scripts\") pod \"manila-api-0\" (UID: \"eed25ff8-c8ee-41d6-bb33-89e4c5f30724\") " pod="openstack/manila-api-0" Feb 02 09:59:05 crc kubenswrapper[4764]: I0202 09:59:05.808103 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lq9d4\" (UniqueName: \"kubernetes.io/projected/eed25ff8-c8ee-41d6-bb33-89e4c5f30724-kube-api-access-lq9d4\") pod \"manila-api-0\" (UID: \"eed25ff8-c8ee-41d6-bb33-89e4c5f30724\") " pod="openstack/manila-api-0" Feb 02 09:59:05 crc kubenswrapper[4764]: I0202 09:59:05.808160 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/eed25ff8-c8ee-41d6-bb33-89e4c5f30724-config-data-custom\") pod \"manila-api-0\" (UID: \"eed25ff8-c8ee-41d6-bb33-89e4c5f30724\") " pod="openstack/manila-api-0" Feb 02 09:59:05 crc kubenswrapper[4764]: I0202 09:59:05.808186 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/eed25ff8-c8ee-41d6-bb33-89e4c5f30724-internal-tls-certs\") pod \"manila-api-0\" (UID: \"eed25ff8-c8ee-41d6-bb33-89e4c5f30724\") " pod="openstack/manila-api-0" Feb 02 09:59:05 crc kubenswrapper[4764]: I0202 09:59:05.808228 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/eed25ff8-c8ee-41d6-bb33-89e4c5f30724-etc-machine-id\") pod \"manila-api-0\" (UID: \"eed25ff8-c8ee-41d6-bb33-89e4c5f30724\") " pod="openstack/manila-api-0" Feb 02 09:59:05 crc kubenswrapper[4764]: I0202 09:59:05.808249 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eed25ff8-c8ee-41d6-bb33-89e4c5f30724-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"eed25ff8-c8ee-41d6-bb33-89e4c5f30724\") " pod="openstack/manila-api-0" Feb 02 09:59:05 crc kubenswrapper[4764]: I0202 09:59:05.840883 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="073343c4-a6d4-494f-8c19-7c27c3434eb8" path="/var/lib/kubelet/pods/073343c4-a6d4-494f-8c19-7c27c3434eb8/volumes" Feb 02 09:59:05 crc kubenswrapper[4764]: I0202 09:59:05.909978 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eed25ff8-c8ee-41d6-bb33-89e4c5f30724-scripts\") pod \"manila-api-0\" (UID: \"eed25ff8-c8ee-41d6-bb33-89e4c5f30724\") " pod="openstack/manila-api-0" Feb 02 09:59:05 crc kubenswrapper[4764]: I0202 09:59:05.910044 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lq9d4\" (UniqueName: \"kubernetes.io/projected/eed25ff8-c8ee-41d6-bb33-89e4c5f30724-kube-api-access-lq9d4\") pod \"manila-api-0\" (UID: \"eed25ff8-c8ee-41d6-bb33-89e4c5f30724\") " pod="openstack/manila-api-0" Feb 02 09:59:05 crc kubenswrapper[4764]: I0202 09:59:05.910077 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/eed25ff8-c8ee-41d6-bb33-89e4c5f30724-config-data-custom\") pod \"manila-api-0\" (UID: \"eed25ff8-c8ee-41d6-bb33-89e4c5f30724\") " pod="openstack/manila-api-0" Feb 02 09:59:05 crc kubenswrapper[4764]: I0202 09:59:05.910099 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/eed25ff8-c8ee-41d6-bb33-89e4c5f30724-internal-tls-certs\") pod \"manila-api-0\" (UID: \"eed25ff8-c8ee-41d6-bb33-89e4c5f30724\") " pod="openstack/manila-api-0" Feb 02 09:59:05 crc kubenswrapper[4764]: I0202 09:59:05.910143 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/eed25ff8-c8ee-41d6-bb33-89e4c5f30724-etc-machine-id\") pod \"manila-api-0\" (UID: \"eed25ff8-c8ee-41d6-bb33-89e4c5f30724\") " pod="openstack/manila-api-0" Feb 02 09:59:05 crc kubenswrapper[4764]: I0202 09:59:05.910160 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eed25ff8-c8ee-41d6-bb33-89e4c5f30724-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"eed25ff8-c8ee-41d6-bb33-89e4c5f30724\") " pod="openstack/manila-api-0" Feb 02 09:59:05 crc kubenswrapper[4764]: I0202 09:59:05.910204 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eed25ff8-c8ee-41d6-bb33-89e4c5f30724-logs\") pod \"manila-api-0\" (UID: \"eed25ff8-c8ee-41d6-bb33-89e4c5f30724\") " pod="openstack/manila-api-0" Feb 02 09:59:05 crc kubenswrapper[4764]: I0202 09:59:05.910286 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/eed25ff8-c8ee-41d6-bb33-89e4c5f30724-public-tls-certs\") pod \"manila-api-0\" (UID: \"eed25ff8-c8ee-41d6-bb33-89e4c5f30724\") " pod="openstack/manila-api-0" Feb 02 09:59:05 crc kubenswrapper[4764]: I0202 09:59:05.910315 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eed25ff8-c8ee-41d6-bb33-89e4c5f30724-config-data\") pod \"manila-api-0\" (UID: \"eed25ff8-c8ee-41d6-bb33-89e4c5f30724\") " pod="openstack/manila-api-0" Feb 02 09:59:05 crc kubenswrapper[4764]: I0202 09:59:05.910345 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/eed25ff8-c8ee-41d6-bb33-89e4c5f30724-etc-machine-id\") pod \"manila-api-0\" (UID: \"eed25ff8-c8ee-41d6-bb33-89e4c5f30724\") " pod="openstack/manila-api-0" Feb 02 09:59:05 crc kubenswrapper[4764]: I0202 09:59:05.911637 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eed25ff8-c8ee-41d6-bb33-89e4c5f30724-logs\") pod \"manila-api-0\" (UID: \"eed25ff8-c8ee-41d6-bb33-89e4c5f30724\") " pod="openstack/manila-api-0" Feb 02 09:59:05 crc kubenswrapper[4764]: I0202 09:59:05.915429 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/eed25ff8-c8ee-41d6-bb33-89e4c5f30724-config-data-custom\") pod \"manila-api-0\" (UID: \"eed25ff8-c8ee-41d6-bb33-89e4c5f30724\") " pod="openstack/manila-api-0" Feb 02 09:59:05 crc kubenswrapper[4764]: I0202 09:59:05.920409 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/eed25ff8-c8ee-41d6-bb33-89e4c5f30724-public-tls-certs\") pod \"manila-api-0\" (UID: \"eed25ff8-c8ee-41d6-bb33-89e4c5f30724\") " pod="openstack/manila-api-0" Feb 02 09:59:05 crc kubenswrapper[4764]: I0202 09:59:05.920628 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eed25ff8-c8ee-41d6-bb33-89e4c5f30724-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"eed25ff8-c8ee-41d6-bb33-89e4c5f30724\") " pod="openstack/manila-api-0" Feb 02 09:59:05 crc kubenswrapper[4764]: I0202 09:59:05.921435 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eed25ff8-c8ee-41d6-bb33-89e4c5f30724-config-data\") pod \"manila-api-0\" (UID: \"eed25ff8-c8ee-41d6-bb33-89e4c5f30724\") " pod="openstack/manila-api-0" Feb 02 09:59:05 crc kubenswrapper[4764]: I0202 09:59:05.924494 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eed25ff8-c8ee-41d6-bb33-89e4c5f30724-scripts\") pod \"manila-api-0\" (UID: \"eed25ff8-c8ee-41d6-bb33-89e4c5f30724\") " pod="openstack/manila-api-0" Feb 02 09:59:05 crc kubenswrapper[4764]: I0202 09:59:05.928374 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lq9d4\" (UniqueName: \"kubernetes.io/projected/eed25ff8-c8ee-41d6-bb33-89e4c5f30724-kube-api-access-lq9d4\") pod \"manila-api-0\" (UID: \"eed25ff8-c8ee-41d6-bb33-89e4c5f30724\") " pod="openstack/manila-api-0" Feb 02 09:59:05 crc kubenswrapper[4764]: I0202 09:59:05.928639 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/eed25ff8-c8ee-41d6-bb33-89e4c5f30724-internal-tls-certs\") pod \"manila-api-0\" (UID: \"eed25ff8-c8ee-41d6-bb33-89e4c5f30724\") " pod="openstack/manila-api-0" Feb 02 09:59:06 crc kubenswrapper[4764]: I0202 09:59:06.067242 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Feb 02 09:59:06 crc kubenswrapper[4764]: I0202 09:59:06.691043 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Feb 02 09:59:06 crc kubenswrapper[4764]: W0202 09:59:06.702982 4764 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeed25ff8_c8ee_41d6_bb33_89e4c5f30724.slice/crio-6080c4d0efd27ffecd9e029fd0f4dfcb999735bfebef40d3174d5211599ce723 WatchSource:0}: Error finding container 6080c4d0efd27ffecd9e029fd0f4dfcb999735bfebef40d3174d5211599ce723: Status 404 returned error can't find the container with id 6080c4d0efd27ffecd9e029fd0f4dfcb999735bfebef40d3174d5211599ce723 Feb 02 09:59:07 crc kubenswrapper[4764]: I0202 09:59:07.655049 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"eed25ff8-c8ee-41d6-bb33-89e4c5f30724","Type":"ContainerStarted","Data":"aa47892d518f587769343170513d61d0cfbbc113360fd2c674ace216f8c695e8"} Feb 02 09:59:07 crc kubenswrapper[4764]: I0202 09:59:07.655323 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"eed25ff8-c8ee-41d6-bb33-89e4c5f30724","Type":"ContainerStarted","Data":"6080c4d0efd27ffecd9e029fd0f4dfcb999735bfebef40d3174d5211599ce723"} Feb 02 09:59:08 crc kubenswrapper[4764]: I0202 09:59:08.225423 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-scheduler-0" Feb 02 09:59:08 crc kubenswrapper[4764]: I0202 09:59:08.277181 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6d94cddcb9-n4wc4" Feb 02 09:59:08 crc kubenswrapper[4764]: I0202 09:59:08.369385 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-77755c997f-6996z"] Feb 02 09:59:08 crc kubenswrapper[4764]: I0202 09:59:08.369839 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-77755c997f-6996z" podUID="33b1de07-e5a5-472a-acfb-102dbfe83e0e" containerName="dnsmasq-dns" containerID="cri-o://b42ec966d617e60cef3c0bd4351d46ef96d749b7d8a53fafb653e8d7beb3a126" gracePeriod=10 Feb 02 09:59:08 crc kubenswrapper[4764]: I0202 09:59:08.665895 4764 generic.go:334] "Generic (PLEG): container finished" podID="33b1de07-e5a5-472a-acfb-102dbfe83e0e" containerID="b42ec966d617e60cef3c0bd4351d46ef96d749b7d8a53fafb653e8d7beb3a126" exitCode=0 Feb 02 09:59:08 crc kubenswrapper[4764]: I0202 09:59:08.665968 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77755c997f-6996z" event={"ID":"33b1de07-e5a5-472a-acfb-102dbfe83e0e","Type":"ContainerDied","Data":"b42ec966d617e60cef3c0bd4351d46ef96d749b7d8a53fafb653e8d7beb3a126"} Feb 02 09:59:08 crc kubenswrapper[4764]: I0202 09:59:08.668061 4764 generic.go:334] "Generic (PLEG): container finished" podID="b79a9079-9d96-4edc-97fc-b9bdf122e4d8" containerID="46b17bc1ee2d2cee2f865d09367965cf4628ec9996c321118f0fea7ca29a0810" exitCode=137 Feb 02 09:59:08 crc kubenswrapper[4764]: I0202 09:59:08.668084 4764 generic.go:334] "Generic (PLEG): container finished" podID="b79a9079-9d96-4edc-97fc-b9bdf122e4d8" containerID="71fc5c9189af4e03a6f6fc0a3c0dbb2e1e002c4c3ba7e050a1b56f82fac1fc29" exitCode=137 Feb 02 09:59:08 crc kubenswrapper[4764]: I0202 09:59:08.668149 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6d4f8f8755-dp7xg" event={"ID":"b79a9079-9d96-4edc-97fc-b9bdf122e4d8","Type":"ContainerDied","Data":"46b17bc1ee2d2cee2f865d09367965cf4628ec9996c321118f0fea7ca29a0810"} Feb 02 09:59:08 crc kubenswrapper[4764]: I0202 09:59:08.668216 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6d4f8f8755-dp7xg" event={"ID":"b79a9079-9d96-4edc-97fc-b9bdf122e4d8","Type":"ContainerDied","Data":"71fc5c9189af4e03a6f6fc0a3c0dbb2e1e002c4c3ba7e050a1b56f82fac1fc29"} Feb 02 09:59:08 crc kubenswrapper[4764]: I0202 09:59:08.670679 4764 generic.go:334] "Generic (PLEG): container finished" podID="165479bf-43a7-47fc-9083-c87cb1ee0f2c" containerID="1bea9ae58a50124665a0eed96e3eb3490c806bc91198c0e7e5d31d67d80fcad6" exitCode=137 Feb 02 09:59:08 crc kubenswrapper[4764]: I0202 09:59:08.670702 4764 generic.go:334] "Generic (PLEG): container finished" podID="165479bf-43a7-47fc-9083-c87cb1ee0f2c" containerID="5c0ef8d37db32c9fb6af9fafb8fd85ea7920d910afd0c5afef3e02e43598b8de" exitCode=137 Feb 02 09:59:08 crc kubenswrapper[4764]: I0202 09:59:08.670715 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7f4b5b9665-qqgjw" event={"ID":"165479bf-43a7-47fc-9083-c87cb1ee0f2c","Type":"ContainerDied","Data":"1bea9ae58a50124665a0eed96e3eb3490c806bc91198c0e7e5d31d67d80fcad6"} Feb 02 09:59:08 crc kubenswrapper[4764]: I0202 09:59:08.670731 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7f4b5b9665-qqgjw" event={"ID":"165479bf-43a7-47fc-9083-c87cb1ee0f2c","Type":"ContainerDied","Data":"5c0ef8d37db32c9fb6af9fafb8fd85ea7920d910afd0c5afef3e02e43598b8de"} Feb 02 09:59:09 crc kubenswrapper[4764]: I0202 09:59:09.826230 4764 scope.go:117] "RemoveContainer" containerID="db4ebd4a8dbee216b70cb175f017f1799f3ca01a887f602ad42b9f295964da5d" Feb 02 09:59:09 crc kubenswrapper[4764]: E0202 09:59:09.826560 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 09:59:11 crc kubenswrapper[4764]: I0202 09:59:11.184029 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6d4f8f8755-dp7xg" Feb 02 09:59:11 crc kubenswrapper[4764]: I0202 09:59:11.349592 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b79a9079-9d96-4edc-97fc-b9bdf122e4d8-config-data\") pod \"b79a9079-9d96-4edc-97fc-b9bdf122e4d8\" (UID: \"b79a9079-9d96-4edc-97fc-b9bdf122e4d8\") " Feb 02 09:59:11 crc kubenswrapper[4764]: I0202 09:59:11.349635 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/b79a9079-9d96-4edc-97fc-b9bdf122e4d8-horizon-secret-key\") pod \"b79a9079-9d96-4edc-97fc-b9bdf122e4d8\" (UID: \"b79a9079-9d96-4edc-97fc-b9bdf122e4d8\") " Feb 02 09:59:11 crc kubenswrapper[4764]: I0202 09:59:11.349794 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nct92\" (UniqueName: \"kubernetes.io/projected/b79a9079-9d96-4edc-97fc-b9bdf122e4d8-kube-api-access-nct92\") pod \"b79a9079-9d96-4edc-97fc-b9bdf122e4d8\" (UID: \"b79a9079-9d96-4edc-97fc-b9bdf122e4d8\") " Feb 02 09:59:11 crc kubenswrapper[4764]: I0202 09:59:11.349837 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b79a9079-9d96-4edc-97fc-b9bdf122e4d8-logs\") pod \"b79a9079-9d96-4edc-97fc-b9bdf122e4d8\" (UID: \"b79a9079-9d96-4edc-97fc-b9bdf122e4d8\") " Feb 02 09:59:11 crc kubenswrapper[4764]: I0202 09:59:11.349958 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b79a9079-9d96-4edc-97fc-b9bdf122e4d8-scripts\") pod \"b79a9079-9d96-4edc-97fc-b9bdf122e4d8\" (UID: \"b79a9079-9d96-4edc-97fc-b9bdf122e4d8\") " Feb 02 09:59:11 crc kubenswrapper[4764]: I0202 09:59:11.350471 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b79a9079-9d96-4edc-97fc-b9bdf122e4d8-logs" (OuterVolumeSpecName: "logs") pod "b79a9079-9d96-4edc-97fc-b9bdf122e4d8" (UID: "b79a9079-9d96-4edc-97fc-b9bdf122e4d8"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 09:59:11 crc kubenswrapper[4764]: I0202 09:59:11.359285 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b79a9079-9d96-4edc-97fc-b9bdf122e4d8-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "b79a9079-9d96-4edc-97fc-b9bdf122e4d8" (UID: "b79a9079-9d96-4edc-97fc-b9bdf122e4d8"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:59:11 crc kubenswrapper[4764]: I0202 09:59:11.367468 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b79a9079-9d96-4edc-97fc-b9bdf122e4d8-kube-api-access-nct92" (OuterVolumeSpecName: "kube-api-access-nct92") pod "b79a9079-9d96-4edc-97fc-b9bdf122e4d8" (UID: "b79a9079-9d96-4edc-97fc-b9bdf122e4d8"). InnerVolumeSpecName "kube-api-access-nct92". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:59:11 crc kubenswrapper[4764]: I0202 09:59:11.381152 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b79a9079-9d96-4edc-97fc-b9bdf122e4d8-scripts" (OuterVolumeSpecName: "scripts") pod "b79a9079-9d96-4edc-97fc-b9bdf122e4d8" (UID: "b79a9079-9d96-4edc-97fc-b9bdf122e4d8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:59:11 crc kubenswrapper[4764]: I0202 09:59:11.422429 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b79a9079-9d96-4edc-97fc-b9bdf122e4d8-config-data" (OuterVolumeSpecName: "config-data") pod "b79a9079-9d96-4edc-97fc-b9bdf122e4d8" (UID: "b79a9079-9d96-4edc-97fc-b9bdf122e4d8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:59:11 crc kubenswrapper[4764]: I0202 09:59:11.438874 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7f4b5b9665-qqgjw" Feb 02 09:59:11 crc kubenswrapper[4764]: I0202 09:59:11.445483 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77755c997f-6996z" Feb 02 09:59:11 crc kubenswrapper[4764]: I0202 09:59:11.462566 4764 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b79a9079-9d96-4edc-97fc-b9bdf122e4d8-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 09:59:11 crc kubenswrapper[4764]: I0202 09:59:11.462600 4764 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b79a9079-9d96-4edc-97fc-b9bdf122e4d8-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 09:59:11 crc kubenswrapper[4764]: I0202 09:59:11.462610 4764 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/b79a9079-9d96-4edc-97fc-b9bdf122e4d8-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Feb 02 09:59:11 crc kubenswrapper[4764]: I0202 09:59:11.462621 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nct92\" (UniqueName: \"kubernetes.io/projected/b79a9079-9d96-4edc-97fc-b9bdf122e4d8-kube-api-access-nct92\") on node \"crc\" DevicePath \"\"" Feb 02 09:59:11 crc kubenswrapper[4764]: I0202 09:59:11.462629 4764 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b79a9079-9d96-4edc-97fc-b9bdf122e4d8-logs\") on node \"crc\" DevicePath \"\"" Feb 02 09:59:11 crc kubenswrapper[4764]: I0202 09:59:11.564170 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/165479bf-43a7-47fc-9083-c87cb1ee0f2c-horizon-secret-key\") pod \"165479bf-43a7-47fc-9083-c87cb1ee0f2c\" (UID: \"165479bf-43a7-47fc-9083-c87cb1ee0f2c\") " Feb 02 09:59:11 crc kubenswrapper[4764]: I0202 09:59:11.564450 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/33b1de07-e5a5-472a-acfb-102dbfe83e0e-ovsdbserver-nb\") pod \"33b1de07-e5a5-472a-acfb-102dbfe83e0e\" (UID: \"33b1de07-e5a5-472a-acfb-102dbfe83e0e\") " Feb 02 09:59:11 crc kubenswrapper[4764]: I0202 09:59:11.565000 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/33b1de07-e5a5-472a-acfb-102dbfe83e0e-ovsdbserver-sb\") pod \"33b1de07-e5a5-472a-acfb-102dbfe83e0e\" (UID: \"33b1de07-e5a5-472a-acfb-102dbfe83e0e\") " Feb 02 09:59:11 crc kubenswrapper[4764]: I0202 09:59:11.565170 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/165479bf-43a7-47fc-9083-c87cb1ee0f2c-config-data\") pod \"165479bf-43a7-47fc-9083-c87cb1ee0f2c\" (UID: \"165479bf-43a7-47fc-9083-c87cb1ee0f2c\") " Feb 02 09:59:11 crc kubenswrapper[4764]: I0202 09:59:11.565275 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/165479bf-43a7-47fc-9083-c87cb1ee0f2c-logs\") pod \"165479bf-43a7-47fc-9083-c87cb1ee0f2c\" (UID: \"165479bf-43a7-47fc-9083-c87cb1ee0f2c\") " Feb 02 09:59:11 crc kubenswrapper[4764]: I0202 09:59:11.565385 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/33b1de07-e5a5-472a-acfb-102dbfe83e0e-dns-svc\") pod \"33b1de07-e5a5-472a-acfb-102dbfe83e0e\" (UID: \"33b1de07-e5a5-472a-acfb-102dbfe83e0e\") " Feb 02 09:59:11 crc kubenswrapper[4764]: I0202 09:59:11.565472 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-496bg\" (UniqueName: \"kubernetes.io/projected/33b1de07-e5a5-472a-acfb-102dbfe83e0e-kube-api-access-496bg\") pod \"33b1de07-e5a5-472a-acfb-102dbfe83e0e\" (UID: \"33b1de07-e5a5-472a-acfb-102dbfe83e0e\") " Feb 02 09:59:11 crc kubenswrapper[4764]: I0202 09:59:11.565549 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/33b1de07-e5a5-472a-acfb-102dbfe83e0e-config\") pod \"33b1de07-e5a5-472a-acfb-102dbfe83e0e\" (UID: \"33b1de07-e5a5-472a-acfb-102dbfe83e0e\") " Feb 02 09:59:11 crc kubenswrapper[4764]: I0202 09:59:11.565611 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ftgmp\" (UniqueName: \"kubernetes.io/projected/165479bf-43a7-47fc-9083-c87cb1ee0f2c-kube-api-access-ftgmp\") pod \"165479bf-43a7-47fc-9083-c87cb1ee0f2c\" (UID: \"165479bf-43a7-47fc-9083-c87cb1ee0f2c\") " Feb 02 09:59:11 crc kubenswrapper[4764]: I0202 09:59:11.565705 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/165479bf-43a7-47fc-9083-c87cb1ee0f2c-scripts\") pod \"165479bf-43a7-47fc-9083-c87cb1ee0f2c\" (UID: \"165479bf-43a7-47fc-9083-c87cb1ee0f2c\") " Feb 02 09:59:11 crc kubenswrapper[4764]: I0202 09:59:11.565793 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/33b1de07-e5a5-472a-acfb-102dbfe83e0e-openstack-edpm-ipam\") pod \"33b1de07-e5a5-472a-acfb-102dbfe83e0e\" (UID: \"33b1de07-e5a5-472a-acfb-102dbfe83e0e\") " Feb 02 09:59:11 crc kubenswrapper[4764]: I0202 09:59:11.566115 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/165479bf-43a7-47fc-9083-c87cb1ee0f2c-logs" (OuterVolumeSpecName: "logs") pod "165479bf-43a7-47fc-9083-c87cb1ee0f2c" (UID: "165479bf-43a7-47fc-9083-c87cb1ee0f2c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 09:59:11 crc kubenswrapper[4764]: I0202 09:59:11.566433 4764 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/165479bf-43a7-47fc-9083-c87cb1ee0f2c-logs\") on node \"crc\" DevicePath \"\"" Feb 02 09:59:11 crc kubenswrapper[4764]: I0202 09:59:11.569734 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/165479bf-43a7-47fc-9083-c87cb1ee0f2c-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "165479bf-43a7-47fc-9083-c87cb1ee0f2c" (UID: "165479bf-43a7-47fc-9083-c87cb1ee0f2c"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:59:11 crc kubenswrapper[4764]: I0202 09:59:11.571267 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/33b1de07-e5a5-472a-acfb-102dbfe83e0e-kube-api-access-496bg" (OuterVolumeSpecName: "kube-api-access-496bg") pod "33b1de07-e5a5-472a-acfb-102dbfe83e0e" (UID: "33b1de07-e5a5-472a-acfb-102dbfe83e0e"). InnerVolumeSpecName "kube-api-access-496bg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:59:11 crc kubenswrapper[4764]: I0202 09:59:11.581444 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/165479bf-43a7-47fc-9083-c87cb1ee0f2c-kube-api-access-ftgmp" (OuterVolumeSpecName: "kube-api-access-ftgmp") pod "165479bf-43a7-47fc-9083-c87cb1ee0f2c" (UID: "165479bf-43a7-47fc-9083-c87cb1ee0f2c"). InnerVolumeSpecName "kube-api-access-ftgmp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:59:11 crc kubenswrapper[4764]: I0202 09:59:11.608635 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/165479bf-43a7-47fc-9083-c87cb1ee0f2c-config-data" (OuterVolumeSpecName: "config-data") pod "165479bf-43a7-47fc-9083-c87cb1ee0f2c" (UID: "165479bf-43a7-47fc-9083-c87cb1ee0f2c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:59:11 crc kubenswrapper[4764]: I0202 09:59:11.616399 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/165479bf-43a7-47fc-9083-c87cb1ee0f2c-scripts" (OuterVolumeSpecName: "scripts") pod "165479bf-43a7-47fc-9083-c87cb1ee0f2c" (UID: "165479bf-43a7-47fc-9083-c87cb1ee0f2c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:59:11 crc kubenswrapper[4764]: I0202 09:59:11.625559 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/33b1de07-e5a5-472a-acfb-102dbfe83e0e-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "33b1de07-e5a5-472a-acfb-102dbfe83e0e" (UID: "33b1de07-e5a5-472a-acfb-102dbfe83e0e"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:59:11 crc kubenswrapper[4764]: I0202 09:59:11.634863 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/33b1de07-e5a5-472a-acfb-102dbfe83e0e-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "33b1de07-e5a5-472a-acfb-102dbfe83e0e" (UID: "33b1de07-e5a5-472a-acfb-102dbfe83e0e"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:59:11 crc kubenswrapper[4764]: I0202 09:59:11.635677 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/33b1de07-e5a5-472a-acfb-102dbfe83e0e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "33b1de07-e5a5-472a-acfb-102dbfe83e0e" (UID: "33b1de07-e5a5-472a-acfb-102dbfe83e0e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:59:11 crc kubenswrapper[4764]: I0202 09:59:11.649381 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/33b1de07-e5a5-472a-acfb-102dbfe83e0e-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "33b1de07-e5a5-472a-acfb-102dbfe83e0e" (UID: "33b1de07-e5a5-472a-acfb-102dbfe83e0e"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:59:11 crc kubenswrapper[4764]: I0202 09:59:11.654989 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/33b1de07-e5a5-472a-acfb-102dbfe83e0e-config" (OuterVolumeSpecName: "config") pod "33b1de07-e5a5-472a-acfb-102dbfe83e0e" (UID: "33b1de07-e5a5-472a-acfb-102dbfe83e0e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:59:11 crc kubenswrapper[4764]: I0202 09:59:11.668683 4764 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/33b1de07-e5a5-472a-acfb-102dbfe83e0e-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 02 09:59:11 crc kubenswrapper[4764]: I0202 09:59:11.668719 4764 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/165479bf-43a7-47fc-9083-c87cb1ee0f2c-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 09:59:11 crc kubenswrapper[4764]: I0202 09:59:11.668731 4764 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/33b1de07-e5a5-472a-acfb-102dbfe83e0e-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 02 09:59:11 crc kubenswrapper[4764]: I0202 09:59:11.668755 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-496bg\" (UniqueName: \"kubernetes.io/projected/33b1de07-e5a5-472a-acfb-102dbfe83e0e-kube-api-access-496bg\") on node \"crc\" DevicePath \"\"" Feb 02 09:59:11 crc kubenswrapper[4764]: I0202 09:59:11.668765 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ftgmp\" (UniqueName: \"kubernetes.io/projected/165479bf-43a7-47fc-9083-c87cb1ee0f2c-kube-api-access-ftgmp\") on node \"crc\" DevicePath \"\"" Feb 02 09:59:11 crc kubenswrapper[4764]: I0202 09:59:11.668773 4764 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/33b1de07-e5a5-472a-acfb-102dbfe83e0e-config\") on node \"crc\" DevicePath \"\"" Feb 02 09:59:11 crc kubenswrapper[4764]: I0202 09:59:11.668780 4764 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/33b1de07-e5a5-472a-acfb-102dbfe83e0e-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 02 09:59:11 crc kubenswrapper[4764]: I0202 09:59:11.668789 4764 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/165479bf-43a7-47fc-9083-c87cb1ee0f2c-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 09:59:11 crc kubenswrapper[4764]: I0202 09:59:11.668796 4764 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/165479bf-43a7-47fc-9083-c87cb1ee0f2c-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Feb 02 09:59:11 crc kubenswrapper[4764]: I0202 09:59:11.668804 4764 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/33b1de07-e5a5-472a-acfb-102dbfe83e0e-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 02 09:59:11 crc kubenswrapper[4764]: I0202 09:59:11.707896 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"eed25ff8-c8ee-41d6-bb33-89e4c5f30724","Type":"ContainerStarted","Data":"1cbdbe3912fc3a95a6f32add6570f26675e80c39d56508f60effdce88e750838"} Feb 02 09:59:11 crc kubenswrapper[4764]: I0202 09:59:11.709430 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77755c997f-6996z" event={"ID":"33b1de07-e5a5-472a-acfb-102dbfe83e0e","Type":"ContainerDied","Data":"9a5a2a5be6affd06eb000aba6a885f6c506bcd48084acc13349fd4216e72945e"} Feb 02 09:59:11 crc kubenswrapper[4764]: I0202 09:59:11.709465 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77755c997f-6996z" Feb 02 09:59:11 crc kubenswrapper[4764]: I0202 09:59:11.709486 4764 scope.go:117] "RemoveContainer" containerID="b42ec966d617e60cef3c0bd4351d46ef96d749b7d8a53fafb653e8d7beb3a126" Feb 02 09:59:11 crc kubenswrapper[4764]: I0202 09:59:11.713383 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6d4f8f8755-dp7xg" event={"ID":"b79a9079-9d96-4edc-97fc-b9bdf122e4d8","Type":"ContainerDied","Data":"d7b5b153040634feda48c21ad1fc5dd45a2e9e376c8028d9de2d75e11fe2d349"} Feb 02 09:59:11 crc kubenswrapper[4764]: I0202 09:59:11.713449 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6d4f8f8755-dp7xg" Feb 02 09:59:11 crc kubenswrapper[4764]: I0202 09:59:11.735665 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7f4b5b9665-qqgjw" event={"ID":"165479bf-43a7-47fc-9083-c87cb1ee0f2c","Type":"ContainerDied","Data":"5dc251dad9db660ac1714da21d2ad61cb71204b205a96fadba2e570712a989aa"} Feb 02 09:59:11 crc kubenswrapper[4764]: I0202 09:59:11.735753 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7f4b5b9665-qqgjw" Feb 02 09:59:11 crc kubenswrapper[4764]: I0202 09:59:11.777118 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-api-0" podStartSLOduration=6.777098173 podStartE2EDuration="6.777098173s" podCreationTimestamp="2026-02-02 09:59:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:59:11.73488423 +0000 UTC m=+3114.668608318" watchObservedRunningTime="2026-02-02 09:59:11.777098173 +0000 UTC m=+3114.710822261" Feb 02 09:59:11 crc kubenswrapper[4764]: I0202 09:59:11.804721 4764 scope.go:117] "RemoveContainer" containerID="781e28062c683f31972368a29d5067daf908ad95169f41f525d8e9446af75809" Feb 02 09:59:11 crc kubenswrapper[4764]: I0202 09:59:11.879882 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-6d4f8f8755-dp7xg"] Feb 02 09:59:11 crc kubenswrapper[4764]: I0202 09:59:11.879922 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-6d4f8f8755-dp7xg"] Feb 02 09:59:11 crc kubenswrapper[4764]: I0202 09:59:11.879987 4764 scope.go:117] "RemoveContainer" containerID="46b17bc1ee2d2cee2f865d09367965cf4628ec9996c321118f0fea7ca29a0810" Feb 02 09:59:11 crc kubenswrapper[4764]: I0202 09:59:11.885076 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7f4b5b9665-qqgjw"] Feb 02 09:59:11 crc kubenswrapper[4764]: I0202 09:59:11.895259 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-7f4b5b9665-qqgjw"] Feb 02 09:59:11 crc kubenswrapper[4764]: I0202 09:59:11.905983 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-77755c997f-6996z"] Feb 02 09:59:11 crc kubenswrapper[4764]: I0202 09:59:11.911483 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-77755c997f-6996z"] Feb 02 09:59:12 crc kubenswrapper[4764]: I0202 09:59:12.076174 4764 scope.go:117] "RemoveContainer" containerID="71fc5c9189af4e03a6f6fc0a3c0dbb2e1e002c4c3ba7e050a1b56f82fac1fc29" Feb 02 09:59:12 crc kubenswrapper[4764]: I0202 09:59:12.108278 4764 scope.go:117] "RemoveContainer" containerID="1bea9ae58a50124665a0eed96e3eb3490c806bc91198c0e7e5d31d67d80fcad6" Feb 02 09:59:12 crc kubenswrapper[4764]: I0202 09:59:12.747008 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"c0ba1c62-6619-4988-867b-f4a9e4d426dc","Type":"ContainerStarted","Data":"19ceda1a55292b83b269c26e4c0889cf09ba4c0c6ca5a5565dcb3adfc576a20b"} Feb 02 09:59:12 crc kubenswrapper[4764]: I0202 09:59:12.753030 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/manila-api-0" Feb 02 09:59:12 crc kubenswrapper[4764]: I0202 09:59:12.777135 4764 scope.go:117] "RemoveContainer" containerID="5c0ef8d37db32c9fb6af9fafb8fd85ea7920d910afd0c5afef3e02e43598b8de" Feb 02 09:59:13 crc kubenswrapper[4764]: I0202 09:59:13.762387 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"c0ba1c62-6619-4988-867b-f4a9e4d426dc","Type":"ContainerStarted","Data":"1013758ae9afcfb1fcd04a4e0c05610e7b91121bc8c77bda8d1b60f70e2a7927"} Feb 02 09:59:13 crc kubenswrapper[4764]: I0202 09:59:13.787859 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-share-share1-0" podStartSLOduration=5.168423922 podStartE2EDuration="16.787843917s" podCreationTimestamp="2026-02-02 09:58:57 +0000 UTC" firstStartedPulling="2026-02-02 09:58:59.076815478 +0000 UTC m=+3102.010539566" lastFinishedPulling="2026-02-02 09:59:10.696235473 +0000 UTC m=+3113.629959561" observedRunningTime="2026-02-02 09:59:13.781877928 +0000 UTC m=+3116.715602036" watchObservedRunningTime="2026-02-02 09:59:13.787843917 +0000 UTC m=+3116.721568005" Feb 02 09:59:13 crc kubenswrapper[4764]: I0202 09:59:13.837911 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="165479bf-43a7-47fc-9083-c87cb1ee0f2c" path="/var/lib/kubelet/pods/165479bf-43a7-47fc-9083-c87cb1ee0f2c/volumes" Feb 02 09:59:13 crc kubenswrapper[4764]: I0202 09:59:13.838850 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="33b1de07-e5a5-472a-acfb-102dbfe83e0e" path="/var/lib/kubelet/pods/33b1de07-e5a5-472a-acfb-102dbfe83e0e/volumes" Feb 02 09:59:13 crc kubenswrapper[4764]: I0202 09:59:13.839600 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b79a9079-9d96-4edc-97fc-b9bdf122e4d8" path="/var/lib/kubelet/pods/b79a9079-9d96-4edc-97fc-b9bdf122e4d8/volumes" Feb 02 09:59:14 crc kubenswrapper[4764]: I0202 09:59:14.394465 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 02 09:59:14 crc kubenswrapper[4764]: I0202 09:59:14.394718 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ab13eb4e-3576-47f7-bda0-dbac43dd1200" containerName="ceilometer-central-agent" containerID="cri-o://1f234f6678d369cfb7c8faf829aa7ccf331f7c163b39d33b6f9c87ecb5534564" gracePeriod=30 Feb 02 09:59:14 crc kubenswrapper[4764]: I0202 09:59:14.394832 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ab13eb4e-3576-47f7-bda0-dbac43dd1200" containerName="sg-core" containerID="cri-o://9b1db71ed2689d7dbec046dab3a016f1284b7c3571476ce84e62b27a58ae9a6a" gracePeriod=30 Feb 02 09:59:14 crc kubenswrapper[4764]: I0202 09:59:14.394852 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ab13eb4e-3576-47f7-bda0-dbac43dd1200" containerName="ceilometer-notification-agent" containerID="cri-o://1429d52ae9bdcecbc9724b64fe86de6efb703fdea6c88978d63974d8de4cb204" gracePeriod=30 Feb 02 09:59:14 crc kubenswrapper[4764]: I0202 09:59:14.394908 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ab13eb4e-3576-47f7-bda0-dbac43dd1200" containerName="proxy-httpd" containerID="cri-o://024682a0e94744dfba3481e7a6baee331b2323d6d1006332b15cda00582b0e74" gracePeriod=30 Feb 02 09:59:14 crc kubenswrapper[4764]: I0202 09:59:14.641522 4764 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="ab13eb4e-3576-47f7-bda0-dbac43dd1200" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.190:3000/\": dial tcp 10.217.0.190:3000: connect: connection refused" Feb 02 09:59:14 crc kubenswrapper[4764]: I0202 09:59:14.793968 4764 generic.go:334] "Generic (PLEG): container finished" podID="ab13eb4e-3576-47f7-bda0-dbac43dd1200" containerID="024682a0e94744dfba3481e7a6baee331b2323d6d1006332b15cda00582b0e74" exitCode=0 Feb 02 09:59:14 crc kubenswrapper[4764]: I0202 09:59:14.794002 4764 generic.go:334] "Generic (PLEG): container finished" podID="ab13eb4e-3576-47f7-bda0-dbac43dd1200" containerID="9b1db71ed2689d7dbec046dab3a016f1284b7c3571476ce84e62b27a58ae9a6a" exitCode=2 Feb 02 09:59:14 crc kubenswrapper[4764]: I0202 09:59:14.794896 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ab13eb4e-3576-47f7-bda0-dbac43dd1200","Type":"ContainerDied","Data":"024682a0e94744dfba3481e7a6baee331b2323d6d1006332b15cda00582b0e74"} Feb 02 09:59:14 crc kubenswrapper[4764]: I0202 09:59:14.795036 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ab13eb4e-3576-47f7-bda0-dbac43dd1200","Type":"ContainerDied","Data":"9b1db71ed2689d7dbec046dab3a016f1284b7c3571476ce84e62b27a58ae9a6a"} Feb 02 09:59:15 crc kubenswrapper[4764]: I0202 09:59:15.804160 4764 generic.go:334] "Generic (PLEG): container finished" podID="ab13eb4e-3576-47f7-bda0-dbac43dd1200" containerID="1f234f6678d369cfb7c8faf829aa7ccf331f7c163b39d33b6f9c87ecb5534564" exitCode=0 Feb 02 09:59:15 crc kubenswrapper[4764]: I0202 09:59:15.804219 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ab13eb4e-3576-47f7-bda0-dbac43dd1200","Type":"ContainerDied","Data":"1f234f6678d369cfb7c8faf829aa7ccf331f7c163b39d33b6f9c87ecb5534564"} Feb 02 09:59:16 crc kubenswrapper[4764]: I0202 09:59:16.339513 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-698ff6d98d-6h7sk" Feb 02 09:59:16 crc kubenswrapper[4764]: I0202 09:59:16.349991 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-6c8bdc8fbb-gkt26" Feb 02 09:59:16 crc kubenswrapper[4764]: I0202 09:59:16.822878 4764 generic.go:334] "Generic (PLEG): container finished" podID="ab13eb4e-3576-47f7-bda0-dbac43dd1200" containerID="1429d52ae9bdcecbc9724b64fe86de6efb703fdea6c88978d63974d8de4cb204" exitCode=0 Feb 02 09:59:16 crc kubenswrapper[4764]: I0202 09:59:16.822963 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ab13eb4e-3576-47f7-bda0-dbac43dd1200","Type":"ContainerDied","Data":"1429d52ae9bdcecbc9724b64fe86de6efb703fdea6c88978d63974d8de4cb204"} Feb 02 09:59:17 crc kubenswrapper[4764]: I0202 09:59:17.016437 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 02 09:59:17 crc kubenswrapper[4764]: I0202 09:59:17.183537 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wx9sw\" (UniqueName: \"kubernetes.io/projected/ab13eb4e-3576-47f7-bda0-dbac43dd1200-kube-api-access-wx9sw\") pod \"ab13eb4e-3576-47f7-bda0-dbac43dd1200\" (UID: \"ab13eb4e-3576-47f7-bda0-dbac43dd1200\") " Feb 02 09:59:17 crc kubenswrapper[4764]: I0202 09:59:17.183600 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ab13eb4e-3576-47f7-bda0-dbac43dd1200-log-httpd\") pod \"ab13eb4e-3576-47f7-bda0-dbac43dd1200\" (UID: \"ab13eb4e-3576-47f7-bda0-dbac43dd1200\") " Feb 02 09:59:17 crc kubenswrapper[4764]: I0202 09:59:17.183659 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab13eb4e-3576-47f7-bda0-dbac43dd1200-config-data\") pod \"ab13eb4e-3576-47f7-bda0-dbac43dd1200\" (UID: \"ab13eb4e-3576-47f7-bda0-dbac43dd1200\") " Feb 02 09:59:17 crc kubenswrapper[4764]: I0202 09:59:17.183737 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab13eb4e-3576-47f7-bda0-dbac43dd1200-combined-ca-bundle\") pod \"ab13eb4e-3576-47f7-bda0-dbac43dd1200\" (UID: \"ab13eb4e-3576-47f7-bda0-dbac43dd1200\") " Feb 02 09:59:17 crc kubenswrapper[4764]: I0202 09:59:17.183788 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ab13eb4e-3576-47f7-bda0-dbac43dd1200-scripts\") pod \"ab13eb4e-3576-47f7-bda0-dbac43dd1200\" (UID: \"ab13eb4e-3576-47f7-bda0-dbac43dd1200\") " Feb 02 09:59:17 crc kubenswrapper[4764]: I0202 09:59:17.183821 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ab13eb4e-3576-47f7-bda0-dbac43dd1200-run-httpd\") pod \"ab13eb4e-3576-47f7-bda0-dbac43dd1200\" (UID: \"ab13eb4e-3576-47f7-bda0-dbac43dd1200\") " Feb 02 09:59:17 crc kubenswrapper[4764]: I0202 09:59:17.183844 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab13eb4e-3576-47f7-bda0-dbac43dd1200-ceilometer-tls-certs\") pod \"ab13eb4e-3576-47f7-bda0-dbac43dd1200\" (UID: \"ab13eb4e-3576-47f7-bda0-dbac43dd1200\") " Feb 02 09:59:17 crc kubenswrapper[4764]: I0202 09:59:17.183878 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ab13eb4e-3576-47f7-bda0-dbac43dd1200-sg-core-conf-yaml\") pod \"ab13eb4e-3576-47f7-bda0-dbac43dd1200\" (UID: \"ab13eb4e-3576-47f7-bda0-dbac43dd1200\") " Feb 02 09:59:17 crc kubenswrapper[4764]: I0202 09:59:17.184341 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ab13eb4e-3576-47f7-bda0-dbac43dd1200-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "ab13eb4e-3576-47f7-bda0-dbac43dd1200" (UID: "ab13eb4e-3576-47f7-bda0-dbac43dd1200"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 09:59:17 crc kubenswrapper[4764]: I0202 09:59:17.184451 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ab13eb4e-3576-47f7-bda0-dbac43dd1200-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "ab13eb4e-3576-47f7-bda0-dbac43dd1200" (UID: "ab13eb4e-3576-47f7-bda0-dbac43dd1200"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 09:59:17 crc kubenswrapper[4764]: I0202 09:59:17.185003 4764 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ab13eb4e-3576-47f7-bda0-dbac43dd1200-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 02 09:59:17 crc kubenswrapper[4764]: I0202 09:59:17.185024 4764 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ab13eb4e-3576-47f7-bda0-dbac43dd1200-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 02 09:59:17 crc kubenswrapper[4764]: I0202 09:59:17.193076 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab13eb4e-3576-47f7-bda0-dbac43dd1200-scripts" (OuterVolumeSpecName: "scripts") pod "ab13eb4e-3576-47f7-bda0-dbac43dd1200" (UID: "ab13eb4e-3576-47f7-bda0-dbac43dd1200"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:59:17 crc kubenswrapper[4764]: I0202 09:59:17.199199 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ab13eb4e-3576-47f7-bda0-dbac43dd1200-kube-api-access-wx9sw" (OuterVolumeSpecName: "kube-api-access-wx9sw") pod "ab13eb4e-3576-47f7-bda0-dbac43dd1200" (UID: "ab13eb4e-3576-47f7-bda0-dbac43dd1200"). InnerVolumeSpecName "kube-api-access-wx9sw". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:59:17 crc kubenswrapper[4764]: I0202 09:59:17.225916 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab13eb4e-3576-47f7-bda0-dbac43dd1200-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "ab13eb4e-3576-47f7-bda0-dbac43dd1200" (UID: "ab13eb4e-3576-47f7-bda0-dbac43dd1200"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:59:17 crc kubenswrapper[4764]: I0202 09:59:17.263139 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab13eb4e-3576-47f7-bda0-dbac43dd1200-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "ab13eb4e-3576-47f7-bda0-dbac43dd1200" (UID: "ab13eb4e-3576-47f7-bda0-dbac43dd1200"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:59:17 crc kubenswrapper[4764]: I0202 09:59:17.287326 4764 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ab13eb4e-3576-47f7-bda0-dbac43dd1200-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 09:59:17 crc kubenswrapper[4764]: I0202 09:59:17.287355 4764 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab13eb4e-3576-47f7-bda0-dbac43dd1200-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 02 09:59:17 crc kubenswrapper[4764]: I0202 09:59:17.287367 4764 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ab13eb4e-3576-47f7-bda0-dbac43dd1200-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 02 09:59:17 crc kubenswrapper[4764]: I0202 09:59:17.287375 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wx9sw\" (UniqueName: \"kubernetes.io/projected/ab13eb4e-3576-47f7-bda0-dbac43dd1200-kube-api-access-wx9sw\") on node \"crc\" DevicePath \"\"" Feb 02 09:59:17 crc kubenswrapper[4764]: I0202 09:59:17.289195 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab13eb4e-3576-47f7-bda0-dbac43dd1200-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ab13eb4e-3576-47f7-bda0-dbac43dd1200" (UID: "ab13eb4e-3576-47f7-bda0-dbac43dd1200"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:59:17 crc kubenswrapper[4764]: I0202 09:59:17.325854 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab13eb4e-3576-47f7-bda0-dbac43dd1200-config-data" (OuterVolumeSpecName: "config-data") pod "ab13eb4e-3576-47f7-bda0-dbac43dd1200" (UID: "ab13eb4e-3576-47f7-bda0-dbac43dd1200"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:59:17 crc kubenswrapper[4764]: I0202 09:59:17.389526 4764 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab13eb4e-3576-47f7-bda0-dbac43dd1200-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 09:59:17 crc kubenswrapper[4764]: I0202 09:59:17.389566 4764 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab13eb4e-3576-47f7-bda0-dbac43dd1200-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 09:59:17 crc kubenswrapper[4764]: I0202 09:59:17.859721 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 02 09:59:17 crc kubenswrapper[4764]: I0202 09:59:17.877831 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ab13eb4e-3576-47f7-bda0-dbac43dd1200","Type":"ContainerDied","Data":"26d9f36a2d01e51dbe8e4b17a4ea22ca650308bfb4f9f1aaad22cd30ffe2f479"} Feb 02 09:59:17 crc kubenswrapper[4764]: I0202 09:59:17.877987 4764 scope.go:117] "RemoveContainer" containerID="024682a0e94744dfba3481e7a6baee331b2323d6d1006332b15cda00582b0e74" Feb 02 09:59:18 crc kubenswrapper[4764]: I0202 09:59:18.007637 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 02 09:59:18 crc kubenswrapper[4764]: I0202 09:59:18.023698 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 02 09:59:18 crc kubenswrapper[4764]: I0202 09:59:18.031988 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 02 09:59:18 crc kubenswrapper[4764]: E0202 09:59:18.035834 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b79a9079-9d96-4edc-97fc-b9bdf122e4d8" containerName="horizon-log" Feb 02 09:59:18 crc kubenswrapper[4764]: I0202 09:59:18.035856 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="b79a9079-9d96-4edc-97fc-b9bdf122e4d8" containerName="horizon-log" Feb 02 09:59:18 crc kubenswrapper[4764]: E0202 09:59:18.035869 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="165479bf-43a7-47fc-9083-c87cb1ee0f2c" containerName="horizon" Feb 02 09:59:18 crc kubenswrapper[4764]: I0202 09:59:18.035896 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="165479bf-43a7-47fc-9083-c87cb1ee0f2c" containerName="horizon" Feb 02 09:59:18 crc kubenswrapper[4764]: E0202 09:59:18.035906 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab13eb4e-3576-47f7-bda0-dbac43dd1200" containerName="proxy-httpd" Feb 02 09:59:18 crc kubenswrapper[4764]: I0202 09:59:18.035911 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab13eb4e-3576-47f7-bda0-dbac43dd1200" containerName="proxy-httpd" Feb 02 09:59:18 crc kubenswrapper[4764]: E0202 09:59:18.035926 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab13eb4e-3576-47f7-bda0-dbac43dd1200" containerName="sg-core" Feb 02 09:59:18 crc kubenswrapper[4764]: I0202 09:59:18.035956 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab13eb4e-3576-47f7-bda0-dbac43dd1200" containerName="sg-core" Feb 02 09:59:18 crc kubenswrapper[4764]: E0202 09:59:18.035968 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab13eb4e-3576-47f7-bda0-dbac43dd1200" containerName="ceilometer-central-agent" Feb 02 09:59:18 crc kubenswrapper[4764]: I0202 09:59:18.035974 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab13eb4e-3576-47f7-bda0-dbac43dd1200" containerName="ceilometer-central-agent" Feb 02 09:59:18 crc kubenswrapper[4764]: E0202 09:59:18.035995 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="165479bf-43a7-47fc-9083-c87cb1ee0f2c" containerName="horizon-log" Feb 02 09:59:18 crc kubenswrapper[4764]: I0202 09:59:18.036001 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="165479bf-43a7-47fc-9083-c87cb1ee0f2c" containerName="horizon-log" Feb 02 09:59:18 crc kubenswrapper[4764]: E0202 09:59:18.036016 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33b1de07-e5a5-472a-acfb-102dbfe83e0e" containerName="dnsmasq-dns" Feb 02 09:59:18 crc kubenswrapper[4764]: I0202 09:59:18.036021 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="33b1de07-e5a5-472a-acfb-102dbfe83e0e" containerName="dnsmasq-dns" Feb 02 09:59:18 crc kubenswrapper[4764]: E0202 09:59:18.036030 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33b1de07-e5a5-472a-acfb-102dbfe83e0e" containerName="init" Feb 02 09:59:18 crc kubenswrapper[4764]: I0202 09:59:18.036036 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="33b1de07-e5a5-472a-acfb-102dbfe83e0e" containerName="init" Feb 02 09:59:18 crc kubenswrapper[4764]: E0202 09:59:18.036043 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab13eb4e-3576-47f7-bda0-dbac43dd1200" containerName="ceilometer-notification-agent" Feb 02 09:59:18 crc kubenswrapper[4764]: I0202 09:59:18.036049 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab13eb4e-3576-47f7-bda0-dbac43dd1200" containerName="ceilometer-notification-agent" Feb 02 09:59:18 crc kubenswrapper[4764]: E0202 09:59:18.036055 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b79a9079-9d96-4edc-97fc-b9bdf122e4d8" containerName="horizon" Feb 02 09:59:18 crc kubenswrapper[4764]: I0202 09:59:18.036061 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="b79a9079-9d96-4edc-97fc-b9bdf122e4d8" containerName="horizon" Feb 02 09:59:18 crc kubenswrapper[4764]: I0202 09:59:18.036260 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="165479bf-43a7-47fc-9083-c87cb1ee0f2c" containerName="horizon-log" Feb 02 09:59:18 crc kubenswrapper[4764]: I0202 09:59:18.036270 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab13eb4e-3576-47f7-bda0-dbac43dd1200" containerName="ceilometer-notification-agent" Feb 02 09:59:18 crc kubenswrapper[4764]: I0202 09:59:18.036282 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="33b1de07-e5a5-472a-acfb-102dbfe83e0e" containerName="dnsmasq-dns" Feb 02 09:59:18 crc kubenswrapper[4764]: I0202 09:59:18.036291 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab13eb4e-3576-47f7-bda0-dbac43dd1200" containerName="proxy-httpd" Feb 02 09:59:18 crc kubenswrapper[4764]: I0202 09:59:18.036302 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="165479bf-43a7-47fc-9083-c87cb1ee0f2c" containerName="horizon" Feb 02 09:59:18 crc kubenswrapper[4764]: I0202 09:59:18.036310 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="b79a9079-9d96-4edc-97fc-b9bdf122e4d8" containerName="horizon" Feb 02 09:59:18 crc kubenswrapper[4764]: I0202 09:59:18.036318 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab13eb4e-3576-47f7-bda0-dbac43dd1200" containerName="ceilometer-central-agent" Feb 02 09:59:18 crc kubenswrapper[4764]: I0202 09:59:18.036331 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="b79a9079-9d96-4edc-97fc-b9bdf122e4d8" containerName="horizon-log" Feb 02 09:59:18 crc kubenswrapper[4764]: I0202 09:59:18.036341 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab13eb4e-3576-47f7-bda0-dbac43dd1200" containerName="sg-core" Feb 02 09:59:18 crc kubenswrapper[4764]: I0202 09:59:18.038090 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 02 09:59:18 crc kubenswrapper[4764]: I0202 09:59:18.041801 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 02 09:59:18 crc kubenswrapper[4764]: I0202 09:59:18.041974 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 02 09:59:18 crc kubenswrapper[4764]: I0202 09:59:18.042141 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Feb 02 09:59:18 crc kubenswrapper[4764]: I0202 09:59:18.046717 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 02 09:59:18 crc kubenswrapper[4764]: I0202 09:59:18.103625 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/242b61fa-f577-449a-86b2-4cafa5175bd6-run-httpd\") pod \"ceilometer-0\" (UID: \"242b61fa-f577-449a-86b2-4cafa5175bd6\") " pod="openstack/ceilometer-0" Feb 02 09:59:18 crc kubenswrapper[4764]: I0202 09:59:18.103793 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/242b61fa-f577-449a-86b2-4cafa5175bd6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"242b61fa-f577-449a-86b2-4cafa5175bd6\") " pod="openstack/ceilometer-0" Feb 02 09:59:18 crc kubenswrapper[4764]: I0202 09:59:18.103873 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/242b61fa-f577-449a-86b2-4cafa5175bd6-scripts\") pod \"ceilometer-0\" (UID: \"242b61fa-f577-449a-86b2-4cafa5175bd6\") " pod="openstack/ceilometer-0" Feb 02 09:59:18 crc kubenswrapper[4764]: I0202 09:59:18.104027 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/242b61fa-f577-449a-86b2-4cafa5175bd6-config-data\") pod \"ceilometer-0\" (UID: \"242b61fa-f577-449a-86b2-4cafa5175bd6\") " pod="openstack/ceilometer-0" Feb 02 09:59:18 crc kubenswrapper[4764]: I0202 09:59:18.104085 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jwpts\" (UniqueName: \"kubernetes.io/projected/242b61fa-f577-449a-86b2-4cafa5175bd6-kube-api-access-jwpts\") pod \"ceilometer-0\" (UID: \"242b61fa-f577-449a-86b2-4cafa5175bd6\") " pod="openstack/ceilometer-0" Feb 02 09:59:18 crc kubenswrapper[4764]: I0202 09:59:18.104314 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/242b61fa-f577-449a-86b2-4cafa5175bd6-log-httpd\") pod \"ceilometer-0\" (UID: \"242b61fa-f577-449a-86b2-4cafa5175bd6\") " pod="openstack/ceilometer-0" Feb 02 09:59:18 crc kubenswrapper[4764]: I0202 09:59:18.104386 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/242b61fa-f577-449a-86b2-4cafa5175bd6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"242b61fa-f577-449a-86b2-4cafa5175bd6\") " pod="openstack/ceilometer-0" Feb 02 09:59:18 crc kubenswrapper[4764]: I0202 09:59:18.104572 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/242b61fa-f577-449a-86b2-4cafa5175bd6-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"242b61fa-f577-449a-86b2-4cafa5175bd6\") " pod="openstack/ceilometer-0" Feb 02 09:59:18 crc kubenswrapper[4764]: I0202 09:59:18.183355 4764 scope.go:117] "RemoveContainer" containerID="9b1db71ed2689d7dbec046dab3a016f1284b7c3571476ce84e62b27a58ae9a6a" Feb 02 09:59:18 crc kubenswrapper[4764]: I0202 09:59:18.201986 4764 scope.go:117] "RemoveContainer" containerID="1429d52ae9bdcecbc9724b64fe86de6efb703fdea6c88978d63974d8de4cb204" Feb 02 09:59:18 crc kubenswrapper[4764]: I0202 09:59:18.205455 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-share-share1-0" Feb 02 09:59:18 crc kubenswrapper[4764]: I0202 09:59:18.206704 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/242b61fa-f577-449a-86b2-4cafa5175bd6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"242b61fa-f577-449a-86b2-4cafa5175bd6\") " pod="openstack/ceilometer-0" Feb 02 09:59:18 crc kubenswrapper[4764]: I0202 09:59:18.206785 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/242b61fa-f577-449a-86b2-4cafa5175bd6-scripts\") pod \"ceilometer-0\" (UID: \"242b61fa-f577-449a-86b2-4cafa5175bd6\") " pod="openstack/ceilometer-0" Feb 02 09:59:18 crc kubenswrapper[4764]: I0202 09:59:18.206862 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/242b61fa-f577-449a-86b2-4cafa5175bd6-config-data\") pod \"ceilometer-0\" (UID: \"242b61fa-f577-449a-86b2-4cafa5175bd6\") " pod="openstack/ceilometer-0" Feb 02 09:59:18 crc kubenswrapper[4764]: I0202 09:59:18.206904 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jwpts\" (UniqueName: \"kubernetes.io/projected/242b61fa-f577-449a-86b2-4cafa5175bd6-kube-api-access-jwpts\") pod \"ceilometer-0\" (UID: \"242b61fa-f577-449a-86b2-4cafa5175bd6\") " pod="openstack/ceilometer-0" Feb 02 09:59:18 crc kubenswrapper[4764]: I0202 09:59:18.207111 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/242b61fa-f577-449a-86b2-4cafa5175bd6-log-httpd\") pod \"ceilometer-0\" (UID: \"242b61fa-f577-449a-86b2-4cafa5175bd6\") " pod="openstack/ceilometer-0" Feb 02 09:59:18 crc kubenswrapper[4764]: I0202 09:59:18.207180 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/242b61fa-f577-449a-86b2-4cafa5175bd6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"242b61fa-f577-449a-86b2-4cafa5175bd6\") " pod="openstack/ceilometer-0" Feb 02 09:59:18 crc kubenswrapper[4764]: I0202 09:59:18.207442 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/242b61fa-f577-449a-86b2-4cafa5175bd6-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"242b61fa-f577-449a-86b2-4cafa5175bd6\") " pod="openstack/ceilometer-0" Feb 02 09:59:18 crc kubenswrapper[4764]: I0202 09:59:18.207546 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/242b61fa-f577-449a-86b2-4cafa5175bd6-run-httpd\") pod \"ceilometer-0\" (UID: \"242b61fa-f577-449a-86b2-4cafa5175bd6\") " pod="openstack/ceilometer-0" Feb 02 09:59:18 crc kubenswrapper[4764]: I0202 09:59:18.207670 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/242b61fa-f577-449a-86b2-4cafa5175bd6-log-httpd\") pod \"ceilometer-0\" (UID: \"242b61fa-f577-449a-86b2-4cafa5175bd6\") " pod="openstack/ceilometer-0" Feb 02 09:59:18 crc kubenswrapper[4764]: I0202 09:59:18.213387 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/242b61fa-f577-449a-86b2-4cafa5175bd6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"242b61fa-f577-449a-86b2-4cafa5175bd6\") " pod="openstack/ceilometer-0" Feb 02 09:59:18 crc kubenswrapper[4764]: I0202 09:59:18.213989 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/242b61fa-f577-449a-86b2-4cafa5175bd6-scripts\") pod \"ceilometer-0\" (UID: \"242b61fa-f577-449a-86b2-4cafa5175bd6\") " pod="openstack/ceilometer-0" Feb 02 09:59:18 crc kubenswrapper[4764]: I0202 09:59:18.219291 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/242b61fa-f577-449a-86b2-4cafa5175bd6-run-httpd\") pod \"ceilometer-0\" (UID: \"242b61fa-f577-449a-86b2-4cafa5175bd6\") " pod="openstack/ceilometer-0" Feb 02 09:59:18 crc kubenswrapper[4764]: I0202 09:59:18.220345 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/242b61fa-f577-449a-86b2-4cafa5175bd6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"242b61fa-f577-449a-86b2-4cafa5175bd6\") " pod="openstack/ceilometer-0" Feb 02 09:59:18 crc kubenswrapper[4764]: I0202 09:59:18.224507 4764 scope.go:117] "RemoveContainer" containerID="1f234f6678d369cfb7c8faf829aa7ccf331f7c163b39d33b6f9c87ecb5534564" Feb 02 09:59:18 crc kubenswrapper[4764]: I0202 09:59:18.225329 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/242b61fa-f577-449a-86b2-4cafa5175bd6-config-data\") pod \"ceilometer-0\" (UID: \"242b61fa-f577-449a-86b2-4cafa5175bd6\") " pod="openstack/ceilometer-0" Feb 02 09:59:18 crc kubenswrapper[4764]: I0202 09:59:18.227138 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/242b61fa-f577-449a-86b2-4cafa5175bd6-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"242b61fa-f577-449a-86b2-4cafa5175bd6\") " pod="openstack/ceilometer-0" Feb 02 09:59:18 crc kubenswrapper[4764]: I0202 09:59:18.230166 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jwpts\" (UniqueName: \"kubernetes.io/projected/242b61fa-f577-449a-86b2-4cafa5175bd6-kube-api-access-jwpts\") pod \"ceilometer-0\" (UID: \"242b61fa-f577-449a-86b2-4cafa5175bd6\") " pod="openstack/ceilometer-0" Feb 02 09:59:18 crc kubenswrapper[4764]: I0202 09:59:18.333994 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-698ff6d98d-6h7sk" Feb 02 09:59:18 crc kubenswrapper[4764]: I0202 09:59:18.352906 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 02 09:59:18 crc kubenswrapper[4764]: I0202 09:59:18.413844 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-6c8bdc8fbb-gkt26"] Feb 02 09:59:18 crc kubenswrapper[4764]: I0202 09:59:18.414060 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-6c8bdc8fbb-gkt26" podUID="bd8ea4a3-c3ab-4f61-8257-40adfbbaf5df" containerName="horizon-log" containerID="cri-o://1522fd5feafe9818e7378965fcd9371ca91da359f83c68850824668df3553089" gracePeriod=30 Feb 02 09:59:18 crc kubenswrapper[4764]: I0202 09:59:18.414429 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-6c8bdc8fbb-gkt26" podUID="bd8ea4a3-c3ab-4f61-8257-40adfbbaf5df" containerName="horizon" containerID="cri-o://a9ba8889704978e8cdf3c04a2ed778ba56d8623240747403033dcaf0c56a278a" gracePeriod=30 Feb 02 09:59:18 crc kubenswrapper[4764]: I0202 09:59:18.430347 4764 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-6c8bdc8fbb-gkt26" podUID="bd8ea4a3-c3ab-4f61-8257-40adfbbaf5df" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.245:8443/dashboard/auth/login/?next=/dashboard/\": EOF" Feb 02 09:59:18 crc kubenswrapper[4764]: I0202 09:59:18.491833 4764 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-6c8bdc8fbb-gkt26" podUID="bd8ea4a3-c3ab-4f61-8257-40adfbbaf5df" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.245:8443/dashboard/auth/login/?next=/dashboard/\": EOF" Feb 02 09:59:18 crc kubenswrapper[4764]: I0202 09:59:18.964086 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 02 09:59:19 crc kubenswrapper[4764]: I0202 09:59:19.836736 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ab13eb4e-3576-47f7-bda0-dbac43dd1200" path="/var/lib/kubelet/pods/ab13eb4e-3576-47f7-bda0-dbac43dd1200/volumes" Feb 02 09:59:19 crc kubenswrapper[4764]: I0202 09:59:19.876091 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"242b61fa-f577-449a-86b2-4cafa5175bd6","Type":"ContainerStarted","Data":"b0f5a97a8d1e1186ff1ab04fdf5614e3e5017b2d21ea89a3faf033f2626478e1"} Feb 02 09:59:19 crc kubenswrapper[4764]: I0202 09:59:19.876142 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"242b61fa-f577-449a-86b2-4cafa5175bd6","Type":"ContainerStarted","Data":"2f69340b1e2d99241a7add0502a8fb4767c8e67d11518095e369b8d02da1ec6f"} Feb 02 09:59:19 crc kubenswrapper[4764]: I0202 09:59:19.931266 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-scheduler-0" Feb 02 09:59:19 crc kubenswrapper[4764]: I0202 09:59:19.970031 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-scheduler-0"] Feb 02 09:59:20 crc kubenswrapper[4764]: I0202 09:59:20.885281 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"242b61fa-f577-449a-86b2-4cafa5175bd6","Type":"ContainerStarted","Data":"26d990f679b1cc7fbf5fe4be9cb0cc2d0e798cd931fa844406103856a88c0b67"} Feb 02 09:59:20 crc kubenswrapper[4764]: I0202 09:59:20.885423 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-scheduler-0" podUID="b78ef5a7-a2a0-494d-b687-18384876989b" containerName="manila-scheduler" containerID="cri-o://68750e3d5d783461dcd76ac064febdaec06ae407bc81f56036014b1305a00528" gracePeriod=30 Feb 02 09:59:20 crc kubenswrapper[4764]: I0202 09:59:20.885464 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-scheduler-0" podUID="b78ef5a7-a2a0-494d-b687-18384876989b" containerName="probe" containerID="cri-o://c4978209530c86d62dededf2405f3a1ef19fe1ccb720a7d1b81bfc6027c1da2b" gracePeriod=30 Feb 02 09:59:21 crc kubenswrapper[4764]: I0202 09:59:21.899378 4764 generic.go:334] "Generic (PLEG): container finished" podID="bd8ea4a3-c3ab-4f61-8257-40adfbbaf5df" containerID="a9ba8889704978e8cdf3c04a2ed778ba56d8623240747403033dcaf0c56a278a" exitCode=0 Feb 02 09:59:21 crc kubenswrapper[4764]: I0202 09:59:21.899553 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6c8bdc8fbb-gkt26" event={"ID":"bd8ea4a3-c3ab-4f61-8257-40adfbbaf5df","Type":"ContainerDied","Data":"a9ba8889704978e8cdf3c04a2ed778ba56d8623240747403033dcaf0c56a278a"} Feb 02 09:59:21 crc kubenswrapper[4764]: I0202 09:59:21.902471 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"242b61fa-f577-449a-86b2-4cafa5175bd6","Type":"ContainerStarted","Data":"9f40ff3a0e9acddb38b3fca481a1d039579f5a4c7dbd0e4e608a3717a6c3c34e"} Feb 02 09:59:21 crc kubenswrapper[4764]: I0202 09:59:21.904788 4764 generic.go:334] "Generic (PLEG): container finished" podID="b78ef5a7-a2a0-494d-b687-18384876989b" containerID="c4978209530c86d62dededf2405f3a1ef19fe1ccb720a7d1b81bfc6027c1da2b" exitCode=0 Feb 02 09:59:21 crc kubenswrapper[4764]: I0202 09:59:21.904820 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"b78ef5a7-a2a0-494d-b687-18384876989b","Type":"ContainerDied","Data":"c4978209530c86d62dededf2405f3a1ef19fe1ccb720a7d1b81bfc6027c1da2b"} Feb 02 09:59:22 crc kubenswrapper[4764]: E0202 09:59:22.752464 4764 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb78ef5a7_a2a0_494d_b687_18384876989b.slice/crio-68750e3d5d783461dcd76ac064febdaec06ae407bc81f56036014b1305a00528.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb78ef5a7_a2a0_494d_b687_18384876989b.slice/crio-conmon-68750e3d5d783461dcd76ac064febdaec06ae407bc81f56036014b1305a00528.scope\": RecentStats: unable to find data in memory cache]" Feb 02 09:59:22 crc kubenswrapper[4764]: I0202 09:59:22.916728 4764 generic.go:334] "Generic (PLEG): container finished" podID="b78ef5a7-a2a0-494d-b687-18384876989b" containerID="68750e3d5d783461dcd76ac064febdaec06ae407bc81f56036014b1305a00528" exitCode=0 Feb 02 09:59:22 crc kubenswrapper[4764]: I0202 09:59:22.917018 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"b78ef5a7-a2a0-494d-b687-18384876989b","Type":"ContainerDied","Data":"68750e3d5d783461dcd76ac064febdaec06ae407bc81f56036014b1305a00528"} Feb 02 09:59:22 crc kubenswrapper[4764]: I0202 09:59:22.917044 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"b78ef5a7-a2a0-494d-b687-18384876989b","Type":"ContainerDied","Data":"6a92e0901d5e2d0f76df49de40945cbc8df9429b1736cec50e4bc4ed45a221c5"} Feb 02 09:59:22 crc kubenswrapper[4764]: I0202 09:59:22.917055 4764 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6a92e0901d5e2d0f76df49de40945cbc8df9429b1736cec50e4bc4ed45a221c5" Feb 02 09:59:22 crc kubenswrapper[4764]: I0202 09:59:22.961589 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Feb 02 09:59:23 crc kubenswrapper[4764]: I0202 09:59:23.113270 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b78ef5a7-a2a0-494d-b687-18384876989b-scripts\") pod \"b78ef5a7-a2a0-494d-b687-18384876989b\" (UID: \"b78ef5a7-a2a0-494d-b687-18384876989b\") " Feb 02 09:59:23 crc kubenswrapper[4764]: I0202 09:59:23.113368 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b78ef5a7-a2a0-494d-b687-18384876989b-config-data\") pod \"b78ef5a7-a2a0-494d-b687-18384876989b\" (UID: \"b78ef5a7-a2a0-494d-b687-18384876989b\") " Feb 02 09:59:23 crc kubenswrapper[4764]: I0202 09:59:23.113477 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b78ef5a7-a2a0-494d-b687-18384876989b-config-data-custom\") pod \"b78ef5a7-a2a0-494d-b687-18384876989b\" (UID: \"b78ef5a7-a2a0-494d-b687-18384876989b\") " Feb 02 09:59:23 crc kubenswrapper[4764]: I0202 09:59:23.113556 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b78ef5a7-a2a0-494d-b687-18384876989b-combined-ca-bundle\") pod \"b78ef5a7-a2a0-494d-b687-18384876989b\" (UID: \"b78ef5a7-a2a0-494d-b687-18384876989b\") " Feb 02 09:59:23 crc kubenswrapper[4764]: I0202 09:59:23.113582 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b78ef5a7-a2a0-494d-b687-18384876989b-etc-machine-id\") pod \"b78ef5a7-a2a0-494d-b687-18384876989b\" (UID: \"b78ef5a7-a2a0-494d-b687-18384876989b\") " Feb 02 09:59:23 crc kubenswrapper[4764]: I0202 09:59:23.113691 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r5tbq\" (UniqueName: \"kubernetes.io/projected/b78ef5a7-a2a0-494d-b687-18384876989b-kube-api-access-r5tbq\") pod \"b78ef5a7-a2a0-494d-b687-18384876989b\" (UID: \"b78ef5a7-a2a0-494d-b687-18384876989b\") " Feb 02 09:59:23 crc kubenswrapper[4764]: I0202 09:59:23.114346 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b78ef5a7-a2a0-494d-b687-18384876989b-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "b78ef5a7-a2a0-494d-b687-18384876989b" (UID: "b78ef5a7-a2a0-494d-b687-18384876989b"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 09:59:23 crc kubenswrapper[4764]: I0202 09:59:23.115147 4764 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b78ef5a7-a2a0-494d-b687-18384876989b-etc-machine-id\") on node \"crc\" DevicePath \"\"" Feb 02 09:59:23 crc kubenswrapper[4764]: I0202 09:59:23.120095 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b78ef5a7-a2a0-494d-b687-18384876989b-kube-api-access-r5tbq" (OuterVolumeSpecName: "kube-api-access-r5tbq") pod "b78ef5a7-a2a0-494d-b687-18384876989b" (UID: "b78ef5a7-a2a0-494d-b687-18384876989b"). InnerVolumeSpecName "kube-api-access-r5tbq". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:59:23 crc kubenswrapper[4764]: I0202 09:59:23.123467 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b78ef5a7-a2a0-494d-b687-18384876989b-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "b78ef5a7-a2a0-494d-b687-18384876989b" (UID: "b78ef5a7-a2a0-494d-b687-18384876989b"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:59:23 crc kubenswrapper[4764]: I0202 09:59:23.123625 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b78ef5a7-a2a0-494d-b687-18384876989b-scripts" (OuterVolumeSpecName: "scripts") pod "b78ef5a7-a2a0-494d-b687-18384876989b" (UID: "b78ef5a7-a2a0-494d-b687-18384876989b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:59:23 crc kubenswrapper[4764]: I0202 09:59:23.212561 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b78ef5a7-a2a0-494d-b687-18384876989b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b78ef5a7-a2a0-494d-b687-18384876989b" (UID: "b78ef5a7-a2a0-494d-b687-18384876989b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:59:23 crc kubenswrapper[4764]: I0202 09:59:23.217413 4764 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b78ef5a7-a2a0-494d-b687-18384876989b-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 09:59:23 crc kubenswrapper[4764]: I0202 09:59:23.217440 4764 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b78ef5a7-a2a0-494d-b687-18384876989b-config-data-custom\") on node \"crc\" DevicePath \"\"" Feb 02 09:59:23 crc kubenswrapper[4764]: I0202 09:59:23.217453 4764 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b78ef5a7-a2a0-494d-b687-18384876989b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 09:59:23 crc kubenswrapper[4764]: I0202 09:59:23.217461 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r5tbq\" (UniqueName: \"kubernetes.io/projected/b78ef5a7-a2a0-494d-b687-18384876989b-kube-api-access-r5tbq\") on node \"crc\" DevicePath \"\"" Feb 02 09:59:23 crc kubenswrapper[4764]: I0202 09:59:23.248397 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b78ef5a7-a2a0-494d-b687-18384876989b-config-data" (OuterVolumeSpecName: "config-data") pod "b78ef5a7-a2a0-494d-b687-18384876989b" (UID: "b78ef5a7-a2a0-494d-b687-18384876989b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:59:23 crc kubenswrapper[4764]: I0202 09:59:23.320208 4764 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b78ef5a7-a2a0-494d-b687-18384876989b-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 09:59:23 crc kubenswrapper[4764]: I0202 09:59:23.638918 4764 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-6c8bdc8fbb-gkt26" podUID="bd8ea4a3-c3ab-4f61-8257-40adfbbaf5df" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.245:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.245:8443: connect: connection refused" Feb 02 09:59:23 crc kubenswrapper[4764]: I0202 09:59:23.826018 4764 scope.go:117] "RemoveContainer" containerID="db4ebd4a8dbee216b70cb175f017f1799f3ca01a887f602ad42b9f295964da5d" Feb 02 09:59:23 crc kubenswrapper[4764]: E0202 09:59:23.826728 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 09:59:23 crc kubenswrapper[4764]: I0202 09:59:23.927364 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Feb 02 09:59:23 crc kubenswrapper[4764]: I0202 09:59:23.928756 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"242b61fa-f577-449a-86b2-4cafa5175bd6","Type":"ContainerStarted","Data":"615aa4b97ca82b72ed239b9b4d1d9a9cbb9eaee50ceebf2363d72c7c7863570e"} Feb 02 09:59:23 crc kubenswrapper[4764]: I0202 09:59:23.928791 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 02 09:59:23 crc kubenswrapper[4764]: I0202 09:59:23.955898 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.635347899 podStartE2EDuration="6.955881753s" podCreationTimestamp="2026-02-02 09:59:17 +0000 UTC" firstStartedPulling="2026-02-02 09:59:18.973255302 +0000 UTC m=+3121.906979400" lastFinishedPulling="2026-02-02 09:59:23.293789166 +0000 UTC m=+3126.227513254" observedRunningTime="2026-02-02 09:59:23.953667454 +0000 UTC m=+3126.887391552" watchObservedRunningTime="2026-02-02 09:59:23.955881753 +0000 UTC m=+3126.889605851" Feb 02 09:59:24 crc kubenswrapper[4764]: I0202 09:59:24.002306 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-scheduler-0"] Feb 02 09:59:24 crc kubenswrapper[4764]: I0202 09:59:24.052376 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-scheduler-0"] Feb 02 09:59:24 crc kubenswrapper[4764]: I0202 09:59:24.067981 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-scheduler-0"] Feb 02 09:59:24 crc kubenswrapper[4764]: E0202 09:59:24.068406 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b78ef5a7-a2a0-494d-b687-18384876989b" containerName="probe" Feb 02 09:59:24 crc kubenswrapper[4764]: I0202 09:59:24.068426 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="b78ef5a7-a2a0-494d-b687-18384876989b" containerName="probe" Feb 02 09:59:24 crc kubenswrapper[4764]: E0202 09:59:24.068448 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b78ef5a7-a2a0-494d-b687-18384876989b" containerName="manila-scheduler" Feb 02 09:59:24 crc kubenswrapper[4764]: I0202 09:59:24.068455 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="b78ef5a7-a2a0-494d-b687-18384876989b" containerName="manila-scheduler" Feb 02 09:59:24 crc kubenswrapper[4764]: I0202 09:59:24.068665 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="b78ef5a7-a2a0-494d-b687-18384876989b" containerName="manila-scheduler" Feb 02 09:59:24 crc kubenswrapper[4764]: I0202 09:59:24.068697 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="b78ef5a7-a2a0-494d-b687-18384876989b" containerName="probe" Feb 02 09:59:24 crc kubenswrapper[4764]: I0202 09:59:24.073067 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Feb 02 09:59:24 crc kubenswrapper[4764]: I0202 09:59:24.073173 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Feb 02 09:59:24 crc kubenswrapper[4764]: I0202 09:59:24.076429 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-scheduler-config-data" Feb 02 09:59:24 crc kubenswrapper[4764]: I0202 09:59:24.241998 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/47fbbd57-48b1-4cbf-b857-8ef14b8c8529-scripts\") pod \"manila-scheduler-0\" (UID: \"47fbbd57-48b1-4cbf-b857-8ef14b8c8529\") " pod="openstack/manila-scheduler-0" Feb 02 09:59:24 crc kubenswrapper[4764]: I0202 09:59:24.242320 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9nz5b\" (UniqueName: \"kubernetes.io/projected/47fbbd57-48b1-4cbf-b857-8ef14b8c8529-kube-api-access-9nz5b\") pod \"manila-scheduler-0\" (UID: \"47fbbd57-48b1-4cbf-b857-8ef14b8c8529\") " pod="openstack/manila-scheduler-0" Feb 02 09:59:24 crc kubenswrapper[4764]: I0202 09:59:24.242387 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/47fbbd57-48b1-4cbf-b857-8ef14b8c8529-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"47fbbd57-48b1-4cbf-b857-8ef14b8c8529\") " pod="openstack/manila-scheduler-0" Feb 02 09:59:24 crc kubenswrapper[4764]: I0202 09:59:24.242624 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/47fbbd57-48b1-4cbf-b857-8ef14b8c8529-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"47fbbd57-48b1-4cbf-b857-8ef14b8c8529\") " pod="openstack/manila-scheduler-0" Feb 02 09:59:24 crc kubenswrapper[4764]: I0202 09:59:24.242889 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47fbbd57-48b1-4cbf-b857-8ef14b8c8529-config-data\") pod \"manila-scheduler-0\" (UID: \"47fbbd57-48b1-4cbf-b857-8ef14b8c8529\") " pod="openstack/manila-scheduler-0" Feb 02 09:59:24 crc kubenswrapper[4764]: I0202 09:59:24.243118 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47fbbd57-48b1-4cbf-b857-8ef14b8c8529-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"47fbbd57-48b1-4cbf-b857-8ef14b8c8529\") " pod="openstack/manila-scheduler-0" Feb 02 09:59:24 crc kubenswrapper[4764]: I0202 09:59:24.344792 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/47fbbd57-48b1-4cbf-b857-8ef14b8c8529-scripts\") pod \"manila-scheduler-0\" (UID: \"47fbbd57-48b1-4cbf-b857-8ef14b8c8529\") " pod="openstack/manila-scheduler-0" Feb 02 09:59:24 crc kubenswrapper[4764]: I0202 09:59:24.344956 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9nz5b\" (UniqueName: \"kubernetes.io/projected/47fbbd57-48b1-4cbf-b857-8ef14b8c8529-kube-api-access-9nz5b\") pod \"manila-scheduler-0\" (UID: \"47fbbd57-48b1-4cbf-b857-8ef14b8c8529\") " pod="openstack/manila-scheduler-0" Feb 02 09:59:24 crc kubenswrapper[4764]: I0202 09:59:24.344985 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/47fbbd57-48b1-4cbf-b857-8ef14b8c8529-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"47fbbd57-48b1-4cbf-b857-8ef14b8c8529\") " pod="openstack/manila-scheduler-0" Feb 02 09:59:24 crc kubenswrapper[4764]: I0202 09:59:24.345051 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/47fbbd57-48b1-4cbf-b857-8ef14b8c8529-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"47fbbd57-48b1-4cbf-b857-8ef14b8c8529\") " pod="openstack/manila-scheduler-0" Feb 02 09:59:24 crc kubenswrapper[4764]: I0202 09:59:24.345109 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47fbbd57-48b1-4cbf-b857-8ef14b8c8529-config-data\") pod \"manila-scheduler-0\" (UID: \"47fbbd57-48b1-4cbf-b857-8ef14b8c8529\") " pod="openstack/manila-scheduler-0" Feb 02 09:59:24 crc kubenswrapper[4764]: I0202 09:59:24.345140 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47fbbd57-48b1-4cbf-b857-8ef14b8c8529-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"47fbbd57-48b1-4cbf-b857-8ef14b8c8529\") " pod="openstack/manila-scheduler-0" Feb 02 09:59:24 crc kubenswrapper[4764]: I0202 09:59:24.346446 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/47fbbd57-48b1-4cbf-b857-8ef14b8c8529-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"47fbbd57-48b1-4cbf-b857-8ef14b8c8529\") " pod="openstack/manila-scheduler-0" Feb 02 09:59:24 crc kubenswrapper[4764]: I0202 09:59:24.365652 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47fbbd57-48b1-4cbf-b857-8ef14b8c8529-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"47fbbd57-48b1-4cbf-b857-8ef14b8c8529\") " pod="openstack/manila-scheduler-0" Feb 02 09:59:24 crc kubenswrapper[4764]: I0202 09:59:24.366094 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/47fbbd57-48b1-4cbf-b857-8ef14b8c8529-scripts\") pod \"manila-scheduler-0\" (UID: \"47fbbd57-48b1-4cbf-b857-8ef14b8c8529\") " pod="openstack/manila-scheduler-0" Feb 02 09:59:24 crc kubenswrapper[4764]: I0202 09:59:24.366518 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47fbbd57-48b1-4cbf-b857-8ef14b8c8529-config-data\") pod \"manila-scheduler-0\" (UID: \"47fbbd57-48b1-4cbf-b857-8ef14b8c8529\") " pod="openstack/manila-scheduler-0" Feb 02 09:59:24 crc kubenswrapper[4764]: I0202 09:59:24.374671 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/47fbbd57-48b1-4cbf-b857-8ef14b8c8529-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"47fbbd57-48b1-4cbf-b857-8ef14b8c8529\") " pod="openstack/manila-scheduler-0" Feb 02 09:59:24 crc kubenswrapper[4764]: I0202 09:59:24.374825 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9nz5b\" (UniqueName: \"kubernetes.io/projected/47fbbd57-48b1-4cbf-b857-8ef14b8c8529-kube-api-access-9nz5b\") pod \"manila-scheduler-0\" (UID: \"47fbbd57-48b1-4cbf-b857-8ef14b8c8529\") " pod="openstack/manila-scheduler-0" Feb 02 09:59:24 crc kubenswrapper[4764]: I0202 09:59:24.400698 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Feb 02 09:59:24 crc kubenswrapper[4764]: I0202 09:59:24.926868 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Feb 02 09:59:24 crc kubenswrapper[4764]: I0202 09:59:24.944579 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"47fbbd57-48b1-4cbf-b857-8ef14b8c8529","Type":"ContainerStarted","Data":"e5c2aff91e594a8dde759cf453a3820ec3f1c44596db5b9640ffb79861555ef3"} Feb 02 09:59:25 crc kubenswrapper[4764]: I0202 09:59:25.840805 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b78ef5a7-a2a0-494d-b687-18384876989b" path="/var/lib/kubelet/pods/b78ef5a7-a2a0-494d-b687-18384876989b/volumes" Feb 02 09:59:25 crc kubenswrapper[4764]: I0202 09:59:25.957229 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"47fbbd57-48b1-4cbf-b857-8ef14b8c8529","Type":"ContainerStarted","Data":"acda51562e259dc72921d2569689ad86e1d9bc5400bcabe09adf9d7e146eb3d8"} Feb 02 09:59:25 crc kubenswrapper[4764]: I0202 09:59:25.957295 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"47fbbd57-48b1-4cbf-b857-8ef14b8c8529","Type":"ContainerStarted","Data":"43a5ab07ccba6d5add41b78ce7c7b051ff13101e30145e46d908a77c990c8640"} Feb 02 09:59:25 crc kubenswrapper[4764]: I0202 09:59:25.988355 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-scheduler-0" podStartSLOduration=2.988338724 podStartE2EDuration="2.988338724s" podCreationTimestamp="2026-02-02 09:59:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:59:25.987162262 +0000 UTC m=+3128.920886380" watchObservedRunningTime="2026-02-02 09:59:25.988338724 +0000 UTC m=+3128.922062802" Feb 02 09:59:27 crc kubenswrapper[4764]: I0202 09:59:27.446362 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/manila-api-0" Feb 02 09:59:29 crc kubenswrapper[4764]: I0202 09:59:29.599212 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-share-share1-0" Feb 02 09:59:29 crc kubenswrapper[4764]: I0202 09:59:29.681293 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-share-share1-0"] Feb 02 09:59:30 crc kubenswrapper[4764]: I0202 09:59:30.009564 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-share-share1-0" podUID="c0ba1c62-6619-4988-867b-f4a9e4d426dc" containerName="probe" containerID="cri-o://1013758ae9afcfb1fcd04a4e0c05610e7b91121bc8c77bda8d1b60f70e2a7927" gracePeriod=30 Feb 02 09:59:30 crc kubenswrapper[4764]: I0202 09:59:30.009923 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-share-share1-0" podUID="c0ba1c62-6619-4988-867b-f4a9e4d426dc" containerName="manila-share" containerID="cri-o://19ceda1a55292b83b269c26e4c0889cf09ba4c0c6ca5a5565dcb3adfc576a20b" gracePeriod=30 Feb 02 09:59:31 crc kubenswrapper[4764]: I0202 09:59:31.020739 4764 generic.go:334] "Generic (PLEG): container finished" podID="c0ba1c62-6619-4988-867b-f4a9e4d426dc" containerID="1013758ae9afcfb1fcd04a4e0c05610e7b91121bc8c77bda8d1b60f70e2a7927" exitCode=0 Feb 02 09:59:31 crc kubenswrapper[4764]: I0202 09:59:31.021062 4764 generic.go:334] "Generic (PLEG): container finished" podID="c0ba1c62-6619-4988-867b-f4a9e4d426dc" containerID="19ceda1a55292b83b269c26e4c0889cf09ba4c0c6ca5a5565dcb3adfc576a20b" exitCode=1 Feb 02 09:59:31 crc kubenswrapper[4764]: I0202 09:59:31.020817 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"c0ba1c62-6619-4988-867b-f4a9e4d426dc","Type":"ContainerDied","Data":"1013758ae9afcfb1fcd04a4e0c05610e7b91121bc8c77bda8d1b60f70e2a7927"} Feb 02 09:59:31 crc kubenswrapper[4764]: I0202 09:59:31.021119 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"c0ba1c62-6619-4988-867b-f4a9e4d426dc","Type":"ContainerDied","Data":"19ceda1a55292b83b269c26e4c0889cf09ba4c0c6ca5a5565dcb3adfc576a20b"} Feb 02 09:59:31 crc kubenswrapper[4764]: I0202 09:59:31.021133 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"c0ba1c62-6619-4988-867b-f4a9e4d426dc","Type":"ContainerDied","Data":"bf8b0c387315154e7bc4f1001ea4a7383094f96c8feb20161a8568e6decec43b"} Feb 02 09:59:31 crc kubenswrapper[4764]: I0202 09:59:31.021144 4764 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bf8b0c387315154e7bc4f1001ea4a7383094f96c8feb20161a8568e6decec43b" Feb 02 09:59:31 crc kubenswrapper[4764]: I0202 09:59:31.079146 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Feb 02 09:59:31 crc kubenswrapper[4764]: I0202 09:59:31.185984 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/c0ba1c62-6619-4988-867b-f4a9e4d426dc-var-lib-manila\") pod \"c0ba1c62-6619-4988-867b-f4a9e4d426dc\" (UID: \"c0ba1c62-6619-4988-867b-f4a9e4d426dc\") " Feb 02 09:59:31 crc kubenswrapper[4764]: I0202 09:59:31.186212 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c0ba1c62-6619-4988-867b-f4a9e4d426dc-var-lib-manila" (OuterVolumeSpecName: "var-lib-manila") pod "c0ba1c62-6619-4988-867b-f4a9e4d426dc" (UID: "c0ba1c62-6619-4988-867b-f4a9e4d426dc"). InnerVolumeSpecName "var-lib-manila". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 09:59:31 crc kubenswrapper[4764]: I0202 09:59:31.186387 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c0ba1c62-6619-4988-867b-f4a9e4d426dc-etc-machine-id\") pod \"c0ba1c62-6619-4988-867b-f4a9e4d426dc\" (UID: \"c0ba1c62-6619-4988-867b-f4a9e4d426dc\") " Feb 02 09:59:31 crc kubenswrapper[4764]: I0202 09:59:31.186436 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c0ba1c62-6619-4988-867b-f4a9e4d426dc-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "c0ba1c62-6619-4988-867b-f4a9e4d426dc" (UID: "c0ba1c62-6619-4988-867b-f4a9e4d426dc"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 09:59:31 crc kubenswrapper[4764]: I0202 09:59:31.186519 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c0ba1c62-6619-4988-867b-f4a9e4d426dc-scripts\") pod \"c0ba1c62-6619-4988-867b-f4a9e4d426dc\" (UID: \"c0ba1c62-6619-4988-867b-f4a9e4d426dc\") " Feb 02 09:59:31 crc kubenswrapper[4764]: I0202 09:59:31.186566 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0ba1c62-6619-4988-867b-f4a9e4d426dc-config-data\") pod \"c0ba1c62-6619-4988-867b-f4a9e4d426dc\" (UID: \"c0ba1c62-6619-4988-867b-f4a9e4d426dc\") " Feb 02 09:59:31 crc kubenswrapper[4764]: I0202 09:59:31.186633 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-78gmk\" (UniqueName: \"kubernetes.io/projected/c0ba1c62-6619-4988-867b-f4a9e4d426dc-kube-api-access-78gmk\") pod \"c0ba1c62-6619-4988-867b-f4a9e4d426dc\" (UID: \"c0ba1c62-6619-4988-867b-f4a9e4d426dc\") " Feb 02 09:59:31 crc kubenswrapper[4764]: I0202 09:59:31.187680 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/c0ba1c62-6619-4988-867b-f4a9e4d426dc-ceph\") pod \"c0ba1c62-6619-4988-867b-f4a9e4d426dc\" (UID: \"c0ba1c62-6619-4988-867b-f4a9e4d426dc\") " Feb 02 09:59:31 crc kubenswrapper[4764]: I0202 09:59:31.187811 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0ba1c62-6619-4988-867b-f4a9e4d426dc-combined-ca-bundle\") pod \"c0ba1c62-6619-4988-867b-f4a9e4d426dc\" (UID: \"c0ba1c62-6619-4988-867b-f4a9e4d426dc\") " Feb 02 09:59:31 crc kubenswrapper[4764]: I0202 09:59:31.188245 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c0ba1c62-6619-4988-867b-f4a9e4d426dc-config-data-custom\") pod \"c0ba1c62-6619-4988-867b-f4a9e4d426dc\" (UID: \"c0ba1c62-6619-4988-867b-f4a9e4d426dc\") " Feb 02 09:59:31 crc kubenswrapper[4764]: I0202 09:59:31.189547 4764 reconciler_common.go:293] "Volume detached for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/c0ba1c62-6619-4988-867b-f4a9e4d426dc-var-lib-manila\") on node \"crc\" DevicePath \"\"" Feb 02 09:59:31 crc kubenswrapper[4764]: I0202 09:59:31.189581 4764 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c0ba1c62-6619-4988-867b-f4a9e4d426dc-etc-machine-id\") on node \"crc\" DevicePath \"\"" Feb 02 09:59:31 crc kubenswrapper[4764]: I0202 09:59:31.192399 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c0ba1c62-6619-4988-867b-f4a9e4d426dc-kube-api-access-78gmk" (OuterVolumeSpecName: "kube-api-access-78gmk") pod "c0ba1c62-6619-4988-867b-f4a9e4d426dc" (UID: "c0ba1c62-6619-4988-867b-f4a9e4d426dc"). InnerVolumeSpecName "kube-api-access-78gmk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:59:31 crc kubenswrapper[4764]: I0202 09:59:31.204573 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c0ba1c62-6619-4988-867b-f4a9e4d426dc-scripts" (OuterVolumeSpecName: "scripts") pod "c0ba1c62-6619-4988-867b-f4a9e4d426dc" (UID: "c0ba1c62-6619-4988-867b-f4a9e4d426dc"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:59:31 crc kubenswrapper[4764]: I0202 09:59:31.204874 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c0ba1c62-6619-4988-867b-f4a9e4d426dc-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "c0ba1c62-6619-4988-867b-f4a9e4d426dc" (UID: "c0ba1c62-6619-4988-867b-f4a9e4d426dc"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:59:31 crc kubenswrapper[4764]: I0202 09:59:31.212533 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c0ba1c62-6619-4988-867b-f4a9e4d426dc-ceph" (OuterVolumeSpecName: "ceph") pod "c0ba1c62-6619-4988-867b-f4a9e4d426dc" (UID: "c0ba1c62-6619-4988-867b-f4a9e4d426dc"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:59:31 crc kubenswrapper[4764]: I0202 09:59:31.249575 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c0ba1c62-6619-4988-867b-f4a9e4d426dc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c0ba1c62-6619-4988-867b-f4a9e4d426dc" (UID: "c0ba1c62-6619-4988-867b-f4a9e4d426dc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:59:31 crc kubenswrapper[4764]: I0202 09:59:31.291612 4764 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c0ba1c62-6619-4988-867b-f4a9e4d426dc-config-data-custom\") on node \"crc\" DevicePath \"\"" Feb 02 09:59:31 crc kubenswrapper[4764]: I0202 09:59:31.291649 4764 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c0ba1c62-6619-4988-867b-f4a9e4d426dc-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 09:59:31 crc kubenswrapper[4764]: I0202 09:59:31.291658 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-78gmk\" (UniqueName: \"kubernetes.io/projected/c0ba1c62-6619-4988-867b-f4a9e4d426dc-kube-api-access-78gmk\") on node \"crc\" DevicePath \"\"" Feb 02 09:59:31 crc kubenswrapper[4764]: I0202 09:59:31.291671 4764 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/c0ba1c62-6619-4988-867b-f4a9e4d426dc-ceph\") on node \"crc\" DevicePath \"\"" Feb 02 09:59:31 crc kubenswrapper[4764]: I0202 09:59:31.291679 4764 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0ba1c62-6619-4988-867b-f4a9e4d426dc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 09:59:31 crc kubenswrapper[4764]: I0202 09:59:31.310767 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c0ba1c62-6619-4988-867b-f4a9e4d426dc-config-data" (OuterVolumeSpecName: "config-data") pod "c0ba1c62-6619-4988-867b-f4a9e4d426dc" (UID: "c0ba1c62-6619-4988-867b-f4a9e4d426dc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:59:31 crc kubenswrapper[4764]: I0202 09:59:31.393685 4764 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0ba1c62-6619-4988-867b-f4a9e4d426dc-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 09:59:32 crc kubenswrapper[4764]: I0202 09:59:32.028789 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Feb 02 09:59:32 crc kubenswrapper[4764]: I0202 09:59:32.065124 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-share-share1-0"] Feb 02 09:59:32 crc kubenswrapper[4764]: I0202 09:59:32.077457 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-share-share1-0"] Feb 02 09:59:32 crc kubenswrapper[4764]: I0202 09:59:32.104661 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-share-share1-0"] Feb 02 09:59:32 crc kubenswrapper[4764]: E0202 09:59:32.105334 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c0ba1c62-6619-4988-867b-f4a9e4d426dc" containerName="probe" Feb 02 09:59:32 crc kubenswrapper[4764]: I0202 09:59:32.105369 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="c0ba1c62-6619-4988-867b-f4a9e4d426dc" containerName="probe" Feb 02 09:59:32 crc kubenswrapper[4764]: E0202 09:59:32.105406 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c0ba1c62-6619-4988-867b-f4a9e4d426dc" containerName="manila-share" Feb 02 09:59:32 crc kubenswrapper[4764]: I0202 09:59:32.105420 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="c0ba1c62-6619-4988-867b-f4a9e4d426dc" containerName="manila-share" Feb 02 09:59:32 crc kubenswrapper[4764]: I0202 09:59:32.105776 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="c0ba1c62-6619-4988-867b-f4a9e4d426dc" containerName="manila-share" Feb 02 09:59:32 crc kubenswrapper[4764]: I0202 09:59:32.105806 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="c0ba1c62-6619-4988-867b-f4a9e4d426dc" containerName="probe" Feb 02 09:59:32 crc kubenswrapper[4764]: I0202 09:59:32.107740 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Feb 02 09:59:32 crc kubenswrapper[4764]: I0202 09:59:32.110471 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-share-share1-config-data" Feb 02 09:59:32 crc kubenswrapper[4764]: I0202 09:59:32.120764 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Feb 02 09:59:32 crc kubenswrapper[4764]: I0202 09:59:32.209381 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b9fbc0ac-9686-4433-ac79-79b9e0cfb54b-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"b9fbc0ac-9686-4433-ac79-79b9e0cfb54b\") " pod="openstack/manila-share-share1-0" Feb 02 09:59:32 crc kubenswrapper[4764]: I0202 09:59:32.209625 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/b9fbc0ac-9686-4433-ac79-79b9e0cfb54b-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"b9fbc0ac-9686-4433-ac79-79b9e0cfb54b\") " pod="openstack/manila-share-share1-0" Feb 02 09:59:32 crc kubenswrapper[4764]: I0202 09:59:32.209738 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-54lh5\" (UniqueName: \"kubernetes.io/projected/b9fbc0ac-9686-4433-ac79-79b9e0cfb54b-kube-api-access-54lh5\") pod \"manila-share-share1-0\" (UID: \"b9fbc0ac-9686-4433-ac79-79b9e0cfb54b\") " pod="openstack/manila-share-share1-0" Feb 02 09:59:32 crc kubenswrapper[4764]: I0202 09:59:32.209808 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b9fbc0ac-9686-4433-ac79-79b9e0cfb54b-scripts\") pod \"manila-share-share1-0\" (UID: \"b9fbc0ac-9686-4433-ac79-79b9e0cfb54b\") " pod="openstack/manila-share-share1-0" Feb 02 09:59:32 crc kubenswrapper[4764]: I0202 09:59:32.209902 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b9fbc0ac-9686-4433-ac79-79b9e0cfb54b-config-data\") pod \"manila-share-share1-0\" (UID: \"b9fbc0ac-9686-4433-ac79-79b9e0cfb54b\") " pod="openstack/manila-share-share1-0" Feb 02 09:59:32 crc kubenswrapper[4764]: I0202 09:59:32.210025 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/b9fbc0ac-9686-4433-ac79-79b9e0cfb54b-ceph\") pod \"manila-share-share1-0\" (UID: \"b9fbc0ac-9686-4433-ac79-79b9e0cfb54b\") " pod="openstack/manila-share-share1-0" Feb 02 09:59:32 crc kubenswrapper[4764]: I0202 09:59:32.210097 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b9fbc0ac-9686-4433-ac79-79b9e0cfb54b-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"b9fbc0ac-9686-4433-ac79-79b9e0cfb54b\") " pod="openstack/manila-share-share1-0" Feb 02 09:59:32 crc kubenswrapper[4764]: I0202 09:59:32.210182 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b9fbc0ac-9686-4433-ac79-79b9e0cfb54b-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"b9fbc0ac-9686-4433-ac79-79b9e0cfb54b\") " pod="openstack/manila-share-share1-0" Feb 02 09:59:32 crc kubenswrapper[4764]: I0202 09:59:32.312322 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b9fbc0ac-9686-4433-ac79-79b9e0cfb54b-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"b9fbc0ac-9686-4433-ac79-79b9e0cfb54b\") " pod="openstack/manila-share-share1-0" Feb 02 09:59:32 crc kubenswrapper[4764]: I0202 09:59:32.312620 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/b9fbc0ac-9686-4433-ac79-79b9e0cfb54b-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"b9fbc0ac-9686-4433-ac79-79b9e0cfb54b\") " pod="openstack/manila-share-share1-0" Feb 02 09:59:32 crc kubenswrapper[4764]: I0202 09:59:32.312654 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-54lh5\" (UniqueName: \"kubernetes.io/projected/b9fbc0ac-9686-4433-ac79-79b9e0cfb54b-kube-api-access-54lh5\") pod \"manila-share-share1-0\" (UID: \"b9fbc0ac-9686-4433-ac79-79b9e0cfb54b\") " pod="openstack/manila-share-share1-0" Feb 02 09:59:32 crc kubenswrapper[4764]: I0202 09:59:32.312672 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b9fbc0ac-9686-4433-ac79-79b9e0cfb54b-scripts\") pod \"manila-share-share1-0\" (UID: \"b9fbc0ac-9686-4433-ac79-79b9e0cfb54b\") " pod="openstack/manila-share-share1-0" Feb 02 09:59:32 crc kubenswrapper[4764]: I0202 09:59:32.312727 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b9fbc0ac-9686-4433-ac79-79b9e0cfb54b-config-data\") pod \"manila-share-share1-0\" (UID: \"b9fbc0ac-9686-4433-ac79-79b9e0cfb54b\") " pod="openstack/manila-share-share1-0" Feb 02 09:59:32 crc kubenswrapper[4764]: I0202 09:59:32.312785 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/b9fbc0ac-9686-4433-ac79-79b9e0cfb54b-ceph\") pod \"manila-share-share1-0\" (UID: \"b9fbc0ac-9686-4433-ac79-79b9e0cfb54b\") " pod="openstack/manila-share-share1-0" Feb 02 09:59:32 crc kubenswrapper[4764]: I0202 09:59:32.312804 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b9fbc0ac-9686-4433-ac79-79b9e0cfb54b-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"b9fbc0ac-9686-4433-ac79-79b9e0cfb54b\") " pod="openstack/manila-share-share1-0" Feb 02 09:59:32 crc kubenswrapper[4764]: I0202 09:59:32.312820 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b9fbc0ac-9686-4433-ac79-79b9e0cfb54b-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"b9fbc0ac-9686-4433-ac79-79b9e0cfb54b\") " pod="openstack/manila-share-share1-0" Feb 02 09:59:32 crc kubenswrapper[4764]: I0202 09:59:32.312895 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b9fbc0ac-9686-4433-ac79-79b9e0cfb54b-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"b9fbc0ac-9686-4433-ac79-79b9e0cfb54b\") " pod="openstack/manila-share-share1-0" Feb 02 09:59:32 crc kubenswrapper[4764]: I0202 09:59:32.312975 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/b9fbc0ac-9686-4433-ac79-79b9e0cfb54b-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"b9fbc0ac-9686-4433-ac79-79b9e0cfb54b\") " pod="openstack/manila-share-share1-0" Feb 02 09:59:32 crc kubenswrapper[4764]: I0202 09:59:32.316259 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/b9fbc0ac-9686-4433-ac79-79b9e0cfb54b-ceph\") pod \"manila-share-share1-0\" (UID: \"b9fbc0ac-9686-4433-ac79-79b9e0cfb54b\") " pod="openstack/manila-share-share1-0" Feb 02 09:59:32 crc kubenswrapper[4764]: I0202 09:59:32.318741 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b9fbc0ac-9686-4433-ac79-79b9e0cfb54b-config-data\") pod \"manila-share-share1-0\" (UID: \"b9fbc0ac-9686-4433-ac79-79b9e0cfb54b\") " pod="openstack/manila-share-share1-0" Feb 02 09:59:32 crc kubenswrapper[4764]: I0202 09:59:32.319152 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b9fbc0ac-9686-4433-ac79-79b9e0cfb54b-scripts\") pod \"manila-share-share1-0\" (UID: \"b9fbc0ac-9686-4433-ac79-79b9e0cfb54b\") " pod="openstack/manila-share-share1-0" Feb 02 09:59:32 crc kubenswrapper[4764]: I0202 09:59:32.320632 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b9fbc0ac-9686-4433-ac79-79b9e0cfb54b-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"b9fbc0ac-9686-4433-ac79-79b9e0cfb54b\") " pod="openstack/manila-share-share1-0" Feb 02 09:59:32 crc kubenswrapper[4764]: I0202 09:59:32.320817 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b9fbc0ac-9686-4433-ac79-79b9e0cfb54b-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"b9fbc0ac-9686-4433-ac79-79b9e0cfb54b\") " pod="openstack/manila-share-share1-0" Feb 02 09:59:32 crc kubenswrapper[4764]: I0202 09:59:32.336176 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-54lh5\" (UniqueName: \"kubernetes.io/projected/b9fbc0ac-9686-4433-ac79-79b9e0cfb54b-kube-api-access-54lh5\") pod \"manila-share-share1-0\" (UID: \"b9fbc0ac-9686-4433-ac79-79b9e0cfb54b\") " pod="openstack/manila-share-share1-0" Feb 02 09:59:32 crc kubenswrapper[4764]: I0202 09:59:32.429438 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Feb 02 09:59:32 crc kubenswrapper[4764]: I0202 09:59:32.998857 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Feb 02 09:59:33 crc kubenswrapper[4764]: W0202 09:59:33.000251 4764 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb9fbc0ac_9686_4433_ac79_79b9e0cfb54b.slice/crio-8b6aa2ccea9a3959d4d66872468e927c5c2a797cd3e984f43d08b5d4b1c744c4 WatchSource:0}: Error finding container 8b6aa2ccea9a3959d4d66872468e927c5c2a797cd3e984f43d08b5d4b1c744c4: Status 404 returned error can't find the container with id 8b6aa2ccea9a3959d4d66872468e927c5c2a797cd3e984f43d08b5d4b1c744c4 Feb 02 09:59:33 crc kubenswrapper[4764]: I0202 09:59:33.051022 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"b9fbc0ac-9686-4433-ac79-79b9e0cfb54b","Type":"ContainerStarted","Data":"8b6aa2ccea9a3959d4d66872468e927c5c2a797cd3e984f43d08b5d4b1c744c4"} Feb 02 09:59:33 crc kubenswrapper[4764]: I0202 09:59:33.638222 4764 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-6c8bdc8fbb-gkt26" podUID="bd8ea4a3-c3ab-4f61-8257-40adfbbaf5df" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.245:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.245:8443: connect: connection refused" Feb 02 09:59:33 crc kubenswrapper[4764]: I0202 09:59:33.838066 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c0ba1c62-6619-4988-867b-f4a9e4d426dc" path="/var/lib/kubelet/pods/c0ba1c62-6619-4988-867b-f4a9e4d426dc/volumes" Feb 02 09:59:34 crc kubenswrapper[4764]: I0202 09:59:34.062609 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"b9fbc0ac-9686-4433-ac79-79b9e0cfb54b","Type":"ContainerStarted","Data":"393c86e8499084e122b7efe434b651801bdecc48e691e1e46a29a34550eda7b7"} Feb 02 09:59:34 crc kubenswrapper[4764]: I0202 09:59:34.062649 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"b9fbc0ac-9686-4433-ac79-79b9e0cfb54b","Type":"ContainerStarted","Data":"03b111976c0b5e40af13d766eaa9702d940f306e18dd53e37ebffed99807dd84"} Feb 02 09:59:34 crc kubenswrapper[4764]: I0202 09:59:34.090561 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-share-share1-0" podStartSLOduration=2.090545352 podStartE2EDuration="2.090545352s" podCreationTimestamp="2026-02-02 09:59:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 09:59:34.083177896 +0000 UTC m=+3137.016901984" watchObservedRunningTime="2026-02-02 09:59:34.090545352 +0000 UTC m=+3137.024269440" Feb 02 09:59:34 crc kubenswrapper[4764]: I0202 09:59:34.402107 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-scheduler-0" Feb 02 09:59:37 crc kubenswrapper[4764]: I0202 09:59:37.840014 4764 scope.go:117] "RemoveContainer" containerID="db4ebd4a8dbee216b70cb175f017f1799f3ca01a887f602ad42b9f295964da5d" Feb 02 09:59:37 crc kubenswrapper[4764]: E0202 09:59:37.841170 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 09:59:42 crc kubenswrapper[4764]: I0202 09:59:42.430476 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-share-share1-0" Feb 02 09:59:43 crc kubenswrapper[4764]: I0202 09:59:43.638837 4764 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-6c8bdc8fbb-gkt26" podUID="bd8ea4a3-c3ab-4f61-8257-40adfbbaf5df" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.245:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.245:8443: connect: connection refused" Feb 02 09:59:45 crc kubenswrapper[4764]: I0202 09:59:45.767244 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-scheduler-0" Feb 02 09:59:48 crc kubenswrapper[4764]: I0202 09:59:48.373592 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Feb 02 09:59:48 crc kubenswrapper[4764]: I0202 09:59:48.838242 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6c8bdc8fbb-gkt26" Feb 02 09:59:48 crc kubenswrapper[4764]: I0202 09:59:48.937249 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/bd8ea4a3-c3ab-4f61-8257-40adfbbaf5df-horizon-tls-certs\") pod \"bd8ea4a3-c3ab-4f61-8257-40adfbbaf5df\" (UID: \"bd8ea4a3-c3ab-4f61-8257-40adfbbaf5df\") " Feb 02 09:59:48 crc kubenswrapper[4764]: I0202 09:59:48.937404 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bd8ea4a3-c3ab-4f61-8257-40adfbbaf5df-scripts\") pod \"bd8ea4a3-c3ab-4f61-8257-40adfbbaf5df\" (UID: \"bd8ea4a3-c3ab-4f61-8257-40adfbbaf5df\") " Feb 02 09:59:48 crc kubenswrapper[4764]: I0202 09:59:48.937488 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bd8ea4a3-c3ab-4f61-8257-40adfbbaf5df-logs\") pod \"bd8ea4a3-c3ab-4f61-8257-40adfbbaf5df\" (UID: \"bd8ea4a3-c3ab-4f61-8257-40adfbbaf5df\") " Feb 02 09:59:48 crc kubenswrapper[4764]: I0202 09:59:48.937549 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd8ea4a3-c3ab-4f61-8257-40adfbbaf5df-combined-ca-bundle\") pod \"bd8ea4a3-c3ab-4f61-8257-40adfbbaf5df\" (UID: \"bd8ea4a3-c3ab-4f61-8257-40adfbbaf5df\") " Feb 02 09:59:48 crc kubenswrapper[4764]: I0202 09:59:48.937727 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bd8ea4a3-c3ab-4f61-8257-40adfbbaf5df-config-data\") pod \"bd8ea4a3-c3ab-4f61-8257-40adfbbaf5df\" (UID: \"bd8ea4a3-c3ab-4f61-8257-40adfbbaf5df\") " Feb 02 09:59:48 crc kubenswrapper[4764]: I0202 09:59:48.937874 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/bd8ea4a3-c3ab-4f61-8257-40adfbbaf5df-horizon-secret-key\") pod \"bd8ea4a3-c3ab-4f61-8257-40adfbbaf5df\" (UID: \"bd8ea4a3-c3ab-4f61-8257-40adfbbaf5df\") " Feb 02 09:59:48 crc kubenswrapper[4764]: I0202 09:59:48.937974 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mhfvj\" (UniqueName: \"kubernetes.io/projected/bd8ea4a3-c3ab-4f61-8257-40adfbbaf5df-kube-api-access-mhfvj\") pod \"bd8ea4a3-c3ab-4f61-8257-40adfbbaf5df\" (UID: \"bd8ea4a3-c3ab-4f61-8257-40adfbbaf5df\") " Feb 02 09:59:48 crc kubenswrapper[4764]: I0202 09:59:48.951234 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bd8ea4a3-c3ab-4f61-8257-40adfbbaf5df-logs" (OuterVolumeSpecName: "logs") pod "bd8ea4a3-c3ab-4f61-8257-40adfbbaf5df" (UID: "bd8ea4a3-c3ab-4f61-8257-40adfbbaf5df"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 09:59:48 crc kubenswrapper[4764]: I0202 09:59:48.956569 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd8ea4a3-c3ab-4f61-8257-40adfbbaf5df-kube-api-access-mhfvj" (OuterVolumeSpecName: "kube-api-access-mhfvj") pod "bd8ea4a3-c3ab-4f61-8257-40adfbbaf5df" (UID: "bd8ea4a3-c3ab-4f61-8257-40adfbbaf5df"). InnerVolumeSpecName "kube-api-access-mhfvj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 09:59:48 crc kubenswrapper[4764]: I0202 09:59:48.965080 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd8ea4a3-c3ab-4f61-8257-40adfbbaf5df-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "bd8ea4a3-c3ab-4f61-8257-40adfbbaf5df" (UID: "bd8ea4a3-c3ab-4f61-8257-40adfbbaf5df"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:59:48 crc kubenswrapper[4764]: I0202 09:59:48.978557 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bd8ea4a3-c3ab-4f61-8257-40adfbbaf5df-scripts" (OuterVolumeSpecName: "scripts") pod "bd8ea4a3-c3ab-4f61-8257-40adfbbaf5df" (UID: "bd8ea4a3-c3ab-4f61-8257-40adfbbaf5df"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:59:48 crc kubenswrapper[4764]: I0202 09:59:48.987663 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd8ea4a3-c3ab-4f61-8257-40adfbbaf5df-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bd8ea4a3-c3ab-4f61-8257-40adfbbaf5df" (UID: "bd8ea4a3-c3ab-4f61-8257-40adfbbaf5df"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:59:48 crc kubenswrapper[4764]: I0202 09:59:48.995041 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bd8ea4a3-c3ab-4f61-8257-40adfbbaf5df-config-data" (OuterVolumeSpecName: "config-data") pod "bd8ea4a3-c3ab-4f61-8257-40adfbbaf5df" (UID: "bd8ea4a3-c3ab-4f61-8257-40adfbbaf5df"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 09:59:49 crc kubenswrapper[4764]: I0202 09:59:49.012190 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd8ea4a3-c3ab-4f61-8257-40adfbbaf5df-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "bd8ea4a3-c3ab-4f61-8257-40adfbbaf5df" (UID: "bd8ea4a3-c3ab-4f61-8257-40adfbbaf5df"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 09:59:49 crc kubenswrapper[4764]: I0202 09:59:49.042245 4764 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/bd8ea4a3-c3ab-4f61-8257-40adfbbaf5df-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 02 09:59:49 crc kubenswrapper[4764]: I0202 09:59:49.042291 4764 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bd8ea4a3-c3ab-4f61-8257-40adfbbaf5df-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 09:59:49 crc kubenswrapper[4764]: I0202 09:59:49.042303 4764 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bd8ea4a3-c3ab-4f61-8257-40adfbbaf5df-logs\") on node \"crc\" DevicePath \"\"" Feb 02 09:59:49 crc kubenswrapper[4764]: I0202 09:59:49.042313 4764 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd8ea4a3-c3ab-4f61-8257-40adfbbaf5df-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 09:59:49 crc kubenswrapper[4764]: I0202 09:59:49.042322 4764 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bd8ea4a3-c3ab-4f61-8257-40adfbbaf5df-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 09:59:49 crc kubenswrapper[4764]: I0202 09:59:49.042330 4764 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/bd8ea4a3-c3ab-4f61-8257-40adfbbaf5df-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Feb 02 09:59:49 crc kubenswrapper[4764]: I0202 09:59:49.042338 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mhfvj\" (UniqueName: \"kubernetes.io/projected/bd8ea4a3-c3ab-4f61-8257-40adfbbaf5df-kube-api-access-mhfvj\") on node \"crc\" DevicePath \"\"" Feb 02 09:59:49 crc kubenswrapper[4764]: I0202 09:59:49.263519 4764 generic.go:334] "Generic (PLEG): container finished" podID="bd8ea4a3-c3ab-4f61-8257-40adfbbaf5df" containerID="1522fd5feafe9818e7378965fcd9371ca91da359f83c68850824668df3553089" exitCode=137 Feb 02 09:59:49 crc kubenswrapper[4764]: I0202 09:59:49.263557 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6c8bdc8fbb-gkt26" event={"ID":"bd8ea4a3-c3ab-4f61-8257-40adfbbaf5df","Type":"ContainerDied","Data":"1522fd5feafe9818e7378965fcd9371ca91da359f83c68850824668df3553089"} Feb 02 09:59:49 crc kubenswrapper[4764]: I0202 09:59:49.263582 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6c8bdc8fbb-gkt26" event={"ID":"bd8ea4a3-c3ab-4f61-8257-40adfbbaf5df","Type":"ContainerDied","Data":"da638134da8211a36a9d8683d288afbf1a33da4f89967b1503e15dab9a594e74"} Feb 02 09:59:49 crc kubenswrapper[4764]: I0202 09:59:49.263599 4764 scope.go:117] "RemoveContainer" containerID="a9ba8889704978e8cdf3c04a2ed778ba56d8623240747403033dcaf0c56a278a" Feb 02 09:59:49 crc kubenswrapper[4764]: I0202 09:59:49.263707 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6c8bdc8fbb-gkt26" Feb 02 09:59:49 crc kubenswrapper[4764]: I0202 09:59:49.310517 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-6c8bdc8fbb-gkt26"] Feb 02 09:59:49 crc kubenswrapper[4764]: I0202 09:59:49.318066 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-6c8bdc8fbb-gkt26"] Feb 02 09:59:49 crc kubenswrapper[4764]: I0202 09:59:49.494248 4764 scope.go:117] "RemoveContainer" containerID="1522fd5feafe9818e7378965fcd9371ca91da359f83c68850824668df3553089" Feb 02 09:59:49 crc kubenswrapper[4764]: I0202 09:59:49.541776 4764 scope.go:117] "RemoveContainer" containerID="a9ba8889704978e8cdf3c04a2ed778ba56d8623240747403033dcaf0c56a278a" Feb 02 09:59:49 crc kubenswrapper[4764]: E0202 09:59:49.542324 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a9ba8889704978e8cdf3c04a2ed778ba56d8623240747403033dcaf0c56a278a\": container with ID starting with a9ba8889704978e8cdf3c04a2ed778ba56d8623240747403033dcaf0c56a278a not found: ID does not exist" containerID="a9ba8889704978e8cdf3c04a2ed778ba56d8623240747403033dcaf0c56a278a" Feb 02 09:59:49 crc kubenswrapper[4764]: I0202 09:59:49.542357 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a9ba8889704978e8cdf3c04a2ed778ba56d8623240747403033dcaf0c56a278a"} err="failed to get container status \"a9ba8889704978e8cdf3c04a2ed778ba56d8623240747403033dcaf0c56a278a\": rpc error: code = NotFound desc = could not find container \"a9ba8889704978e8cdf3c04a2ed778ba56d8623240747403033dcaf0c56a278a\": container with ID starting with a9ba8889704978e8cdf3c04a2ed778ba56d8623240747403033dcaf0c56a278a not found: ID does not exist" Feb 02 09:59:49 crc kubenswrapper[4764]: I0202 09:59:49.542377 4764 scope.go:117] "RemoveContainer" containerID="1522fd5feafe9818e7378965fcd9371ca91da359f83c68850824668df3553089" Feb 02 09:59:49 crc kubenswrapper[4764]: E0202 09:59:49.542749 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1522fd5feafe9818e7378965fcd9371ca91da359f83c68850824668df3553089\": container with ID starting with 1522fd5feafe9818e7378965fcd9371ca91da359f83c68850824668df3553089 not found: ID does not exist" containerID="1522fd5feafe9818e7378965fcd9371ca91da359f83c68850824668df3553089" Feb 02 09:59:49 crc kubenswrapper[4764]: I0202 09:59:49.542776 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1522fd5feafe9818e7378965fcd9371ca91da359f83c68850824668df3553089"} err="failed to get container status \"1522fd5feafe9818e7378965fcd9371ca91da359f83c68850824668df3553089\": rpc error: code = NotFound desc = could not find container \"1522fd5feafe9818e7378965fcd9371ca91da359f83c68850824668df3553089\": container with ID starting with 1522fd5feafe9818e7378965fcd9371ca91da359f83c68850824668df3553089 not found: ID does not exist" Feb 02 09:59:49 crc kubenswrapper[4764]: I0202 09:59:49.826171 4764 scope.go:117] "RemoveContainer" containerID="db4ebd4a8dbee216b70cb175f017f1799f3ca01a887f602ad42b9f295964da5d" Feb 02 09:59:49 crc kubenswrapper[4764]: E0202 09:59:49.826600 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 09:59:49 crc kubenswrapper[4764]: I0202 09:59:49.839299 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd8ea4a3-c3ab-4f61-8257-40adfbbaf5df" path="/var/lib/kubelet/pods/bd8ea4a3-c3ab-4f61-8257-40adfbbaf5df/volumes" Feb 02 09:59:53 crc kubenswrapper[4764]: I0202 09:59:53.843174 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-share-share1-0" Feb 02 10:00:00 crc kubenswrapper[4764]: I0202 10:00:00.159493 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29500440-55bml"] Feb 02 10:00:00 crc kubenswrapper[4764]: E0202 10:00:00.160417 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd8ea4a3-c3ab-4f61-8257-40adfbbaf5df" containerName="horizon" Feb 02 10:00:00 crc kubenswrapper[4764]: I0202 10:00:00.160430 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd8ea4a3-c3ab-4f61-8257-40adfbbaf5df" containerName="horizon" Feb 02 10:00:00 crc kubenswrapper[4764]: E0202 10:00:00.160444 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd8ea4a3-c3ab-4f61-8257-40adfbbaf5df" containerName="horizon-log" Feb 02 10:00:00 crc kubenswrapper[4764]: I0202 10:00:00.160450 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd8ea4a3-c3ab-4f61-8257-40adfbbaf5df" containerName="horizon-log" Feb 02 10:00:00 crc kubenswrapper[4764]: I0202 10:00:00.160622 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd8ea4a3-c3ab-4f61-8257-40adfbbaf5df" containerName="horizon" Feb 02 10:00:00 crc kubenswrapper[4764]: I0202 10:00:00.160641 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd8ea4a3-c3ab-4f61-8257-40adfbbaf5df" containerName="horizon-log" Feb 02 10:00:00 crc kubenswrapper[4764]: I0202 10:00:00.161209 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29500440-55bml" Feb 02 10:00:00 crc kubenswrapper[4764]: I0202 10:00:00.168091 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 02 10:00:00 crc kubenswrapper[4764]: I0202 10:00:00.168168 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 02 10:00:00 crc kubenswrapper[4764]: I0202 10:00:00.176473 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29500440-55bml"] Feb 02 10:00:00 crc kubenswrapper[4764]: I0202 10:00:00.186247 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sgbjh\" (UniqueName: \"kubernetes.io/projected/bfcf1702-589a-42fd-bf3a-64d313cac8ff-kube-api-access-sgbjh\") pod \"collect-profiles-29500440-55bml\" (UID: \"bfcf1702-589a-42fd-bf3a-64d313cac8ff\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500440-55bml" Feb 02 10:00:00 crc kubenswrapper[4764]: I0202 10:00:00.186395 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bfcf1702-589a-42fd-bf3a-64d313cac8ff-secret-volume\") pod \"collect-profiles-29500440-55bml\" (UID: \"bfcf1702-589a-42fd-bf3a-64d313cac8ff\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500440-55bml" Feb 02 10:00:00 crc kubenswrapper[4764]: I0202 10:00:00.186456 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bfcf1702-589a-42fd-bf3a-64d313cac8ff-config-volume\") pod \"collect-profiles-29500440-55bml\" (UID: \"bfcf1702-589a-42fd-bf3a-64d313cac8ff\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500440-55bml" Feb 02 10:00:00 crc kubenswrapper[4764]: I0202 10:00:00.288329 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bfcf1702-589a-42fd-bf3a-64d313cac8ff-secret-volume\") pod \"collect-profiles-29500440-55bml\" (UID: \"bfcf1702-589a-42fd-bf3a-64d313cac8ff\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500440-55bml" Feb 02 10:00:00 crc kubenswrapper[4764]: I0202 10:00:00.288385 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bfcf1702-589a-42fd-bf3a-64d313cac8ff-config-volume\") pod \"collect-profiles-29500440-55bml\" (UID: \"bfcf1702-589a-42fd-bf3a-64d313cac8ff\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500440-55bml" Feb 02 10:00:00 crc kubenswrapper[4764]: I0202 10:00:00.288470 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sgbjh\" (UniqueName: \"kubernetes.io/projected/bfcf1702-589a-42fd-bf3a-64d313cac8ff-kube-api-access-sgbjh\") pod \"collect-profiles-29500440-55bml\" (UID: \"bfcf1702-589a-42fd-bf3a-64d313cac8ff\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500440-55bml" Feb 02 10:00:00 crc kubenswrapper[4764]: I0202 10:00:00.289468 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bfcf1702-589a-42fd-bf3a-64d313cac8ff-config-volume\") pod \"collect-profiles-29500440-55bml\" (UID: \"bfcf1702-589a-42fd-bf3a-64d313cac8ff\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500440-55bml" Feb 02 10:00:00 crc kubenswrapper[4764]: I0202 10:00:00.306768 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bfcf1702-589a-42fd-bf3a-64d313cac8ff-secret-volume\") pod \"collect-profiles-29500440-55bml\" (UID: \"bfcf1702-589a-42fd-bf3a-64d313cac8ff\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500440-55bml" Feb 02 10:00:00 crc kubenswrapper[4764]: I0202 10:00:00.307383 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sgbjh\" (UniqueName: \"kubernetes.io/projected/bfcf1702-589a-42fd-bf3a-64d313cac8ff-kube-api-access-sgbjh\") pod \"collect-profiles-29500440-55bml\" (UID: \"bfcf1702-589a-42fd-bf3a-64d313cac8ff\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500440-55bml" Feb 02 10:00:00 crc kubenswrapper[4764]: I0202 10:00:00.486597 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29500440-55bml" Feb 02 10:00:00 crc kubenswrapper[4764]: I0202 10:00:00.968304 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29500440-55bml"] Feb 02 10:00:00 crc kubenswrapper[4764]: W0202 10:00:00.988599 4764 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbfcf1702_589a_42fd_bf3a_64d313cac8ff.slice/crio-9fa49f04e909962489ba0b8201b75be700931896fd1d627e9196673f466fb851 WatchSource:0}: Error finding container 9fa49f04e909962489ba0b8201b75be700931896fd1d627e9196673f466fb851: Status 404 returned error can't find the container with id 9fa49f04e909962489ba0b8201b75be700931896fd1d627e9196673f466fb851 Feb 02 10:00:01 crc kubenswrapper[4764]: I0202 10:00:01.404277 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29500440-55bml" event={"ID":"bfcf1702-589a-42fd-bf3a-64d313cac8ff","Type":"ContainerStarted","Data":"5ad577e62b253c766dad586a2afa5a2d2b881ede7b447d5c6b582c1e2b791476"} Feb 02 10:00:01 crc kubenswrapper[4764]: I0202 10:00:01.404641 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29500440-55bml" event={"ID":"bfcf1702-589a-42fd-bf3a-64d313cac8ff","Type":"ContainerStarted","Data":"9fa49f04e909962489ba0b8201b75be700931896fd1d627e9196673f466fb851"} Feb 02 10:00:01 crc kubenswrapper[4764]: I0202 10:00:01.825712 4764 scope.go:117] "RemoveContainer" containerID="db4ebd4a8dbee216b70cb175f017f1799f3ca01a887f602ad42b9f295964da5d" Feb 02 10:00:01 crc kubenswrapper[4764]: E0202 10:00:01.826234 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 10:00:02 crc kubenswrapper[4764]: I0202 10:00:02.416864 4764 generic.go:334] "Generic (PLEG): container finished" podID="bfcf1702-589a-42fd-bf3a-64d313cac8ff" containerID="5ad577e62b253c766dad586a2afa5a2d2b881ede7b447d5c6b582c1e2b791476" exitCode=0 Feb 02 10:00:02 crc kubenswrapper[4764]: I0202 10:00:02.416910 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29500440-55bml" event={"ID":"bfcf1702-589a-42fd-bf3a-64d313cac8ff","Type":"ContainerDied","Data":"5ad577e62b253c766dad586a2afa5a2d2b881ede7b447d5c6b582c1e2b791476"} Feb 02 10:00:03 crc kubenswrapper[4764]: I0202 10:00:03.744463 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29500440-55bml" Feb 02 10:00:03 crc kubenswrapper[4764]: I0202 10:00:03.849787 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bfcf1702-589a-42fd-bf3a-64d313cac8ff-config-volume\") pod \"bfcf1702-589a-42fd-bf3a-64d313cac8ff\" (UID: \"bfcf1702-589a-42fd-bf3a-64d313cac8ff\") " Feb 02 10:00:03 crc kubenswrapper[4764]: I0202 10:00:03.850216 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sgbjh\" (UniqueName: \"kubernetes.io/projected/bfcf1702-589a-42fd-bf3a-64d313cac8ff-kube-api-access-sgbjh\") pod \"bfcf1702-589a-42fd-bf3a-64d313cac8ff\" (UID: \"bfcf1702-589a-42fd-bf3a-64d313cac8ff\") " Feb 02 10:00:03 crc kubenswrapper[4764]: I0202 10:00:03.850495 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bfcf1702-589a-42fd-bf3a-64d313cac8ff-secret-volume\") pod \"bfcf1702-589a-42fd-bf3a-64d313cac8ff\" (UID: \"bfcf1702-589a-42fd-bf3a-64d313cac8ff\") " Feb 02 10:00:03 crc kubenswrapper[4764]: I0202 10:00:03.850581 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bfcf1702-589a-42fd-bf3a-64d313cac8ff-config-volume" (OuterVolumeSpecName: "config-volume") pod "bfcf1702-589a-42fd-bf3a-64d313cac8ff" (UID: "bfcf1702-589a-42fd-bf3a-64d313cac8ff"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 10:00:03 crc kubenswrapper[4764]: I0202 10:00:03.851627 4764 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bfcf1702-589a-42fd-bf3a-64d313cac8ff-config-volume\") on node \"crc\" DevicePath \"\"" Feb 02 10:00:03 crc kubenswrapper[4764]: I0202 10:00:03.857549 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bfcf1702-589a-42fd-bf3a-64d313cac8ff-kube-api-access-sgbjh" (OuterVolumeSpecName: "kube-api-access-sgbjh") pod "bfcf1702-589a-42fd-bf3a-64d313cac8ff" (UID: "bfcf1702-589a-42fd-bf3a-64d313cac8ff"). InnerVolumeSpecName "kube-api-access-sgbjh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 10:00:03 crc kubenswrapper[4764]: I0202 10:00:03.858694 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bfcf1702-589a-42fd-bf3a-64d313cac8ff-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "bfcf1702-589a-42fd-bf3a-64d313cac8ff" (UID: "bfcf1702-589a-42fd-bf3a-64d313cac8ff"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 10:00:03 crc kubenswrapper[4764]: I0202 10:00:03.953397 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sgbjh\" (UniqueName: \"kubernetes.io/projected/bfcf1702-589a-42fd-bf3a-64d313cac8ff-kube-api-access-sgbjh\") on node \"crc\" DevicePath \"\"" Feb 02 10:00:03 crc kubenswrapper[4764]: I0202 10:00:03.953434 4764 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bfcf1702-589a-42fd-bf3a-64d313cac8ff-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 02 10:00:04 crc kubenswrapper[4764]: I0202 10:00:04.439086 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29500440-55bml" event={"ID":"bfcf1702-589a-42fd-bf3a-64d313cac8ff","Type":"ContainerDied","Data":"9fa49f04e909962489ba0b8201b75be700931896fd1d627e9196673f466fb851"} Feb 02 10:00:04 crc kubenswrapper[4764]: I0202 10:00:04.439488 4764 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9fa49f04e909962489ba0b8201b75be700931896fd1d627e9196673f466fb851" Feb 02 10:00:04 crc kubenswrapper[4764]: I0202 10:00:04.439152 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29500440-55bml" Feb 02 10:00:04 crc kubenswrapper[4764]: I0202 10:00:04.517541 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29500395-v7kj2"] Feb 02 10:00:04 crc kubenswrapper[4764]: I0202 10:00:04.526635 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29500395-v7kj2"] Feb 02 10:00:05 crc kubenswrapper[4764]: I0202 10:00:05.847303 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="60049dcd-be0a-4eda-800a-1ce0c6da3195" path="/var/lib/kubelet/pods/60049dcd-be0a-4eda-800a-1ce0c6da3195/volumes" Feb 02 10:00:12 crc kubenswrapper[4764]: I0202 10:00:12.826015 4764 scope.go:117] "RemoveContainer" containerID="db4ebd4a8dbee216b70cb175f017f1799f3ca01a887f602ad42b9f295964da5d" Feb 02 10:00:12 crc kubenswrapper[4764]: E0202 10:00:12.826744 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 10:00:24 crc kubenswrapper[4764]: E0202 10:00:24.558897 4764 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.217:46276->38.102.83.217:34003: write tcp 38.102.83.217:46276->38.102.83.217:34003: write: broken pipe Feb 02 10:00:25 crc kubenswrapper[4764]: I0202 10:00:25.826282 4764 scope.go:117] "RemoveContainer" containerID="db4ebd4a8dbee216b70cb175f017f1799f3ca01a887f602ad42b9f295964da5d" Feb 02 10:00:25 crc kubenswrapper[4764]: E0202 10:00:25.826857 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 10:00:28 crc kubenswrapper[4764]: I0202 10:00:28.217054 4764 scope.go:117] "RemoveContainer" containerID="a86e06b425266081de6db5fd5c39780eeb69d1c8f9406a1f104bd10689e15b82" Feb 02 10:00:37 crc kubenswrapper[4764]: I0202 10:00:37.835646 4764 scope.go:117] "RemoveContainer" containerID="db4ebd4a8dbee216b70cb175f017f1799f3ca01a887f602ad42b9f295964da5d" Feb 02 10:00:37 crc kubenswrapper[4764]: E0202 10:00:37.836637 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 10:00:44 crc kubenswrapper[4764]: I0202 10:00:44.310167 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Feb 02 10:00:44 crc kubenswrapper[4764]: E0202 10:00:44.311149 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bfcf1702-589a-42fd-bf3a-64d313cac8ff" containerName="collect-profiles" Feb 02 10:00:44 crc kubenswrapper[4764]: I0202 10:00:44.311168 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="bfcf1702-589a-42fd-bf3a-64d313cac8ff" containerName="collect-profiles" Feb 02 10:00:44 crc kubenswrapper[4764]: I0202 10:00:44.311410 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="bfcf1702-589a-42fd-bf3a-64d313cac8ff" containerName="collect-profiles" Feb 02 10:00:44 crc kubenswrapper[4764]: I0202 10:00:44.312317 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Feb 02 10:00:44 crc kubenswrapper[4764]: I0202 10:00:44.317802 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Feb 02 10:00:44 crc kubenswrapper[4764]: I0202 10:00:44.317898 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Feb 02 10:00:44 crc kubenswrapper[4764]: I0202 10:00:44.318094 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Feb 02 10:00:44 crc kubenswrapper[4764]: I0202 10:00:44.319054 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-kgk2v" Feb 02 10:00:44 crc kubenswrapper[4764]: I0202 10:00:44.333492 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Feb 02 10:00:44 crc kubenswrapper[4764]: I0202 10:00:44.478719 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/09adc919-28cf-4d2d-b40d-183a2f51f06f-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"09adc919-28cf-4d2d-b40d-183a2f51f06f\") " pod="openstack/tempest-tests-tempest" Feb 02 10:00:44 crc kubenswrapper[4764]: I0202 10:00:44.479018 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/09adc919-28cf-4d2d-b40d-183a2f51f06f-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"09adc919-28cf-4d2d-b40d-183a2f51f06f\") " pod="openstack/tempest-tests-tempest" Feb 02 10:00:44 crc kubenswrapper[4764]: I0202 10:00:44.479161 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/09adc919-28cf-4d2d-b40d-183a2f51f06f-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"09adc919-28cf-4d2d-b40d-183a2f51f06f\") " pod="openstack/tempest-tests-tempest" Feb 02 10:00:44 crc kubenswrapper[4764]: I0202 10:00:44.479250 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mjdsp\" (UniqueName: \"kubernetes.io/projected/09adc919-28cf-4d2d-b40d-183a2f51f06f-kube-api-access-mjdsp\") pod \"tempest-tests-tempest\" (UID: \"09adc919-28cf-4d2d-b40d-183a2f51f06f\") " pod="openstack/tempest-tests-tempest" Feb 02 10:00:44 crc kubenswrapper[4764]: I0202 10:00:44.479343 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"tempest-tests-tempest\" (UID: \"09adc919-28cf-4d2d-b40d-183a2f51f06f\") " pod="openstack/tempest-tests-tempest" Feb 02 10:00:44 crc kubenswrapper[4764]: I0202 10:00:44.479447 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/09adc919-28cf-4d2d-b40d-183a2f51f06f-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"09adc919-28cf-4d2d-b40d-183a2f51f06f\") " pod="openstack/tempest-tests-tempest" Feb 02 10:00:44 crc kubenswrapper[4764]: I0202 10:00:44.479552 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/09adc919-28cf-4d2d-b40d-183a2f51f06f-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"09adc919-28cf-4d2d-b40d-183a2f51f06f\") " pod="openstack/tempest-tests-tempest" Feb 02 10:00:44 crc kubenswrapper[4764]: I0202 10:00:44.479623 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/09adc919-28cf-4d2d-b40d-183a2f51f06f-config-data\") pod \"tempest-tests-tempest\" (UID: \"09adc919-28cf-4d2d-b40d-183a2f51f06f\") " pod="openstack/tempest-tests-tempest" Feb 02 10:00:44 crc kubenswrapper[4764]: I0202 10:00:44.479723 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/09adc919-28cf-4d2d-b40d-183a2f51f06f-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"09adc919-28cf-4d2d-b40d-183a2f51f06f\") " pod="openstack/tempest-tests-tempest" Feb 02 10:00:44 crc kubenswrapper[4764]: I0202 10:00:44.581853 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/09adc919-28cf-4d2d-b40d-183a2f51f06f-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"09adc919-28cf-4d2d-b40d-183a2f51f06f\") " pod="openstack/tempest-tests-tempest" Feb 02 10:00:44 crc kubenswrapper[4764]: I0202 10:00:44.582178 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/09adc919-28cf-4d2d-b40d-183a2f51f06f-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"09adc919-28cf-4d2d-b40d-183a2f51f06f\") " pod="openstack/tempest-tests-tempest" Feb 02 10:00:44 crc kubenswrapper[4764]: I0202 10:00:44.582338 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/09adc919-28cf-4d2d-b40d-183a2f51f06f-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"09adc919-28cf-4d2d-b40d-183a2f51f06f\") " pod="openstack/tempest-tests-tempest" Feb 02 10:00:44 crc kubenswrapper[4764]: I0202 10:00:44.582458 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mjdsp\" (UniqueName: \"kubernetes.io/projected/09adc919-28cf-4d2d-b40d-183a2f51f06f-kube-api-access-mjdsp\") pod \"tempest-tests-tempest\" (UID: \"09adc919-28cf-4d2d-b40d-183a2f51f06f\") " pod="openstack/tempest-tests-tempest" Feb 02 10:00:44 crc kubenswrapper[4764]: I0202 10:00:44.582573 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"tempest-tests-tempest\" (UID: \"09adc919-28cf-4d2d-b40d-183a2f51f06f\") " pod="openstack/tempest-tests-tempest" Feb 02 10:00:44 crc kubenswrapper[4764]: I0202 10:00:44.582713 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/09adc919-28cf-4d2d-b40d-183a2f51f06f-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"09adc919-28cf-4d2d-b40d-183a2f51f06f\") " pod="openstack/tempest-tests-tempest" Feb 02 10:00:44 crc kubenswrapper[4764]: I0202 10:00:44.582846 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/09adc919-28cf-4d2d-b40d-183a2f51f06f-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"09adc919-28cf-4d2d-b40d-183a2f51f06f\") " pod="openstack/tempest-tests-tempest" Feb 02 10:00:44 crc kubenswrapper[4764]: I0202 10:00:44.582964 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/09adc919-28cf-4d2d-b40d-183a2f51f06f-config-data\") pod \"tempest-tests-tempest\" (UID: \"09adc919-28cf-4d2d-b40d-183a2f51f06f\") " pod="openstack/tempest-tests-tempest" Feb 02 10:00:44 crc kubenswrapper[4764]: I0202 10:00:44.583099 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/09adc919-28cf-4d2d-b40d-183a2f51f06f-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"09adc919-28cf-4d2d-b40d-183a2f51f06f\") " pod="openstack/tempest-tests-tempest" Feb 02 10:00:44 crc kubenswrapper[4764]: I0202 10:00:44.584901 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/09adc919-28cf-4d2d-b40d-183a2f51f06f-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"09adc919-28cf-4d2d-b40d-183a2f51f06f\") " pod="openstack/tempest-tests-tempest" Feb 02 10:00:44 crc kubenswrapper[4764]: I0202 10:00:44.585348 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/09adc919-28cf-4d2d-b40d-183a2f51f06f-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"09adc919-28cf-4d2d-b40d-183a2f51f06f\") " pod="openstack/tempest-tests-tempest" Feb 02 10:00:44 crc kubenswrapper[4764]: I0202 10:00:44.586566 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/09adc919-28cf-4d2d-b40d-183a2f51f06f-config-data\") pod \"tempest-tests-tempest\" (UID: \"09adc919-28cf-4d2d-b40d-183a2f51f06f\") " pod="openstack/tempest-tests-tempest" Feb 02 10:00:44 crc kubenswrapper[4764]: I0202 10:00:44.586767 4764 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"tempest-tests-tempest\" (UID: \"09adc919-28cf-4d2d-b40d-183a2f51f06f\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/tempest-tests-tempest" Feb 02 10:00:44 crc kubenswrapper[4764]: I0202 10:00:44.587196 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/09adc919-28cf-4d2d-b40d-183a2f51f06f-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"09adc919-28cf-4d2d-b40d-183a2f51f06f\") " pod="openstack/tempest-tests-tempest" Feb 02 10:00:44 crc kubenswrapper[4764]: I0202 10:00:44.590595 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/09adc919-28cf-4d2d-b40d-183a2f51f06f-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"09adc919-28cf-4d2d-b40d-183a2f51f06f\") " pod="openstack/tempest-tests-tempest" Feb 02 10:00:44 crc kubenswrapper[4764]: I0202 10:00:44.592090 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/09adc919-28cf-4d2d-b40d-183a2f51f06f-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"09adc919-28cf-4d2d-b40d-183a2f51f06f\") " pod="openstack/tempest-tests-tempest" Feb 02 10:00:44 crc kubenswrapper[4764]: I0202 10:00:44.592682 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/09adc919-28cf-4d2d-b40d-183a2f51f06f-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"09adc919-28cf-4d2d-b40d-183a2f51f06f\") " pod="openstack/tempest-tests-tempest" Feb 02 10:00:44 crc kubenswrapper[4764]: I0202 10:00:44.603855 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mjdsp\" (UniqueName: \"kubernetes.io/projected/09adc919-28cf-4d2d-b40d-183a2f51f06f-kube-api-access-mjdsp\") pod \"tempest-tests-tempest\" (UID: \"09adc919-28cf-4d2d-b40d-183a2f51f06f\") " pod="openstack/tempest-tests-tempest" Feb 02 10:00:44 crc kubenswrapper[4764]: I0202 10:00:44.624102 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"tempest-tests-tempest\" (UID: \"09adc919-28cf-4d2d-b40d-183a2f51f06f\") " pod="openstack/tempest-tests-tempest" Feb 02 10:00:44 crc kubenswrapper[4764]: I0202 10:00:44.663578 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Feb 02 10:00:45 crc kubenswrapper[4764]: I0202 10:00:45.113079 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Feb 02 10:00:45 crc kubenswrapper[4764]: I0202 10:00:45.859485 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"09adc919-28cf-4d2d-b40d-183a2f51f06f","Type":"ContainerStarted","Data":"d58a3a44bda779d0463c24ceb60684b3b9fe76a68689ebe918bb60483db8fc8a"} Feb 02 10:00:52 crc kubenswrapper[4764]: I0202 10:00:52.826206 4764 scope.go:117] "RemoveContainer" containerID="db4ebd4a8dbee216b70cb175f017f1799f3ca01a887f602ad42b9f295964da5d" Feb 02 10:00:52 crc kubenswrapper[4764]: E0202 10:00:52.827023 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 10:01:00 crc kubenswrapper[4764]: I0202 10:01:00.145149 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29500441-qpcvv"] Feb 02 10:01:00 crc kubenswrapper[4764]: I0202 10:01:00.147514 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29500441-qpcvv" Feb 02 10:01:00 crc kubenswrapper[4764]: I0202 10:01:00.155768 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29500441-qpcvv"] Feb 02 10:01:00 crc kubenswrapper[4764]: I0202 10:01:00.235030 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b77a0084-b391-4221-86d1-39629566a940-fernet-keys\") pod \"keystone-cron-29500441-qpcvv\" (UID: \"b77a0084-b391-4221-86d1-39629566a940\") " pod="openstack/keystone-cron-29500441-qpcvv" Feb 02 10:01:00 crc kubenswrapper[4764]: I0202 10:01:00.235159 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b77a0084-b391-4221-86d1-39629566a940-config-data\") pod \"keystone-cron-29500441-qpcvv\" (UID: \"b77a0084-b391-4221-86d1-39629566a940\") " pod="openstack/keystone-cron-29500441-qpcvv" Feb 02 10:01:00 crc kubenswrapper[4764]: I0202 10:01:00.235266 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7jg5b\" (UniqueName: \"kubernetes.io/projected/b77a0084-b391-4221-86d1-39629566a940-kube-api-access-7jg5b\") pod \"keystone-cron-29500441-qpcvv\" (UID: \"b77a0084-b391-4221-86d1-39629566a940\") " pod="openstack/keystone-cron-29500441-qpcvv" Feb 02 10:01:00 crc kubenswrapper[4764]: I0202 10:01:00.235429 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b77a0084-b391-4221-86d1-39629566a940-combined-ca-bundle\") pod \"keystone-cron-29500441-qpcvv\" (UID: \"b77a0084-b391-4221-86d1-39629566a940\") " pod="openstack/keystone-cron-29500441-qpcvv" Feb 02 10:01:00 crc kubenswrapper[4764]: I0202 10:01:00.337337 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b77a0084-b391-4221-86d1-39629566a940-combined-ca-bundle\") pod \"keystone-cron-29500441-qpcvv\" (UID: \"b77a0084-b391-4221-86d1-39629566a940\") " pod="openstack/keystone-cron-29500441-qpcvv" Feb 02 10:01:00 crc kubenswrapper[4764]: I0202 10:01:00.337605 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b77a0084-b391-4221-86d1-39629566a940-fernet-keys\") pod \"keystone-cron-29500441-qpcvv\" (UID: \"b77a0084-b391-4221-86d1-39629566a940\") " pod="openstack/keystone-cron-29500441-qpcvv" Feb 02 10:01:00 crc kubenswrapper[4764]: I0202 10:01:00.337631 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b77a0084-b391-4221-86d1-39629566a940-config-data\") pod \"keystone-cron-29500441-qpcvv\" (UID: \"b77a0084-b391-4221-86d1-39629566a940\") " pod="openstack/keystone-cron-29500441-qpcvv" Feb 02 10:01:00 crc kubenswrapper[4764]: I0202 10:01:00.337680 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7jg5b\" (UniqueName: \"kubernetes.io/projected/b77a0084-b391-4221-86d1-39629566a940-kube-api-access-7jg5b\") pod \"keystone-cron-29500441-qpcvv\" (UID: \"b77a0084-b391-4221-86d1-39629566a940\") " pod="openstack/keystone-cron-29500441-qpcvv" Feb 02 10:01:00 crc kubenswrapper[4764]: I0202 10:01:00.343030 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b77a0084-b391-4221-86d1-39629566a940-combined-ca-bundle\") pod \"keystone-cron-29500441-qpcvv\" (UID: \"b77a0084-b391-4221-86d1-39629566a940\") " pod="openstack/keystone-cron-29500441-qpcvv" Feb 02 10:01:00 crc kubenswrapper[4764]: I0202 10:01:00.343303 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b77a0084-b391-4221-86d1-39629566a940-fernet-keys\") pod \"keystone-cron-29500441-qpcvv\" (UID: \"b77a0084-b391-4221-86d1-39629566a940\") " pod="openstack/keystone-cron-29500441-qpcvv" Feb 02 10:01:00 crc kubenswrapper[4764]: I0202 10:01:00.353559 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b77a0084-b391-4221-86d1-39629566a940-config-data\") pod \"keystone-cron-29500441-qpcvv\" (UID: \"b77a0084-b391-4221-86d1-39629566a940\") " pod="openstack/keystone-cron-29500441-qpcvv" Feb 02 10:01:00 crc kubenswrapper[4764]: I0202 10:01:00.353840 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7jg5b\" (UniqueName: \"kubernetes.io/projected/b77a0084-b391-4221-86d1-39629566a940-kube-api-access-7jg5b\") pod \"keystone-cron-29500441-qpcvv\" (UID: \"b77a0084-b391-4221-86d1-39629566a940\") " pod="openstack/keystone-cron-29500441-qpcvv" Feb 02 10:01:00 crc kubenswrapper[4764]: I0202 10:01:00.478898 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29500441-qpcvv" Feb 02 10:01:03 crc kubenswrapper[4764]: I0202 10:01:03.827055 4764 scope.go:117] "RemoveContainer" containerID="db4ebd4a8dbee216b70cb175f017f1799f3ca01a887f602ad42b9f295964da5d" Feb 02 10:01:03 crc kubenswrapper[4764]: E0202 10:01:03.829122 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 10:01:17 crc kubenswrapper[4764]: I0202 10:01:17.848072 4764 scope.go:117] "RemoveContainer" containerID="db4ebd4a8dbee216b70cb175f017f1799f3ca01a887f602ad42b9f295964da5d" Feb 02 10:01:18 crc kubenswrapper[4764]: E0202 10:01:18.809851 4764 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified" Feb 02 10:01:18 crc kubenswrapper[4764]: E0202 10:01:18.814326 4764 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-mjdsp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest_openstack(09adc919-28cf-4d2d-b40d-183a2f51f06f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 02 10:01:18 crc kubenswrapper[4764]: E0202 10:01:18.815642 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest" podUID="09adc919-28cf-4d2d-b40d-183a2f51f06f" Feb 02 10:01:19 crc kubenswrapper[4764]: I0202 10:01:19.199827 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" event={"ID":"d192f670-9f9d-4539-9641-e4bed73acdd4","Type":"ContainerStarted","Data":"86ea8d894bee98a50d949cdf649a606039569f7d321c8baff9abb00746af44d2"} Feb 02 10:01:19 crc kubenswrapper[4764]: E0202 10:01:19.201444 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified\\\"\"" pod="openstack/tempest-tests-tempest" podUID="09adc919-28cf-4d2d-b40d-183a2f51f06f" Feb 02 10:01:19 crc kubenswrapper[4764]: I0202 10:01:19.262187 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29500441-qpcvv"] Feb 02 10:01:20 crc kubenswrapper[4764]: I0202 10:01:20.214087 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29500441-qpcvv" event={"ID":"b77a0084-b391-4221-86d1-39629566a940","Type":"ContainerStarted","Data":"b130aa190a38dd850e326142e8218bfd04771adb34a924c631a40ffbc96bdabe"} Feb 02 10:01:20 crc kubenswrapper[4764]: I0202 10:01:20.214883 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29500441-qpcvv" event={"ID":"b77a0084-b391-4221-86d1-39629566a940","Type":"ContainerStarted","Data":"4b17b5153dd8d35d6f321b269fad4fcae0f9a9c5586ddfec8cf30910287d9832"} Feb 02 10:01:20 crc kubenswrapper[4764]: I0202 10:01:20.243532 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29500441-qpcvv" podStartSLOduration=20.24351023 podStartE2EDuration="20.24351023s" podCreationTimestamp="2026-02-02 10:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 10:01:20.233642387 +0000 UTC m=+3243.167366495" watchObservedRunningTime="2026-02-02 10:01:20.24351023 +0000 UTC m=+3243.177234328" Feb 02 10:01:23 crc kubenswrapper[4764]: I0202 10:01:23.245364 4764 generic.go:334] "Generic (PLEG): container finished" podID="b77a0084-b391-4221-86d1-39629566a940" containerID="b130aa190a38dd850e326142e8218bfd04771adb34a924c631a40ffbc96bdabe" exitCode=0 Feb 02 10:01:23 crc kubenswrapper[4764]: I0202 10:01:23.245477 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29500441-qpcvv" event={"ID":"b77a0084-b391-4221-86d1-39629566a940","Type":"ContainerDied","Data":"b130aa190a38dd850e326142e8218bfd04771adb34a924c631a40ffbc96bdabe"} Feb 02 10:01:24 crc kubenswrapper[4764]: I0202 10:01:24.685973 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29500441-qpcvv" Feb 02 10:01:24 crc kubenswrapper[4764]: I0202 10:01:24.693784 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b77a0084-b391-4221-86d1-39629566a940-fernet-keys\") pod \"b77a0084-b391-4221-86d1-39629566a940\" (UID: \"b77a0084-b391-4221-86d1-39629566a940\") " Feb 02 10:01:24 crc kubenswrapper[4764]: I0202 10:01:24.694096 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b77a0084-b391-4221-86d1-39629566a940-config-data\") pod \"b77a0084-b391-4221-86d1-39629566a940\" (UID: \"b77a0084-b391-4221-86d1-39629566a940\") " Feb 02 10:01:24 crc kubenswrapper[4764]: I0202 10:01:24.694310 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b77a0084-b391-4221-86d1-39629566a940-combined-ca-bundle\") pod \"b77a0084-b391-4221-86d1-39629566a940\" (UID: \"b77a0084-b391-4221-86d1-39629566a940\") " Feb 02 10:01:24 crc kubenswrapper[4764]: I0202 10:01:24.694677 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7jg5b\" (UniqueName: \"kubernetes.io/projected/b77a0084-b391-4221-86d1-39629566a940-kube-api-access-7jg5b\") pod \"b77a0084-b391-4221-86d1-39629566a940\" (UID: \"b77a0084-b391-4221-86d1-39629566a940\") " Feb 02 10:01:24 crc kubenswrapper[4764]: I0202 10:01:24.739753 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b77a0084-b391-4221-86d1-39629566a940-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "b77a0084-b391-4221-86d1-39629566a940" (UID: "b77a0084-b391-4221-86d1-39629566a940"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 10:01:24 crc kubenswrapper[4764]: I0202 10:01:24.750178 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b77a0084-b391-4221-86d1-39629566a940-kube-api-access-7jg5b" (OuterVolumeSpecName: "kube-api-access-7jg5b") pod "b77a0084-b391-4221-86d1-39629566a940" (UID: "b77a0084-b391-4221-86d1-39629566a940"). InnerVolumeSpecName "kube-api-access-7jg5b". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 10:01:24 crc kubenswrapper[4764]: I0202 10:01:24.814872 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b77a0084-b391-4221-86d1-39629566a940-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b77a0084-b391-4221-86d1-39629566a940" (UID: "b77a0084-b391-4221-86d1-39629566a940"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 10:01:24 crc kubenswrapper[4764]: I0202 10:01:24.815393 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b77a0084-b391-4221-86d1-39629566a940-config-data" (OuterVolumeSpecName: "config-data") pod "b77a0084-b391-4221-86d1-39629566a940" (UID: "b77a0084-b391-4221-86d1-39629566a940"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 10:01:24 crc kubenswrapper[4764]: I0202 10:01:24.817619 4764 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b77a0084-b391-4221-86d1-39629566a940-fernet-keys\") on node \"crc\" DevicePath \"\"" Feb 02 10:01:24 crc kubenswrapper[4764]: I0202 10:01:24.817644 4764 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b77a0084-b391-4221-86d1-39629566a940-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 10:01:24 crc kubenswrapper[4764]: I0202 10:01:24.817653 4764 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b77a0084-b391-4221-86d1-39629566a940-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 10:01:24 crc kubenswrapper[4764]: I0202 10:01:24.817665 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7jg5b\" (UniqueName: \"kubernetes.io/projected/b77a0084-b391-4221-86d1-39629566a940-kube-api-access-7jg5b\") on node \"crc\" DevicePath \"\"" Feb 02 10:01:25 crc kubenswrapper[4764]: I0202 10:01:25.272822 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29500441-qpcvv" event={"ID":"b77a0084-b391-4221-86d1-39629566a940","Type":"ContainerDied","Data":"4b17b5153dd8d35d6f321b269fad4fcae0f9a9c5586ddfec8cf30910287d9832"} Feb 02 10:01:25 crc kubenswrapper[4764]: I0202 10:01:25.272865 4764 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4b17b5153dd8d35d6f321b269fad4fcae0f9a9c5586ddfec8cf30910287d9832" Feb 02 10:01:25 crc kubenswrapper[4764]: I0202 10:01:25.272963 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29500441-qpcvv" Feb 02 10:01:29 crc kubenswrapper[4764]: I0202 10:01:29.831282 4764 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 02 10:01:30 crc kubenswrapper[4764]: I0202 10:01:30.266874 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Feb 02 10:01:31 crc kubenswrapper[4764]: I0202 10:01:31.368765 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"09adc919-28cf-4d2d-b40d-183a2f51f06f","Type":"ContainerStarted","Data":"198eecad7865982369d75b72c24fc8ba50d3f10a9f576c3f0ebabf9a1cb73e22"} Feb 02 10:01:31 crc kubenswrapper[4764]: I0202 10:01:31.397730 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=3.2423030170000002 podStartE2EDuration="48.397710269s" podCreationTimestamp="2026-02-02 10:00:43 +0000 UTC" firstStartedPulling="2026-02-02 10:00:45.106961257 +0000 UTC m=+3208.040685345" lastFinishedPulling="2026-02-02 10:01:30.262368469 +0000 UTC m=+3253.196092597" observedRunningTime="2026-02-02 10:01:31.388419342 +0000 UTC m=+3254.322143480" watchObservedRunningTime="2026-02-02 10:01:31.397710269 +0000 UTC m=+3254.331434367" Feb 02 10:01:43 crc kubenswrapper[4764]: I0202 10:01:43.297858 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-2c5nd"] Feb 02 10:01:43 crc kubenswrapper[4764]: E0202 10:01:43.305271 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b77a0084-b391-4221-86d1-39629566a940" containerName="keystone-cron" Feb 02 10:01:43 crc kubenswrapper[4764]: I0202 10:01:43.305310 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="b77a0084-b391-4221-86d1-39629566a940" containerName="keystone-cron" Feb 02 10:01:43 crc kubenswrapper[4764]: I0202 10:01:43.305547 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="b77a0084-b391-4221-86d1-39629566a940" containerName="keystone-cron" Feb 02 10:01:43 crc kubenswrapper[4764]: I0202 10:01:43.312409 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2c5nd" Feb 02 10:01:43 crc kubenswrapper[4764]: I0202 10:01:43.341112 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/89fb05bd-ea44-4d00-b4ea-0fd4c5190995-utilities\") pod \"community-operators-2c5nd\" (UID: \"89fb05bd-ea44-4d00-b4ea-0fd4c5190995\") " pod="openshift-marketplace/community-operators-2c5nd" Feb 02 10:01:43 crc kubenswrapper[4764]: I0202 10:01:43.341897 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wmvqp\" (UniqueName: \"kubernetes.io/projected/89fb05bd-ea44-4d00-b4ea-0fd4c5190995-kube-api-access-wmvqp\") pod \"community-operators-2c5nd\" (UID: \"89fb05bd-ea44-4d00-b4ea-0fd4c5190995\") " pod="openshift-marketplace/community-operators-2c5nd" Feb 02 10:01:43 crc kubenswrapper[4764]: I0202 10:01:43.342070 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/89fb05bd-ea44-4d00-b4ea-0fd4c5190995-catalog-content\") pod \"community-operators-2c5nd\" (UID: \"89fb05bd-ea44-4d00-b4ea-0fd4c5190995\") " pod="openshift-marketplace/community-operators-2c5nd" Feb 02 10:01:43 crc kubenswrapper[4764]: I0202 10:01:43.356003 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-2c5nd"] Feb 02 10:01:43 crc kubenswrapper[4764]: I0202 10:01:43.444701 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/89fb05bd-ea44-4d00-b4ea-0fd4c5190995-utilities\") pod \"community-operators-2c5nd\" (UID: \"89fb05bd-ea44-4d00-b4ea-0fd4c5190995\") " pod="openshift-marketplace/community-operators-2c5nd" Feb 02 10:01:43 crc kubenswrapper[4764]: I0202 10:01:43.444759 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wmvqp\" (UniqueName: \"kubernetes.io/projected/89fb05bd-ea44-4d00-b4ea-0fd4c5190995-kube-api-access-wmvqp\") pod \"community-operators-2c5nd\" (UID: \"89fb05bd-ea44-4d00-b4ea-0fd4c5190995\") " pod="openshift-marketplace/community-operators-2c5nd" Feb 02 10:01:43 crc kubenswrapper[4764]: I0202 10:01:43.444833 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/89fb05bd-ea44-4d00-b4ea-0fd4c5190995-catalog-content\") pod \"community-operators-2c5nd\" (UID: \"89fb05bd-ea44-4d00-b4ea-0fd4c5190995\") " pod="openshift-marketplace/community-operators-2c5nd" Feb 02 10:01:43 crc kubenswrapper[4764]: I0202 10:01:43.445339 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/89fb05bd-ea44-4d00-b4ea-0fd4c5190995-catalog-content\") pod \"community-operators-2c5nd\" (UID: \"89fb05bd-ea44-4d00-b4ea-0fd4c5190995\") " pod="openshift-marketplace/community-operators-2c5nd" Feb 02 10:01:43 crc kubenswrapper[4764]: I0202 10:01:43.445339 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/89fb05bd-ea44-4d00-b4ea-0fd4c5190995-utilities\") pod \"community-operators-2c5nd\" (UID: \"89fb05bd-ea44-4d00-b4ea-0fd4c5190995\") " pod="openshift-marketplace/community-operators-2c5nd" Feb 02 10:01:43 crc kubenswrapper[4764]: I0202 10:01:43.480453 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wmvqp\" (UniqueName: \"kubernetes.io/projected/89fb05bd-ea44-4d00-b4ea-0fd4c5190995-kube-api-access-wmvqp\") pod \"community-operators-2c5nd\" (UID: \"89fb05bd-ea44-4d00-b4ea-0fd4c5190995\") " pod="openshift-marketplace/community-operators-2c5nd" Feb 02 10:01:43 crc kubenswrapper[4764]: I0202 10:01:43.658475 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2c5nd" Feb 02 10:01:44 crc kubenswrapper[4764]: W0202 10:01:44.106963 4764 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod89fb05bd_ea44_4d00_b4ea_0fd4c5190995.slice/crio-c98e4dc69e44a69ed1cd056926702d1bb6a5cdabe165a44816eb45f39dcf5735 WatchSource:0}: Error finding container c98e4dc69e44a69ed1cd056926702d1bb6a5cdabe165a44816eb45f39dcf5735: Status 404 returned error can't find the container with id c98e4dc69e44a69ed1cd056926702d1bb6a5cdabe165a44816eb45f39dcf5735 Feb 02 10:01:44 crc kubenswrapper[4764]: I0202 10:01:44.111257 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-2c5nd"] Feb 02 10:01:44 crc kubenswrapper[4764]: I0202 10:01:44.516456 4764 generic.go:334] "Generic (PLEG): container finished" podID="89fb05bd-ea44-4d00-b4ea-0fd4c5190995" containerID="6d3930fe876570910583815f6ae3dd4cdb935da0ca07ca0e8cdb5bc9166906fc" exitCode=0 Feb 02 10:01:44 crc kubenswrapper[4764]: I0202 10:01:44.516502 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2c5nd" event={"ID":"89fb05bd-ea44-4d00-b4ea-0fd4c5190995","Type":"ContainerDied","Data":"6d3930fe876570910583815f6ae3dd4cdb935da0ca07ca0e8cdb5bc9166906fc"} Feb 02 10:01:44 crc kubenswrapper[4764]: I0202 10:01:44.516545 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2c5nd" event={"ID":"89fb05bd-ea44-4d00-b4ea-0fd4c5190995","Type":"ContainerStarted","Data":"c98e4dc69e44a69ed1cd056926702d1bb6a5cdabe165a44816eb45f39dcf5735"} Feb 02 10:01:46 crc kubenswrapper[4764]: I0202 10:01:46.536375 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2c5nd" event={"ID":"89fb05bd-ea44-4d00-b4ea-0fd4c5190995","Type":"ContainerStarted","Data":"13f89b0d61381bd77b2569a5c330825b07d424f09bc8e1d457d211872dd41418"} Feb 02 10:01:47 crc kubenswrapper[4764]: I0202 10:01:47.551001 4764 generic.go:334] "Generic (PLEG): container finished" podID="89fb05bd-ea44-4d00-b4ea-0fd4c5190995" containerID="13f89b0d61381bd77b2569a5c330825b07d424f09bc8e1d457d211872dd41418" exitCode=0 Feb 02 10:01:47 crc kubenswrapper[4764]: I0202 10:01:47.551066 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2c5nd" event={"ID":"89fb05bd-ea44-4d00-b4ea-0fd4c5190995","Type":"ContainerDied","Data":"13f89b0d61381bd77b2569a5c330825b07d424f09bc8e1d457d211872dd41418"} Feb 02 10:01:48 crc kubenswrapper[4764]: I0202 10:01:48.563210 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2c5nd" event={"ID":"89fb05bd-ea44-4d00-b4ea-0fd4c5190995","Type":"ContainerStarted","Data":"7f150dd6cb0f8a4705a2261b68b7b62943edecef67ea5c57465891cda565b388"} Feb 02 10:01:48 crc kubenswrapper[4764]: I0202 10:01:48.594579 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-2c5nd" podStartSLOduration=2.160589898 podStartE2EDuration="5.594555111s" podCreationTimestamp="2026-02-02 10:01:43 +0000 UTC" firstStartedPulling="2026-02-02 10:01:44.518716009 +0000 UTC m=+3267.452440107" lastFinishedPulling="2026-02-02 10:01:47.952681192 +0000 UTC m=+3270.886405320" observedRunningTime="2026-02-02 10:01:48.583035625 +0000 UTC m=+3271.516759723" watchObservedRunningTime="2026-02-02 10:01:48.594555111 +0000 UTC m=+3271.528279209" Feb 02 10:01:53 crc kubenswrapper[4764]: I0202 10:01:53.659386 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-2c5nd" Feb 02 10:01:53 crc kubenswrapper[4764]: I0202 10:01:53.659916 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-2c5nd" Feb 02 10:01:54 crc kubenswrapper[4764]: I0202 10:01:54.712343 4764 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-2c5nd" podUID="89fb05bd-ea44-4d00-b4ea-0fd4c5190995" containerName="registry-server" probeResult="failure" output=< Feb 02 10:01:54 crc kubenswrapper[4764]: timeout: failed to connect service ":50051" within 1s Feb 02 10:01:54 crc kubenswrapper[4764]: > Feb 02 10:02:03 crc kubenswrapper[4764]: I0202 10:02:03.742482 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-2c5nd" Feb 02 10:02:03 crc kubenswrapper[4764]: I0202 10:02:03.811348 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-2c5nd" Feb 02 10:02:03 crc kubenswrapper[4764]: I0202 10:02:03.983123 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-2c5nd"] Feb 02 10:02:05 crc kubenswrapper[4764]: I0202 10:02:05.727468 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-2c5nd" podUID="89fb05bd-ea44-4d00-b4ea-0fd4c5190995" containerName="registry-server" containerID="cri-o://7f150dd6cb0f8a4705a2261b68b7b62943edecef67ea5c57465891cda565b388" gracePeriod=2 Feb 02 10:02:06 crc kubenswrapper[4764]: I0202 10:02:06.708632 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2c5nd" Feb 02 10:02:06 crc kubenswrapper[4764]: I0202 10:02:06.755432 4764 generic.go:334] "Generic (PLEG): container finished" podID="89fb05bd-ea44-4d00-b4ea-0fd4c5190995" containerID="7f150dd6cb0f8a4705a2261b68b7b62943edecef67ea5c57465891cda565b388" exitCode=0 Feb 02 10:02:06 crc kubenswrapper[4764]: I0202 10:02:06.755488 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2c5nd" event={"ID":"89fb05bd-ea44-4d00-b4ea-0fd4c5190995","Type":"ContainerDied","Data":"7f150dd6cb0f8a4705a2261b68b7b62943edecef67ea5c57465891cda565b388"} Feb 02 10:02:06 crc kubenswrapper[4764]: I0202 10:02:06.755520 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2c5nd" event={"ID":"89fb05bd-ea44-4d00-b4ea-0fd4c5190995","Type":"ContainerDied","Data":"c98e4dc69e44a69ed1cd056926702d1bb6a5cdabe165a44816eb45f39dcf5735"} Feb 02 10:02:06 crc kubenswrapper[4764]: I0202 10:02:06.755541 4764 scope.go:117] "RemoveContainer" containerID="7f150dd6cb0f8a4705a2261b68b7b62943edecef67ea5c57465891cda565b388" Feb 02 10:02:06 crc kubenswrapper[4764]: I0202 10:02:06.755717 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2c5nd" Feb 02 10:02:06 crc kubenswrapper[4764]: I0202 10:02:06.792361 4764 scope.go:117] "RemoveContainer" containerID="13f89b0d61381bd77b2569a5c330825b07d424f09bc8e1d457d211872dd41418" Feb 02 10:02:06 crc kubenswrapper[4764]: I0202 10:02:06.805553 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/89fb05bd-ea44-4d00-b4ea-0fd4c5190995-utilities\") pod \"89fb05bd-ea44-4d00-b4ea-0fd4c5190995\" (UID: \"89fb05bd-ea44-4d00-b4ea-0fd4c5190995\") " Feb 02 10:02:06 crc kubenswrapper[4764]: I0202 10:02:06.805691 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/89fb05bd-ea44-4d00-b4ea-0fd4c5190995-catalog-content\") pod \"89fb05bd-ea44-4d00-b4ea-0fd4c5190995\" (UID: \"89fb05bd-ea44-4d00-b4ea-0fd4c5190995\") " Feb 02 10:02:06 crc kubenswrapper[4764]: I0202 10:02:06.805755 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wmvqp\" (UniqueName: \"kubernetes.io/projected/89fb05bd-ea44-4d00-b4ea-0fd4c5190995-kube-api-access-wmvqp\") pod \"89fb05bd-ea44-4d00-b4ea-0fd4c5190995\" (UID: \"89fb05bd-ea44-4d00-b4ea-0fd4c5190995\") " Feb 02 10:02:06 crc kubenswrapper[4764]: I0202 10:02:06.814985 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/89fb05bd-ea44-4d00-b4ea-0fd4c5190995-utilities" (OuterVolumeSpecName: "utilities") pod "89fb05bd-ea44-4d00-b4ea-0fd4c5190995" (UID: "89fb05bd-ea44-4d00-b4ea-0fd4c5190995"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 10:02:06 crc kubenswrapper[4764]: I0202 10:02:06.827921 4764 scope.go:117] "RemoveContainer" containerID="6d3930fe876570910583815f6ae3dd4cdb935da0ca07ca0e8cdb5bc9166906fc" Feb 02 10:02:06 crc kubenswrapper[4764]: I0202 10:02:06.841561 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/89fb05bd-ea44-4d00-b4ea-0fd4c5190995-kube-api-access-wmvqp" (OuterVolumeSpecName: "kube-api-access-wmvqp") pod "89fb05bd-ea44-4d00-b4ea-0fd4c5190995" (UID: "89fb05bd-ea44-4d00-b4ea-0fd4c5190995"). InnerVolumeSpecName "kube-api-access-wmvqp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 10:02:06 crc kubenswrapper[4764]: I0202 10:02:06.849845 4764 scope.go:117] "RemoveContainer" containerID="7f150dd6cb0f8a4705a2261b68b7b62943edecef67ea5c57465891cda565b388" Feb 02 10:02:06 crc kubenswrapper[4764]: E0202 10:02:06.850332 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7f150dd6cb0f8a4705a2261b68b7b62943edecef67ea5c57465891cda565b388\": container with ID starting with 7f150dd6cb0f8a4705a2261b68b7b62943edecef67ea5c57465891cda565b388 not found: ID does not exist" containerID="7f150dd6cb0f8a4705a2261b68b7b62943edecef67ea5c57465891cda565b388" Feb 02 10:02:06 crc kubenswrapper[4764]: I0202 10:02:06.850373 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7f150dd6cb0f8a4705a2261b68b7b62943edecef67ea5c57465891cda565b388"} err="failed to get container status \"7f150dd6cb0f8a4705a2261b68b7b62943edecef67ea5c57465891cda565b388\": rpc error: code = NotFound desc = could not find container \"7f150dd6cb0f8a4705a2261b68b7b62943edecef67ea5c57465891cda565b388\": container with ID starting with 7f150dd6cb0f8a4705a2261b68b7b62943edecef67ea5c57465891cda565b388 not found: ID does not exist" Feb 02 10:02:06 crc kubenswrapper[4764]: I0202 10:02:06.850397 4764 scope.go:117] "RemoveContainer" containerID="13f89b0d61381bd77b2569a5c330825b07d424f09bc8e1d457d211872dd41418" Feb 02 10:02:06 crc kubenswrapper[4764]: E0202 10:02:06.851008 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"13f89b0d61381bd77b2569a5c330825b07d424f09bc8e1d457d211872dd41418\": container with ID starting with 13f89b0d61381bd77b2569a5c330825b07d424f09bc8e1d457d211872dd41418 not found: ID does not exist" containerID="13f89b0d61381bd77b2569a5c330825b07d424f09bc8e1d457d211872dd41418" Feb 02 10:02:06 crc kubenswrapper[4764]: I0202 10:02:06.851171 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"13f89b0d61381bd77b2569a5c330825b07d424f09bc8e1d457d211872dd41418"} err="failed to get container status \"13f89b0d61381bd77b2569a5c330825b07d424f09bc8e1d457d211872dd41418\": rpc error: code = NotFound desc = could not find container \"13f89b0d61381bd77b2569a5c330825b07d424f09bc8e1d457d211872dd41418\": container with ID starting with 13f89b0d61381bd77b2569a5c330825b07d424f09bc8e1d457d211872dd41418 not found: ID does not exist" Feb 02 10:02:06 crc kubenswrapper[4764]: I0202 10:02:06.851312 4764 scope.go:117] "RemoveContainer" containerID="6d3930fe876570910583815f6ae3dd4cdb935da0ca07ca0e8cdb5bc9166906fc" Feb 02 10:02:06 crc kubenswrapper[4764]: E0202 10:02:06.851824 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6d3930fe876570910583815f6ae3dd4cdb935da0ca07ca0e8cdb5bc9166906fc\": container with ID starting with 6d3930fe876570910583815f6ae3dd4cdb935da0ca07ca0e8cdb5bc9166906fc not found: ID does not exist" containerID="6d3930fe876570910583815f6ae3dd4cdb935da0ca07ca0e8cdb5bc9166906fc" Feb 02 10:02:06 crc kubenswrapper[4764]: I0202 10:02:06.851861 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6d3930fe876570910583815f6ae3dd4cdb935da0ca07ca0e8cdb5bc9166906fc"} err="failed to get container status \"6d3930fe876570910583815f6ae3dd4cdb935da0ca07ca0e8cdb5bc9166906fc\": rpc error: code = NotFound desc = could not find container \"6d3930fe876570910583815f6ae3dd4cdb935da0ca07ca0e8cdb5bc9166906fc\": container with ID starting with 6d3930fe876570910583815f6ae3dd4cdb935da0ca07ca0e8cdb5bc9166906fc not found: ID does not exist" Feb 02 10:02:06 crc kubenswrapper[4764]: I0202 10:02:06.881810 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/89fb05bd-ea44-4d00-b4ea-0fd4c5190995-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "89fb05bd-ea44-4d00-b4ea-0fd4c5190995" (UID: "89fb05bd-ea44-4d00-b4ea-0fd4c5190995"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 10:02:06 crc kubenswrapper[4764]: I0202 10:02:06.907782 4764 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/89fb05bd-ea44-4d00-b4ea-0fd4c5190995-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 10:02:06 crc kubenswrapper[4764]: I0202 10:02:06.907816 4764 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/89fb05bd-ea44-4d00-b4ea-0fd4c5190995-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 10:02:06 crc kubenswrapper[4764]: I0202 10:02:06.907831 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wmvqp\" (UniqueName: \"kubernetes.io/projected/89fb05bd-ea44-4d00-b4ea-0fd4c5190995-kube-api-access-wmvqp\") on node \"crc\" DevicePath \"\"" Feb 02 10:02:07 crc kubenswrapper[4764]: I0202 10:02:07.092671 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-2c5nd"] Feb 02 10:02:07 crc kubenswrapper[4764]: I0202 10:02:07.105037 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-2c5nd"] Feb 02 10:02:07 crc kubenswrapper[4764]: I0202 10:02:07.852208 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="89fb05bd-ea44-4d00-b4ea-0fd4c5190995" path="/var/lib/kubelet/pods/89fb05bd-ea44-4d00-b4ea-0fd4c5190995/volumes" Feb 02 10:03:07 crc kubenswrapper[4764]: I0202 10:03:07.963230 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-zcll4"] Feb 02 10:03:07 crc kubenswrapper[4764]: E0202 10:03:07.965135 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89fb05bd-ea44-4d00-b4ea-0fd4c5190995" containerName="extract-utilities" Feb 02 10:03:07 crc kubenswrapper[4764]: I0202 10:03:07.965239 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="89fb05bd-ea44-4d00-b4ea-0fd4c5190995" containerName="extract-utilities" Feb 02 10:03:07 crc kubenswrapper[4764]: E0202 10:03:07.965325 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89fb05bd-ea44-4d00-b4ea-0fd4c5190995" containerName="extract-content" Feb 02 10:03:07 crc kubenswrapper[4764]: I0202 10:03:07.965400 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="89fb05bd-ea44-4d00-b4ea-0fd4c5190995" containerName="extract-content" Feb 02 10:03:07 crc kubenswrapper[4764]: E0202 10:03:07.965495 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89fb05bd-ea44-4d00-b4ea-0fd4c5190995" containerName="registry-server" Feb 02 10:03:07 crc kubenswrapper[4764]: I0202 10:03:07.965572 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="89fb05bd-ea44-4d00-b4ea-0fd4c5190995" containerName="registry-server" Feb 02 10:03:07 crc kubenswrapper[4764]: I0202 10:03:07.965887 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="89fb05bd-ea44-4d00-b4ea-0fd4c5190995" containerName="registry-server" Feb 02 10:03:07 crc kubenswrapper[4764]: I0202 10:03:07.967675 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zcll4" Feb 02 10:03:07 crc kubenswrapper[4764]: I0202 10:03:07.975465 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-zcll4"] Feb 02 10:03:08 crc kubenswrapper[4764]: I0202 10:03:08.054902 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/521092c4-ad2a-419f-85b4-8dd47ddc88f3-utilities\") pod \"redhat-operators-zcll4\" (UID: \"521092c4-ad2a-419f-85b4-8dd47ddc88f3\") " pod="openshift-marketplace/redhat-operators-zcll4" Feb 02 10:03:08 crc kubenswrapper[4764]: I0202 10:03:08.054987 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/521092c4-ad2a-419f-85b4-8dd47ddc88f3-catalog-content\") pod \"redhat-operators-zcll4\" (UID: \"521092c4-ad2a-419f-85b4-8dd47ddc88f3\") " pod="openshift-marketplace/redhat-operators-zcll4" Feb 02 10:03:08 crc kubenswrapper[4764]: I0202 10:03:08.055041 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cnngr\" (UniqueName: \"kubernetes.io/projected/521092c4-ad2a-419f-85b4-8dd47ddc88f3-kube-api-access-cnngr\") pod \"redhat-operators-zcll4\" (UID: \"521092c4-ad2a-419f-85b4-8dd47ddc88f3\") " pod="openshift-marketplace/redhat-operators-zcll4" Feb 02 10:03:08 crc kubenswrapper[4764]: I0202 10:03:08.156149 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cnngr\" (UniqueName: \"kubernetes.io/projected/521092c4-ad2a-419f-85b4-8dd47ddc88f3-kube-api-access-cnngr\") pod \"redhat-operators-zcll4\" (UID: \"521092c4-ad2a-419f-85b4-8dd47ddc88f3\") " pod="openshift-marketplace/redhat-operators-zcll4" Feb 02 10:03:08 crc kubenswrapper[4764]: I0202 10:03:08.156319 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/521092c4-ad2a-419f-85b4-8dd47ddc88f3-utilities\") pod \"redhat-operators-zcll4\" (UID: \"521092c4-ad2a-419f-85b4-8dd47ddc88f3\") " pod="openshift-marketplace/redhat-operators-zcll4" Feb 02 10:03:08 crc kubenswrapper[4764]: I0202 10:03:08.156386 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/521092c4-ad2a-419f-85b4-8dd47ddc88f3-catalog-content\") pod \"redhat-operators-zcll4\" (UID: \"521092c4-ad2a-419f-85b4-8dd47ddc88f3\") " pod="openshift-marketplace/redhat-operators-zcll4" Feb 02 10:03:08 crc kubenswrapper[4764]: I0202 10:03:08.156916 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/521092c4-ad2a-419f-85b4-8dd47ddc88f3-utilities\") pod \"redhat-operators-zcll4\" (UID: \"521092c4-ad2a-419f-85b4-8dd47ddc88f3\") " pod="openshift-marketplace/redhat-operators-zcll4" Feb 02 10:03:08 crc kubenswrapper[4764]: I0202 10:03:08.157302 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/521092c4-ad2a-419f-85b4-8dd47ddc88f3-catalog-content\") pod \"redhat-operators-zcll4\" (UID: \"521092c4-ad2a-419f-85b4-8dd47ddc88f3\") " pod="openshift-marketplace/redhat-operators-zcll4" Feb 02 10:03:08 crc kubenswrapper[4764]: I0202 10:03:08.174541 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cnngr\" (UniqueName: \"kubernetes.io/projected/521092c4-ad2a-419f-85b4-8dd47ddc88f3-kube-api-access-cnngr\") pod \"redhat-operators-zcll4\" (UID: \"521092c4-ad2a-419f-85b4-8dd47ddc88f3\") " pod="openshift-marketplace/redhat-operators-zcll4" Feb 02 10:03:08 crc kubenswrapper[4764]: I0202 10:03:08.321453 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zcll4" Feb 02 10:03:08 crc kubenswrapper[4764]: I0202 10:03:08.854002 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-zcll4"] Feb 02 10:03:09 crc kubenswrapper[4764]: I0202 10:03:09.314189 4764 generic.go:334] "Generic (PLEG): container finished" podID="521092c4-ad2a-419f-85b4-8dd47ddc88f3" containerID="b1f8e677e03ca355861f8c19e430a9068df6b5365582d204972d97809e714bf8" exitCode=0 Feb 02 10:03:09 crc kubenswrapper[4764]: I0202 10:03:09.314287 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zcll4" event={"ID":"521092c4-ad2a-419f-85b4-8dd47ddc88f3","Type":"ContainerDied","Data":"b1f8e677e03ca355861f8c19e430a9068df6b5365582d204972d97809e714bf8"} Feb 02 10:03:09 crc kubenswrapper[4764]: I0202 10:03:09.314555 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zcll4" event={"ID":"521092c4-ad2a-419f-85b4-8dd47ddc88f3","Type":"ContainerStarted","Data":"580c37bb2abdf2258b550875dbe26bb935386600cc1693f077a9247b773fbb0e"} Feb 02 10:03:10 crc kubenswrapper[4764]: I0202 10:03:10.325406 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zcll4" event={"ID":"521092c4-ad2a-419f-85b4-8dd47ddc88f3","Type":"ContainerStarted","Data":"b69af71691876d37654e53e6951553f5f195d4d96bfc6d8262d2838fd9163e8e"} Feb 02 10:03:16 crc kubenswrapper[4764]: I0202 10:03:16.379656 4764 generic.go:334] "Generic (PLEG): container finished" podID="521092c4-ad2a-419f-85b4-8dd47ddc88f3" containerID="b69af71691876d37654e53e6951553f5f195d4d96bfc6d8262d2838fd9163e8e" exitCode=0 Feb 02 10:03:16 crc kubenswrapper[4764]: I0202 10:03:16.379734 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zcll4" event={"ID":"521092c4-ad2a-419f-85b4-8dd47ddc88f3","Type":"ContainerDied","Data":"b69af71691876d37654e53e6951553f5f195d4d96bfc6d8262d2838fd9163e8e"} Feb 02 10:03:17 crc kubenswrapper[4764]: I0202 10:03:17.392007 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zcll4" event={"ID":"521092c4-ad2a-419f-85b4-8dd47ddc88f3","Type":"ContainerStarted","Data":"661fc9a45bc777beb2c8017a0147f5c6670ddf04a96ba322a337a975c17fd33e"} Feb 02 10:03:18 crc kubenswrapper[4764]: I0202 10:03:18.323103 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-zcll4" Feb 02 10:03:18 crc kubenswrapper[4764]: I0202 10:03:18.323701 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-zcll4" Feb 02 10:03:19 crc kubenswrapper[4764]: I0202 10:03:19.389692 4764 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-zcll4" podUID="521092c4-ad2a-419f-85b4-8dd47ddc88f3" containerName="registry-server" probeResult="failure" output=< Feb 02 10:03:19 crc kubenswrapper[4764]: timeout: failed to connect service ":50051" within 1s Feb 02 10:03:19 crc kubenswrapper[4764]: > Feb 02 10:03:29 crc kubenswrapper[4764]: I0202 10:03:29.383300 4764 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-zcll4" podUID="521092c4-ad2a-419f-85b4-8dd47ddc88f3" containerName="registry-server" probeResult="failure" output=< Feb 02 10:03:29 crc kubenswrapper[4764]: timeout: failed to connect service ":50051" within 1s Feb 02 10:03:29 crc kubenswrapper[4764]: > Feb 02 10:03:38 crc kubenswrapper[4764]: I0202 10:03:38.401459 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-zcll4" Feb 02 10:03:38 crc kubenswrapper[4764]: I0202 10:03:38.452552 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-zcll4" Feb 02 10:03:38 crc kubenswrapper[4764]: I0202 10:03:38.456437 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-zcll4" podStartSLOduration=23.97316071 podStartE2EDuration="31.456406699s" podCreationTimestamp="2026-02-02 10:03:07 +0000 UTC" firstStartedPulling="2026-02-02 10:03:09.316915649 +0000 UTC m=+3352.250639737" lastFinishedPulling="2026-02-02 10:03:16.800161638 +0000 UTC m=+3359.733885726" observedRunningTime="2026-02-02 10:03:17.423128243 +0000 UTC m=+3360.356852351" watchObservedRunningTime="2026-02-02 10:03:38.456406699 +0000 UTC m=+3381.390130797" Feb 02 10:03:39 crc kubenswrapper[4764]: I0202 10:03:39.186768 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-zcll4"] Feb 02 10:03:39 crc kubenswrapper[4764]: I0202 10:03:39.599029 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-zcll4" podUID="521092c4-ad2a-419f-85b4-8dd47ddc88f3" containerName="registry-server" containerID="cri-o://661fc9a45bc777beb2c8017a0147f5c6670ddf04a96ba322a337a975c17fd33e" gracePeriod=2 Feb 02 10:03:40 crc kubenswrapper[4764]: I0202 10:03:40.159400 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zcll4" Feb 02 10:03:40 crc kubenswrapper[4764]: I0202 10:03:40.228753 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/521092c4-ad2a-419f-85b4-8dd47ddc88f3-catalog-content\") pod \"521092c4-ad2a-419f-85b4-8dd47ddc88f3\" (UID: \"521092c4-ad2a-419f-85b4-8dd47ddc88f3\") " Feb 02 10:03:40 crc kubenswrapper[4764]: I0202 10:03:40.228915 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/521092c4-ad2a-419f-85b4-8dd47ddc88f3-utilities\") pod \"521092c4-ad2a-419f-85b4-8dd47ddc88f3\" (UID: \"521092c4-ad2a-419f-85b4-8dd47ddc88f3\") " Feb 02 10:03:40 crc kubenswrapper[4764]: I0202 10:03:40.229070 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cnngr\" (UniqueName: \"kubernetes.io/projected/521092c4-ad2a-419f-85b4-8dd47ddc88f3-kube-api-access-cnngr\") pod \"521092c4-ad2a-419f-85b4-8dd47ddc88f3\" (UID: \"521092c4-ad2a-419f-85b4-8dd47ddc88f3\") " Feb 02 10:03:40 crc kubenswrapper[4764]: I0202 10:03:40.229832 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/521092c4-ad2a-419f-85b4-8dd47ddc88f3-utilities" (OuterVolumeSpecName: "utilities") pod "521092c4-ad2a-419f-85b4-8dd47ddc88f3" (UID: "521092c4-ad2a-419f-85b4-8dd47ddc88f3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 10:03:40 crc kubenswrapper[4764]: I0202 10:03:40.236131 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/521092c4-ad2a-419f-85b4-8dd47ddc88f3-kube-api-access-cnngr" (OuterVolumeSpecName: "kube-api-access-cnngr") pod "521092c4-ad2a-419f-85b4-8dd47ddc88f3" (UID: "521092c4-ad2a-419f-85b4-8dd47ddc88f3"). InnerVolumeSpecName "kube-api-access-cnngr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 10:03:40 crc kubenswrapper[4764]: I0202 10:03:40.332570 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cnngr\" (UniqueName: \"kubernetes.io/projected/521092c4-ad2a-419f-85b4-8dd47ddc88f3-kube-api-access-cnngr\") on node \"crc\" DevicePath \"\"" Feb 02 10:03:40 crc kubenswrapper[4764]: I0202 10:03:40.332627 4764 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/521092c4-ad2a-419f-85b4-8dd47ddc88f3-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 10:03:40 crc kubenswrapper[4764]: I0202 10:03:40.356140 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/521092c4-ad2a-419f-85b4-8dd47ddc88f3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "521092c4-ad2a-419f-85b4-8dd47ddc88f3" (UID: "521092c4-ad2a-419f-85b4-8dd47ddc88f3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 10:03:40 crc kubenswrapper[4764]: I0202 10:03:40.434712 4764 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/521092c4-ad2a-419f-85b4-8dd47ddc88f3-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 10:03:40 crc kubenswrapper[4764]: I0202 10:03:40.609078 4764 generic.go:334] "Generic (PLEG): container finished" podID="521092c4-ad2a-419f-85b4-8dd47ddc88f3" containerID="661fc9a45bc777beb2c8017a0147f5c6670ddf04a96ba322a337a975c17fd33e" exitCode=0 Feb 02 10:03:40 crc kubenswrapper[4764]: I0202 10:03:40.609287 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zcll4" event={"ID":"521092c4-ad2a-419f-85b4-8dd47ddc88f3","Type":"ContainerDied","Data":"661fc9a45bc777beb2c8017a0147f5c6670ddf04a96ba322a337a975c17fd33e"} Feb 02 10:03:40 crc kubenswrapper[4764]: I0202 10:03:40.610564 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zcll4" event={"ID":"521092c4-ad2a-419f-85b4-8dd47ddc88f3","Type":"ContainerDied","Data":"580c37bb2abdf2258b550875dbe26bb935386600cc1693f077a9247b773fbb0e"} Feb 02 10:03:40 crc kubenswrapper[4764]: I0202 10:03:40.610649 4764 scope.go:117] "RemoveContainer" containerID="661fc9a45bc777beb2c8017a0147f5c6670ddf04a96ba322a337a975c17fd33e" Feb 02 10:03:40 crc kubenswrapper[4764]: I0202 10:03:40.609373 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zcll4" Feb 02 10:03:40 crc kubenswrapper[4764]: I0202 10:03:40.659150 4764 scope.go:117] "RemoveContainer" containerID="b69af71691876d37654e53e6951553f5f195d4d96bfc6d8262d2838fd9163e8e" Feb 02 10:03:40 crc kubenswrapper[4764]: I0202 10:03:40.659812 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-zcll4"] Feb 02 10:03:40 crc kubenswrapper[4764]: I0202 10:03:40.674113 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-zcll4"] Feb 02 10:03:40 crc kubenswrapper[4764]: I0202 10:03:40.697230 4764 scope.go:117] "RemoveContainer" containerID="b1f8e677e03ca355861f8c19e430a9068df6b5365582d204972d97809e714bf8" Feb 02 10:03:40 crc kubenswrapper[4764]: I0202 10:03:40.766629 4764 scope.go:117] "RemoveContainer" containerID="661fc9a45bc777beb2c8017a0147f5c6670ddf04a96ba322a337a975c17fd33e" Feb 02 10:03:40 crc kubenswrapper[4764]: E0202 10:03:40.766996 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"661fc9a45bc777beb2c8017a0147f5c6670ddf04a96ba322a337a975c17fd33e\": container with ID starting with 661fc9a45bc777beb2c8017a0147f5c6670ddf04a96ba322a337a975c17fd33e not found: ID does not exist" containerID="661fc9a45bc777beb2c8017a0147f5c6670ddf04a96ba322a337a975c17fd33e" Feb 02 10:03:40 crc kubenswrapper[4764]: I0202 10:03:40.767026 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"661fc9a45bc777beb2c8017a0147f5c6670ddf04a96ba322a337a975c17fd33e"} err="failed to get container status \"661fc9a45bc777beb2c8017a0147f5c6670ddf04a96ba322a337a975c17fd33e\": rpc error: code = NotFound desc = could not find container \"661fc9a45bc777beb2c8017a0147f5c6670ddf04a96ba322a337a975c17fd33e\": container with ID starting with 661fc9a45bc777beb2c8017a0147f5c6670ddf04a96ba322a337a975c17fd33e not found: ID does not exist" Feb 02 10:03:40 crc kubenswrapper[4764]: I0202 10:03:40.767045 4764 scope.go:117] "RemoveContainer" containerID="b69af71691876d37654e53e6951553f5f195d4d96bfc6d8262d2838fd9163e8e" Feb 02 10:03:40 crc kubenswrapper[4764]: E0202 10:03:40.767271 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b69af71691876d37654e53e6951553f5f195d4d96bfc6d8262d2838fd9163e8e\": container with ID starting with b69af71691876d37654e53e6951553f5f195d4d96bfc6d8262d2838fd9163e8e not found: ID does not exist" containerID="b69af71691876d37654e53e6951553f5f195d4d96bfc6d8262d2838fd9163e8e" Feb 02 10:03:40 crc kubenswrapper[4764]: I0202 10:03:40.767290 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b69af71691876d37654e53e6951553f5f195d4d96bfc6d8262d2838fd9163e8e"} err="failed to get container status \"b69af71691876d37654e53e6951553f5f195d4d96bfc6d8262d2838fd9163e8e\": rpc error: code = NotFound desc = could not find container \"b69af71691876d37654e53e6951553f5f195d4d96bfc6d8262d2838fd9163e8e\": container with ID starting with b69af71691876d37654e53e6951553f5f195d4d96bfc6d8262d2838fd9163e8e not found: ID does not exist" Feb 02 10:03:40 crc kubenswrapper[4764]: I0202 10:03:40.767304 4764 scope.go:117] "RemoveContainer" containerID="b1f8e677e03ca355861f8c19e430a9068df6b5365582d204972d97809e714bf8" Feb 02 10:03:40 crc kubenswrapper[4764]: E0202 10:03:40.767739 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b1f8e677e03ca355861f8c19e430a9068df6b5365582d204972d97809e714bf8\": container with ID starting with b1f8e677e03ca355861f8c19e430a9068df6b5365582d204972d97809e714bf8 not found: ID does not exist" containerID="b1f8e677e03ca355861f8c19e430a9068df6b5365582d204972d97809e714bf8" Feb 02 10:03:40 crc kubenswrapper[4764]: I0202 10:03:40.767764 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b1f8e677e03ca355861f8c19e430a9068df6b5365582d204972d97809e714bf8"} err="failed to get container status \"b1f8e677e03ca355861f8c19e430a9068df6b5365582d204972d97809e714bf8\": rpc error: code = NotFound desc = could not find container \"b1f8e677e03ca355861f8c19e430a9068df6b5365582d204972d97809e714bf8\": container with ID starting with b1f8e677e03ca355861f8c19e430a9068df6b5365582d204972d97809e714bf8 not found: ID does not exist" Feb 02 10:03:41 crc kubenswrapper[4764]: I0202 10:03:41.839001 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="521092c4-ad2a-419f-85b4-8dd47ddc88f3" path="/var/lib/kubelet/pods/521092c4-ad2a-419f-85b4-8dd47ddc88f3/volumes" Feb 02 10:03:43 crc kubenswrapper[4764]: I0202 10:03:43.531915 4764 patch_prober.go:28] interesting pod/machine-config-daemon-4ndm4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 10:03:43 crc kubenswrapper[4764]: I0202 10:03:43.533574 4764 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 10:04:13 crc kubenswrapper[4764]: I0202 10:04:13.524196 4764 patch_prober.go:28] interesting pod/machine-config-daemon-4ndm4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 10:04:13 crc kubenswrapper[4764]: I0202 10:04:13.524809 4764 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 10:04:43 crc kubenswrapper[4764]: I0202 10:04:43.523388 4764 patch_prober.go:28] interesting pod/machine-config-daemon-4ndm4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 10:04:43 crc kubenswrapper[4764]: I0202 10:04:43.524808 4764 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 10:04:43 crc kubenswrapper[4764]: I0202 10:04:43.524909 4764 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" Feb 02 10:04:43 crc kubenswrapper[4764]: I0202 10:04:43.525670 4764 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"86ea8d894bee98a50d949cdf649a606039569f7d321c8baff9abb00746af44d2"} pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 02 10:04:43 crc kubenswrapper[4764]: I0202 10:04:43.525783 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" containerName="machine-config-daemon" containerID="cri-o://86ea8d894bee98a50d949cdf649a606039569f7d321c8baff9abb00746af44d2" gracePeriod=600 Feb 02 10:04:43 crc kubenswrapper[4764]: I0202 10:04:43.683008 4764 generic.go:334] "Generic (PLEG): container finished" podID="d192f670-9f9d-4539-9641-e4bed73acdd4" containerID="86ea8d894bee98a50d949cdf649a606039569f7d321c8baff9abb00746af44d2" exitCode=0 Feb 02 10:04:43 crc kubenswrapper[4764]: I0202 10:04:43.683093 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" event={"ID":"d192f670-9f9d-4539-9641-e4bed73acdd4","Type":"ContainerDied","Data":"86ea8d894bee98a50d949cdf649a606039569f7d321c8baff9abb00746af44d2"} Feb 02 10:04:43 crc kubenswrapper[4764]: I0202 10:04:43.683329 4764 scope.go:117] "RemoveContainer" containerID="db4ebd4a8dbee216b70cb175f017f1799f3ca01a887f602ad42b9f295964da5d" Feb 02 10:04:44 crc kubenswrapper[4764]: I0202 10:04:44.693024 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" event={"ID":"d192f670-9f9d-4539-9641-e4bed73acdd4","Type":"ContainerStarted","Data":"4b1cf5e6a0ac319e2831449adb6329644931ba722d3592c4cde96fd7de3b310e"} Feb 02 10:05:23 crc kubenswrapper[4764]: E0202 10:05:23.658411 4764 kubelet.go:2526] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="1.833s" Feb 02 10:05:28 crc kubenswrapper[4764]: I0202 10:05:28.490330 4764 scope.go:117] "RemoveContainer" containerID="c4978209530c86d62dededf2405f3a1ef19fe1ccb720a7d1b81bfc6027c1da2b" Feb 02 10:05:28 crc kubenswrapper[4764]: I0202 10:05:28.521562 4764 scope.go:117] "RemoveContainer" containerID="1013758ae9afcfb1fcd04a4e0c05610e7b91121bc8c77bda8d1b60f70e2a7927" Feb 02 10:05:28 crc kubenswrapper[4764]: I0202 10:05:28.549568 4764 scope.go:117] "RemoveContainer" containerID="19ceda1a55292b83b269c26e4c0889cf09ba4c0c6ca5a5565dcb3adfc576a20b" Feb 02 10:05:28 crc kubenswrapper[4764]: I0202 10:05:28.586133 4764 scope.go:117] "RemoveContainer" containerID="68750e3d5d783461dcd76ac064febdaec06ae407bc81f56036014b1305a00528" Feb 02 10:06:18 crc kubenswrapper[4764]: I0202 10:06:18.783900 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-xzbn4"] Feb 02 10:06:18 crc kubenswrapper[4764]: E0202 10:06:18.785050 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="521092c4-ad2a-419f-85b4-8dd47ddc88f3" containerName="extract-utilities" Feb 02 10:06:18 crc kubenswrapper[4764]: I0202 10:06:18.785067 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="521092c4-ad2a-419f-85b4-8dd47ddc88f3" containerName="extract-utilities" Feb 02 10:06:18 crc kubenswrapper[4764]: E0202 10:06:18.785100 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="521092c4-ad2a-419f-85b4-8dd47ddc88f3" containerName="extract-content" Feb 02 10:06:18 crc kubenswrapper[4764]: I0202 10:06:18.785108 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="521092c4-ad2a-419f-85b4-8dd47ddc88f3" containerName="extract-content" Feb 02 10:06:18 crc kubenswrapper[4764]: E0202 10:06:18.785133 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="521092c4-ad2a-419f-85b4-8dd47ddc88f3" containerName="registry-server" Feb 02 10:06:18 crc kubenswrapper[4764]: I0202 10:06:18.785140 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="521092c4-ad2a-419f-85b4-8dd47ddc88f3" containerName="registry-server" Feb 02 10:06:18 crc kubenswrapper[4764]: I0202 10:06:18.785354 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="521092c4-ad2a-419f-85b4-8dd47ddc88f3" containerName="registry-server" Feb 02 10:06:18 crc kubenswrapper[4764]: I0202 10:06:18.786959 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xzbn4" Feb 02 10:06:18 crc kubenswrapper[4764]: I0202 10:06:18.810341 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xzbn4"] Feb 02 10:06:18 crc kubenswrapper[4764]: I0202 10:06:18.957310 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bsxk5\" (UniqueName: \"kubernetes.io/projected/ab7463a8-e37f-4653-afa8-55a78090c0b6-kube-api-access-bsxk5\") pod \"certified-operators-xzbn4\" (UID: \"ab7463a8-e37f-4653-afa8-55a78090c0b6\") " pod="openshift-marketplace/certified-operators-xzbn4" Feb 02 10:06:18 crc kubenswrapper[4764]: I0202 10:06:18.957374 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ab7463a8-e37f-4653-afa8-55a78090c0b6-catalog-content\") pod \"certified-operators-xzbn4\" (UID: \"ab7463a8-e37f-4653-afa8-55a78090c0b6\") " pod="openshift-marketplace/certified-operators-xzbn4" Feb 02 10:06:18 crc kubenswrapper[4764]: I0202 10:06:18.957682 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ab7463a8-e37f-4653-afa8-55a78090c0b6-utilities\") pod \"certified-operators-xzbn4\" (UID: \"ab7463a8-e37f-4653-afa8-55a78090c0b6\") " pod="openshift-marketplace/certified-operators-xzbn4" Feb 02 10:06:19 crc kubenswrapper[4764]: I0202 10:06:19.060240 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ab7463a8-e37f-4653-afa8-55a78090c0b6-utilities\") pod \"certified-operators-xzbn4\" (UID: \"ab7463a8-e37f-4653-afa8-55a78090c0b6\") " pod="openshift-marketplace/certified-operators-xzbn4" Feb 02 10:06:19 crc kubenswrapper[4764]: I0202 10:06:19.060344 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bsxk5\" (UniqueName: \"kubernetes.io/projected/ab7463a8-e37f-4653-afa8-55a78090c0b6-kube-api-access-bsxk5\") pod \"certified-operators-xzbn4\" (UID: \"ab7463a8-e37f-4653-afa8-55a78090c0b6\") " pod="openshift-marketplace/certified-operators-xzbn4" Feb 02 10:06:19 crc kubenswrapper[4764]: I0202 10:06:19.060411 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ab7463a8-e37f-4653-afa8-55a78090c0b6-catalog-content\") pod \"certified-operators-xzbn4\" (UID: \"ab7463a8-e37f-4653-afa8-55a78090c0b6\") " pod="openshift-marketplace/certified-operators-xzbn4" Feb 02 10:06:19 crc kubenswrapper[4764]: I0202 10:06:19.060906 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ab7463a8-e37f-4653-afa8-55a78090c0b6-utilities\") pod \"certified-operators-xzbn4\" (UID: \"ab7463a8-e37f-4653-afa8-55a78090c0b6\") " pod="openshift-marketplace/certified-operators-xzbn4" Feb 02 10:06:19 crc kubenswrapper[4764]: I0202 10:06:19.061010 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ab7463a8-e37f-4653-afa8-55a78090c0b6-catalog-content\") pod \"certified-operators-xzbn4\" (UID: \"ab7463a8-e37f-4653-afa8-55a78090c0b6\") " pod="openshift-marketplace/certified-operators-xzbn4" Feb 02 10:06:19 crc kubenswrapper[4764]: I0202 10:06:19.092520 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bsxk5\" (UniqueName: \"kubernetes.io/projected/ab7463a8-e37f-4653-afa8-55a78090c0b6-kube-api-access-bsxk5\") pod \"certified-operators-xzbn4\" (UID: \"ab7463a8-e37f-4653-afa8-55a78090c0b6\") " pod="openshift-marketplace/certified-operators-xzbn4" Feb 02 10:06:19 crc kubenswrapper[4764]: I0202 10:06:19.115312 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xzbn4" Feb 02 10:06:19 crc kubenswrapper[4764]: I0202 10:06:19.596288 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xzbn4"] Feb 02 10:06:20 crc kubenswrapper[4764]: I0202 10:06:20.212590 4764 generic.go:334] "Generic (PLEG): container finished" podID="ab7463a8-e37f-4653-afa8-55a78090c0b6" containerID="521f9eec448f4e470e4dcb317dc0f466360043c8c95d9062bc25a0733406c420" exitCode=0 Feb 02 10:06:20 crc kubenswrapper[4764]: I0202 10:06:20.212643 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xzbn4" event={"ID":"ab7463a8-e37f-4653-afa8-55a78090c0b6","Type":"ContainerDied","Data":"521f9eec448f4e470e4dcb317dc0f466360043c8c95d9062bc25a0733406c420"} Feb 02 10:06:20 crc kubenswrapper[4764]: I0202 10:06:20.213162 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xzbn4" event={"ID":"ab7463a8-e37f-4653-afa8-55a78090c0b6","Type":"ContainerStarted","Data":"56bfa936a6936e3259daded9e77c676e1d8bc8032b63b63cf4ef6e0609f5d30d"} Feb 02 10:06:21 crc kubenswrapper[4764]: I0202 10:06:21.224347 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xzbn4" event={"ID":"ab7463a8-e37f-4653-afa8-55a78090c0b6","Type":"ContainerStarted","Data":"69f51bd221c1a50e2e775bdb3a0716384febecff2101b08ed04f8c1974f6e808"} Feb 02 10:06:23 crc kubenswrapper[4764]: I0202 10:06:23.244759 4764 generic.go:334] "Generic (PLEG): container finished" podID="ab7463a8-e37f-4653-afa8-55a78090c0b6" containerID="69f51bd221c1a50e2e775bdb3a0716384febecff2101b08ed04f8c1974f6e808" exitCode=0 Feb 02 10:06:23 crc kubenswrapper[4764]: I0202 10:06:23.244814 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xzbn4" event={"ID":"ab7463a8-e37f-4653-afa8-55a78090c0b6","Type":"ContainerDied","Data":"69f51bd221c1a50e2e775bdb3a0716384febecff2101b08ed04f8c1974f6e808"} Feb 02 10:06:23 crc kubenswrapper[4764]: I0202 10:06:23.997131 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-v9bl7"] Feb 02 10:06:24 crc kubenswrapper[4764]: I0202 10:06:23.999430 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-v9bl7" Feb 02 10:06:24 crc kubenswrapper[4764]: I0202 10:06:24.019382 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-v9bl7"] Feb 02 10:06:24 crc kubenswrapper[4764]: I0202 10:06:24.167225 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zcgzb\" (UniqueName: \"kubernetes.io/projected/46650772-f820-40a9-bf79-314d783d0200-kube-api-access-zcgzb\") pod \"redhat-marketplace-v9bl7\" (UID: \"46650772-f820-40a9-bf79-314d783d0200\") " pod="openshift-marketplace/redhat-marketplace-v9bl7" Feb 02 10:06:24 crc kubenswrapper[4764]: I0202 10:06:24.167738 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/46650772-f820-40a9-bf79-314d783d0200-catalog-content\") pod \"redhat-marketplace-v9bl7\" (UID: \"46650772-f820-40a9-bf79-314d783d0200\") " pod="openshift-marketplace/redhat-marketplace-v9bl7" Feb 02 10:06:24 crc kubenswrapper[4764]: I0202 10:06:24.167801 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/46650772-f820-40a9-bf79-314d783d0200-utilities\") pod \"redhat-marketplace-v9bl7\" (UID: \"46650772-f820-40a9-bf79-314d783d0200\") " pod="openshift-marketplace/redhat-marketplace-v9bl7" Feb 02 10:06:24 crc kubenswrapper[4764]: I0202 10:06:24.256222 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xzbn4" event={"ID":"ab7463a8-e37f-4653-afa8-55a78090c0b6","Type":"ContainerStarted","Data":"37bf20845ea2844c1a74d0505627f3cfe1151634b57f94d8eb2b33ba8cd17350"} Feb 02 10:06:24 crc kubenswrapper[4764]: I0202 10:06:24.269445 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/46650772-f820-40a9-bf79-314d783d0200-utilities\") pod \"redhat-marketplace-v9bl7\" (UID: \"46650772-f820-40a9-bf79-314d783d0200\") " pod="openshift-marketplace/redhat-marketplace-v9bl7" Feb 02 10:06:24 crc kubenswrapper[4764]: I0202 10:06:24.269553 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zcgzb\" (UniqueName: \"kubernetes.io/projected/46650772-f820-40a9-bf79-314d783d0200-kube-api-access-zcgzb\") pod \"redhat-marketplace-v9bl7\" (UID: \"46650772-f820-40a9-bf79-314d783d0200\") " pod="openshift-marketplace/redhat-marketplace-v9bl7" Feb 02 10:06:24 crc kubenswrapper[4764]: I0202 10:06:24.269651 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/46650772-f820-40a9-bf79-314d783d0200-catalog-content\") pod \"redhat-marketplace-v9bl7\" (UID: \"46650772-f820-40a9-bf79-314d783d0200\") " pod="openshift-marketplace/redhat-marketplace-v9bl7" Feb 02 10:06:24 crc kubenswrapper[4764]: I0202 10:06:24.270005 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/46650772-f820-40a9-bf79-314d783d0200-utilities\") pod \"redhat-marketplace-v9bl7\" (UID: \"46650772-f820-40a9-bf79-314d783d0200\") " pod="openshift-marketplace/redhat-marketplace-v9bl7" Feb 02 10:06:24 crc kubenswrapper[4764]: I0202 10:06:24.270166 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/46650772-f820-40a9-bf79-314d783d0200-catalog-content\") pod \"redhat-marketplace-v9bl7\" (UID: \"46650772-f820-40a9-bf79-314d783d0200\") " pod="openshift-marketplace/redhat-marketplace-v9bl7" Feb 02 10:06:24 crc kubenswrapper[4764]: I0202 10:06:24.287405 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-xzbn4" podStartSLOduration=2.769715187 podStartE2EDuration="6.287386577s" podCreationTimestamp="2026-02-02 10:06:18 +0000 UTC" firstStartedPulling="2026-02-02 10:06:20.218143731 +0000 UTC m=+3543.151867839" lastFinishedPulling="2026-02-02 10:06:23.735815141 +0000 UTC m=+3546.669539229" observedRunningTime="2026-02-02 10:06:24.273922479 +0000 UTC m=+3547.207646557" watchObservedRunningTime="2026-02-02 10:06:24.287386577 +0000 UTC m=+3547.221110665" Feb 02 10:06:24 crc kubenswrapper[4764]: I0202 10:06:24.296075 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zcgzb\" (UniqueName: \"kubernetes.io/projected/46650772-f820-40a9-bf79-314d783d0200-kube-api-access-zcgzb\") pod \"redhat-marketplace-v9bl7\" (UID: \"46650772-f820-40a9-bf79-314d783d0200\") " pod="openshift-marketplace/redhat-marketplace-v9bl7" Feb 02 10:06:24 crc kubenswrapper[4764]: I0202 10:06:24.318404 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-v9bl7" Feb 02 10:06:24 crc kubenswrapper[4764]: I0202 10:06:24.898215 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-v9bl7"] Feb 02 10:06:24 crc kubenswrapper[4764]: W0202 10:06:24.900244 4764 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod46650772_f820_40a9_bf79_314d783d0200.slice/crio-af05f04136af603538995c637c0c7e461e4e23a413a09c1ec3536f6912e0b226 WatchSource:0}: Error finding container af05f04136af603538995c637c0c7e461e4e23a413a09c1ec3536f6912e0b226: Status 404 returned error can't find the container with id af05f04136af603538995c637c0c7e461e4e23a413a09c1ec3536f6912e0b226 Feb 02 10:06:25 crc kubenswrapper[4764]: I0202 10:06:25.266399 4764 generic.go:334] "Generic (PLEG): container finished" podID="46650772-f820-40a9-bf79-314d783d0200" containerID="79b7496c49ae297c6e64eec611a21282df48cb9bef4f61922a49baf39c9e71a3" exitCode=0 Feb 02 10:06:25 crc kubenswrapper[4764]: I0202 10:06:25.266477 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v9bl7" event={"ID":"46650772-f820-40a9-bf79-314d783d0200","Type":"ContainerDied","Data":"79b7496c49ae297c6e64eec611a21282df48cb9bef4f61922a49baf39c9e71a3"} Feb 02 10:06:25 crc kubenswrapper[4764]: I0202 10:06:25.266761 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v9bl7" event={"ID":"46650772-f820-40a9-bf79-314d783d0200","Type":"ContainerStarted","Data":"af05f04136af603538995c637c0c7e461e4e23a413a09c1ec3536f6912e0b226"} Feb 02 10:06:26 crc kubenswrapper[4764]: I0202 10:06:26.279314 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v9bl7" event={"ID":"46650772-f820-40a9-bf79-314d783d0200","Type":"ContainerStarted","Data":"7a784eaf3a6665c475d02a4a8bc3015ec6e184a56bb32ab6e3586c4e1f5a665f"} Feb 02 10:06:28 crc kubenswrapper[4764]: I0202 10:06:28.314315 4764 generic.go:334] "Generic (PLEG): container finished" podID="46650772-f820-40a9-bf79-314d783d0200" containerID="7a784eaf3a6665c475d02a4a8bc3015ec6e184a56bb32ab6e3586c4e1f5a665f" exitCode=0 Feb 02 10:06:28 crc kubenswrapper[4764]: I0202 10:06:28.314681 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v9bl7" event={"ID":"46650772-f820-40a9-bf79-314d783d0200","Type":"ContainerDied","Data":"7a784eaf3a6665c475d02a4a8bc3015ec6e184a56bb32ab6e3586c4e1f5a665f"} Feb 02 10:06:29 crc kubenswrapper[4764]: I0202 10:06:29.115928 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-xzbn4" Feb 02 10:06:29 crc kubenswrapper[4764]: I0202 10:06:29.116314 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-xzbn4" Feb 02 10:06:29 crc kubenswrapper[4764]: I0202 10:06:29.179667 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-xzbn4" Feb 02 10:06:29 crc kubenswrapper[4764]: I0202 10:06:29.325810 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v9bl7" event={"ID":"46650772-f820-40a9-bf79-314d783d0200","Type":"ContainerStarted","Data":"0f411cc0a0186ca2c7bf191e3ca998ca9d0f5c8bd62c3a847449dc48702d3dd4"} Feb 02 10:06:29 crc kubenswrapper[4764]: I0202 10:06:29.356190 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-v9bl7" podStartSLOduration=2.846562141 podStartE2EDuration="6.356168651s" podCreationTimestamp="2026-02-02 10:06:23 +0000 UTC" firstStartedPulling="2026-02-02 10:06:25.268852363 +0000 UTC m=+3548.202576451" lastFinishedPulling="2026-02-02 10:06:28.778458863 +0000 UTC m=+3551.712182961" observedRunningTime="2026-02-02 10:06:29.349718984 +0000 UTC m=+3552.283443092" watchObservedRunningTime="2026-02-02 10:06:29.356168651 +0000 UTC m=+3552.289892749" Feb 02 10:06:29 crc kubenswrapper[4764]: I0202 10:06:29.375311 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-xzbn4" Feb 02 10:06:30 crc kubenswrapper[4764]: I0202 10:06:30.765280 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xzbn4"] Feb 02 10:06:31 crc kubenswrapper[4764]: I0202 10:06:31.343063 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-xzbn4" podUID="ab7463a8-e37f-4653-afa8-55a78090c0b6" containerName="registry-server" containerID="cri-o://37bf20845ea2844c1a74d0505627f3cfe1151634b57f94d8eb2b33ba8cd17350" gracePeriod=2 Feb 02 10:06:31 crc kubenswrapper[4764]: I0202 10:06:31.873621 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xzbn4" Feb 02 10:06:32 crc kubenswrapper[4764]: I0202 10:06:32.030349 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ab7463a8-e37f-4653-afa8-55a78090c0b6-catalog-content\") pod \"ab7463a8-e37f-4653-afa8-55a78090c0b6\" (UID: \"ab7463a8-e37f-4653-afa8-55a78090c0b6\") " Feb 02 10:06:32 crc kubenswrapper[4764]: I0202 10:06:32.030504 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bsxk5\" (UniqueName: \"kubernetes.io/projected/ab7463a8-e37f-4653-afa8-55a78090c0b6-kube-api-access-bsxk5\") pod \"ab7463a8-e37f-4653-afa8-55a78090c0b6\" (UID: \"ab7463a8-e37f-4653-afa8-55a78090c0b6\") " Feb 02 10:06:32 crc kubenswrapper[4764]: I0202 10:06:32.030613 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ab7463a8-e37f-4653-afa8-55a78090c0b6-utilities\") pod \"ab7463a8-e37f-4653-afa8-55a78090c0b6\" (UID: \"ab7463a8-e37f-4653-afa8-55a78090c0b6\") " Feb 02 10:06:32 crc kubenswrapper[4764]: I0202 10:06:32.031357 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ab7463a8-e37f-4653-afa8-55a78090c0b6-utilities" (OuterVolumeSpecName: "utilities") pod "ab7463a8-e37f-4653-afa8-55a78090c0b6" (UID: "ab7463a8-e37f-4653-afa8-55a78090c0b6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 10:06:32 crc kubenswrapper[4764]: I0202 10:06:32.031574 4764 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ab7463a8-e37f-4653-afa8-55a78090c0b6-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 10:06:32 crc kubenswrapper[4764]: I0202 10:06:32.037595 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ab7463a8-e37f-4653-afa8-55a78090c0b6-kube-api-access-bsxk5" (OuterVolumeSpecName: "kube-api-access-bsxk5") pod "ab7463a8-e37f-4653-afa8-55a78090c0b6" (UID: "ab7463a8-e37f-4653-afa8-55a78090c0b6"). InnerVolumeSpecName "kube-api-access-bsxk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 10:06:32 crc kubenswrapper[4764]: I0202 10:06:32.073846 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ab7463a8-e37f-4653-afa8-55a78090c0b6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ab7463a8-e37f-4653-afa8-55a78090c0b6" (UID: "ab7463a8-e37f-4653-afa8-55a78090c0b6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 10:06:32 crc kubenswrapper[4764]: I0202 10:06:32.133234 4764 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ab7463a8-e37f-4653-afa8-55a78090c0b6-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 10:06:32 crc kubenswrapper[4764]: I0202 10:06:32.133578 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bsxk5\" (UniqueName: \"kubernetes.io/projected/ab7463a8-e37f-4653-afa8-55a78090c0b6-kube-api-access-bsxk5\") on node \"crc\" DevicePath \"\"" Feb 02 10:06:32 crc kubenswrapper[4764]: I0202 10:06:32.355468 4764 generic.go:334] "Generic (PLEG): container finished" podID="ab7463a8-e37f-4653-afa8-55a78090c0b6" containerID="37bf20845ea2844c1a74d0505627f3cfe1151634b57f94d8eb2b33ba8cd17350" exitCode=0 Feb 02 10:06:32 crc kubenswrapper[4764]: I0202 10:06:32.355527 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xzbn4" event={"ID":"ab7463a8-e37f-4653-afa8-55a78090c0b6","Type":"ContainerDied","Data":"37bf20845ea2844c1a74d0505627f3cfe1151634b57f94d8eb2b33ba8cd17350"} Feb 02 10:06:32 crc kubenswrapper[4764]: I0202 10:06:32.355545 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xzbn4" Feb 02 10:06:32 crc kubenswrapper[4764]: I0202 10:06:32.355563 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xzbn4" event={"ID":"ab7463a8-e37f-4653-afa8-55a78090c0b6","Type":"ContainerDied","Data":"56bfa936a6936e3259daded9e77c676e1d8bc8032b63b63cf4ef6e0609f5d30d"} Feb 02 10:06:32 crc kubenswrapper[4764]: I0202 10:06:32.355615 4764 scope.go:117] "RemoveContainer" containerID="37bf20845ea2844c1a74d0505627f3cfe1151634b57f94d8eb2b33ba8cd17350" Feb 02 10:06:32 crc kubenswrapper[4764]: I0202 10:06:32.393097 4764 scope.go:117] "RemoveContainer" containerID="69f51bd221c1a50e2e775bdb3a0716384febecff2101b08ed04f8c1974f6e808" Feb 02 10:06:32 crc kubenswrapper[4764]: I0202 10:06:32.413205 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xzbn4"] Feb 02 10:06:32 crc kubenswrapper[4764]: I0202 10:06:32.425727 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-xzbn4"] Feb 02 10:06:32 crc kubenswrapper[4764]: I0202 10:06:32.426354 4764 scope.go:117] "RemoveContainer" containerID="521f9eec448f4e470e4dcb317dc0f466360043c8c95d9062bc25a0733406c420" Feb 02 10:06:32 crc kubenswrapper[4764]: I0202 10:06:32.467451 4764 scope.go:117] "RemoveContainer" containerID="37bf20845ea2844c1a74d0505627f3cfe1151634b57f94d8eb2b33ba8cd17350" Feb 02 10:06:32 crc kubenswrapper[4764]: E0202 10:06:32.468101 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"37bf20845ea2844c1a74d0505627f3cfe1151634b57f94d8eb2b33ba8cd17350\": container with ID starting with 37bf20845ea2844c1a74d0505627f3cfe1151634b57f94d8eb2b33ba8cd17350 not found: ID does not exist" containerID="37bf20845ea2844c1a74d0505627f3cfe1151634b57f94d8eb2b33ba8cd17350" Feb 02 10:06:32 crc kubenswrapper[4764]: I0202 10:06:32.468145 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"37bf20845ea2844c1a74d0505627f3cfe1151634b57f94d8eb2b33ba8cd17350"} err="failed to get container status \"37bf20845ea2844c1a74d0505627f3cfe1151634b57f94d8eb2b33ba8cd17350\": rpc error: code = NotFound desc = could not find container \"37bf20845ea2844c1a74d0505627f3cfe1151634b57f94d8eb2b33ba8cd17350\": container with ID starting with 37bf20845ea2844c1a74d0505627f3cfe1151634b57f94d8eb2b33ba8cd17350 not found: ID does not exist" Feb 02 10:06:32 crc kubenswrapper[4764]: I0202 10:06:32.468171 4764 scope.go:117] "RemoveContainer" containerID="69f51bd221c1a50e2e775bdb3a0716384febecff2101b08ed04f8c1974f6e808" Feb 02 10:06:32 crc kubenswrapper[4764]: E0202 10:06:32.468658 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"69f51bd221c1a50e2e775bdb3a0716384febecff2101b08ed04f8c1974f6e808\": container with ID starting with 69f51bd221c1a50e2e775bdb3a0716384febecff2101b08ed04f8c1974f6e808 not found: ID does not exist" containerID="69f51bd221c1a50e2e775bdb3a0716384febecff2101b08ed04f8c1974f6e808" Feb 02 10:06:32 crc kubenswrapper[4764]: I0202 10:06:32.468801 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"69f51bd221c1a50e2e775bdb3a0716384febecff2101b08ed04f8c1974f6e808"} err="failed to get container status \"69f51bd221c1a50e2e775bdb3a0716384febecff2101b08ed04f8c1974f6e808\": rpc error: code = NotFound desc = could not find container \"69f51bd221c1a50e2e775bdb3a0716384febecff2101b08ed04f8c1974f6e808\": container with ID starting with 69f51bd221c1a50e2e775bdb3a0716384febecff2101b08ed04f8c1974f6e808 not found: ID does not exist" Feb 02 10:06:32 crc kubenswrapper[4764]: I0202 10:06:32.468916 4764 scope.go:117] "RemoveContainer" containerID="521f9eec448f4e470e4dcb317dc0f466360043c8c95d9062bc25a0733406c420" Feb 02 10:06:32 crc kubenswrapper[4764]: E0202 10:06:32.469478 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"521f9eec448f4e470e4dcb317dc0f466360043c8c95d9062bc25a0733406c420\": container with ID starting with 521f9eec448f4e470e4dcb317dc0f466360043c8c95d9062bc25a0733406c420 not found: ID does not exist" containerID="521f9eec448f4e470e4dcb317dc0f466360043c8c95d9062bc25a0733406c420" Feb 02 10:06:32 crc kubenswrapper[4764]: I0202 10:06:32.469506 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"521f9eec448f4e470e4dcb317dc0f466360043c8c95d9062bc25a0733406c420"} err="failed to get container status \"521f9eec448f4e470e4dcb317dc0f466360043c8c95d9062bc25a0733406c420\": rpc error: code = NotFound desc = could not find container \"521f9eec448f4e470e4dcb317dc0f466360043c8c95d9062bc25a0733406c420\": container with ID starting with 521f9eec448f4e470e4dcb317dc0f466360043c8c95d9062bc25a0733406c420 not found: ID does not exist" Feb 02 10:06:33 crc kubenswrapper[4764]: I0202 10:06:33.839505 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ab7463a8-e37f-4653-afa8-55a78090c0b6" path="/var/lib/kubelet/pods/ab7463a8-e37f-4653-afa8-55a78090c0b6/volumes" Feb 02 10:06:34 crc kubenswrapper[4764]: I0202 10:06:34.319675 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-v9bl7" Feb 02 10:06:34 crc kubenswrapper[4764]: I0202 10:06:34.320617 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-v9bl7" Feb 02 10:06:34 crc kubenswrapper[4764]: I0202 10:06:34.389091 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-v9bl7" Feb 02 10:06:35 crc kubenswrapper[4764]: I0202 10:06:35.447627 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-v9bl7" Feb 02 10:06:35 crc kubenswrapper[4764]: I0202 10:06:35.970762 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-v9bl7"] Feb 02 10:06:37 crc kubenswrapper[4764]: I0202 10:06:37.415132 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-v9bl7" podUID="46650772-f820-40a9-bf79-314d783d0200" containerName="registry-server" containerID="cri-o://0f411cc0a0186ca2c7bf191e3ca998ca9d0f5c8bd62c3a847449dc48702d3dd4" gracePeriod=2 Feb 02 10:06:37 crc kubenswrapper[4764]: I0202 10:06:37.991549 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-v9bl7" Feb 02 10:06:38 crc kubenswrapper[4764]: I0202 10:06:38.172208 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zcgzb\" (UniqueName: \"kubernetes.io/projected/46650772-f820-40a9-bf79-314d783d0200-kube-api-access-zcgzb\") pod \"46650772-f820-40a9-bf79-314d783d0200\" (UID: \"46650772-f820-40a9-bf79-314d783d0200\") " Feb 02 10:06:38 crc kubenswrapper[4764]: I0202 10:06:38.172323 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/46650772-f820-40a9-bf79-314d783d0200-catalog-content\") pod \"46650772-f820-40a9-bf79-314d783d0200\" (UID: \"46650772-f820-40a9-bf79-314d783d0200\") " Feb 02 10:06:38 crc kubenswrapper[4764]: I0202 10:06:38.172372 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/46650772-f820-40a9-bf79-314d783d0200-utilities\") pod \"46650772-f820-40a9-bf79-314d783d0200\" (UID: \"46650772-f820-40a9-bf79-314d783d0200\") " Feb 02 10:06:38 crc kubenswrapper[4764]: I0202 10:06:38.173984 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/46650772-f820-40a9-bf79-314d783d0200-utilities" (OuterVolumeSpecName: "utilities") pod "46650772-f820-40a9-bf79-314d783d0200" (UID: "46650772-f820-40a9-bf79-314d783d0200"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 10:06:38 crc kubenswrapper[4764]: I0202 10:06:38.191313 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/46650772-f820-40a9-bf79-314d783d0200-kube-api-access-zcgzb" (OuterVolumeSpecName: "kube-api-access-zcgzb") pod "46650772-f820-40a9-bf79-314d783d0200" (UID: "46650772-f820-40a9-bf79-314d783d0200"). InnerVolumeSpecName "kube-api-access-zcgzb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 10:06:38 crc kubenswrapper[4764]: I0202 10:06:38.196894 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/46650772-f820-40a9-bf79-314d783d0200-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "46650772-f820-40a9-bf79-314d783d0200" (UID: "46650772-f820-40a9-bf79-314d783d0200"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 10:06:38 crc kubenswrapper[4764]: I0202 10:06:38.275664 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zcgzb\" (UniqueName: \"kubernetes.io/projected/46650772-f820-40a9-bf79-314d783d0200-kube-api-access-zcgzb\") on node \"crc\" DevicePath \"\"" Feb 02 10:06:38 crc kubenswrapper[4764]: I0202 10:06:38.275800 4764 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/46650772-f820-40a9-bf79-314d783d0200-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 10:06:38 crc kubenswrapper[4764]: I0202 10:06:38.275816 4764 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/46650772-f820-40a9-bf79-314d783d0200-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 10:06:38 crc kubenswrapper[4764]: I0202 10:06:38.425875 4764 generic.go:334] "Generic (PLEG): container finished" podID="46650772-f820-40a9-bf79-314d783d0200" containerID="0f411cc0a0186ca2c7bf191e3ca998ca9d0f5c8bd62c3a847449dc48702d3dd4" exitCode=0 Feb 02 10:06:38 crc kubenswrapper[4764]: I0202 10:06:38.425920 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v9bl7" event={"ID":"46650772-f820-40a9-bf79-314d783d0200","Type":"ContainerDied","Data":"0f411cc0a0186ca2c7bf191e3ca998ca9d0f5c8bd62c3a847449dc48702d3dd4"} Feb 02 10:06:38 crc kubenswrapper[4764]: I0202 10:06:38.425970 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v9bl7" event={"ID":"46650772-f820-40a9-bf79-314d783d0200","Type":"ContainerDied","Data":"af05f04136af603538995c637c0c7e461e4e23a413a09c1ec3536f6912e0b226"} Feb 02 10:06:38 crc kubenswrapper[4764]: I0202 10:06:38.425975 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-v9bl7" Feb 02 10:06:38 crc kubenswrapper[4764]: I0202 10:06:38.425988 4764 scope.go:117] "RemoveContainer" containerID="0f411cc0a0186ca2c7bf191e3ca998ca9d0f5c8bd62c3a847449dc48702d3dd4" Feb 02 10:06:38 crc kubenswrapper[4764]: I0202 10:06:38.455133 4764 scope.go:117] "RemoveContainer" containerID="7a784eaf3a6665c475d02a4a8bc3015ec6e184a56bb32ab6e3586c4e1f5a665f" Feb 02 10:06:38 crc kubenswrapper[4764]: I0202 10:06:38.472108 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-v9bl7"] Feb 02 10:06:38 crc kubenswrapper[4764]: I0202 10:06:38.480912 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-v9bl7"] Feb 02 10:06:38 crc kubenswrapper[4764]: I0202 10:06:38.484635 4764 scope.go:117] "RemoveContainer" containerID="79b7496c49ae297c6e64eec611a21282df48cb9bef4f61922a49baf39c9e71a3" Feb 02 10:06:38 crc kubenswrapper[4764]: I0202 10:06:38.527104 4764 scope.go:117] "RemoveContainer" containerID="0f411cc0a0186ca2c7bf191e3ca998ca9d0f5c8bd62c3a847449dc48702d3dd4" Feb 02 10:06:38 crc kubenswrapper[4764]: E0202 10:06:38.527570 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0f411cc0a0186ca2c7bf191e3ca998ca9d0f5c8bd62c3a847449dc48702d3dd4\": container with ID starting with 0f411cc0a0186ca2c7bf191e3ca998ca9d0f5c8bd62c3a847449dc48702d3dd4 not found: ID does not exist" containerID="0f411cc0a0186ca2c7bf191e3ca998ca9d0f5c8bd62c3a847449dc48702d3dd4" Feb 02 10:06:38 crc kubenswrapper[4764]: I0202 10:06:38.527618 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0f411cc0a0186ca2c7bf191e3ca998ca9d0f5c8bd62c3a847449dc48702d3dd4"} err="failed to get container status \"0f411cc0a0186ca2c7bf191e3ca998ca9d0f5c8bd62c3a847449dc48702d3dd4\": rpc error: code = NotFound desc = could not find container \"0f411cc0a0186ca2c7bf191e3ca998ca9d0f5c8bd62c3a847449dc48702d3dd4\": container with ID starting with 0f411cc0a0186ca2c7bf191e3ca998ca9d0f5c8bd62c3a847449dc48702d3dd4 not found: ID does not exist" Feb 02 10:06:38 crc kubenswrapper[4764]: I0202 10:06:38.527644 4764 scope.go:117] "RemoveContainer" containerID="7a784eaf3a6665c475d02a4a8bc3015ec6e184a56bb32ab6e3586c4e1f5a665f" Feb 02 10:06:38 crc kubenswrapper[4764]: E0202 10:06:38.528073 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7a784eaf3a6665c475d02a4a8bc3015ec6e184a56bb32ab6e3586c4e1f5a665f\": container with ID starting with 7a784eaf3a6665c475d02a4a8bc3015ec6e184a56bb32ab6e3586c4e1f5a665f not found: ID does not exist" containerID="7a784eaf3a6665c475d02a4a8bc3015ec6e184a56bb32ab6e3586c4e1f5a665f" Feb 02 10:06:38 crc kubenswrapper[4764]: I0202 10:06:38.528131 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7a784eaf3a6665c475d02a4a8bc3015ec6e184a56bb32ab6e3586c4e1f5a665f"} err="failed to get container status \"7a784eaf3a6665c475d02a4a8bc3015ec6e184a56bb32ab6e3586c4e1f5a665f\": rpc error: code = NotFound desc = could not find container \"7a784eaf3a6665c475d02a4a8bc3015ec6e184a56bb32ab6e3586c4e1f5a665f\": container with ID starting with 7a784eaf3a6665c475d02a4a8bc3015ec6e184a56bb32ab6e3586c4e1f5a665f not found: ID does not exist" Feb 02 10:06:38 crc kubenswrapper[4764]: I0202 10:06:38.528167 4764 scope.go:117] "RemoveContainer" containerID="79b7496c49ae297c6e64eec611a21282df48cb9bef4f61922a49baf39c9e71a3" Feb 02 10:06:38 crc kubenswrapper[4764]: E0202 10:06:38.528527 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"79b7496c49ae297c6e64eec611a21282df48cb9bef4f61922a49baf39c9e71a3\": container with ID starting with 79b7496c49ae297c6e64eec611a21282df48cb9bef4f61922a49baf39c9e71a3 not found: ID does not exist" containerID="79b7496c49ae297c6e64eec611a21282df48cb9bef4f61922a49baf39c9e71a3" Feb 02 10:06:38 crc kubenswrapper[4764]: I0202 10:06:38.528564 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"79b7496c49ae297c6e64eec611a21282df48cb9bef4f61922a49baf39c9e71a3"} err="failed to get container status \"79b7496c49ae297c6e64eec611a21282df48cb9bef4f61922a49baf39c9e71a3\": rpc error: code = NotFound desc = could not find container \"79b7496c49ae297c6e64eec611a21282df48cb9bef4f61922a49baf39c9e71a3\": container with ID starting with 79b7496c49ae297c6e64eec611a21282df48cb9bef4f61922a49baf39c9e71a3 not found: ID does not exist" Feb 02 10:06:39 crc kubenswrapper[4764]: I0202 10:06:39.860718 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="46650772-f820-40a9-bf79-314d783d0200" path="/var/lib/kubelet/pods/46650772-f820-40a9-bf79-314d783d0200/volumes" Feb 02 10:06:43 crc kubenswrapper[4764]: I0202 10:06:43.522491 4764 patch_prober.go:28] interesting pod/machine-config-daemon-4ndm4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 10:06:43 crc kubenswrapper[4764]: I0202 10:06:43.524072 4764 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 10:07:13 crc kubenswrapper[4764]: I0202 10:07:13.537187 4764 patch_prober.go:28] interesting pod/machine-config-daemon-4ndm4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 10:07:13 crc kubenswrapper[4764]: I0202 10:07:13.537988 4764 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 10:07:43 crc kubenswrapper[4764]: I0202 10:07:43.523290 4764 patch_prober.go:28] interesting pod/machine-config-daemon-4ndm4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 10:07:43 crc kubenswrapper[4764]: I0202 10:07:43.523882 4764 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 10:07:43 crc kubenswrapper[4764]: I0202 10:07:43.523963 4764 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" Feb 02 10:07:43 crc kubenswrapper[4764]: I0202 10:07:43.524815 4764 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"4b1cf5e6a0ac319e2831449adb6329644931ba722d3592c4cde96fd7de3b310e"} pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 02 10:07:43 crc kubenswrapper[4764]: I0202 10:07:43.524875 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" containerName="machine-config-daemon" containerID="cri-o://4b1cf5e6a0ac319e2831449adb6329644931ba722d3592c4cde96fd7de3b310e" gracePeriod=600 Feb 02 10:07:43 crc kubenswrapper[4764]: E0202 10:07:43.656279 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 10:07:44 crc kubenswrapper[4764]: I0202 10:07:44.371481 4764 generic.go:334] "Generic (PLEG): container finished" podID="d192f670-9f9d-4539-9641-e4bed73acdd4" containerID="4b1cf5e6a0ac319e2831449adb6329644931ba722d3592c4cde96fd7de3b310e" exitCode=0 Feb 02 10:07:44 crc kubenswrapper[4764]: I0202 10:07:44.371554 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" event={"ID":"d192f670-9f9d-4539-9641-e4bed73acdd4","Type":"ContainerDied","Data":"4b1cf5e6a0ac319e2831449adb6329644931ba722d3592c4cde96fd7de3b310e"} Feb 02 10:07:44 crc kubenswrapper[4764]: I0202 10:07:44.371871 4764 scope.go:117] "RemoveContainer" containerID="86ea8d894bee98a50d949cdf649a606039569f7d321c8baff9abb00746af44d2" Feb 02 10:07:44 crc kubenswrapper[4764]: I0202 10:07:44.372946 4764 scope.go:117] "RemoveContainer" containerID="4b1cf5e6a0ac319e2831449adb6329644931ba722d3592c4cde96fd7de3b310e" Feb 02 10:07:44 crc kubenswrapper[4764]: E0202 10:07:44.373272 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 10:07:57 crc kubenswrapper[4764]: I0202 10:07:57.833937 4764 scope.go:117] "RemoveContainer" containerID="4b1cf5e6a0ac319e2831449adb6329644931ba722d3592c4cde96fd7de3b310e" Feb 02 10:07:57 crc kubenswrapper[4764]: E0202 10:07:57.835325 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 10:08:10 crc kubenswrapper[4764]: I0202 10:08:10.826513 4764 scope.go:117] "RemoveContainer" containerID="4b1cf5e6a0ac319e2831449adb6329644931ba722d3592c4cde96fd7de3b310e" Feb 02 10:08:10 crc kubenswrapper[4764]: E0202 10:08:10.827512 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 10:08:22 crc kubenswrapper[4764]: I0202 10:08:22.826454 4764 scope.go:117] "RemoveContainer" containerID="4b1cf5e6a0ac319e2831449adb6329644931ba722d3592c4cde96fd7de3b310e" Feb 02 10:08:22 crc kubenswrapper[4764]: E0202 10:08:22.827686 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 10:08:26 crc kubenswrapper[4764]: I0202 10:08:26.050114 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-db-create-v6n5l"] Feb 02 10:08:26 crc kubenswrapper[4764]: I0202 10:08:26.072243 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-5e6b-account-create-update-shmv8"] Feb 02 10:08:26 crc kubenswrapper[4764]: I0202 10:08:26.082194 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-5e6b-account-create-update-shmv8"] Feb 02 10:08:26 crc kubenswrapper[4764]: I0202 10:08:26.090803 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-db-create-v6n5l"] Feb 02 10:08:27 crc kubenswrapper[4764]: I0202 10:08:27.840317 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ca0252ea-2836-433b-b26f-bee25fd3adb2" path="/var/lib/kubelet/pods/ca0252ea-2836-433b-b26f-bee25fd3adb2/volumes" Feb 02 10:08:27 crc kubenswrapper[4764]: I0202 10:08:27.843219 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f001efec-ca89-4c17-be01-3f42fd33e5c0" path="/var/lib/kubelet/pods/f001efec-ca89-4c17-be01-3f42fd33e5c0/volumes" Feb 02 10:08:28 crc kubenswrapper[4764]: I0202 10:08:28.742090 4764 scope.go:117] "RemoveContainer" containerID="73fe62e3ef90007d6d8d663dcf54065a81911cd5c1fb0755c45efd55d939ee35" Feb 02 10:08:28 crc kubenswrapper[4764]: I0202 10:08:28.780382 4764 scope.go:117] "RemoveContainer" containerID="f5c1b7838458bf480c48d4f8862d78d8d2ad95451f151514fc66ce939fa6a89b" Feb 02 10:08:37 crc kubenswrapper[4764]: I0202 10:08:37.831906 4764 scope.go:117] "RemoveContainer" containerID="4b1cf5e6a0ac319e2831449adb6329644931ba722d3592c4cde96fd7de3b310e" Feb 02 10:08:37 crc kubenswrapper[4764]: E0202 10:08:37.832679 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 10:08:49 crc kubenswrapper[4764]: I0202 10:08:49.827038 4764 scope.go:117] "RemoveContainer" containerID="4b1cf5e6a0ac319e2831449adb6329644931ba722d3592c4cde96fd7de3b310e" Feb 02 10:08:49 crc kubenswrapper[4764]: E0202 10:08:49.827866 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 10:08:57 crc kubenswrapper[4764]: I0202 10:08:57.049161 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-db-sync-xvkkk"] Feb 02 10:08:57 crc kubenswrapper[4764]: I0202 10:08:57.063514 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-db-sync-xvkkk"] Feb 02 10:08:57 crc kubenswrapper[4764]: I0202 10:08:57.845250 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7db82-4332-485f-acbf-3edbf4269f50" path="/var/lib/kubelet/pods/1bf7db82-4332-485f-acbf-3edbf4269f50/volumes" Feb 02 10:09:02 crc kubenswrapper[4764]: I0202 10:09:02.826579 4764 scope.go:117] "RemoveContainer" containerID="4b1cf5e6a0ac319e2831449adb6329644931ba722d3592c4cde96fd7de3b310e" Feb 02 10:09:02 crc kubenswrapper[4764]: E0202 10:09:02.827406 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 10:09:15 crc kubenswrapper[4764]: I0202 10:09:15.881962 4764 scope.go:117] "RemoveContainer" containerID="4b1cf5e6a0ac319e2831449adb6329644931ba722d3592c4cde96fd7de3b310e" Feb 02 10:09:15 crc kubenswrapper[4764]: E0202 10:09:15.882985 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 10:09:27 crc kubenswrapper[4764]: I0202 10:09:27.832663 4764 scope.go:117] "RemoveContainer" containerID="4b1cf5e6a0ac319e2831449adb6329644931ba722d3592c4cde96fd7de3b310e" Feb 02 10:09:27 crc kubenswrapper[4764]: E0202 10:09:27.833466 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 10:09:28 crc kubenswrapper[4764]: I0202 10:09:28.858061 4764 scope.go:117] "RemoveContainer" containerID="d8a10c4c9ac37c4ec6829a95cae27abf3c0ab4bbb1602a4daf3f9b9a087b5478" Feb 02 10:09:41 crc kubenswrapper[4764]: I0202 10:09:41.825996 4764 scope.go:117] "RemoveContainer" containerID="4b1cf5e6a0ac319e2831449adb6329644931ba722d3592c4cde96fd7de3b310e" Feb 02 10:09:41 crc kubenswrapper[4764]: E0202 10:09:41.826923 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 10:09:53 crc kubenswrapper[4764]: I0202 10:09:53.826519 4764 scope.go:117] "RemoveContainer" containerID="4b1cf5e6a0ac319e2831449adb6329644931ba722d3592c4cde96fd7de3b310e" Feb 02 10:09:53 crc kubenswrapper[4764]: E0202 10:09:53.827194 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 10:10:08 crc kubenswrapper[4764]: I0202 10:10:08.825865 4764 scope.go:117] "RemoveContainer" containerID="4b1cf5e6a0ac319e2831449adb6329644931ba722d3592c4cde96fd7de3b310e" Feb 02 10:10:08 crc kubenswrapper[4764]: E0202 10:10:08.826554 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 10:10:19 crc kubenswrapper[4764]: I0202 10:10:19.826251 4764 scope.go:117] "RemoveContainer" containerID="4b1cf5e6a0ac319e2831449adb6329644931ba722d3592c4cde96fd7de3b310e" Feb 02 10:10:19 crc kubenswrapper[4764]: E0202 10:10:19.827248 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 10:10:33 crc kubenswrapper[4764]: I0202 10:10:33.825525 4764 scope.go:117] "RemoveContainer" containerID="4b1cf5e6a0ac319e2831449adb6329644931ba722d3592c4cde96fd7de3b310e" Feb 02 10:10:33 crc kubenswrapper[4764]: E0202 10:10:33.827176 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 10:10:45 crc kubenswrapper[4764]: I0202 10:10:45.825705 4764 scope.go:117] "RemoveContainer" containerID="4b1cf5e6a0ac319e2831449adb6329644931ba722d3592c4cde96fd7de3b310e" Feb 02 10:10:45 crc kubenswrapper[4764]: E0202 10:10:45.826800 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 10:10:58 crc kubenswrapper[4764]: I0202 10:10:58.825591 4764 scope.go:117] "RemoveContainer" containerID="4b1cf5e6a0ac319e2831449adb6329644931ba722d3592c4cde96fd7de3b310e" Feb 02 10:10:58 crc kubenswrapper[4764]: E0202 10:10:58.826374 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 10:11:13 crc kubenswrapper[4764]: I0202 10:11:13.826861 4764 scope.go:117] "RemoveContainer" containerID="4b1cf5e6a0ac319e2831449adb6329644931ba722d3592c4cde96fd7de3b310e" Feb 02 10:11:13 crc kubenswrapper[4764]: E0202 10:11:13.827696 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 10:11:27 crc kubenswrapper[4764]: I0202 10:11:27.831572 4764 scope.go:117] "RemoveContainer" containerID="4b1cf5e6a0ac319e2831449adb6329644931ba722d3592c4cde96fd7de3b310e" Feb 02 10:11:27 crc kubenswrapper[4764]: E0202 10:11:27.832441 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 10:11:41 crc kubenswrapper[4764]: I0202 10:11:41.826075 4764 scope.go:117] "RemoveContainer" containerID="4b1cf5e6a0ac319e2831449adb6329644931ba722d3592c4cde96fd7de3b310e" Feb 02 10:11:41 crc kubenswrapper[4764]: E0202 10:11:41.829092 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 10:11:45 crc kubenswrapper[4764]: I0202 10:11:45.580146 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-bfwt6"] Feb 02 10:11:45 crc kubenswrapper[4764]: E0202 10:11:45.580836 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab7463a8-e37f-4653-afa8-55a78090c0b6" containerName="extract-content" Feb 02 10:11:45 crc kubenswrapper[4764]: I0202 10:11:45.580851 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab7463a8-e37f-4653-afa8-55a78090c0b6" containerName="extract-content" Feb 02 10:11:45 crc kubenswrapper[4764]: E0202 10:11:45.580866 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46650772-f820-40a9-bf79-314d783d0200" containerName="extract-content" Feb 02 10:11:45 crc kubenswrapper[4764]: I0202 10:11:45.580872 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="46650772-f820-40a9-bf79-314d783d0200" containerName="extract-content" Feb 02 10:11:45 crc kubenswrapper[4764]: E0202 10:11:45.580881 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46650772-f820-40a9-bf79-314d783d0200" containerName="extract-utilities" Feb 02 10:11:45 crc kubenswrapper[4764]: I0202 10:11:45.580887 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="46650772-f820-40a9-bf79-314d783d0200" containerName="extract-utilities" Feb 02 10:11:45 crc kubenswrapper[4764]: E0202 10:11:45.580901 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab7463a8-e37f-4653-afa8-55a78090c0b6" containerName="registry-server" Feb 02 10:11:45 crc kubenswrapper[4764]: I0202 10:11:45.580908 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab7463a8-e37f-4653-afa8-55a78090c0b6" containerName="registry-server" Feb 02 10:11:45 crc kubenswrapper[4764]: E0202 10:11:45.580919 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46650772-f820-40a9-bf79-314d783d0200" containerName="registry-server" Feb 02 10:11:45 crc kubenswrapper[4764]: I0202 10:11:45.580926 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="46650772-f820-40a9-bf79-314d783d0200" containerName="registry-server" Feb 02 10:11:45 crc kubenswrapper[4764]: E0202 10:11:45.580978 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab7463a8-e37f-4653-afa8-55a78090c0b6" containerName="extract-utilities" Feb 02 10:11:45 crc kubenswrapper[4764]: I0202 10:11:45.580986 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab7463a8-e37f-4653-afa8-55a78090c0b6" containerName="extract-utilities" Feb 02 10:11:45 crc kubenswrapper[4764]: I0202 10:11:45.581359 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab7463a8-e37f-4653-afa8-55a78090c0b6" containerName="registry-server" Feb 02 10:11:45 crc kubenswrapper[4764]: I0202 10:11:45.581372 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="46650772-f820-40a9-bf79-314d783d0200" containerName="registry-server" Feb 02 10:11:45 crc kubenswrapper[4764]: I0202 10:11:45.583359 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bfwt6" Feb 02 10:11:45 crc kubenswrapper[4764]: I0202 10:11:45.671714 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-bfwt6"] Feb 02 10:11:45 crc kubenswrapper[4764]: I0202 10:11:45.684663 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/78b58f9f-3b0b-41a9-aed7-82341bbe9442-catalog-content\") pod \"community-operators-bfwt6\" (UID: \"78b58f9f-3b0b-41a9-aed7-82341bbe9442\") " pod="openshift-marketplace/community-operators-bfwt6" Feb 02 10:11:45 crc kubenswrapper[4764]: I0202 10:11:45.684840 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/78b58f9f-3b0b-41a9-aed7-82341bbe9442-utilities\") pod \"community-operators-bfwt6\" (UID: \"78b58f9f-3b0b-41a9-aed7-82341bbe9442\") " pod="openshift-marketplace/community-operators-bfwt6" Feb 02 10:11:45 crc kubenswrapper[4764]: I0202 10:11:45.684962 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hg8rf\" (UniqueName: \"kubernetes.io/projected/78b58f9f-3b0b-41a9-aed7-82341bbe9442-kube-api-access-hg8rf\") pod \"community-operators-bfwt6\" (UID: \"78b58f9f-3b0b-41a9-aed7-82341bbe9442\") " pod="openshift-marketplace/community-operators-bfwt6" Feb 02 10:11:45 crc kubenswrapper[4764]: I0202 10:11:45.786552 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/78b58f9f-3b0b-41a9-aed7-82341bbe9442-catalog-content\") pod \"community-operators-bfwt6\" (UID: \"78b58f9f-3b0b-41a9-aed7-82341bbe9442\") " pod="openshift-marketplace/community-operators-bfwt6" Feb 02 10:11:45 crc kubenswrapper[4764]: I0202 10:11:45.786757 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/78b58f9f-3b0b-41a9-aed7-82341bbe9442-utilities\") pod \"community-operators-bfwt6\" (UID: \"78b58f9f-3b0b-41a9-aed7-82341bbe9442\") " pod="openshift-marketplace/community-operators-bfwt6" Feb 02 10:11:45 crc kubenswrapper[4764]: I0202 10:11:45.786870 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hg8rf\" (UniqueName: \"kubernetes.io/projected/78b58f9f-3b0b-41a9-aed7-82341bbe9442-kube-api-access-hg8rf\") pod \"community-operators-bfwt6\" (UID: \"78b58f9f-3b0b-41a9-aed7-82341bbe9442\") " pod="openshift-marketplace/community-operators-bfwt6" Feb 02 10:11:45 crc kubenswrapper[4764]: I0202 10:11:45.787089 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/78b58f9f-3b0b-41a9-aed7-82341bbe9442-catalog-content\") pod \"community-operators-bfwt6\" (UID: \"78b58f9f-3b0b-41a9-aed7-82341bbe9442\") " pod="openshift-marketplace/community-operators-bfwt6" Feb 02 10:11:45 crc kubenswrapper[4764]: I0202 10:11:45.787190 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/78b58f9f-3b0b-41a9-aed7-82341bbe9442-utilities\") pod \"community-operators-bfwt6\" (UID: \"78b58f9f-3b0b-41a9-aed7-82341bbe9442\") " pod="openshift-marketplace/community-operators-bfwt6" Feb 02 10:11:45 crc kubenswrapper[4764]: I0202 10:11:45.809135 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hg8rf\" (UniqueName: \"kubernetes.io/projected/78b58f9f-3b0b-41a9-aed7-82341bbe9442-kube-api-access-hg8rf\") pod \"community-operators-bfwt6\" (UID: \"78b58f9f-3b0b-41a9-aed7-82341bbe9442\") " pod="openshift-marketplace/community-operators-bfwt6" Feb 02 10:11:45 crc kubenswrapper[4764]: I0202 10:11:45.902590 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bfwt6" Feb 02 10:11:46 crc kubenswrapper[4764]: I0202 10:11:46.427338 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-bfwt6"] Feb 02 10:11:46 crc kubenswrapper[4764]: I0202 10:11:46.775092 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bfwt6" event={"ID":"78b58f9f-3b0b-41a9-aed7-82341bbe9442","Type":"ContainerStarted","Data":"07d628a4f82fd8ca502bdbb01eb5ecd570d084a7608136b0f06961e52c5e4199"} Feb 02 10:11:47 crc kubenswrapper[4764]: I0202 10:11:47.785829 4764 generic.go:334] "Generic (PLEG): container finished" podID="78b58f9f-3b0b-41a9-aed7-82341bbe9442" containerID="841292066b9618bc445d091ef0bc4b4ffe6198c9e1ae7bc4c0504b3f090a83d3" exitCode=0 Feb 02 10:11:47 crc kubenswrapper[4764]: I0202 10:11:47.785884 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bfwt6" event={"ID":"78b58f9f-3b0b-41a9-aed7-82341bbe9442","Type":"ContainerDied","Data":"841292066b9618bc445d091ef0bc4b4ffe6198c9e1ae7bc4c0504b3f090a83d3"} Feb 02 10:11:47 crc kubenswrapper[4764]: I0202 10:11:47.787705 4764 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 02 10:11:49 crc kubenswrapper[4764]: I0202 10:11:49.805365 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bfwt6" event={"ID":"78b58f9f-3b0b-41a9-aed7-82341bbe9442","Type":"ContainerStarted","Data":"bb47742204fd368d9762a2e8f971a55fbd049f947334ed8ccae1adad8e1762d9"} Feb 02 10:11:53 crc kubenswrapper[4764]: I0202 10:11:53.826063 4764 scope.go:117] "RemoveContainer" containerID="4b1cf5e6a0ac319e2831449adb6329644931ba722d3592c4cde96fd7de3b310e" Feb 02 10:11:53 crc kubenswrapper[4764]: E0202 10:11:53.826808 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 10:11:55 crc kubenswrapper[4764]: I0202 10:11:55.867966 4764 generic.go:334] "Generic (PLEG): container finished" podID="78b58f9f-3b0b-41a9-aed7-82341bbe9442" containerID="bb47742204fd368d9762a2e8f971a55fbd049f947334ed8ccae1adad8e1762d9" exitCode=0 Feb 02 10:11:55 crc kubenswrapper[4764]: I0202 10:11:55.868065 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bfwt6" event={"ID":"78b58f9f-3b0b-41a9-aed7-82341bbe9442","Type":"ContainerDied","Data":"bb47742204fd368d9762a2e8f971a55fbd049f947334ed8ccae1adad8e1762d9"} Feb 02 10:11:57 crc kubenswrapper[4764]: I0202 10:11:57.884496 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bfwt6" event={"ID":"78b58f9f-3b0b-41a9-aed7-82341bbe9442","Type":"ContainerStarted","Data":"fb4157c8abe808c4595949480710a1f86c196baccc3d9ec642ea8b1ac3c11fbf"} Feb 02 10:11:57 crc kubenswrapper[4764]: I0202 10:11:57.910663 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-bfwt6" podStartSLOduration=4.445771309 podStartE2EDuration="12.910645175s" podCreationTimestamp="2026-02-02 10:11:45 +0000 UTC" firstStartedPulling="2026-02-02 10:11:47.787499845 +0000 UTC m=+3870.721223933" lastFinishedPulling="2026-02-02 10:11:56.252373701 +0000 UTC m=+3879.186097799" observedRunningTime="2026-02-02 10:11:57.904805993 +0000 UTC m=+3880.838530081" watchObservedRunningTime="2026-02-02 10:11:57.910645175 +0000 UTC m=+3880.844369263" Feb 02 10:12:05 crc kubenswrapper[4764]: I0202 10:12:05.903349 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-bfwt6" Feb 02 10:12:05 crc kubenswrapper[4764]: I0202 10:12:05.904165 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-bfwt6" Feb 02 10:12:05 crc kubenswrapper[4764]: I0202 10:12:05.963263 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-bfwt6" Feb 02 10:12:06 crc kubenswrapper[4764]: I0202 10:12:06.020102 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-bfwt6" Feb 02 10:12:06 crc kubenswrapper[4764]: I0202 10:12:06.212355 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-bfwt6"] Feb 02 10:12:06 crc kubenswrapper[4764]: I0202 10:12:06.826396 4764 scope.go:117] "RemoveContainer" containerID="4b1cf5e6a0ac319e2831449adb6329644931ba722d3592c4cde96fd7de3b310e" Feb 02 10:12:06 crc kubenswrapper[4764]: E0202 10:12:06.826889 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 10:12:07 crc kubenswrapper[4764]: I0202 10:12:07.976890 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-bfwt6" podUID="78b58f9f-3b0b-41a9-aed7-82341bbe9442" containerName="registry-server" containerID="cri-o://fb4157c8abe808c4595949480710a1f86c196baccc3d9ec642ea8b1ac3c11fbf" gracePeriod=2 Feb 02 10:12:08 crc kubenswrapper[4764]: I0202 10:12:08.581561 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bfwt6" Feb 02 10:12:08 crc kubenswrapper[4764]: I0202 10:12:08.677718 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/78b58f9f-3b0b-41a9-aed7-82341bbe9442-catalog-content\") pod \"78b58f9f-3b0b-41a9-aed7-82341bbe9442\" (UID: \"78b58f9f-3b0b-41a9-aed7-82341bbe9442\") " Feb 02 10:12:08 crc kubenswrapper[4764]: I0202 10:12:08.677848 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/78b58f9f-3b0b-41a9-aed7-82341bbe9442-utilities\") pod \"78b58f9f-3b0b-41a9-aed7-82341bbe9442\" (UID: \"78b58f9f-3b0b-41a9-aed7-82341bbe9442\") " Feb 02 10:12:08 crc kubenswrapper[4764]: I0202 10:12:08.677908 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hg8rf\" (UniqueName: \"kubernetes.io/projected/78b58f9f-3b0b-41a9-aed7-82341bbe9442-kube-api-access-hg8rf\") pod \"78b58f9f-3b0b-41a9-aed7-82341bbe9442\" (UID: \"78b58f9f-3b0b-41a9-aed7-82341bbe9442\") " Feb 02 10:12:08 crc kubenswrapper[4764]: I0202 10:12:08.678711 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/78b58f9f-3b0b-41a9-aed7-82341bbe9442-utilities" (OuterVolumeSpecName: "utilities") pod "78b58f9f-3b0b-41a9-aed7-82341bbe9442" (UID: "78b58f9f-3b0b-41a9-aed7-82341bbe9442"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 10:12:08 crc kubenswrapper[4764]: I0202 10:12:08.683793 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/78b58f9f-3b0b-41a9-aed7-82341bbe9442-kube-api-access-hg8rf" (OuterVolumeSpecName: "kube-api-access-hg8rf") pod "78b58f9f-3b0b-41a9-aed7-82341bbe9442" (UID: "78b58f9f-3b0b-41a9-aed7-82341bbe9442"). InnerVolumeSpecName "kube-api-access-hg8rf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 10:12:08 crc kubenswrapper[4764]: I0202 10:12:08.751072 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/78b58f9f-3b0b-41a9-aed7-82341bbe9442-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "78b58f9f-3b0b-41a9-aed7-82341bbe9442" (UID: "78b58f9f-3b0b-41a9-aed7-82341bbe9442"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 10:12:08 crc kubenswrapper[4764]: I0202 10:12:08.781147 4764 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/78b58f9f-3b0b-41a9-aed7-82341bbe9442-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 10:12:08 crc kubenswrapper[4764]: I0202 10:12:08.781186 4764 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/78b58f9f-3b0b-41a9-aed7-82341bbe9442-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 10:12:08 crc kubenswrapper[4764]: I0202 10:12:08.781203 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hg8rf\" (UniqueName: \"kubernetes.io/projected/78b58f9f-3b0b-41a9-aed7-82341bbe9442-kube-api-access-hg8rf\") on node \"crc\" DevicePath \"\"" Feb 02 10:12:09 crc kubenswrapper[4764]: I0202 10:12:08.997021 4764 generic.go:334] "Generic (PLEG): container finished" podID="78b58f9f-3b0b-41a9-aed7-82341bbe9442" containerID="fb4157c8abe808c4595949480710a1f86c196baccc3d9ec642ea8b1ac3c11fbf" exitCode=0 Feb 02 10:12:09 crc kubenswrapper[4764]: I0202 10:12:08.997073 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bfwt6" event={"ID":"78b58f9f-3b0b-41a9-aed7-82341bbe9442","Type":"ContainerDied","Data":"fb4157c8abe808c4595949480710a1f86c196baccc3d9ec642ea8b1ac3c11fbf"} Feb 02 10:12:09 crc kubenswrapper[4764]: I0202 10:12:08.997115 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bfwt6" event={"ID":"78b58f9f-3b0b-41a9-aed7-82341bbe9442","Type":"ContainerDied","Data":"07d628a4f82fd8ca502bdbb01eb5ecd570d084a7608136b0f06961e52c5e4199"} Feb 02 10:12:09 crc kubenswrapper[4764]: I0202 10:12:08.997115 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bfwt6" Feb 02 10:12:09 crc kubenswrapper[4764]: I0202 10:12:08.997134 4764 scope.go:117] "RemoveContainer" containerID="fb4157c8abe808c4595949480710a1f86c196baccc3d9ec642ea8b1ac3c11fbf" Feb 02 10:12:09 crc kubenswrapper[4764]: I0202 10:12:09.039692 4764 scope.go:117] "RemoveContainer" containerID="bb47742204fd368d9762a2e8f971a55fbd049f947334ed8ccae1adad8e1762d9" Feb 02 10:12:09 crc kubenswrapper[4764]: I0202 10:12:09.050150 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-bfwt6"] Feb 02 10:12:09 crc kubenswrapper[4764]: I0202 10:12:09.072597 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-bfwt6"] Feb 02 10:12:09 crc kubenswrapper[4764]: I0202 10:12:09.079734 4764 scope.go:117] "RemoveContainer" containerID="841292066b9618bc445d091ef0bc4b4ffe6198c9e1ae7bc4c0504b3f090a83d3" Feb 02 10:12:09 crc kubenswrapper[4764]: I0202 10:12:09.111843 4764 scope.go:117] "RemoveContainer" containerID="fb4157c8abe808c4595949480710a1f86c196baccc3d9ec642ea8b1ac3c11fbf" Feb 02 10:12:09 crc kubenswrapper[4764]: E0202 10:12:09.112298 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fb4157c8abe808c4595949480710a1f86c196baccc3d9ec642ea8b1ac3c11fbf\": container with ID starting with fb4157c8abe808c4595949480710a1f86c196baccc3d9ec642ea8b1ac3c11fbf not found: ID does not exist" containerID="fb4157c8abe808c4595949480710a1f86c196baccc3d9ec642ea8b1ac3c11fbf" Feb 02 10:12:09 crc kubenswrapper[4764]: I0202 10:12:09.112338 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fb4157c8abe808c4595949480710a1f86c196baccc3d9ec642ea8b1ac3c11fbf"} err="failed to get container status \"fb4157c8abe808c4595949480710a1f86c196baccc3d9ec642ea8b1ac3c11fbf\": rpc error: code = NotFound desc = could not find container \"fb4157c8abe808c4595949480710a1f86c196baccc3d9ec642ea8b1ac3c11fbf\": container with ID starting with fb4157c8abe808c4595949480710a1f86c196baccc3d9ec642ea8b1ac3c11fbf not found: ID does not exist" Feb 02 10:12:09 crc kubenswrapper[4764]: I0202 10:12:09.112365 4764 scope.go:117] "RemoveContainer" containerID="bb47742204fd368d9762a2e8f971a55fbd049f947334ed8ccae1adad8e1762d9" Feb 02 10:12:09 crc kubenswrapper[4764]: E0202 10:12:09.112738 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bb47742204fd368d9762a2e8f971a55fbd049f947334ed8ccae1adad8e1762d9\": container with ID starting with bb47742204fd368d9762a2e8f971a55fbd049f947334ed8ccae1adad8e1762d9 not found: ID does not exist" containerID="bb47742204fd368d9762a2e8f971a55fbd049f947334ed8ccae1adad8e1762d9" Feb 02 10:12:09 crc kubenswrapper[4764]: I0202 10:12:09.112778 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bb47742204fd368d9762a2e8f971a55fbd049f947334ed8ccae1adad8e1762d9"} err="failed to get container status \"bb47742204fd368d9762a2e8f971a55fbd049f947334ed8ccae1adad8e1762d9\": rpc error: code = NotFound desc = could not find container \"bb47742204fd368d9762a2e8f971a55fbd049f947334ed8ccae1adad8e1762d9\": container with ID starting with bb47742204fd368d9762a2e8f971a55fbd049f947334ed8ccae1adad8e1762d9 not found: ID does not exist" Feb 02 10:12:09 crc kubenswrapper[4764]: I0202 10:12:09.112800 4764 scope.go:117] "RemoveContainer" containerID="841292066b9618bc445d091ef0bc4b4ffe6198c9e1ae7bc4c0504b3f090a83d3" Feb 02 10:12:09 crc kubenswrapper[4764]: E0202 10:12:09.113021 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"841292066b9618bc445d091ef0bc4b4ffe6198c9e1ae7bc4c0504b3f090a83d3\": container with ID starting with 841292066b9618bc445d091ef0bc4b4ffe6198c9e1ae7bc4c0504b3f090a83d3 not found: ID does not exist" containerID="841292066b9618bc445d091ef0bc4b4ffe6198c9e1ae7bc4c0504b3f090a83d3" Feb 02 10:12:09 crc kubenswrapper[4764]: I0202 10:12:09.113044 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"841292066b9618bc445d091ef0bc4b4ffe6198c9e1ae7bc4c0504b3f090a83d3"} err="failed to get container status \"841292066b9618bc445d091ef0bc4b4ffe6198c9e1ae7bc4c0504b3f090a83d3\": rpc error: code = NotFound desc = could not find container \"841292066b9618bc445d091ef0bc4b4ffe6198c9e1ae7bc4c0504b3f090a83d3\": container with ID starting with 841292066b9618bc445d091ef0bc4b4ffe6198c9e1ae7bc4c0504b3f090a83d3 not found: ID does not exist" Feb 02 10:12:09 crc kubenswrapper[4764]: I0202 10:12:09.838244 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="78b58f9f-3b0b-41a9-aed7-82341bbe9442" path="/var/lib/kubelet/pods/78b58f9f-3b0b-41a9-aed7-82341bbe9442/volumes" Feb 02 10:12:21 crc kubenswrapper[4764]: I0202 10:12:21.825174 4764 scope.go:117] "RemoveContainer" containerID="4b1cf5e6a0ac319e2831449adb6329644931ba722d3592c4cde96fd7de3b310e" Feb 02 10:12:21 crc kubenswrapper[4764]: E0202 10:12:21.825870 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 10:12:34 crc kubenswrapper[4764]: I0202 10:12:34.825630 4764 scope.go:117] "RemoveContainer" containerID="4b1cf5e6a0ac319e2831449adb6329644931ba722d3592c4cde96fd7de3b310e" Feb 02 10:12:34 crc kubenswrapper[4764]: E0202 10:12:34.826609 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 10:12:49 crc kubenswrapper[4764]: I0202 10:12:49.826241 4764 scope.go:117] "RemoveContainer" containerID="4b1cf5e6a0ac319e2831449adb6329644931ba722d3592c4cde96fd7de3b310e" Feb 02 10:12:50 crc kubenswrapper[4764]: I0202 10:12:50.408652 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" event={"ID":"d192f670-9f9d-4539-9641-e4bed73acdd4","Type":"ContainerStarted","Data":"ee21cc1574aec6b1e483445c9c7de0962b1f3fa4c95e8c7525277156a9bf1c39"} Feb 02 10:13:20 crc kubenswrapper[4764]: I0202 10:13:20.711654 4764 generic.go:334] "Generic (PLEG): container finished" podID="09adc919-28cf-4d2d-b40d-183a2f51f06f" containerID="198eecad7865982369d75b72c24fc8ba50d3f10a9f576c3f0ebabf9a1cb73e22" exitCode=0 Feb 02 10:13:20 crc kubenswrapper[4764]: I0202 10:13:20.711736 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"09adc919-28cf-4d2d-b40d-183a2f51f06f","Type":"ContainerDied","Data":"198eecad7865982369d75b72c24fc8ba50d3f10a9f576c3f0ebabf9a1cb73e22"} Feb 02 10:13:22 crc kubenswrapper[4764]: I0202 10:13:22.148112 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Feb 02 10:13:22 crc kubenswrapper[4764]: I0202 10:13:22.265435 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/09adc919-28cf-4d2d-b40d-183a2f51f06f-openstack-config\") pod \"09adc919-28cf-4d2d-b40d-183a2f51f06f\" (UID: \"09adc919-28cf-4d2d-b40d-183a2f51f06f\") " Feb 02 10:13:22 crc kubenswrapper[4764]: I0202 10:13:22.265514 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mjdsp\" (UniqueName: \"kubernetes.io/projected/09adc919-28cf-4d2d-b40d-183a2f51f06f-kube-api-access-mjdsp\") pod \"09adc919-28cf-4d2d-b40d-183a2f51f06f\" (UID: \"09adc919-28cf-4d2d-b40d-183a2f51f06f\") " Feb 02 10:13:22 crc kubenswrapper[4764]: I0202 10:13:22.265651 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/09adc919-28cf-4d2d-b40d-183a2f51f06f-test-operator-ephemeral-workdir\") pod \"09adc919-28cf-4d2d-b40d-183a2f51f06f\" (UID: \"09adc919-28cf-4d2d-b40d-183a2f51f06f\") " Feb 02 10:13:22 crc kubenswrapper[4764]: I0202 10:13:22.269180 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/09adc919-28cf-4d2d-b40d-183a2f51f06f-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "09adc919-28cf-4d2d-b40d-183a2f51f06f" (UID: "09adc919-28cf-4d2d-b40d-183a2f51f06f"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 10:13:22 crc kubenswrapper[4764]: I0202 10:13:22.269259 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/09adc919-28cf-4d2d-b40d-183a2f51f06f-openstack-config-secret\") pod \"09adc919-28cf-4d2d-b40d-183a2f51f06f\" (UID: \"09adc919-28cf-4d2d-b40d-183a2f51f06f\") " Feb 02 10:13:22 crc kubenswrapper[4764]: I0202 10:13:22.269593 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/09adc919-28cf-4d2d-b40d-183a2f51f06f-config-data\") pod \"09adc919-28cf-4d2d-b40d-183a2f51f06f\" (UID: \"09adc919-28cf-4d2d-b40d-183a2f51f06f\") " Feb 02 10:13:22 crc kubenswrapper[4764]: I0202 10:13:22.269667 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/09adc919-28cf-4d2d-b40d-183a2f51f06f-test-operator-ephemeral-temporary\") pod \"09adc919-28cf-4d2d-b40d-183a2f51f06f\" (UID: \"09adc919-28cf-4d2d-b40d-183a2f51f06f\") " Feb 02 10:13:22 crc kubenswrapper[4764]: I0202 10:13:22.269739 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/09adc919-28cf-4d2d-b40d-183a2f51f06f-ssh-key\") pod \"09adc919-28cf-4d2d-b40d-183a2f51f06f\" (UID: \"09adc919-28cf-4d2d-b40d-183a2f51f06f\") " Feb 02 10:13:22 crc kubenswrapper[4764]: I0202 10:13:22.269786 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/09adc919-28cf-4d2d-b40d-183a2f51f06f-ca-certs\") pod \"09adc919-28cf-4d2d-b40d-183a2f51f06f\" (UID: \"09adc919-28cf-4d2d-b40d-183a2f51f06f\") " Feb 02 10:13:22 crc kubenswrapper[4764]: I0202 10:13:22.270148 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/09adc919-28cf-4d2d-b40d-183a2f51f06f-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "09adc919-28cf-4d2d-b40d-183a2f51f06f" (UID: "09adc919-28cf-4d2d-b40d-183a2f51f06f"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 10:13:22 crc kubenswrapper[4764]: I0202 10:13:22.270201 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09adc919-28cf-4d2d-b40d-183a2f51f06f-config-data" (OuterVolumeSpecName: "config-data") pod "09adc919-28cf-4d2d-b40d-183a2f51f06f" (UID: "09adc919-28cf-4d2d-b40d-183a2f51f06f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 10:13:22 crc kubenswrapper[4764]: I0202 10:13:22.270910 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"09adc919-28cf-4d2d-b40d-183a2f51f06f\" (UID: \"09adc919-28cf-4d2d-b40d-183a2f51f06f\") " Feb 02 10:13:22 crc kubenswrapper[4764]: I0202 10:13:22.274270 4764 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/09adc919-28cf-4d2d-b40d-183a2f51f06f-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Feb 02 10:13:22 crc kubenswrapper[4764]: I0202 10:13:22.274465 4764 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/09adc919-28cf-4d2d-b40d-183a2f51f06f-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 10:13:22 crc kubenswrapper[4764]: I0202 10:13:22.279231 4764 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/09adc919-28cf-4d2d-b40d-183a2f51f06f-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Feb 02 10:13:22 crc kubenswrapper[4764]: I0202 10:13:22.275188 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09adc919-28cf-4d2d-b40d-183a2f51f06f-kube-api-access-mjdsp" (OuterVolumeSpecName: "kube-api-access-mjdsp") pod "09adc919-28cf-4d2d-b40d-183a2f51f06f" (UID: "09adc919-28cf-4d2d-b40d-183a2f51f06f"). InnerVolumeSpecName "kube-api-access-mjdsp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 10:13:22 crc kubenswrapper[4764]: I0202 10:13:22.275590 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "test-operator-logs") pod "09adc919-28cf-4d2d-b40d-183a2f51f06f" (UID: "09adc919-28cf-4d2d-b40d-183a2f51f06f"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 02 10:13:22 crc kubenswrapper[4764]: I0202 10:13:22.298287 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09adc919-28cf-4d2d-b40d-183a2f51f06f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "09adc919-28cf-4d2d-b40d-183a2f51f06f" (UID: "09adc919-28cf-4d2d-b40d-183a2f51f06f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 10:13:22 crc kubenswrapper[4764]: I0202 10:13:22.312426 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09adc919-28cf-4d2d-b40d-183a2f51f06f-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "09adc919-28cf-4d2d-b40d-183a2f51f06f" (UID: "09adc919-28cf-4d2d-b40d-183a2f51f06f"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 10:13:22 crc kubenswrapper[4764]: I0202 10:13:22.317616 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09adc919-28cf-4d2d-b40d-183a2f51f06f-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "09adc919-28cf-4d2d-b40d-183a2f51f06f" (UID: "09adc919-28cf-4d2d-b40d-183a2f51f06f"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 10:13:22 crc kubenswrapper[4764]: I0202 10:13:22.322825 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09adc919-28cf-4d2d-b40d-183a2f51f06f-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "09adc919-28cf-4d2d-b40d-183a2f51f06f" (UID: "09adc919-28cf-4d2d-b40d-183a2f51f06f"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 10:13:22 crc kubenswrapper[4764]: I0202 10:13:22.384912 4764 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/09adc919-28cf-4d2d-b40d-183a2f51f06f-openstack-config\") on node \"crc\" DevicePath \"\"" Feb 02 10:13:22 crc kubenswrapper[4764]: I0202 10:13:22.384962 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mjdsp\" (UniqueName: \"kubernetes.io/projected/09adc919-28cf-4d2d-b40d-183a2f51f06f-kube-api-access-mjdsp\") on node \"crc\" DevicePath \"\"" Feb 02 10:13:22 crc kubenswrapper[4764]: I0202 10:13:22.384975 4764 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/09adc919-28cf-4d2d-b40d-183a2f51f06f-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Feb 02 10:13:22 crc kubenswrapper[4764]: I0202 10:13:22.384987 4764 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/09adc919-28cf-4d2d-b40d-183a2f51f06f-ssh-key\") on node \"crc\" DevicePath \"\"" Feb 02 10:13:22 crc kubenswrapper[4764]: I0202 10:13:22.384998 4764 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/09adc919-28cf-4d2d-b40d-183a2f51f06f-ca-certs\") on node \"crc\" DevicePath \"\"" Feb 02 10:13:22 crc kubenswrapper[4764]: I0202 10:13:22.385623 4764 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Feb 02 10:13:22 crc kubenswrapper[4764]: I0202 10:13:22.404575 4764 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Feb 02 10:13:22 crc kubenswrapper[4764]: I0202 10:13:22.487610 4764 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Feb 02 10:13:22 crc kubenswrapper[4764]: I0202 10:13:22.730836 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"09adc919-28cf-4d2d-b40d-183a2f51f06f","Type":"ContainerDied","Data":"d58a3a44bda779d0463c24ceb60684b3b9fe76a68689ebe918bb60483db8fc8a"} Feb 02 10:13:22 crc kubenswrapper[4764]: I0202 10:13:22.730914 4764 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d58a3a44bda779d0463c24ceb60684b3b9fe76a68689ebe918bb60483db8fc8a" Feb 02 10:13:22 crc kubenswrapper[4764]: I0202 10:13:22.731001 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Feb 02 10:13:28 crc kubenswrapper[4764]: I0202 10:13:28.720187 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Feb 02 10:13:28 crc kubenswrapper[4764]: E0202 10:13:28.721174 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09adc919-28cf-4d2d-b40d-183a2f51f06f" containerName="tempest-tests-tempest-tests-runner" Feb 02 10:13:28 crc kubenswrapper[4764]: I0202 10:13:28.721192 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="09adc919-28cf-4d2d-b40d-183a2f51f06f" containerName="tempest-tests-tempest-tests-runner" Feb 02 10:13:28 crc kubenswrapper[4764]: E0202 10:13:28.721215 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78b58f9f-3b0b-41a9-aed7-82341bbe9442" containerName="extract-content" Feb 02 10:13:28 crc kubenswrapper[4764]: I0202 10:13:28.721226 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="78b58f9f-3b0b-41a9-aed7-82341bbe9442" containerName="extract-content" Feb 02 10:13:28 crc kubenswrapper[4764]: E0202 10:13:28.721237 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78b58f9f-3b0b-41a9-aed7-82341bbe9442" containerName="registry-server" Feb 02 10:13:28 crc kubenswrapper[4764]: I0202 10:13:28.721245 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="78b58f9f-3b0b-41a9-aed7-82341bbe9442" containerName="registry-server" Feb 02 10:13:28 crc kubenswrapper[4764]: E0202 10:13:28.721268 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78b58f9f-3b0b-41a9-aed7-82341bbe9442" containerName="extract-utilities" Feb 02 10:13:28 crc kubenswrapper[4764]: I0202 10:13:28.721276 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="78b58f9f-3b0b-41a9-aed7-82341bbe9442" containerName="extract-utilities" Feb 02 10:13:28 crc kubenswrapper[4764]: I0202 10:13:28.721514 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="09adc919-28cf-4d2d-b40d-183a2f51f06f" containerName="tempest-tests-tempest-tests-runner" Feb 02 10:13:28 crc kubenswrapper[4764]: I0202 10:13:28.721530 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="78b58f9f-3b0b-41a9-aed7-82341bbe9442" containerName="registry-server" Feb 02 10:13:28 crc kubenswrapper[4764]: I0202 10:13:28.723878 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 02 10:13:28 crc kubenswrapper[4764]: I0202 10:13:28.731737 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-kgk2v" Feb 02 10:13:28 crc kubenswrapper[4764]: I0202 10:13:28.738308 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Feb 02 10:13:28 crc kubenswrapper[4764]: I0202 10:13:28.813235 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v7q4k\" (UniqueName: \"kubernetes.io/projected/7576fa25-4c66-4f89-90a5-c2979b1c0eaa-kube-api-access-v7q4k\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"7576fa25-4c66-4f89-90a5-c2979b1c0eaa\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 02 10:13:28 crc kubenswrapper[4764]: I0202 10:13:28.813406 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"7576fa25-4c66-4f89-90a5-c2979b1c0eaa\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 02 10:13:28 crc kubenswrapper[4764]: I0202 10:13:28.915432 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v7q4k\" (UniqueName: \"kubernetes.io/projected/7576fa25-4c66-4f89-90a5-c2979b1c0eaa-kube-api-access-v7q4k\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"7576fa25-4c66-4f89-90a5-c2979b1c0eaa\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 02 10:13:28 crc kubenswrapper[4764]: I0202 10:13:28.915622 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"7576fa25-4c66-4f89-90a5-c2979b1c0eaa\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 02 10:13:28 crc kubenswrapper[4764]: I0202 10:13:28.916133 4764 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"7576fa25-4c66-4f89-90a5-c2979b1c0eaa\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 02 10:13:28 crc kubenswrapper[4764]: I0202 10:13:28.940129 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v7q4k\" (UniqueName: \"kubernetes.io/projected/7576fa25-4c66-4f89-90a5-c2979b1c0eaa-kube-api-access-v7q4k\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"7576fa25-4c66-4f89-90a5-c2979b1c0eaa\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 02 10:13:28 crc kubenswrapper[4764]: I0202 10:13:28.943030 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"7576fa25-4c66-4f89-90a5-c2979b1c0eaa\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 02 10:13:29 crc kubenswrapper[4764]: I0202 10:13:29.051013 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 02 10:13:29 crc kubenswrapper[4764]: I0202 10:13:29.527038 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Feb 02 10:13:29 crc kubenswrapper[4764]: I0202 10:13:29.788231 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"7576fa25-4c66-4f89-90a5-c2979b1c0eaa","Type":"ContainerStarted","Data":"f5f1e3b83aec104b647226e19af6a25604be0c9b5f42cd7ff15afc694d99b40b"} Feb 02 10:13:31 crc kubenswrapper[4764]: I0202 10:13:31.901629 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"7576fa25-4c66-4f89-90a5-c2979b1c0eaa","Type":"ContainerStarted","Data":"7d3896b8540db509531e558c7bf793ee35daf1d0c78c1832185f1b09cbc2f380"} Feb 02 10:13:31 crc kubenswrapper[4764]: I0202 10:13:31.931837 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=2.858714063 podStartE2EDuration="3.931821057s" podCreationTimestamp="2026-02-02 10:13:28 +0000 UTC" firstStartedPulling="2026-02-02 10:13:29.544954615 +0000 UTC m=+3972.478678703" lastFinishedPulling="2026-02-02 10:13:30.618061609 +0000 UTC m=+3973.551785697" observedRunningTime="2026-02-02 10:13:31.919670372 +0000 UTC m=+3974.853394460" watchObservedRunningTime="2026-02-02 10:13:31.931821057 +0000 UTC m=+3974.865545145" Feb 02 10:13:54 crc kubenswrapper[4764]: I0202 10:13:54.370366 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-tvk74/must-gather-4qvng"] Feb 02 10:13:54 crc kubenswrapper[4764]: I0202 10:13:54.372245 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-tvk74/must-gather-4qvng" Feb 02 10:13:54 crc kubenswrapper[4764]: I0202 10:13:54.375117 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-tvk74"/"kube-root-ca.crt" Feb 02 10:13:54 crc kubenswrapper[4764]: I0202 10:13:54.377149 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-tvk74"/"openshift-service-ca.crt" Feb 02 10:13:54 crc kubenswrapper[4764]: I0202 10:13:54.377894 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-tvk74"/"default-dockercfg-bbb76" Feb 02 10:13:54 crc kubenswrapper[4764]: I0202 10:13:54.397422 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-tvk74/must-gather-4qvng"] Feb 02 10:13:54 crc kubenswrapper[4764]: I0202 10:13:54.506445 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-njs28\" (UniqueName: \"kubernetes.io/projected/bc598811-a908-46d9-aadc-073b575a864d-kube-api-access-njs28\") pod \"must-gather-4qvng\" (UID: \"bc598811-a908-46d9-aadc-073b575a864d\") " pod="openshift-must-gather-tvk74/must-gather-4qvng" Feb 02 10:13:54 crc kubenswrapper[4764]: I0202 10:13:54.506490 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/bc598811-a908-46d9-aadc-073b575a864d-must-gather-output\") pod \"must-gather-4qvng\" (UID: \"bc598811-a908-46d9-aadc-073b575a864d\") " pod="openshift-must-gather-tvk74/must-gather-4qvng" Feb 02 10:13:54 crc kubenswrapper[4764]: I0202 10:13:54.608376 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-njs28\" (UniqueName: \"kubernetes.io/projected/bc598811-a908-46d9-aadc-073b575a864d-kube-api-access-njs28\") pod \"must-gather-4qvng\" (UID: \"bc598811-a908-46d9-aadc-073b575a864d\") " pod="openshift-must-gather-tvk74/must-gather-4qvng" Feb 02 10:13:54 crc kubenswrapper[4764]: I0202 10:13:54.608420 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/bc598811-a908-46d9-aadc-073b575a864d-must-gather-output\") pod \"must-gather-4qvng\" (UID: \"bc598811-a908-46d9-aadc-073b575a864d\") " pod="openshift-must-gather-tvk74/must-gather-4qvng" Feb 02 10:13:54 crc kubenswrapper[4764]: I0202 10:13:54.609059 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/bc598811-a908-46d9-aadc-073b575a864d-must-gather-output\") pod \"must-gather-4qvng\" (UID: \"bc598811-a908-46d9-aadc-073b575a864d\") " pod="openshift-must-gather-tvk74/must-gather-4qvng" Feb 02 10:13:54 crc kubenswrapper[4764]: I0202 10:13:54.628172 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-njs28\" (UniqueName: \"kubernetes.io/projected/bc598811-a908-46d9-aadc-073b575a864d-kube-api-access-njs28\") pod \"must-gather-4qvng\" (UID: \"bc598811-a908-46d9-aadc-073b575a864d\") " pod="openshift-must-gather-tvk74/must-gather-4qvng" Feb 02 10:13:54 crc kubenswrapper[4764]: I0202 10:13:54.699461 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-tvk74/must-gather-4qvng" Feb 02 10:13:55 crc kubenswrapper[4764]: I0202 10:13:55.176281 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-tvk74/must-gather-4qvng"] Feb 02 10:13:55 crc kubenswrapper[4764]: W0202 10:13:55.188273 4764 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbc598811_a908_46d9_aadc_073b575a864d.slice/crio-ea07d975a552bdfd1c6a1d4aaf7b536c9866599a4aef53acf2d8757d2186e73e WatchSource:0}: Error finding container ea07d975a552bdfd1c6a1d4aaf7b536c9866599a4aef53acf2d8757d2186e73e: Status 404 returned error can't find the container with id ea07d975a552bdfd1c6a1d4aaf7b536c9866599a4aef53acf2d8757d2186e73e Feb 02 10:13:56 crc kubenswrapper[4764]: I0202 10:13:56.117341 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-tvk74/must-gather-4qvng" event={"ID":"bc598811-a908-46d9-aadc-073b575a864d","Type":"ContainerStarted","Data":"ea07d975a552bdfd1c6a1d4aaf7b536c9866599a4aef53acf2d8757d2186e73e"} Feb 02 10:14:01 crc kubenswrapper[4764]: I0202 10:14:01.192567 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-tvk74/must-gather-4qvng" event={"ID":"bc598811-a908-46d9-aadc-073b575a864d","Type":"ContainerStarted","Data":"b3b304bdbe9119c2d0bf8545eb25701dcaa008bc8ef7b700a2e20300165bb7d6"} Feb 02 10:14:01 crc kubenswrapper[4764]: I0202 10:14:01.193499 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-tvk74/must-gather-4qvng" event={"ID":"bc598811-a908-46d9-aadc-073b575a864d","Type":"ContainerStarted","Data":"f6d1ba99e6bbcb588b915ff9b5f81569c48a8feca4a9c9ace1f50a53385ea4f7"} Feb 02 10:14:01 crc kubenswrapper[4764]: I0202 10:14:01.221314 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-tvk74/must-gather-4qvng" podStartSLOduration=2.21733924 podStartE2EDuration="7.221274807s" podCreationTimestamp="2026-02-02 10:13:54 +0000 UTC" firstStartedPulling="2026-02-02 10:13:55.190360192 +0000 UTC m=+3998.124084280" lastFinishedPulling="2026-02-02 10:14:00.194295759 +0000 UTC m=+4003.128019847" observedRunningTime="2026-02-02 10:14:01.209532413 +0000 UTC m=+4004.143256521" watchObservedRunningTime="2026-02-02 10:14:01.221274807 +0000 UTC m=+4004.154998895" Feb 02 10:14:05 crc kubenswrapper[4764]: E0202 10:14:05.603695 4764 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.217:57158->38.102.83.217:34003: write tcp 38.102.83.217:57158->38.102.83.217:34003: write: broken pipe Feb 02 10:14:07 crc kubenswrapper[4764]: I0202 10:14:07.077213 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-tvk74/crc-debug-58t2z"] Feb 02 10:14:07 crc kubenswrapper[4764]: I0202 10:14:07.091838 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-tvk74/crc-debug-58t2z" Feb 02 10:14:07 crc kubenswrapper[4764]: I0202 10:14:07.191432 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3fc31820-38e5-42a5-824c-e0f62fe8484a-host\") pod \"crc-debug-58t2z\" (UID: \"3fc31820-38e5-42a5-824c-e0f62fe8484a\") " pod="openshift-must-gather-tvk74/crc-debug-58t2z" Feb 02 10:14:07 crc kubenswrapper[4764]: I0202 10:14:07.191725 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bdcvd\" (UniqueName: \"kubernetes.io/projected/3fc31820-38e5-42a5-824c-e0f62fe8484a-kube-api-access-bdcvd\") pod \"crc-debug-58t2z\" (UID: \"3fc31820-38e5-42a5-824c-e0f62fe8484a\") " pod="openshift-must-gather-tvk74/crc-debug-58t2z" Feb 02 10:14:07 crc kubenswrapper[4764]: I0202 10:14:07.293645 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bdcvd\" (UniqueName: \"kubernetes.io/projected/3fc31820-38e5-42a5-824c-e0f62fe8484a-kube-api-access-bdcvd\") pod \"crc-debug-58t2z\" (UID: \"3fc31820-38e5-42a5-824c-e0f62fe8484a\") " pod="openshift-must-gather-tvk74/crc-debug-58t2z" Feb 02 10:14:07 crc kubenswrapper[4764]: I0202 10:14:07.294043 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3fc31820-38e5-42a5-824c-e0f62fe8484a-host\") pod \"crc-debug-58t2z\" (UID: \"3fc31820-38e5-42a5-824c-e0f62fe8484a\") " pod="openshift-must-gather-tvk74/crc-debug-58t2z" Feb 02 10:14:07 crc kubenswrapper[4764]: I0202 10:14:07.294168 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3fc31820-38e5-42a5-824c-e0f62fe8484a-host\") pod \"crc-debug-58t2z\" (UID: \"3fc31820-38e5-42a5-824c-e0f62fe8484a\") " pod="openshift-must-gather-tvk74/crc-debug-58t2z" Feb 02 10:14:07 crc kubenswrapper[4764]: I0202 10:14:07.315675 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bdcvd\" (UniqueName: \"kubernetes.io/projected/3fc31820-38e5-42a5-824c-e0f62fe8484a-kube-api-access-bdcvd\") pod \"crc-debug-58t2z\" (UID: \"3fc31820-38e5-42a5-824c-e0f62fe8484a\") " pod="openshift-must-gather-tvk74/crc-debug-58t2z" Feb 02 10:14:07 crc kubenswrapper[4764]: I0202 10:14:07.413505 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-tvk74/crc-debug-58t2z" Feb 02 10:14:08 crc kubenswrapper[4764]: I0202 10:14:08.250234 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-tvk74/crc-debug-58t2z" event={"ID":"3fc31820-38e5-42a5-824c-e0f62fe8484a","Type":"ContainerStarted","Data":"23d7e1c76bd6c91c65f8519c77d6fd93ca4941c32e078810f728f008df56b1ef"} Feb 02 10:14:20 crc kubenswrapper[4764]: I0202 10:14:20.351672 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-tvk74/crc-debug-58t2z" event={"ID":"3fc31820-38e5-42a5-824c-e0f62fe8484a","Type":"ContainerStarted","Data":"edf39fcc0244cd782d701fc4743f48b2701ea66017d3e26a58694c71fa146a07"} Feb 02 10:14:20 crc kubenswrapper[4764]: I0202 10:14:20.365499 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-tvk74/crc-debug-58t2z" podStartSLOduration=0.91904056 podStartE2EDuration="13.365482924s" podCreationTimestamp="2026-02-02 10:14:07 +0000 UTC" firstStartedPulling="2026-02-02 10:14:07.466892885 +0000 UTC m=+4010.400616973" lastFinishedPulling="2026-02-02 10:14:19.913335249 +0000 UTC m=+4022.847059337" observedRunningTime="2026-02-02 10:14:20.361301386 +0000 UTC m=+4023.295025474" watchObservedRunningTime="2026-02-02 10:14:20.365482924 +0000 UTC m=+4023.299207012" Feb 02 10:14:34 crc kubenswrapper[4764]: I0202 10:14:34.268838 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-kzw9f"] Feb 02 10:14:34 crc kubenswrapper[4764]: I0202 10:14:34.271775 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kzw9f" Feb 02 10:14:34 crc kubenswrapper[4764]: I0202 10:14:34.333719 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kzw9f"] Feb 02 10:14:34 crc kubenswrapper[4764]: I0202 10:14:34.393344 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/df71a15a-b67b-43b7-9487-7723c62304f3-catalog-content\") pod \"redhat-operators-kzw9f\" (UID: \"df71a15a-b67b-43b7-9487-7723c62304f3\") " pod="openshift-marketplace/redhat-operators-kzw9f" Feb 02 10:14:34 crc kubenswrapper[4764]: I0202 10:14:34.393395 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c8gtp\" (UniqueName: \"kubernetes.io/projected/df71a15a-b67b-43b7-9487-7723c62304f3-kube-api-access-c8gtp\") pod \"redhat-operators-kzw9f\" (UID: \"df71a15a-b67b-43b7-9487-7723c62304f3\") " pod="openshift-marketplace/redhat-operators-kzw9f" Feb 02 10:14:34 crc kubenswrapper[4764]: I0202 10:14:34.393492 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/df71a15a-b67b-43b7-9487-7723c62304f3-utilities\") pod \"redhat-operators-kzw9f\" (UID: \"df71a15a-b67b-43b7-9487-7723c62304f3\") " pod="openshift-marketplace/redhat-operators-kzw9f" Feb 02 10:14:34 crc kubenswrapper[4764]: I0202 10:14:34.496299 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/df71a15a-b67b-43b7-9487-7723c62304f3-utilities\") pod \"redhat-operators-kzw9f\" (UID: \"df71a15a-b67b-43b7-9487-7723c62304f3\") " pod="openshift-marketplace/redhat-operators-kzw9f" Feb 02 10:14:34 crc kubenswrapper[4764]: I0202 10:14:34.496616 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/df71a15a-b67b-43b7-9487-7723c62304f3-catalog-content\") pod \"redhat-operators-kzw9f\" (UID: \"df71a15a-b67b-43b7-9487-7723c62304f3\") " pod="openshift-marketplace/redhat-operators-kzw9f" Feb 02 10:14:34 crc kubenswrapper[4764]: I0202 10:14:34.496680 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c8gtp\" (UniqueName: \"kubernetes.io/projected/df71a15a-b67b-43b7-9487-7723c62304f3-kube-api-access-c8gtp\") pod \"redhat-operators-kzw9f\" (UID: \"df71a15a-b67b-43b7-9487-7723c62304f3\") " pod="openshift-marketplace/redhat-operators-kzw9f" Feb 02 10:14:34 crc kubenswrapper[4764]: I0202 10:14:34.497667 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/df71a15a-b67b-43b7-9487-7723c62304f3-utilities\") pod \"redhat-operators-kzw9f\" (UID: \"df71a15a-b67b-43b7-9487-7723c62304f3\") " pod="openshift-marketplace/redhat-operators-kzw9f" Feb 02 10:14:34 crc kubenswrapper[4764]: I0202 10:14:34.497890 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/df71a15a-b67b-43b7-9487-7723c62304f3-catalog-content\") pod \"redhat-operators-kzw9f\" (UID: \"df71a15a-b67b-43b7-9487-7723c62304f3\") " pod="openshift-marketplace/redhat-operators-kzw9f" Feb 02 10:14:34 crc kubenswrapper[4764]: I0202 10:14:34.529049 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c8gtp\" (UniqueName: \"kubernetes.io/projected/df71a15a-b67b-43b7-9487-7723c62304f3-kube-api-access-c8gtp\") pod \"redhat-operators-kzw9f\" (UID: \"df71a15a-b67b-43b7-9487-7723c62304f3\") " pod="openshift-marketplace/redhat-operators-kzw9f" Feb 02 10:14:34 crc kubenswrapper[4764]: I0202 10:14:34.599747 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kzw9f" Feb 02 10:14:35 crc kubenswrapper[4764]: I0202 10:14:35.315718 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kzw9f"] Feb 02 10:14:36 crc kubenswrapper[4764]: I0202 10:14:36.477468 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kzw9f" event={"ID":"df71a15a-b67b-43b7-9487-7723c62304f3","Type":"ContainerStarted","Data":"b90967997b1ebd0be40b048701adc909d5ad22c6fc2ef07ffdd0900dce639211"} Feb 02 10:14:37 crc kubenswrapper[4764]: I0202 10:14:37.489701 4764 generic.go:334] "Generic (PLEG): container finished" podID="df71a15a-b67b-43b7-9487-7723c62304f3" containerID="0bbfe02af0a90947f464f0389dcecfd894a7cbd8808dd6ee859fc48637a85192" exitCode=0 Feb 02 10:14:37 crc kubenswrapper[4764]: I0202 10:14:37.489805 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kzw9f" event={"ID":"df71a15a-b67b-43b7-9487-7723c62304f3","Type":"ContainerDied","Data":"0bbfe02af0a90947f464f0389dcecfd894a7cbd8808dd6ee859fc48637a85192"} Feb 02 10:14:39 crc kubenswrapper[4764]: I0202 10:14:39.507327 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kzw9f" event={"ID":"df71a15a-b67b-43b7-9487-7723c62304f3","Type":"ContainerStarted","Data":"a79d594f10382d3c2cbe01ccffbe09dc83810081c1b8f7a6640a6107d65175af"} Feb 02 10:14:44 crc kubenswrapper[4764]: I0202 10:14:44.573342 4764 generic.go:334] "Generic (PLEG): container finished" podID="df71a15a-b67b-43b7-9487-7723c62304f3" containerID="a79d594f10382d3c2cbe01ccffbe09dc83810081c1b8f7a6640a6107d65175af" exitCode=0 Feb 02 10:14:44 crc kubenswrapper[4764]: I0202 10:14:44.573475 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kzw9f" event={"ID":"df71a15a-b67b-43b7-9487-7723c62304f3","Type":"ContainerDied","Data":"a79d594f10382d3c2cbe01ccffbe09dc83810081c1b8f7a6640a6107d65175af"} Feb 02 10:14:45 crc kubenswrapper[4764]: I0202 10:14:45.594780 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kzw9f" event={"ID":"df71a15a-b67b-43b7-9487-7723c62304f3","Type":"ContainerStarted","Data":"cb8fe6cfe1faa91cbbea75028829c32ff9c6e8084d80efe3f05dae807ecd9215"} Feb 02 10:14:45 crc kubenswrapper[4764]: I0202 10:14:45.623044 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-kzw9f" podStartSLOduration=4.144437478 podStartE2EDuration="11.623021641s" podCreationTimestamp="2026-02-02 10:14:34 +0000 UTC" firstStartedPulling="2026-02-02 10:14:37.492820975 +0000 UTC m=+4040.426545063" lastFinishedPulling="2026-02-02 10:14:44.971405138 +0000 UTC m=+4047.905129226" observedRunningTime="2026-02-02 10:14:45.613244447 +0000 UTC m=+4048.546968535" watchObservedRunningTime="2026-02-02 10:14:45.623021641 +0000 UTC m=+4048.556745729" Feb 02 10:14:54 crc kubenswrapper[4764]: I0202 10:14:54.601012 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-kzw9f" Feb 02 10:14:54 crc kubenswrapper[4764]: I0202 10:14:54.601581 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-kzw9f" Feb 02 10:14:55 crc kubenswrapper[4764]: I0202 10:14:55.656662 4764 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-kzw9f" podUID="df71a15a-b67b-43b7-9487-7723c62304f3" containerName="registry-server" probeResult="failure" output=< Feb 02 10:14:55 crc kubenswrapper[4764]: timeout: failed to connect service ":50051" within 1s Feb 02 10:14:55 crc kubenswrapper[4764]: > Feb 02 10:15:00 crc kubenswrapper[4764]: I0202 10:15:00.203880 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29500455-h8j8c"] Feb 02 10:15:00 crc kubenswrapper[4764]: I0202 10:15:00.205530 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29500455-h8j8c" Feb 02 10:15:00 crc kubenswrapper[4764]: I0202 10:15:00.207281 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 02 10:15:00 crc kubenswrapper[4764]: I0202 10:15:00.209866 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 02 10:15:00 crc kubenswrapper[4764]: I0202 10:15:00.220417 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29500455-h8j8c"] Feb 02 10:15:00 crc kubenswrapper[4764]: I0202 10:15:00.227376 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9pw8r\" (UniqueName: \"kubernetes.io/projected/8d47415a-a9c4-4f36-9b76-2491f748505e-kube-api-access-9pw8r\") pod \"collect-profiles-29500455-h8j8c\" (UID: \"8d47415a-a9c4-4f36-9b76-2491f748505e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500455-h8j8c" Feb 02 10:15:00 crc kubenswrapper[4764]: I0202 10:15:00.227443 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8d47415a-a9c4-4f36-9b76-2491f748505e-secret-volume\") pod \"collect-profiles-29500455-h8j8c\" (UID: \"8d47415a-a9c4-4f36-9b76-2491f748505e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500455-h8j8c" Feb 02 10:15:00 crc kubenswrapper[4764]: I0202 10:15:00.227557 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8d47415a-a9c4-4f36-9b76-2491f748505e-config-volume\") pod \"collect-profiles-29500455-h8j8c\" (UID: \"8d47415a-a9c4-4f36-9b76-2491f748505e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500455-h8j8c" Feb 02 10:15:00 crc kubenswrapper[4764]: I0202 10:15:00.328854 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8d47415a-a9c4-4f36-9b76-2491f748505e-secret-volume\") pod \"collect-profiles-29500455-h8j8c\" (UID: \"8d47415a-a9c4-4f36-9b76-2491f748505e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500455-h8j8c" Feb 02 10:15:00 crc kubenswrapper[4764]: I0202 10:15:00.329178 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8d47415a-a9c4-4f36-9b76-2491f748505e-config-volume\") pod \"collect-profiles-29500455-h8j8c\" (UID: \"8d47415a-a9c4-4f36-9b76-2491f748505e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500455-h8j8c" Feb 02 10:15:00 crc kubenswrapper[4764]: I0202 10:15:00.329355 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9pw8r\" (UniqueName: \"kubernetes.io/projected/8d47415a-a9c4-4f36-9b76-2491f748505e-kube-api-access-9pw8r\") pod \"collect-profiles-29500455-h8j8c\" (UID: \"8d47415a-a9c4-4f36-9b76-2491f748505e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500455-h8j8c" Feb 02 10:15:00 crc kubenswrapper[4764]: I0202 10:15:00.330079 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8d47415a-a9c4-4f36-9b76-2491f748505e-config-volume\") pod \"collect-profiles-29500455-h8j8c\" (UID: \"8d47415a-a9c4-4f36-9b76-2491f748505e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500455-h8j8c" Feb 02 10:15:00 crc kubenswrapper[4764]: I0202 10:15:00.337038 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8d47415a-a9c4-4f36-9b76-2491f748505e-secret-volume\") pod \"collect-profiles-29500455-h8j8c\" (UID: \"8d47415a-a9c4-4f36-9b76-2491f748505e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500455-h8j8c" Feb 02 10:15:00 crc kubenswrapper[4764]: I0202 10:15:00.350778 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9pw8r\" (UniqueName: \"kubernetes.io/projected/8d47415a-a9c4-4f36-9b76-2491f748505e-kube-api-access-9pw8r\") pod \"collect-profiles-29500455-h8j8c\" (UID: \"8d47415a-a9c4-4f36-9b76-2491f748505e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500455-h8j8c" Feb 02 10:15:00 crc kubenswrapper[4764]: I0202 10:15:00.526796 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29500455-h8j8c" Feb 02 10:15:01 crc kubenswrapper[4764]: I0202 10:15:01.080992 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29500455-h8j8c"] Feb 02 10:15:01 crc kubenswrapper[4764]: I0202 10:15:01.722683 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29500455-h8j8c" event={"ID":"8d47415a-a9c4-4f36-9b76-2491f748505e","Type":"ContainerStarted","Data":"d30f136e2d473255d412cbce7feda330ac895b40ef2653da7a01d20b32594793"} Feb 02 10:15:01 crc kubenswrapper[4764]: I0202 10:15:01.723896 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29500455-h8j8c" event={"ID":"8d47415a-a9c4-4f36-9b76-2491f748505e","Type":"ContainerStarted","Data":"1de061ffb132211d3028b2bd822e0a80442e6d160d2ec83367e96700084a6d6e"} Feb 02 10:15:02 crc kubenswrapper[4764]: I0202 10:15:02.731541 4764 generic.go:334] "Generic (PLEG): container finished" podID="8d47415a-a9c4-4f36-9b76-2491f748505e" containerID="d30f136e2d473255d412cbce7feda330ac895b40ef2653da7a01d20b32594793" exitCode=0 Feb 02 10:15:02 crc kubenswrapper[4764]: I0202 10:15:02.731661 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29500455-h8j8c" event={"ID":"8d47415a-a9c4-4f36-9b76-2491f748505e","Type":"ContainerDied","Data":"d30f136e2d473255d412cbce7feda330ac895b40ef2653da7a01d20b32594793"} Feb 02 10:15:04 crc kubenswrapper[4764]: I0202 10:15:04.205170 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29500455-h8j8c" Feb 02 10:15:04 crc kubenswrapper[4764]: I0202 10:15:04.313644 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8d47415a-a9c4-4f36-9b76-2491f748505e-secret-volume\") pod \"8d47415a-a9c4-4f36-9b76-2491f748505e\" (UID: \"8d47415a-a9c4-4f36-9b76-2491f748505e\") " Feb 02 10:15:04 crc kubenswrapper[4764]: I0202 10:15:04.314316 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9pw8r\" (UniqueName: \"kubernetes.io/projected/8d47415a-a9c4-4f36-9b76-2491f748505e-kube-api-access-9pw8r\") pod \"8d47415a-a9c4-4f36-9b76-2491f748505e\" (UID: \"8d47415a-a9c4-4f36-9b76-2491f748505e\") " Feb 02 10:15:04 crc kubenswrapper[4764]: I0202 10:15:04.314403 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8d47415a-a9c4-4f36-9b76-2491f748505e-config-volume\") pod \"8d47415a-a9c4-4f36-9b76-2491f748505e\" (UID: \"8d47415a-a9c4-4f36-9b76-2491f748505e\") " Feb 02 10:15:04 crc kubenswrapper[4764]: I0202 10:15:04.315143 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8d47415a-a9c4-4f36-9b76-2491f748505e-config-volume" (OuterVolumeSpecName: "config-volume") pod "8d47415a-a9c4-4f36-9b76-2491f748505e" (UID: "8d47415a-a9c4-4f36-9b76-2491f748505e"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 10:15:04 crc kubenswrapper[4764]: I0202 10:15:04.320614 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8d47415a-a9c4-4f36-9b76-2491f748505e-kube-api-access-9pw8r" (OuterVolumeSpecName: "kube-api-access-9pw8r") pod "8d47415a-a9c4-4f36-9b76-2491f748505e" (UID: "8d47415a-a9c4-4f36-9b76-2491f748505e"). InnerVolumeSpecName "kube-api-access-9pw8r". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 10:15:04 crc kubenswrapper[4764]: I0202 10:15:04.322506 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d47415a-a9c4-4f36-9b76-2491f748505e-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "8d47415a-a9c4-4f36-9b76-2491f748505e" (UID: "8d47415a-a9c4-4f36-9b76-2491f748505e"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 10:15:04 crc kubenswrapper[4764]: I0202 10:15:04.416458 4764 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8d47415a-a9c4-4f36-9b76-2491f748505e-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 02 10:15:04 crc kubenswrapper[4764]: I0202 10:15:04.416490 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9pw8r\" (UniqueName: \"kubernetes.io/projected/8d47415a-a9c4-4f36-9b76-2491f748505e-kube-api-access-9pw8r\") on node \"crc\" DevicePath \"\"" Feb 02 10:15:04 crc kubenswrapper[4764]: I0202 10:15:04.416501 4764 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8d47415a-a9c4-4f36-9b76-2491f748505e-config-volume\") on node \"crc\" DevicePath \"\"" Feb 02 10:15:04 crc kubenswrapper[4764]: I0202 10:15:04.752370 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29500455-h8j8c" event={"ID":"8d47415a-a9c4-4f36-9b76-2491f748505e","Type":"ContainerDied","Data":"1de061ffb132211d3028b2bd822e0a80442e6d160d2ec83367e96700084a6d6e"} Feb 02 10:15:04 crc kubenswrapper[4764]: I0202 10:15:04.752445 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29500455-h8j8c" Feb 02 10:15:04 crc kubenswrapper[4764]: I0202 10:15:04.752450 4764 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1de061ffb132211d3028b2bd822e0a80442e6d160d2ec83367e96700084a6d6e" Feb 02 10:15:04 crc kubenswrapper[4764]: I0202 10:15:04.816816 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29500410-b8znn"] Feb 02 10:15:04 crc kubenswrapper[4764]: I0202 10:15:04.827013 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29500410-b8znn"] Feb 02 10:15:05 crc kubenswrapper[4764]: I0202 10:15:05.650127 4764 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-kzw9f" podUID="df71a15a-b67b-43b7-9487-7723c62304f3" containerName="registry-server" probeResult="failure" output=< Feb 02 10:15:05 crc kubenswrapper[4764]: timeout: failed to connect service ":50051" within 1s Feb 02 10:15:05 crc kubenswrapper[4764]: > Feb 02 10:15:05 crc kubenswrapper[4764]: I0202 10:15:05.841356 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fd431fc8-7193-4ad6-b7f6-eb83b4dbe824" path="/var/lib/kubelet/pods/fd431fc8-7193-4ad6-b7f6-eb83b4dbe824/volumes" Feb 02 10:15:06 crc kubenswrapper[4764]: I0202 10:15:06.772128 4764 generic.go:334] "Generic (PLEG): container finished" podID="3fc31820-38e5-42a5-824c-e0f62fe8484a" containerID="edf39fcc0244cd782d701fc4743f48b2701ea66017d3e26a58694c71fa146a07" exitCode=0 Feb 02 10:15:06 crc kubenswrapper[4764]: I0202 10:15:06.772170 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-tvk74/crc-debug-58t2z" event={"ID":"3fc31820-38e5-42a5-824c-e0f62fe8484a","Type":"ContainerDied","Data":"edf39fcc0244cd782d701fc4743f48b2701ea66017d3e26a58694c71fa146a07"} Feb 02 10:15:07 crc kubenswrapper[4764]: I0202 10:15:07.880968 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-tvk74/crc-debug-58t2z" Feb 02 10:15:07 crc kubenswrapper[4764]: I0202 10:15:07.916086 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-tvk74/crc-debug-58t2z"] Feb 02 10:15:07 crc kubenswrapper[4764]: I0202 10:15:07.922691 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-tvk74/crc-debug-58t2z"] Feb 02 10:15:07 crc kubenswrapper[4764]: I0202 10:15:07.983204 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3fc31820-38e5-42a5-824c-e0f62fe8484a-host\") pod \"3fc31820-38e5-42a5-824c-e0f62fe8484a\" (UID: \"3fc31820-38e5-42a5-824c-e0f62fe8484a\") " Feb 02 10:15:07 crc kubenswrapper[4764]: I0202 10:15:07.983296 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3fc31820-38e5-42a5-824c-e0f62fe8484a-host" (OuterVolumeSpecName: "host") pod "3fc31820-38e5-42a5-824c-e0f62fe8484a" (UID: "3fc31820-38e5-42a5-824c-e0f62fe8484a"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 10:15:07 crc kubenswrapper[4764]: I0202 10:15:07.983357 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bdcvd\" (UniqueName: \"kubernetes.io/projected/3fc31820-38e5-42a5-824c-e0f62fe8484a-kube-api-access-bdcvd\") pod \"3fc31820-38e5-42a5-824c-e0f62fe8484a\" (UID: \"3fc31820-38e5-42a5-824c-e0f62fe8484a\") " Feb 02 10:15:07 crc kubenswrapper[4764]: I0202 10:15:07.983887 4764 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3fc31820-38e5-42a5-824c-e0f62fe8484a-host\") on node \"crc\" DevicePath \"\"" Feb 02 10:15:07 crc kubenswrapper[4764]: I0202 10:15:07.995181 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3fc31820-38e5-42a5-824c-e0f62fe8484a-kube-api-access-bdcvd" (OuterVolumeSpecName: "kube-api-access-bdcvd") pod "3fc31820-38e5-42a5-824c-e0f62fe8484a" (UID: "3fc31820-38e5-42a5-824c-e0f62fe8484a"). InnerVolumeSpecName "kube-api-access-bdcvd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 10:15:08 crc kubenswrapper[4764]: I0202 10:15:08.085715 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bdcvd\" (UniqueName: \"kubernetes.io/projected/3fc31820-38e5-42a5-824c-e0f62fe8484a-kube-api-access-bdcvd\") on node \"crc\" DevicePath \"\"" Feb 02 10:15:08 crc kubenswrapper[4764]: I0202 10:15:08.790348 4764 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="23d7e1c76bd6c91c65f8519c77d6fd93ca4941c32e078810f728f008df56b1ef" Feb 02 10:15:08 crc kubenswrapper[4764]: I0202 10:15:08.790387 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-tvk74/crc-debug-58t2z" Feb 02 10:15:09 crc kubenswrapper[4764]: I0202 10:15:09.128093 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-tvk74/crc-debug-wghjs"] Feb 02 10:15:09 crc kubenswrapper[4764]: E0202 10:15:09.128442 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d47415a-a9c4-4f36-9b76-2491f748505e" containerName="collect-profiles" Feb 02 10:15:09 crc kubenswrapper[4764]: I0202 10:15:09.128454 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d47415a-a9c4-4f36-9b76-2491f748505e" containerName="collect-profiles" Feb 02 10:15:09 crc kubenswrapper[4764]: E0202 10:15:09.128490 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3fc31820-38e5-42a5-824c-e0f62fe8484a" containerName="container-00" Feb 02 10:15:09 crc kubenswrapper[4764]: I0202 10:15:09.128496 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="3fc31820-38e5-42a5-824c-e0f62fe8484a" containerName="container-00" Feb 02 10:15:09 crc kubenswrapper[4764]: I0202 10:15:09.128677 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d47415a-a9c4-4f36-9b76-2491f748505e" containerName="collect-profiles" Feb 02 10:15:09 crc kubenswrapper[4764]: I0202 10:15:09.128699 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="3fc31820-38e5-42a5-824c-e0f62fe8484a" containerName="container-00" Feb 02 10:15:09 crc kubenswrapper[4764]: I0202 10:15:09.129362 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-tvk74/crc-debug-wghjs" Feb 02 10:15:09 crc kubenswrapper[4764]: I0202 10:15:09.206284 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ee2652ad-3316-4bfd-824f-3dfc84e7a8d8-host\") pod \"crc-debug-wghjs\" (UID: \"ee2652ad-3316-4bfd-824f-3dfc84e7a8d8\") " pod="openshift-must-gather-tvk74/crc-debug-wghjs" Feb 02 10:15:09 crc kubenswrapper[4764]: I0202 10:15:09.206351 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qhp4k\" (UniqueName: \"kubernetes.io/projected/ee2652ad-3316-4bfd-824f-3dfc84e7a8d8-kube-api-access-qhp4k\") pod \"crc-debug-wghjs\" (UID: \"ee2652ad-3316-4bfd-824f-3dfc84e7a8d8\") " pod="openshift-must-gather-tvk74/crc-debug-wghjs" Feb 02 10:15:09 crc kubenswrapper[4764]: I0202 10:15:09.308277 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ee2652ad-3316-4bfd-824f-3dfc84e7a8d8-host\") pod \"crc-debug-wghjs\" (UID: \"ee2652ad-3316-4bfd-824f-3dfc84e7a8d8\") " pod="openshift-must-gather-tvk74/crc-debug-wghjs" Feb 02 10:15:09 crc kubenswrapper[4764]: I0202 10:15:09.308584 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qhp4k\" (UniqueName: \"kubernetes.io/projected/ee2652ad-3316-4bfd-824f-3dfc84e7a8d8-kube-api-access-qhp4k\") pod \"crc-debug-wghjs\" (UID: \"ee2652ad-3316-4bfd-824f-3dfc84e7a8d8\") " pod="openshift-must-gather-tvk74/crc-debug-wghjs" Feb 02 10:15:09 crc kubenswrapper[4764]: I0202 10:15:09.308452 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ee2652ad-3316-4bfd-824f-3dfc84e7a8d8-host\") pod \"crc-debug-wghjs\" (UID: \"ee2652ad-3316-4bfd-824f-3dfc84e7a8d8\") " pod="openshift-must-gather-tvk74/crc-debug-wghjs" Feb 02 10:15:09 crc kubenswrapper[4764]: I0202 10:15:09.329772 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qhp4k\" (UniqueName: \"kubernetes.io/projected/ee2652ad-3316-4bfd-824f-3dfc84e7a8d8-kube-api-access-qhp4k\") pod \"crc-debug-wghjs\" (UID: \"ee2652ad-3316-4bfd-824f-3dfc84e7a8d8\") " pod="openshift-must-gather-tvk74/crc-debug-wghjs" Feb 02 10:15:09 crc kubenswrapper[4764]: I0202 10:15:09.446443 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-tvk74/crc-debug-wghjs" Feb 02 10:15:09 crc kubenswrapper[4764]: I0202 10:15:09.800777 4764 generic.go:334] "Generic (PLEG): container finished" podID="ee2652ad-3316-4bfd-824f-3dfc84e7a8d8" containerID="10ef5001d0ddba0ef1876237d7bfd3920318548f77f9984bbe903f23353a6771" exitCode=0 Feb 02 10:15:09 crc kubenswrapper[4764]: I0202 10:15:09.800860 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-tvk74/crc-debug-wghjs" event={"ID":"ee2652ad-3316-4bfd-824f-3dfc84e7a8d8","Type":"ContainerDied","Data":"10ef5001d0ddba0ef1876237d7bfd3920318548f77f9984bbe903f23353a6771"} Feb 02 10:15:09 crc kubenswrapper[4764]: I0202 10:15:09.801189 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-tvk74/crc-debug-wghjs" event={"ID":"ee2652ad-3316-4bfd-824f-3dfc84e7a8d8","Type":"ContainerStarted","Data":"1ef8c8c2de60489d68e00dd896e09652e35bb052f4bfb8c914f4e35f674ca1fa"} Feb 02 10:15:09 crc kubenswrapper[4764]: I0202 10:15:09.892431 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3fc31820-38e5-42a5-824c-e0f62fe8484a" path="/var/lib/kubelet/pods/3fc31820-38e5-42a5-824c-e0f62fe8484a/volumes" Feb 02 10:15:10 crc kubenswrapper[4764]: I0202 10:15:10.209357 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-tvk74/crc-debug-wghjs"] Feb 02 10:15:10 crc kubenswrapper[4764]: I0202 10:15:10.218874 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-tvk74/crc-debug-wghjs"] Feb 02 10:15:10 crc kubenswrapper[4764]: I0202 10:15:10.910074 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-tvk74/crc-debug-wghjs" Feb 02 10:15:10 crc kubenswrapper[4764]: I0202 10:15:10.949884 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qhp4k\" (UniqueName: \"kubernetes.io/projected/ee2652ad-3316-4bfd-824f-3dfc84e7a8d8-kube-api-access-qhp4k\") pod \"ee2652ad-3316-4bfd-824f-3dfc84e7a8d8\" (UID: \"ee2652ad-3316-4bfd-824f-3dfc84e7a8d8\") " Feb 02 10:15:10 crc kubenswrapper[4764]: I0202 10:15:10.950013 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ee2652ad-3316-4bfd-824f-3dfc84e7a8d8-host\") pod \"ee2652ad-3316-4bfd-824f-3dfc84e7a8d8\" (UID: \"ee2652ad-3316-4bfd-824f-3dfc84e7a8d8\") " Feb 02 10:15:10 crc kubenswrapper[4764]: I0202 10:15:10.950136 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ee2652ad-3316-4bfd-824f-3dfc84e7a8d8-host" (OuterVolumeSpecName: "host") pod "ee2652ad-3316-4bfd-824f-3dfc84e7a8d8" (UID: "ee2652ad-3316-4bfd-824f-3dfc84e7a8d8"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 10:15:10 crc kubenswrapper[4764]: I0202 10:15:10.950684 4764 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ee2652ad-3316-4bfd-824f-3dfc84e7a8d8-host\") on node \"crc\" DevicePath \"\"" Feb 02 10:15:10 crc kubenswrapper[4764]: I0202 10:15:10.956160 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee2652ad-3316-4bfd-824f-3dfc84e7a8d8-kube-api-access-qhp4k" (OuterVolumeSpecName: "kube-api-access-qhp4k") pod "ee2652ad-3316-4bfd-824f-3dfc84e7a8d8" (UID: "ee2652ad-3316-4bfd-824f-3dfc84e7a8d8"). InnerVolumeSpecName "kube-api-access-qhp4k". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 10:15:11 crc kubenswrapper[4764]: I0202 10:15:11.052190 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qhp4k\" (UniqueName: \"kubernetes.io/projected/ee2652ad-3316-4bfd-824f-3dfc84e7a8d8-kube-api-access-qhp4k\") on node \"crc\" DevicePath \"\"" Feb 02 10:15:11 crc kubenswrapper[4764]: I0202 10:15:11.489245 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-tvk74/crc-debug-26lgz"] Feb 02 10:15:11 crc kubenswrapper[4764]: E0202 10:15:11.489831 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee2652ad-3316-4bfd-824f-3dfc84e7a8d8" containerName="container-00" Feb 02 10:15:11 crc kubenswrapper[4764]: I0202 10:15:11.489842 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee2652ad-3316-4bfd-824f-3dfc84e7a8d8" containerName="container-00" Feb 02 10:15:11 crc kubenswrapper[4764]: I0202 10:15:11.490040 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee2652ad-3316-4bfd-824f-3dfc84e7a8d8" containerName="container-00" Feb 02 10:15:11 crc kubenswrapper[4764]: I0202 10:15:11.490607 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-tvk74/crc-debug-26lgz" Feb 02 10:15:11 crc kubenswrapper[4764]: I0202 10:15:11.560982 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cchlv\" (UniqueName: \"kubernetes.io/projected/5877a8aa-a597-42cb-8a26-f615d4bf99bf-kube-api-access-cchlv\") pod \"crc-debug-26lgz\" (UID: \"5877a8aa-a597-42cb-8a26-f615d4bf99bf\") " pod="openshift-must-gather-tvk74/crc-debug-26lgz" Feb 02 10:15:11 crc kubenswrapper[4764]: I0202 10:15:11.561137 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5877a8aa-a597-42cb-8a26-f615d4bf99bf-host\") pod \"crc-debug-26lgz\" (UID: \"5877a8aa-a597-42cb-8a26-f615d4bf99bf\") " pod="openshift-must-gather-tvk74/crc-debug-26lgz" Feb 02 10:15:11 crc kubenswrapper[4764]: I0202 10:15:11.662868 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5877a8aa-a597-42cb-8a26-f615d4bf99bf-host\") pod \"crc-debug-26lgz\" (UID: \"5877a8aa-a597-42cb-8a26-f615d4bf99bf\") " pod="openshift-must-gather-tvk74/crc-debug-26lgz" Feb 02 10:15:11 crc kubenswrapper[4764]: I0202 10:15:11.663016 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5877a8aa-a597-42cb-8a26-f615d4bf99bf-host\") pod \"crc-debug-26lgz\" (UID: \"5877a8aa-a597-42cb-8a26-f615d4bf99bf\") " pod="openshift-must-gather-tvk74/crc-debug-26lgz" Feb 02 10:15:11 crc kubenswrapper[4764]: I0202 10:15:11.663157 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cchlv\" (UniqueName: \"kubernetes.io/projected/5877a8aa-a597-42cb-8a26-f615d4bf99bf-kube-api-access-cchlv\") pod \"crc-debug-26lgz\" (UID: \"5877a8aa-a597-42cb-8a26-f615d4bf99bf\") " pod="openshift-must-gather-tvk74/crc-debug-26lgz" Feb 02 10:15:11 crc kubenswrapper[4764]: I0202 10:15:11.690604 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cchlv\" (UniqueName: \"kubernetes.io/projected/5877a8aa-a597-42cb-8a26-f615d4bf99bf-kube-api-access-cchlv\") pod \"crc-debug-26lgz\" (UID: \"5877a8aa-a597-42cb-8a26-f615d4bf99bf\") " pod="openshift-must-gather-tvk74/crc-debug-26lgz" Feb 02 10:15:11 crc kubenswrapper[4764]: I0202 10:15:11.805189 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-tvk74/crc-debug-26lgz" Feb 02 10:15:11 crc kubenswrapper[4764]: I0202 10:15:11.819862 4764 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1ef8c8c2de60489d68e00dd896e09652e35bb052f4bfb8c914f4e35f674ca1fa" Feb 02 10:15:11 crc kubenswrapper[4764]: I0202 10:15:11.819956 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-tvk74/crc-debug-wghjs" Feb 02 10:15:11 crc kubenswrapper[4764]: W0202 10:15:11.831585 4764 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5877a8aa_a597_42cb_8a26_f615d4bf99bf.slice/crio-83d781ba5714e89c655415ada6403607e94a3604c2b8a587e248cba45a802feb WatchSource:0}: Error finding container 83d781ba5714e89c655415ada6403607e94a3604c2b8a587e248cba45a802feb: Status 404 returned error can't find the container with id 83d781ba5714e89c655415ada6403607e94a3604c2b8a587e248cba45a802feb Feb 02 10:15:11 crc kubenswrapper[4764]: I0202 10:15:11.859141 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ee2652ad-3316-4bfd-824f-3dfc84e7a8d8" path="/var/lib/kubelet/pods/ee2652ad-3316-4bfd-824f-3dfc84e7a8d8/volumes" Feb 02 10:15:12 crc kubenswrapper[4764]: I0202 10:15:12.828909 4764 generic.go:334] "Generic (PLEG): container finished" podID="5877a8aa-a597-42cb-8a26-f615d4bf99bf" containerID="5650b147887effa2f4f3b1cf45c52472b4589863cab6e5f05b293b7d9add2382" exitCode=0 Feb 02 10:15:12 crc kubenswrapper[4764]: I0202 10:15:12.829064 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-tvk74/crc-debug-26lgz" event={"ID":"5877a8aa-a597-42cb-8a26-f615d4bf99bf","Type":"ContainerDied","Data":"5650b147887effa2f4f3b1cf45c52472b4589863cab6e5f05b293b7d9add2382"} Feb 02 10:15:12 crc kubenswrapper[4764]: I0202 10:15:12.829233 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-tvk74/crc-debug-26lgz" event={"ID":"5877a8aa-a597-42cb-8a26-f615d4bf99bf","Type":"ContainerStarted","Data":"83d781ba5714e89c655415ada6403607e94a3604c2b8a587e248cba45a802feb"} Feb 02 10:15:12 crc kubenswrapper[4764]: I0202 10:15:12.865038 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-tvk74/crc-debug-26lgz"] Feb 02 10:15:12 crc kubenswrapper[4764]: I0202 10:15:12.874505 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-tvk74/crc-debug-26lgz"] Feb 02 10:15:13 crc kubenswrapper[4764]: I0202 10:15:13.563013 4764 patch_prober.go:28] interesting pod/machine-config-daemon-4ndm4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 10:15:13 crc kubenswrapper[4764]: I0202 10:15:13.563227 4764 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 10:15:13 crc kubenswrapper[4764]: I0202 10:15:13.938032 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-tvk74/crc-debug-26lgz" Feb 02 10:15:14 crc kubenswrapper[4764]: I0202 10:15:14.072503 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cchlv\" (UniqueName: \"kubernetes.io/projected/5877a8aa-a597-42cb-8a26-f615d4bf99bf-kube-api-access-cchlv\") pod \"5877a8aa-a597-42cb-8a26-f615d4bf99bf\" (UID: \"5877a8aa-a597-42cb-8a26-f615d4bf99bf\") " Feb 02 10:15:14 crc kubenswrapper[4764]: I0202 10:15:14.074979 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5877a8aa-a597-42cb-8a26-f615d4bf99bf-host\") pod \"5877a8aa-a597-42cb-8a26-f615d4bf99bf\" (UID: \"5877a8aa-a597-42cb-8a26-f615d4bf99bf\") " Feb 02 10:15:14 crc kubenswrapper[4764]: I0202 10:15:14.075124 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5877a8aa-a597-42cb-8a26-f615d4bf99bf-host" (OuterVolumeSpecName: "host") pod "5877a8aa-a597-42cb-8a26-f615d4bf99bf" (UID: "5877a8aa-a597-42cb-8a26-f615d4bf99bf"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 10:15:14 crc kubenswrapper[4764]: I0202 10:15:14.081254 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5877a8aa-a597-42cb-8a26-f615d4bf99bf-kube-api-access-cchlv" (OuterVolumeSpecName: "kube-api-access-cchlv") pod "5877a8aa-a597-42cb-8a26-f615d4bf99bf" (UID: "5877a8aa-a597-42cb-8a26-f615d4bf99bf"). InnerVolumeSpecName "kube-api-access-cchlv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 10:15:14 crc kubenswrapper[4764]: I0202 10:15:14.176366 4764 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5877a8aa-a597-42cb-8a26-f615d4bf99bf-host\") on node \"crc\" DevicePath \"\"" Feb 02 10:15:14 crc kubenswrapper[4764]: I0202 10:15:14.176398 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cchlv\" (UniqueName: \"kubernetes.io/projected/5877a8aa-a597-42cb-8a26-f615d4bf99bf-kube-api-access-cchlv\") on node \"crc\" DevicePath \"\"" Feb 02 10:15:14 crc kubenswrapper[4764]: I0202 10:15:14.861236 4764 scope.go:117] "RemoveContainer" containerID="5650b147887effa2f4f3b1cf45c52472b4589863cab6e5f05b293b7d9add2382" Feb 02 10:15:14 crc kubenswrapper[4764]: I0202 10:15:14.861316 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-tvk74/crc-debug-26lgz" Feb 02 10:15:15 crc kubenswrapper[4764]: I0202 10:15:15.838285 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5877a8aa-a597-42cb-8a26-f615d4bf99bf" path="/var/lib/kubelet/pods/5877a8aa-a597-42cb-8a26-f615d4bf99bf/volumes" Feb 02 10:15:15 crc kubenswrapper[4764]: I0202 10:15:15.910144 4764 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-kzw9f" podUID="df71a15a-b67b-43b7-9487-7723c62304f3" containerName="registry-server" probeResult="failure" output=< Feb 02 10:15:15 crc kubenswrapper[4764]: timeout: failed to connect service ":50051" within 1s Feb 02 10:15:15 crc kubenswrapper[4764]: > Feb 02 10:15:24 crc kubenswrapper[4764]: I0202 10:15:24.650296 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-kzw9f" Feb 02 10:15:24 crc kubenswrapper[4764]: I0202 10:15:24.706182 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-kzw9f" Feb 02 10:15:28 crc kubenswrapper[4764]: I0202 10:15:28.481030 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-kzw9f"] Feb 02 10:15:28 crc kubenswrapper[4764]: I0202 10:15:28.483021 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-kzw9f" podUID="df71a15a-b67b-43b7-9487-7723c62304f3" containerName="registry-server" containerID="cri-o://cb8fe6cfe1faa91cbbea75028829c32ff9c6e8084d80efe3f05dae807ecd9215" gracePeriod=2 Feb 02 10:15:28 crc kubenswrapper[4764]: I0202 10:15:28.928193 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kzw9f" Feb 02 10:15:28 crc kubenswrapper[4764]: I0202 10:15:28.981641 4764 generic.go:334] "Generic (PLEG): container finished" podID="df71a15a-b67b-43b7-9487-7723c62304f3" containerID="cb8fe6cfe1faa91cbbea75028829c32ff9c6e8084d80efe3f05dae807ecd9215" exitCode=0 Feb 02 10:15:28 crc kubenswrapper[4764]: I0202 10:15:28.981722 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kzw9f" event={"ID":"df71a15a-b67b-43b7-9487-7723c62304f3","Type":"ContainerDied","Data":"cb8fe6cfe1faa91cbbea75028829c32ff9c6e8084d80efe3f05dae807ecd9215"} Feb 02 10:15:28 crc kubenswrapper[4764]: I0202 10:15:28.981748 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kzw9f" event={"ID":"df71a15a-b67b-43b7-9487-7723c62304f3","Type":"ContainerDied","Data":"b90967997b1ebd0be40b048701adc909d5ad22c6fc2ef07ffdd0900dce639211"} Feb 02 10:15:28 crc kubenswrapper[4764]: I0202 10:15:28.981764 4764 scope.go:117] "RemoveContainer" containerID="cb8fe6cfe1faa91cbbea75028829c32ff9c6e8084d80efe3f05dae807ecd9215" Feb 02 10:15:28 crc kubenswrapper[4764]: I0202 10:15:28.981874 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kzw9f" Feb 02 10:15:29 crc kubenswrapper[4764]: I0202 10:15:29.001979 4764 scope.go:117] "RemoveContainer" containerID="a79d594f10382d3c2cbe01ccffbe09dc83810081c1b8f7a6640a6107d65175af" Feb 02 10:15:29 crc kubenswrapper[4764]: I0202 10:15:29.021924 4764 scope.go:117] "RemoveContainer" containerID="0bbfe02af0a90947f464f0389dcecfd894a7cbd8808dd6ee859fc48637a85192" Feb 02 10:15:29 crc kubenswrapper[4764]: I0202 10:15:29.067036 4764 scope.go:117] "RemoveContainer" containerID="cb8fe6cfe1faa91cbbea75028829c32ff9c6e8084d80efe3f05dae807ecd9215" Feb 02 10:15:29 crc kubenswrapper[4764]: E0202 10:15:29.067562 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cb8fe6cfe1faa91cbbea75028829c32ff9c6e8084d80efe3f05dae807ecd9215\": container with ID starting with cb8fe6cfe1faa91cbbea75028829c32ff9c6e8084d80efe3f05dae807ecd9215 not found: ID does not exist" containerID="cb8fe6cfe1faa91cbbea75028829c32ff9c6e8084d80efe3f05dae807ecd9215" Feb 02 10:15:29 crc kubenswrapper[4764]: I0202 10:15:29.067612 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cb8fe6cfe1faa91cbbea75028829c32ff9c6e8084d80efe3f05dae807ecd9215"} err="failed to get container status \"cb8fe6cfe1faa91cbbea75028829c32ff9c6e8084d80efe3f05dae807ecd9215\": rpc error: code = NotFound desc = could not find container \"cb8fe6cfe1faa91cbbea75028829c32ff9c6e8084d80efe3f05dae807ecd9215\": container with ID starting with cb8fe6cfe1faa91cbbea75028829c32ff9c6e8084d80efe3f05dae807ecd9215 not found: ID does not exist" Feb 02 10:15:29 crc kubenswrapper[4764]: I0202 10:15:29.067643 4764 scope.go:117] "RemoveContainer" containerID="a79d594f10382d3c2cbe01ccffbe09dc83810081c1b8f7a6640a6107d65175af" Feb 02 10:15:29 crc kubenswrapper[4764]: E0202 10:15:29.067937 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a79d594f10382d3c2cbe01ccffbe09dc83810081c1b8f7a6640a6107d65175af\": container with ID starting with a79d594f10382d3c2cbe01ccffbe09dc83810081c1b8f7a6640a6107d65175af not found: ID does not exist" containerID="a79d594f10382d3c2cbe01ccffbe09dc83810081c1b8f7a6640a6107d65175af" Feb 02 10:15:29 crc kubenswrapper[4764]: I0202 10:15:29.067975 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a79d594f10382d3c2cbe01ccffbe09dc83810081c1b8f7a6640a6107d65175af"} err="failed to get container status \"a79d594f10382d3c2cbe01ccffbe09dc83810081c1b8f7a6640a6107d65175af\": rpc error: code = NotFound desc = could not find container \"a79d594f10382d3c2cbe01ccffbe09dc83810081c1b8f7a6640a6107d65175af\": container with ID starting with a79d594f10382d3c2cbe01ccffbe09dc83810081c1b8f7a6640a6107d65175af not found: ID does not exist" Feb 02 10:15:29 crc kubenswrapper[4764]: I0202 10:15:29.067991 4764 scope.go:117] "RemoveContainer" containerID="0bbfe02af0a90947f464f0389dcecfd894a7cbd8808dd6ee859fc48637a85192" Feb 02 10:15:29 crc kubenswrapper[4764]: E0202 10:15:29.068230 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0bbfe02af0a90947f464f0389dcecfd894a7cbd8808dd6ee859fc48637a85192\": container with ID starting with 0bbfe02af0a90947f464f0389dcecfd894a7cbd8808dd6ee859fc48637a85192 not found: ID does not exist" containerID="0bbfe02af0a90947f464f0389dcecfd894a7cbd8808dd6ee859fc48637a85192" Feb 02 10:15:29 crc kubenswrapper[4764]: I0202 10:15:29.068257 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0bbfe02af0a90947f464f0389dcecfd894a7cbd8808dd6ee859fc48637a85192"} err="failed to get container status \"0bbfe02af0a90947f464f0389dcecfd894a7cbd8808dd6ee859fc48637a85192\": rpc error: code = NotFound desc = could not find container \"0bbfe02af0a90947f464f0389dcecfd894a7cbd8808dd6ee859fc48637a85192\": container with ID starting with 0bbfe02af0a90947f464f0389dcecfd894a7cbd8808dd6ee859fc48637a85192 not found: ID does not exist" Feb 02 10:15:29 crc kubenswrapper[4764]: I0202 10:15:29.121188 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/df71a15a-b67b-43b7-9487-7723c62304f3-catalog-content\") pod \"df71a15a-b67b-43b7-9487-7723c62304f3\" (UID: \"df71a15a-b67b-43b7-9487-7723c62304f3\") " Feb 02 10:15:29 crc kubenswrapper[4764]: I0202 10:15:29.121382 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/df71a15a-b67b-43b7-9487-7723c62304f3-utilities\") pod \"df71a15a-b67b-43b7-9487-7723c62304f3\" (UID: \"df71a15a-b67b-43b7-9487-7723c62304f3\") " Feb 02 10:15:29 crc kubenswrapper[4764]: I0202 10:15:29.121492 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c8gtp\" (UniqueName: \"kubernetes.io/projected/df71a15a-b67b-43b7-9487-7723c62304f3-kube-api-access-c8gtp\") pod \"df71a15a-b67b-43b7-9487-7723c62304f3\" (UID: \"df71a15a-b67b-43b7-9487-7723c62304f3\") " Feb 02 10:15:29 crc kubenswrapper[4764]: I0202 10:15:29.122091 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/df71a15a-b67b-43b7-9487-7723c62304f3-utilities" (OuterVolumeSpecName: "utilities") pod "df71a15a-b67b-43b7-9487-7723c62304f3" (UID: "df71a15a-b67b-43b7-9487-7723c62304f3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 10:15:29 crc kubenswrapper[4764]: I0202 10:15:29.127803 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/df71a15a-b67b-43b7-9487-7723c62304f3-kube-api-access-c8gtp" (OuterVolumeSpecName: "kube-api-access-c8gtp") pod "df71a15a-b67b-43b7-9487-7723c62304f3" (UID: "df71a15a-b67b-43b7-9487-7723c62304f3"). InnerVolumeSpecName "kube-api-access-c8gtp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 10:15:29 crc kubenswrapper[4764]: I0202 10:15:29.223774 4764 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/df71a15a-b67b-43b7-9487-7723c62304f3-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 10:15:29 crc kubenswrapper[4764]: I0202 10:15:29.223802 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c8gtp\" (UniqueName: \"kubernetes.io/projected/df71a15a-b67b-43b7-9487-7723c62304f3-kube-api-access-c8gtp\") on node \"crc\" DevicePath \"\"" Feb 02 10:15:29 crc kubenswrapper[4764]: I0202 10:15:29.227115 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/df71a15a-b67b-43b7-9487-7723c62304f3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "df71a15a-b67b-43b7-9487-7723c62304f3" (UID: "df71a15a-b67b-43b7-9487-7723c62304f3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 10:15:29 crc kubenswrapper[4764]: I0202 10:15:29.240050 4764 scope.go:117] "RemoveContainer" containerID="5e600332710ba1ec80f0870d729d873b95fe41f319c1d38d3d8d14af9bb59bae" Feb 02 10:15:29 crc kubenswrapper[4764]: I0202 10:15:29.322650 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-kzw9f"] Feb 02 10:15:29 crc kubenswrapper[4764]: I0202 10:15:29.325053 4764 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/df71a15a-b67b-43b7-9487-7723c62304f3-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 10:15:29 crc kubenswrapper[4764]: I0202 10:15:29.332537 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-kzw9f"] Feb 02 10:15:29 crc kubenswrapper[4764]: I0202 10:15:29.838478 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="df71a15a-b67b-43b7-9487-7723c62304f3" path="/var/lib/kubelet/pods/df71a15a-b67b-43b7-9487-7723c62304f3/volumes" Feb 02 10:15:43 crc kubenswrapper[4764]: I0202 10:15:43.523151 4764 patch_prober.go:28] interesting pod/machine-config-daemon-4ndm4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 10:15:43 crc kubenswrapper[4764]: I0202 10:15:43.523736 4764 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 10:16:08 crc kubenswrapper[4764]: I0202 10:16:08.161058 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-797597c45b-9mkpp_f06f976d-5e29-4e78-81f4-81de01ae8391/barbican-api/0.log" Feb 02 10:16:08 crc kubenswrapper[4764]: I0202 10:16:08.349061 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-797597c45b-9mkpp_f06f976d-5e29-4e78-81f4-81de01ae8391/barbican-api-log/0.log" Feb 02 10:16:08 crc kubenswrapper[4764]: I0202 10:16:08.394555 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-855c89b68b-ptbcl_b004a188-196b-4812-a9e6-e398062cf4d3/barbican-keystone-listener/0.log" Feb 02 10:16:08 crc kubenswrapper[4764]: I0202 10:16:08.493481 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-855c89b68b-ptbcl_b004a188-196b-4812-a9e6-e398062cf4d3/barbican-keystone-listener-log/0.log" Feb 02 10:16:08 crc kubenswrapper[4764]: I0202 10:16:08.641117 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-5686dc79cc-sb69r_4e3b3433-3d64-44ce-9a3a-14fcd0bdaa6d/barbican-worker/0.log" Feb 02 10:16:08 crc kubenswrapper[4764]: I0202 10:16:08.695704 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-5686dc79cc-sb69r_4e3b3433-3d64-44ce-9a3a-14fcd0bdaa6d/barbican-worker-log/0.log" Feb 02 10:16:08 crc kubenswrapper[4764]: I0202 10:16:08.823166 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-d9x8m_af835a1d-09cc-49ca-979f-41eaf146e926/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Feb 02 10:16:08 crc kubenswrapper[4764]: I0202 10:16:08.895885 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_242b61fa-f577-449a-86b2-4cafa5175bd6/ceilometer-central-agent/0.log" Feb 02 10:16:08 crc kubenswrapper[4764]: I0202 10:16:08.957318 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_242b61fa-f577-449a-86b2-4cafa5175bd6/ceilometer-notification-agent/0.log" Feb 02 10:16:09 crc kubenswrapper[4764]: I0202 10:16:09.104985 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_242b61fa-f577-449a-86b2-4cafa5175bd6/proxy-httpd/0.log" Feb 02 10:16:09 crc kubenswrapper[4764]: I0202 10:16:09.123116 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_242b61fa-f577-449a-86b2-4cafa5175bd6/sg-core/0.log" Feb 02 10:16:09 crc kubenswrapper[4764]: I0202 10:16:09.210017 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceph-client-edpm-deployment-openstack-edpm-ipam-bs4pz_3858fd68-f1b8-4e73-a0c9-5fb26dcbedd6/ceph-client-edpm-deployment-openstack-edpm-ipam/0.log" Feb 02 10:16:09 crc kubenswrapper[4764]: I0202 10:16:09.530833 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-d4lm7_329c5f09-ca39-4ced-a6b6-02db68c98510/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam/0.log" Feb 02 10:16:09 crc kubenswrapper[4764]: I0202 10:16:09.701065 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_4b8e0fd6-1949-4b4f-bbc5-b663c51f08c4/cinder-api/0.log" Feb 02 10:16:09 crc kubenswrapper[4764]: I0202 10:16:09.810808 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_4b8e0fd6-1949-4b4f-bbc5-b663c51f08c4/cinder-api-log/0.log" Feb 02 10:16:09 crc kubenswrapper[4764]: I0202 10:16:09.900703 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_371fb9ce-9ba4-429b-a68c-62d2e12d688c/probe/0.log" Feb 02 10:16:10 crc kubenswrapper[4764]: I0202 10:16:10.047345 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_371fb9ce-9ba4-429b-a68c-62d2e12d688c/cinder-backup/0.log" Feb 02 10:16:10 crc kubenswrapper[4764]: I0202 10:16:10.194896 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_10cfff03-5b66-41ef-948b-0bad4f6e9a91/cinder-scheduler/0.log" Feb 02 10:16:10 crc kubenswrapper[4764]: I0202 10:16:10.243633 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_10cfff03-5b66-41ef-948b-0bad4f6e9a91/probe/0.log" Feb 02 10:16:10 crc kubenswrapper[4764]: I0202 10:16:10.449427 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_1c67d8f7-21c3-456d-af99-c96d04090915/cinder-volume/0.log" Feb 02 10:16:10 crc kubenswrapper[4764]: I0202 10:16:10.460231 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_1c67d8f7-21c3-456d-af99-c96d04090915/probe/0.log" Feb 02 10:16:10 crc kubenswrapper[4764]: I0202 10:16:10.718255 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-c4mpn_c3b49d2a-ed47-45cb-9940-b5e87ff7a197/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Feb 02 10:16:10 crc kubenswrapper[4764]: I0202 10:16:10.728106 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-rsd54_5a0ee3ef-d01f-4961-8362-8b6c4d3f37fc/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Feb 02 10:16:10 crc kubenswrapper[4764]: I0202 10:16:10.855588 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6d94cddcb9-n4wc4_d61062ba-f497-4f61-9b51-27e62888797e/init/0.log" Feb 02 10:16:11 crc kubenswrapper[4764]: I0202 10:16:11.036727 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6d94cddcb9-n4wc4_d61062ba-f497-4f61-9b51-27e62888797e/init/0.log" Feb 02 10:16:11 crc kubenswrapper[4764]: I0202 10:16:11.145406 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_9f924031-1e49-4009-aeba-a4e5f3476373/glance-httpd/0.log" Feb 02 10:16:11 crc kubenswrapper[4764]: I0202 10:16:11.247628 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6d94cddcb9-n4wc4_d61062ba-f497-4f61-9b51-27e62888797e/dnsmasq-dns/0.log" Feb 02 10:16:11 crc kubenswrapper[4764]: I0202 10:16:11.300050 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_9f924031-1e49-4009-aeba-a4e5f3476373/glance-log/0.log" Feb 02 10:16:11 crc kubenswrapper[4764]: I0202 10:16:11.413658 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_687693a0-eb78-4922-991f-6738de00f468/glance-httpd/0.log" Feb 02 10:16:11 crc kubenswrapper[4764]: I0202 10:16:11.485782 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_687693a0-eb78-4922-991f-6738de00f468/glance-log/0.log" Feb 02 10:16:11 crc kubenswrapper[4764]: I0202 10:16:11.678064 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-698ff6d98d-6h7sk_174bcb22-4370-4c3f-9141-9c1cee0d2542/horizon/0.log" Feb 02 10:16:11 crc kubenswrapper[4764]: I0202 10:16:11.739435 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-698ff6d98d-6h7sk_174bcb22-4370-4c3f-9141-9c1cee0d2542/horizon-log/0.log" Feb 02 10:16:11 crc kubenswrapper[4764]: I0202 10:16:11.847871 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-9q5kq_7d388199-9f34-415c-b83c-d5f54d17d063/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Feb 02 10:16:12 crc kubenswrapper[4764]: I0202 10:16:12.012540 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-5469f_6b2831d8-cc8c-4ded-99e1-45ad94ff9776/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Feb 02 10:16:12 crc kubenswrapper[4764]: I0202 10:16:12.214600 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-b47bfbff6-m265b_45891ac1-d70d-4c3d-af07-6a65855a9281/keystone-api/0.log" Feb 02 10:16:12 crc kubenswrapper[4764]: I0202 10:16:12.246926 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29500441-qpcvv_b77a0084-b391-4221-86d1-39629566a940/keystone-cron/0.log" Feb 02 10:16:12 crc kubenswrapper[4764]: I0202 10:16:12.418414 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_2fce6f0b-cb36-4ff1-ab73-35c40d709502/kube-state-metrics/0.log" Feb 02 10:16:12 crc kubenswrapper[4764]: I0202 10:16:12.568598 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-ccmw8_f232691b-63e5-4e8a-9659-a6591c7dbe61/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Feb 02 10:16:12 crc kubenswrapper[4764]: I0202 10:16:12.690278 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_eed25ff8-c8ee-41d6-bb33-89e4c5f30724/manila-api/0.log" Feb 02 10:16:12 crc kubenswrapper[4764]: I0202 10:16:12.709891 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_eed25ff8-c8ee-41d6-bb33-89e4c5f30724/manila-api-log/0.log" Feb 02 10:16:12 crc kubenswrapper[4764]: I0202 10:16:12.931870 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_47fbbd57-48b1-4cbf-b857-8ef14b8c8529/manila-scheduler/0.log" Feb 02 10:16:12 crc kubenswrapper[4764]: I0202 10:16:12.963541 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_b9fbc0ac-9686-4433-ac79-79b9e0cfb54b/manila-share/0.log" Feb 02 10:16:13 crc kubenswrapper[4764]: I0202 10:16:13.009462 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_47fbbd57-48b1-4cbf-b857-8ef14b8c8529/probe/0.log" Feb 02 10:16:13 crc kubenswrapper[4764]: I0202 10:16:13.069605 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_b9fbc0ac-9686-4433-ac79-79b9e0cfb54b/probe/0.log" Feb 02 10:16:13 crc kubenswrapper[4764]: I0202 10:16:13.523536 4764 patch_prober.go:28] interesting pod/machine-config-daemon-4ndm4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 10:16:13 crc kubenswrapper[4764]: I0202 10:16:13.523910 4764 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 10:16:13 crc kubenswrapper[4764]: I0202 10:16:13.523977 4764 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" Feb 02 10:16:13 crc kubenswrapper[4764]: I0202 10:16:13.524823 4764 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ee21cc1574aec6b1e483445c9c7de0962b1f3fa4c95e8c7525277156a9bf1c39"} pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 02 10:16:13 crc kubenswrapper[4764]: I0202 10:16:13.524886 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" containerName="machine-config-daemon" containerID="cri-o://ee21cc1574aec6b1e483445c9c7de0962b1f3fa4c95e8c7525277156a9bf1c39" gracePeriod=600 Feb 02 10:16:13 crc kubenswrapper[4764]: I0202 10:16:13.554778 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-6b9f774cdf-j8pdr_c7553cbf-5c61-4003-bd4a-aad61da77950/neutron-httpd/0.log" Feb 02 10:16:13 crc kubenswrapper[4764]: I0202 10:16:13.642042 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-6b9f774cdf-j8pdr_c7553cbf-5c61-4003-bd4a-aad61da77950/neutron-api/0.log" Feb 02 10:16:13 crc kubenswrapper[4764]: I0202 10:16:13.698561 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-6pxwj_a74b118e-ee59-4e87-967d-c8332a898a4c/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Feb 02 10:16:14 crc kubenswrapper[4764]: I0202 10:16:14.378184 4764 generic.go:334] "Generic (PLEG): container finished" podID="d192f670-9f9d-4539-9641-e4bed73acdd4" containerID="ee21cc1574aec6b1e483445c9c7de0962b1f3fa4c95e8c7525277156a9bf1c39" exitCode=0 Feb 02 10:16:14 crc kubenswrapper[4764]: I0202 10:16:14.378560 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" event={"ID":"d192f670-9f9d-4539-9641-e4bed73acdd4","Type":"ContainerDied","Data":"ee21cc1574aec6b1e483445c9c7de0962b1f3fa4c95e8c7525277156a9bf1c39"} Feb 02 10:16:14 crc kubenswrapper[4764]: I0202 10:16:14.378602 4764 scope.go:117] "RemoveContainer" containerID="4b1cf5e6a0ac319e2831449adb6329644931ba722d3592c4cde96fd7de3b310e" Feb 02 10:16:14 crc kubenswrapper[4764]: I0202 10:16:14.803168 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_db89c9c9-cad1-4e6b-aaf0-708e5b8b32b4/nova-api-log/0.log" Feb 02 10:16:14 crc kubenswrapper[4764]: I0202 10:16:14.975682 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_82f0b9ff-5694-4e86-acb1-6a8ef7b5b97a/nova-cell0-conductor-conductor/0.log" Feb 02 10:16:15 crc kubenswrapper[4764]: I0202 10:16:15.138391 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_b640955b-f755-48d2-9f0a-751b24424f35/nova-cell1-conductor-conductor/0.log" Feb 02 10:16:15 crc kubenswrapper[4764]: I0202 10:16:15.355715 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_db89c9c9-cad1-4e6b-aaf0-708e5b8b32b4/nova-api-api/0.log" Feb 02 10:16:15 crc kubenswrapper[4764]: I0202 10:16:15.376546 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_41b7a57b-b6a0-4af7-9304-e4c89c8e3812/nova-cell1-novncproxy-novncproxy/0.log" Feb 02 10:16:15 crc kubenswrapper[4764]: I0202 10:16:15.387628 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" event={"ID":"d192f670-9f9d-4539-9641-e4bed73acdd4","Type":"ContainerStarted","Data":"e7f0b09df430c74ab38568e7c16c9b37764975d9232ef539984f9cdd78407ca1"} Feb 02 10:16:15 crc kubenswrapper[4764]: I0202 10:16:15.909594 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-h4h49_b5b2ecb9-0adf-489b-8a98-42cad0682d09/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam/0.log" Feb 02 10:16:15 crc kubenswrapper[4764]: I0202 10:16:15.951285 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_61eec87e-6dfc-41b8-a670-0914cec8b593/nova-metadata-log/0.log" Feb 02 10:16:16 crc kubenswrapper[4764]: I0202 10:16:16.505647 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_dd11a817-6357-4405-983e-181451ce4daf/nova-scheduler-scheduler/0.log" Feb 02 10:16:16 crc kubenswrapper[4764]: I0202 10:16:16.532663 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_a9177511-b834-4e68-9f6f-8b1b4470b274/mysql-bootstrap/0.log" Feb 02 10:16:16 crc kubenswrapper[4764]: I0202 10:16:16.816275 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_a9177511-b834-4e68-9f6f-8b1b4470b274/galera/0.log" Feb 02 10:16:16 crc kubenswrapper[4764]: I0202 10:16:16.818471 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_a9177511-b834-4e68-9f6f-8b1b4470b274/mysql-bootstrap/0.log" Feb 02 10:16:17 crc kubenswrapper[4764]: I0202 10:16:17.036723 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_1d6b6c31-819e-4098-b11b-6073cc751cad/mysql-bootstrap/0.log" Feb 02 10:16:17 crc kubenswrapper[4764]: I0202 10:16:17.256668 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_1d6b6c31-819e-4098-b11b-6073cc751cad/galera/0.log" Feb 02 10:16:17 crc kubenswrapper[4764]: I0202 10:16:17.298627 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_1d6b6c31-819e-4098-b11b-6073cc751cad/mysql-bootstrap/0.log" Feb 02 10:16:17 crc kubenswrapper[4764]: I0202 10:16:17.498694 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_61eec87e-6dfc-41b8-a670-0914cec8b593/nova-metadata-metadata/0.log" Feb 02 10:16:17 crc kubenswrapper[4764]: I0202 10:16:17.520581 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_54c115e4-eceb-49af-96ba-854f48802c73/openstackclient/0.log" Feb 02 10:16:17 crc kubenswrapper[4764]: I0202 10:16:17.625167 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-6v7f6_66662206-d732-404d-a81d-6303d9947d60/ovn-controller/0.log" Feb 02 10:16:17 crc kubenswrapper[4764]: I0202 10:16:17.803140 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-kgc8j_b3becd85-0167-4ffe-802d-67988c486908/openstack-network-exporter/0.log" Feb 02 10:16:17 crc kubenswrapper[4764]: I0202 10:16:17.951210 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-2p2bk_29a433e1-2651-43cd-99d1-95694063c980/ovsdb-server-init/0.log" Feb 02 10:16:18 crc kubenswrapper[4764]: I0202 10:16:18.234087 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-2p2bk_29a433e1-2651-43cd-99d1-95694063c980/ovs-vswitchd/0.log" Feb 02 10:16:18 crc kubenswrapper[4764]: I0202 10:16:18.268785 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-2p2bk_29a433e1-2651-43cd-99d1-95694063c980/ovsdb-server-init/0.log" Feb 02 10:16:18 crc kubenswrapper[4764]: I0202 10:16:18.359631 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-2p2bk_29a433e1-2651-43cd-99d1-95694063c980/ovsdb-server/0.log" Feb 02 10:16:18 crc kubenswrapper[4764]: I0202 10:16:18.539765 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-2qbnb_e6da450f-20b5-41ef-b9bb-a0df89b3d4f3/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Feb 02 10:16:18 crc kubenswrapper[4764]: I0202 10:16:18.583706 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_f56696ad-9508-4f7b-89fb-28b59af303a0/openstack-network-exporter/0.log" Feb 02 10:16:18 crc kubenswrapper[4764]: I0202 10:16:18.660134 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_f56696ad-9508-4f7b-89fb-28b59af303a0/ovn-northd/0.log" Feb 02 10:16:18 crc kubenswrapper[4764]: I0202 10:16:18.852086 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_cba91a53-7d3d-481b-bd55-3f0bbe7ccbb6/openstack-network-exporter/0.log" Feb 02 10:16:18 crc kubenswrapper[4764]: I0202 10:16:18.972678 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_cba91a53-7d3d-481b-bd55-3f0bbe7ccbb6/ovsdbserver-nb/0.log" Feb 02 10:16:19 crc kubenswrapper[4764]: I0202 10:16:19.009457 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_89c39976-ff91-4c6b-b369-5cba94e02db8/openstack-network-exporter/0.log" Feb 02 10:16:19 crc kubenswrapper[4764]: I0202 10:16:19.112166 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_89c39976-ff91-4c6b-b369-5cba94e02db8/ovsdbserver-sb/0.log" Feb 02 10:16:19 crc kubenswrapper[4764]: I0202 10:16:19.374587 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-68d4474c54-98j2g_f7a36a2c-9ade-4bb2-84d2-faa137207385/placement-log/0.log" Feb 02 10:16:19 crc kubenswrapper[4764]: I0202 10:16:19.417810 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-68d4474c54-98j2g_f7a36a2c-9ade-4bb2-84d2-faa137207385/placement-api/0.log" Feb 02 10:16:19 crc kubenswrapper[4764]: I0202 10:16:19.614094 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_c60b198c-58ad-4363-afea-257f51f64e51/setup-container/0.log" Feb 02 10:16:19 crc kubenswrapper[4764]: I0202 10:16:19.850219 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_c60b198c-58ad-4363-afea-257f51f64e51/setup-container/0.log" Feb 02 10:16:19 crc kubenswrapper[4764]: I0202 10:16:19.915188 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_c60b198c-58ad-4363-afea-257f51f64e51/rabbitmq/0.log" Feb 02 10:16:19 crc kubenswrapper[4764]: I0202 10:16:19.997989 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_ec28045f-e1d4-4b22-bb7d-3c6d2a1f1a38/setup-container/0.log" Feb 02 10:16:20 crc kubenswrapper[4764]: I0202 10:16:20.257012 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_ec28045f-e1d4-4b22-bb7d-3c6d2a1f1a38/rabbitmq/0.log" Feb 02 10:16:20 crc kubenswrapper[4764]: I0202 10:16:20.319473 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_ec28045f-e1d4-4b22-bb7d-3c6d2a1f1a38/setup-container/0.log" Feb 02 10:16:20 crc kubenswrapper[4764]: I0202 10:16:20.328302 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-hrmzd_a79016d5-22d2-4894-829c-7570224b6f69/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Feb 02 10:16:20 crc kubenswrapper[4764]: I0202 10:16:20.646152 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-8f7dl_feb92d24-e755-4837-8461-15e7e3a7b0e8/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Feb 02 10:16:20 crc kubenswrapper[4764]: I0202 10:16:20.716835 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-t5j4g_8649507c-417e-40fc-9a67-b6d6831f3ab9/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Feb 02 10:16:20 crc kubenswrapper[4764]: I0202 10:16:20.986808 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-97dd2_409c3e31-e6bf-409c-9bd8-7bcf433c0ce8/ssh-known-hosts-edpm-deployment/0.log" Feb 02 10:16:21 crc kubenswrapper[4764]: I0202 10:16:21.024431 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_09adc919-28cf-4d2d-b40d-183a2f51f06f/tempest-tests-tempest-tests-runner/0.log" Feb 02 10:16:21 crc kubenswrapper[4764]: I0202 10:16:21.226034 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_7576fa25-4c66-4f89-90a5-c2979b1c0eaa/test-operator-logs-container/0.log" Feb 02 10:16:21 crc kubenswrapper[4764]: I0202 10:16:21.343471 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-nwdxw_ddae0daa-153b-425b-92dd-d6c315ace4e5/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Feb 02 10:16:33 crc kubenswrapper[4764]: I0202 10:16:33.732504 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_cd76e53a-13fc-4293-b5ca-2324224850cc/memcached/0.log" Feb 02 10:16:42 crc kubenswrapper[4764]: I0202 10:16:42.191045 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-n8597"] Feb 02 10:16:42 crc kubenswrapper[4764]: E0202 10:16:42.192706 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df71a15a-b67b-43b7-9487-7723c62304f3" containerName="extract-content" Feb 02 10:16:42 crc kubenswrapper[4764]: I0202 10:16:42.192790 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="df71a15a-b67b-43b7-9487-7723c62304f3" containerName="extract-content" Feb 02 10:16:42 crc kubenswrapper[4764]: E0202 10:16:42.192892 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5877a8aa-a597-42cb-8a26-f615d4bf99bf" containerName="container-00" Feb 02 10:16:42 crc kubenswrapper[4764]: I0202 10:16:42.192969 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="5877a8aa-a597-42cb-8a26-f615d4bf99bf" containerName="container-00" Feb 02 10:16:42 crc kubenswrapper[4764]: E0202 10:16:42.193042 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df71a15a-b67b-43b7-9487-7723c62304f3" containerName="registry-server" Feb 02 10:16:42 crc kubenswrapper[4764]: I0202 10:16:42.193101 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="df71a15a-b67b-43b7-9487-7723c62304f3" containerName="registry-server" Feb 02 10:16:42 crc kubenswrapper[4764]: E0202 10:16:42.193158 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df71a15a-b67b-43b7-9487-7723c62304f3" containerName="extract-utilities" Feb 02 10:16:42 crc kubenswrapper[4764]: I0202 10:16:42.193209 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="df71a15a-b67b-43b7-9487-7723c62304f3" containerName="extract-utilities" Feb 02 10:16:42 crc kubenswrapper[4764]: I0202 10:16:42.193448 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="5877a8aa-a597-42cb-8a26-f615d4bf99bf" containerName="container-00" Feb 02 10:16:42 crc kubenswrapper[4764]: I0202 10:16:42.193513 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="df71a15a-b67b-43b7-9487-7723c62304f3" containerName="registry-server" Feb 02 10:16:42 crc kubenswrapper[4764]: I0202 10:16:42.194822 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-n8597" Feb 02 10:16:42 crc kubenswrapper[4764]: I0202 10:16:42.213765 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-n8597"] Feb 02 10:16:42 crc kubenswrapper[4764]: I0202 10:16:42.328268 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d7582198-460d-45f8-a121-04b97151ac51-catalog-content\") pod \"redhat-marketplace-n8597\" (UID: \"d7582198-460d-45f8-a121-04b97151ac51\") " pod="openshift-marketplace/redhat-marketplace-n8597" Feb 02 10:16:42 crc kubenswrapper[4764]: I0202 10:16:42.328400 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fwvk5\" (UniqueName: \"kubernetes.io/projected/d7582198-460d-45f8-a121-04b97151ac51-kube-api-access-fwvk5\") pod \"redhat-marketplace-n8597\" (UID: \"d7582198-460d-45f8-a121-04b97151ac51\") " pod="openshift-marketplace/redhat-marketplace-n8597" Feb 02 10:16:42 crc kubenswrapper[4764]: I0202 10:16:42.328793 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d7582198-460d-45f8-a121-04b97151ac51-utilities\") pod \"redhat-marketplace-n8597\" (UID: \"d7582198-460d-45f8-a121-04b97151ac51\") " pod="openshift-marketplace/redhat-marketplace-n8597" Feb 02 10:16:42 crc kubenswrapper[4764]: I0202 10:16:42.431042 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d7582198-460d-45f8-a121-04b97151ac51-utilities\") pod \"redhat-marketplace-n8597\" (UID: \"d7582198-460d-45f8-a121-04b97151ac51\") " pod="openshift-marketplace/redhat-marketplace-n8597" Feb 02 10:16:42 crc kubenswrapper[4764]: I0202 10:16:42.431138 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d7582198-460d-45f8-a121-04b97151ac51-catalog-content\") pod \"redhat-marketplace-n8597\" (UID: \"d7582198-460d-45f8-a121-04b97151ac51\") " pod="openshift-marketplace/redhat-marketplace-n8597" Feb 02 10:16:42 crc kubenswrapper[4764]: I0202 10:16:42.431219 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fwvk5\" (UniqueName: \"kubernetes.io/projected/d7582198-460d-45f8-a121-04b97151ac51-kube-api-access-fwvk5\") pod \"redhat-marketplace-n8597\" (UID: \"d7582198-460d-45f8-a121-04b97151ac51\") " pod="openshift-marketplace/redhat-marketplace-n8597" Feb 02 10:16:42 crc kubenswrapper[4764]: I0202 10:16:42.431593 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d7582198-460d-45f8-a121-04b97151ac51-catalog-content\") pod \"redhat-marketplace-n8597\" (UID: \"d7582198-460d-45f8-a121-04b97151ac51\") " pod="openshift-marketplace/redhat-marketplace-n8597" Feb 02 10:16:42 crc kubenswrapper[4764]: I0202 10:16:42.431801 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d7582198-460d-45f8-a121-04b97151ac51-utilities\") pod \"redhat-marketplace-n8597\" (UID: \"d7582198-460d-45f8-a121-04b97151ac51\") " pod="openshift-marketplace/redhat-marketplace-n8597" Feb 02 10:16:42 crc kubenswrapper[4764]: I0202 10:16:42.853732 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fwvk5\" (UniqueName: \"kubernetes.io/projected/d7582198-460d-45f8-a121-04b97151ac51-kube-api-access-fwvk5\") pod \"redhat-marketplace-n8597\" (UID: \"d7582198-460d-45f8-a121-04b97151ac51\") " pod="openshift-marketplace/redhat-marketplace-n8597" Feb 02 10:16:43 crc kubenswrapper[4764]: I0202 10:16:43.121653 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-n8597" Feb 02 10:16:43 crc kubenswrapper[4764]: I0202 10:16:43.607814 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-n8597"] Feb 02 10:16:43 crc kubenswrapper[4764]: W0202 10:16:43.616266 4764 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd7582198_460d_45f8_a121_04b97151ac51.slice/crio-ab33b3ec142c2a895c790c1a7ce9dc93310a4ca05a3458a9d30356713ac1d226 WatchSource:0}: Error finding container ab33b3ec142c2a895c790c1a7ce9dc93310a4ca05a3458a9d30356713ac1d226: Status 404 returned error can't find the container with id ab33b3ec142c2a895c790c1a7ce9dc93310a4ca05a3458a9d30356713ac1d226 Feb 02 10:16:43 crc kubenswrapper[4764]: I0202 10:16:43.667578 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-n8597" event={"ID":"d7582198-460d-45f8-a121-04b97151ac51","Type":"ContainerStarted","Data":"ab33b3ec142c2a895c790c1a7ce9dc93310a4ca05a3458a9d30356713ac1d226"} Feb 02 10:16:44 crc kubenswrapper[4764]: I0202 10:16:44.681911 4764 generic.go:334] "Generic (PLEG): container finished" podID="d7582198-460d-45f8-a121-04b97151ac51" containerID="d3499a01661c807f8a38ded34267d02e19162b24f6f88cd4c16563dc5bb76c24" exitCode=0 Feb 02 10:16:44 crc kubenswrapper[4764]: I0202 10:16:44.682303 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-n8597" event={"ID":"d7582198-460d-45f8-a121-04b97151ac51","Type":"ContainerDied","Data":"d3499a01661c807f8a38ded34267d02e19162b24f6f88cd4c16563dc5bb76c24"} Feb 02 10:16:45 crc kubenswrapper[4764]: I0202 10:16:45.691437 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-n8597" event={"ID":"d7582198-460d-45f8-a121-04b97151ac51","Type":"ContainerStarted","Data":"83e16a9139f9350c2f8e7ac9d214d5502b5012f86db76ec5f71bd203ce854cc7"} Feb 02 10:16:46 crc kubenswrapper[4764]: I0202 10:16:46.703768 4764 generic.go:334] "Generic (PLEG): container finished" podID="d7582198-460d-45f8-a121-04b97151ac51" containerID="83e16a9139f9350c2f8e7ac9d214d5502b5012f86db76ec5f71bd203ce854cc7" exitCode=0 Feb 02 10:16:46 crc kubenswrapper[4764]: I0202 10:16:46.704193 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-n8597" event={"ID":"d7582198-460d-45f8-a121-04b97151ac51","Type":"ContainerDied","Data":"83e16a9139f9350c2f8e7ac9d214d5502b5012f86db76ec5f71bd203ce854cc7"} Feb 02 10:16:47 crc kubenswrapper[4764]: I0202 10:16:47.714780 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-n8597" event={"ID":"d7582198-460d-45f8-a121-04b97151ac51","Type":"ContainerStarted","Data":"78f553c78fe4d6efdeddd9c7c989fce6569156594c2406333250330c4dc83a60"} Feb 02 10:16:47 crc kubenswrapper[4764]: I0202 10:16:47.740606 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-n8597" podStartSLOduration=3.2650446300000002 podStartE2EDuration="5.740589271s" podCreationTimestamp="2026-02-02 10:16:42 +0000 UTC" firstStartedPulling="2026-02-02 10:16:44.715432439 +0000 UTC m=+4167.649156537" lastFinishedPulling="2026-02-02 10:16:47.19097709 +0000 UTC m=+4170.124701178" observedRunningTime="2026-02-02 10:16:47.733095643 +0000 UTC m=+4170.666819741" watchObservedRunningTime="2026-02-02 10:16:47.740589271 +0000 UTC m=+4170.674313359" Feb 02 10:16:53 crc kubenswrapper[4764]: I0202 10:16:53.121984 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-n8597" Feb 02 10:16:53 crc kubenswrapper[4764]: I0202 10:16:53.122423 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-n8597" Feb 02 10:16:53 crc kubenswrapper[4764]: I0202 10:16:53.519620 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-n8597" Feb 02 10:16:53 crc kubenswrapper[4764]: I0202 10:16:53.816734 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-n8597" Feb 02 10:16:54 crc kubenswrapper[4764]: I0202 10:16:54.376383 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-mhq9r"] Feb 02 10:16:54 crc kubenswrapper[4764]: I0202 10:16:54.379085 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mhq9r" Feb 02 10:16:54 crc kubenswrapper[4764]: I0202 10:16:54.407052 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mhq9r"] Feb 02 10:16:54 crc kubenswrapper[4764]: I0202 10:16:54.554167 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/66dc0300-f006-4614-a263-812a27ce11f6-utilities\") pod \"certified-operators-mhq9r\" (UID: \"66dc0300-f006-4614-a263-812a27ce11f6\") " pod="openshift-marketplace/certified-operators-mhq9r" Feb 02 10:16:54 crc kubenswrapper[4764]: I0202 10:16:54.554216 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qhcrm\" (UniqueName: \"kubernetes.io/projected/66dc0300-f006-4614-a263-812a27ce11f6-kube-api-access-qhcrm\") pod \"certified-operators-mhq9r\" (UID: \"66dc0300-f006-4614-a263-812a27ce11f6\") " pod="openshift-marketplace/certified-operators-mhq9r" Feb 02 10:16:54 crc kubenswrapper[4764]: I0202 10:16:54.554588 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/66dc0300-f006-4614-a263-812a27ce11f6-catalog-content\") pod \"certified-operators-mhq9r\" (UID: \"66dc0300-f006-4614-a263-812a27ce11f6\") " pod="openshift-marketplace/certified-operators-mhq9r" Feb 02 10:16:54 crc kubenswrapper[4764]: I0202 10:16:54.656557 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/66dc0300-f006-4614-a263-812a27ce11f6-catalog-content\") pod \"certified-operators-mhq9r\" (UID: \"66dc0300-f006-4614-a263-812a27ce11f6\") " pod="openshift-marketplace/certified-operators-mhq9r" Feb 02 10:16:54 crc kubenswrapper[4764]: I0202 10:16:54.656670 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/66dc0300-f006-4614-a263-812a27ce11f6-utilities\") pod \"certified-operators-mhq9r\" (UID: \"66dc0300-f006-4614-a263-812a27ce11f6\") " pod="openshift-marketplace/certified-operators-mhq9r" Feb 02 10:16:54 crc kubenswrapper[4764]: I0202 10:16:54.656704 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qhcrm\" (UniqueName: \"kubernetes.io/projected/66dc0300-f006-4614-a263-812a27ce11f6-kube-api-access-qhcrm\") pod \"certified-operators-mhq9r\" (UID: \"66dc0300-f006-4614-a263-812a27ce11f6\") " pod="openshift-marketplace/certified-operators-mhq9r" Feb 02 10:16:54 crc kubenswrapper[4764]: I0202 10:16:54.657132 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/66dc0300-f006-4614-a263-812a27ce11f6-catalog-content\") pod \"certified-operators-mhq9r\" (UID: \"66dc0300-f006-4614-a263-812a27ce11f6\") " pod="openshift-marketplace/certified-operators-mhq9r" Feb 02 10:16:54 crc kubenswrapper[4764]: I0202 10:16:54.657417 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/66dc0300-f006-4614-a263-812a27ce11f6-utilities\") pod \"certified-operators-mhq9r\" (UID: \"66dc0300-f006-4614-a263-812a27ce11f6\") " pod="openshift-marketplace/certified-operators-mhq9r" Feb 02 10:16:54 crc kubenswrapper[4764]: I0202 10:16:54.678627 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qhcrm\" (UniqueName: \"kubernetes.io/projected/66dc0300-f006-4614-a263-812a27ce11f6-kube-api-access-qhcrm\") pod \"certified-operators-mhq9r\" (UID: \"66dc0300-f006-4614-a263-812a27ce11f6\") " pod="openshift-marketplace/certified-operators-mhq9r" Feb 02 10:16:54 crc kubenswrapper[4764]: I0202 10:16:54.702689 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mhq9r" Feb 02 10:16:55 crc kubenswrapper[4764]: I0202 10:16:55.228756 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mhq9r"] Feb 02 10:16:55 crc kubenswrapper[4764]: I0202 10:16:55.358473 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-n8597"] Feb 02 10:16:55 crc kubenswrapper[4764]: I0202 10:16:55.792462 4764 generic.go:334] "Generic (PLEG): container finished" podID="66dc0300-f006-4614-a263-812a27ce11f6" containerID="2068ba96484fe8c8abc4b7349d8b5423cebb01eb4adec87129c1c40128962fd1" exitCode=0 Feb 02 10:16:55 crc kubenswrapper[4764]: I0202 10:16:55.793797 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-n8597" podUID="d7582198-460d-45f8-a121-04b97151ac51" containerName="registry-server" containerID="cri-o://78f553c78fe4d6efdeddd9c7c989fce6569156594c2406333250330c4dc83a60" gracePeriod=2 Feb 02 10:16:55 crc kubenswrapper[4764]: I0202 10:16:55.792700 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mhq9r" event={"ID":"66dc0300-f006-4614-a263-812a27ce11f6","Type":"ContainerDied","Data":"2068ba96484fe8c8abc4b7349d8b5423cebb01eb4adec87129c1c40128962fd1"} Feb 02 10:16:55 crc kubenswrapper[4764]: I0202 10:16:55.794098 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mhq9r" event={"ID":"66dc0300-f006-4614-a263-812a27ce11f6","Type":"ContainerStarted","Data":"8a529bca5d0592d3171e9ac6715fca6957648c5f39c8fd717f77051ac75d12d4"} Feb 02 10:16:55 crc kubenswrapper[4764]: I0202 10:16:55.796735 4764 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 02 10:16:56 crc kubenswrapper[4764]: I0202 10:16:56.291253 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-n8597" Feb 02 10:16:56 crc kubenswrapper[4764]: I0202 10:16:56.389218 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d7582198-460d-45f8-a121-04b97151ac51-utilities\") pod \"d7582198-460d-45f8-a121-04b97151ac51\" (UID: \"d7582198-460d-45f8-a121-04b97151ac51\") " Feb 02 10:16:56 crc kubenswrapper[4764]: I0202 10:16:56.389530 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d7582198-460d-45f8-a121-04b97151ac51-catalog-content\") pod \"d7582198-460d-45f8-a121-04b97151ac51\" (UID: \"d7582198-460d-45f8-a121-04b97151ac51\") " Feb 02 10:16:56 crc kubenswrapper[4764]: I0202 10:16:56.389852 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fwvk5\" (UniqueName: \"kubernetes.io/projected/d7582198-460d-45f8-a121-04b97151ac51-kube-api-access-fwvk5\") pod \"d7582198-460d-45f8-a121-04b97151ac51\" (UID: \"d7582198-460d-45f8-a121-04b97151ac51\") " Feb 02 10:16:56 crc kubenswrapper[4764]: I0202 10:16:56.392098 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d7582198-460d-45f8-a121-04b97151ac51-utilities" (OuterVolumeSpecName: "utilities") pod "d7582198-460d-45f8-a121-04b97151ac51" (UID: "d7582198-460d-45f8-a121-04b97151ac51"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 10:16:56 crc kubenswrapper[4764]: I0202 10:16:56.412976 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d7582198-460d-45f8-a121-04b97151ac51-kube-api-access-fwvk5" (OuterVolumeSpecName: "kube-api-access-fwvk5") pod "d7582198-460d-45f8-a121-04b97151ac51" (UID: "d7582198-460d-45f8-a121-04b97151ac51"). InnerVolumeSpecName "kube-api-access-fwvk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 10:16:56 crc kubenswrapper[4764]: I0202 10:16:56.439465 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d7582198-460d-45f8-a121-04b97151ac51-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d7582198-460d-45f8-a121-04b97151ac51" (UID: "d7582198-460d-45f8-a121-04b97151ac51"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 10:16:56 crc kubenswrapper[4764]: I0202 10:16:56.493192 4764 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d7582198-460d-45f8-a121-04b97151ac51-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 10:16:56 crc kubenswrapper[4764]: I0202 10:16:56.493231 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fwvk5\" (UniqueName: \"kubernetes.io/projected/d7582198-460d-45f8-a121-04b97151ac51-kube-api-access-fwvk5\") on node \"crc\" DevicePath \"\"" Feb 02 10:16:56 crc kubenswrapper[4764]: I0202 10:16:56.493242 4764 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d7582198-460d-45f8-a121-04b97151ac51-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 10:16:56 crc kubenswrapper[4764]: I0202 10:16:56.804664 4764 generic.go:334] "Generic (PLEG): container finished" podID="d7582198-460d-45f8-a121-04b97151ac51" containerID="78f553c78fe4d6efdeddd9c7c989fce6569156594c2406333250330c4dc83a60" exitCode=0 Feb 02 10:16:56 crc kubenswrapper[4764]: I0202 10:16:56.804739 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-n8597" Feb 02 10:16:56 crc kubenswrapper[4764]: I0202 10:16:56.804740 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-n8597" event={"ID":"d7582198-460d-45f8-a121-04b97151ac51","Type":"ContainerDied","Data":"78f553c78fe4d6efdeddd9c7c989fce6569156594c2406333250330c4dc83a60"} Feb 02 10:16:56 crc kubenswrapper[4764]: I0202 10:16:56.805680 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-n8597" event={"ID":"d7582198-460d-45f8-a121-04b97151ac51","Type":"ContainerDied","Data":"ab33b3ec142c2a895c790c1a7ce9dc93310a4ca05a3458a9d30356713ac1d226"} Feb 02 10:16:56 crc kubenswrapper[4764]: I0202 10:16:56.805774 4764 scope.go:117] "RemoveContainer" containerID="78f553c78fe4d6efdeddd9c7c989fce6569156594c2406333250330c4dc83a60" Feb 02 10:16:56 crc kubenswrapper[4764]: I0202 10:16:56.807558 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mhq9r" event={"ID":"66dc0300-f006-4614-a263-812a27ce11f6","Type":"ContainerStarted","Data":"4a73f68cadfc6fbcff4e3dfe3a65441e400a529b24316f62b1340de534ebf38a"} Feb 02 10:16:56 crc kubenswrapper[4764]: I0202 10:16:56.825513 4764 scope.go:117] "RemoveContainer" containerID="83e16a9139f9350c2f8e7ac9d214d5502b5012f86db76ec5f71bd203ce854cc7" Feb 02 10:16:56 crc kubenswrapper[4764]: I0202 10:16:56.854715 4764 scope.go:117] "RemoveContainer" containerID="d3499a01661c807f8a38ded34267d02e19162b24f6f88cd4c16563dc5bb76c24" Feb 02 10:16:56 crc kubenswrapper[4764]: I0202 10:16:56.866373 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-n8597"] Feb 02 10:16:56 crc kubenswrapper[4764]: I0202 10:16:56.876313 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-n8597"] Feb 02 10:16:56 crc kubenswrapper[4764]: I0202 10:16:56.877234 4764 scope.go:117] "RemoveContainer" containerID="78f553c78fe4d6efdeddd9c7c989fce6569156594c2406333250330c4dc83a60" Feb 02 10:16:56 crc kubenswrapper[4764]: E0202 10:16:56.879035 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"78f553c78fe4d6efdeddd9c7c989fce6569156594c2406333250330c4dc83a60\": container with ID starting with 78f553c78fe4d6efdeddd9c7c989fce6569156594c2406333250330c4dc83a60 not found: ID does not exist" containerID="78f553c78fe4d6efdeddd9c7c989fce6569156594c2406333250330c4dc83a60" Feb 02 10:16:56 crc kubenswrapper[4764]: I0202 10:16:56.879076 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"78f553c78fe4d6efdeddd9c7c989fce6569156594c2406333250330c4dc83a60"} err="failed to get container status \"78f553c78fe4d6efdeddd9c7c989fce6569156594c2406333250330c4dc83a60\": rpc error: code = NotFound desc = could not find container \"78f553c78fe4d6efdeddd9c7c989fce6569156594c2406333250330c4dc83a60\": container with ID starting with 78f553c78fe4d6efdeddd9c7c989fce6569156594c2406333250330c4dc83a60 not found: ID does not exist" Feb 02 10:16:56 crc kubenswrapper[4764]: I0202 10:16:56.879103 4764 scope.go:117] "RemoveContainer" containerID="83e16a9139f9350c2f8e7ac9d214d5502b5012f86db76ec5f71bd203ce854cc7" Feb 02 10:16:56 crc kubenswrapper[4764]: E0202 10:16:56.879400 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"83e16a9139f9350c2f8e7ac9d214d5502b5012f86db76ec5f71bd203ce854cc7\": container with ID starting with 83e16a9139f9350c2f8e7ac9d214d5502b5012f86db76ec5f71bd203ce854cc7 not found: ID does not exist" containerID="83e16a9139f9350c2f8e7ac9d214d5502b5012f86db76ec5f71bd203ce854cc7" Feb 02 10:16:56 crc kubenswrapper[4764]: I0202 10:16:56.879485 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"83e16a9139f9350c2f8e7ac9d214d5502b5012f86db76ec5f71bd203ce854cc7"} err="failed to get container status \"83e16a9139f9350c2f8e7ac9d214d5502b5012f86db76ec5f71bd203ce854cc7\": rpc error: code = NotFound desc = could not find container \"83e16a9139f9350c2f8e7ac9d214d5502b5012f86db76ec5f71bd203ce854cc7\": container with ID starting with 83e16a9139f9350c2f8e7ac9d214d5502b5012f86db76ec5f71bd203ce854cc7 not found: ID does not exist" Feb 02 10:16:56 crc kubenswrapper[4764]: I0202 10:16:56.879564 4764 scope.go:117] "RemoveContainer" containerID="d3499a01661c807f8a38ded34267d02e19162b24f6f88cd4c16563dc5bb76c24" Feb 02 10:16:56 crc kubenswrapper[4764]: E0202 10:16:56.879834 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d3499a01661c807f8a38ded34267d02e19162b24f6f88cd4c16563dc5bb76c24\": container with ID starting with d3499a01661c807f8a38ded34267d02e19162b24f6f88cd4c16563dc5bb76c24 not found: ID does not exist" containerID="d3499a01661c807f8a38ded34267d02e19162b24f6f88cd4c16563dc5bb76c24" Feb 02 10:16:56 crc kubenswrapper[4764]: I0202 10:16:56.879860 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d3499a01661c807f8a38ded34267d02e19162b24f6f88cd4c16563dc5bb76c24"} err="failed to get container status \"d3499a01661c807f8a38ded34267d02e19162b24f6f88cd4c16563dc5bb76c24\": rpc error: code = NotFound desc = could not find container \"d3499a01661c807f8a38ded34267d02e19162b24f6f88cd4c16563dc5bb76c24\": container with ID starting with d3499a01661c807f8a38ded34267d02e19162b24f6f88cd4c16563dc5bb76c24 not found: ID does not exist" Feb 02 10:16:57 crc kubenswrapper[4764]: I0202 10:16:57.757120 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_805de4b7130efd4fbe2290d0c51d78e6295b83d77efc3ff2d1b014ffe7z8r74_894b5d6a-cddd-43e6-a59e-6076f985e5a3/util/0.log" Feb 02 10:16:57 crc kubenswrapper[4764]: I0202 10:16:57.841788 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d7582198-460d-45f8-a121-04b97151ac51" path="/var/lib/kubelet/pods/d7582198-460d-45f8-a121-04b97151ac51/volumes" Feb 02 10:16:57 crc kubenswrapper[4764]: I0202 10:16:57.932876 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_805de4b7130efd4fbe2290d0c51d78e6295b83d77efc3ff2d1b014ffe7z8r74_894b5d6a-cddd-43e6-a59e-6076f985e5a3/pull/0.log" Feb 02 10:16:58 crc kubenswrapper[4764]: I0202 10:16:58.018040 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_805de4b7130efd4fbe2290d0c51d78e6295b83d77efc3ff2d1b014ffe7z8r74_894b5d6a-cddd-43e6-a59e-6076f985e5a3/pull/0.log" Feb 02 10:16:58 crc kubenswrapper[4764]: I0202 10:16:58.036764 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_805de4b7130efd4fbe2290d0c51d78e6295b83d77efc3ff2d1b014ffe7z8r74_894b5d6a-cddd-43e6-a59e-6076f985e5a3/util/0.log" Feb 02 10:16:58 crc kubenswrapper[4764]: I0202 10:16:58.230044 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_805de4b7130efd4fbe2290d0c51d78e6295b83d77efc3ff2d1b014ffe7z8r74_894b5d6a-cddd-43e6-a59e-6076f985e5a3/util/0.log" Feb 02 10:16:58 crc kubenswrapper[4764]: I0202 10:16:58.246563 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_805de4b7130efd4fbe2290d0c51d78e6295b83d77efc3ff2d1b014ffe7z8r74_894b5d6a-cddd-43e6-a59e-6076f985e5a3/pull/0.log" Feb 02 10:16:58 crc kubenswrapper[4764]: I0202 10:16:58.323211 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_805de4b7130efd4fbe2290d0c51d78e6295b83d77efc3ff2d1b014ffe7z8r74_894b5d6a-cddd-43e6-a59e-6076f985e5a3/extract/0.log" Feb 02 10:16:58 crc kubenswrapper[4764]: I0202 10:16:58.545756 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7b6c4d8c5f-wtkdr_9cbd726b-9dbb-46b3-9bcd-f3e9813b4f07/manager/0.log" Feb 02 10:16:58 crc kubenswrapper[4764]: I0202 10:16:58.599070 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-8d874c8fc-czpmn_7c8a2462-306c-43da-a538-9e34a11b5c21/manager/0.log" Feb 02 10:16:58 crc kubenswrapper[4764]: I0202 10:16:58.745192 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-6d9697b7f4-w4qf5_8b0e7cb4-257b-403b-8b92-4578ce77c388/manager/0.log" Feb 02 10:16:58 crc kubenswrapper[4764]: I0202 10:16:58.828742 4764 generic.go:334] "Generic (PLEG): container finished" podID="66dc0300-f006-4614-a263-812a27ce11f6" containerID="4a73f68cadfc6fbcff4e3dfe3a65441e400a529b24316f62b1340de534ebf38a" exitCode=0 Feb 02 10:16:58 crc kubenswrapper[4764]: I0202 10:16:58.829006 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mhq9r" event={"ID":"66dc0300-f006-4614-a263-812a27ce11f6","Type":"ContainerDied","Data":"4a73f68cadfc6fbcff4e3dfe3a65441e400a529b24316f62b1340de534ebf38a"} Feb 02 10:16:58 crc kubenswrapper[4764]: I0202 10:16:58.969284 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-8886f4c47-hvtx6_ee5cb41c-4d3d-4a9d-b1f7-1ede61e7b21c/manager/0.log" Feb 02 10:16:58 crc kubenswrapper[4764]: I0202 10:16:58.998061 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-69d6db494d-5f766_becd2ba3-ccad-4ea9-b3c7-6b39f6f8cb3d/manager/0.log" Feb 02 10:16:59 crc kubenswrapper[4764]: I0202 10:16:59.187258 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-5fb775575f-ntj47_82e16b73-9dfe-413b-b3dd-65da79b3d9dd/manager/0.log" Feb 02 10:16:59 crc kubenswrapper[4764]: I0202 10:16:59.483622 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-5f4b8bd54d-h4n52_ad8e7100-d5b8-4a56-bbda-056483f79113/manager/0.log" Feb 02 10:16:59 crc kubenswrapper[4764]: I0202 10:16:59.534431 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-79955696d6-mxvzm_b593c3f3-a225-4cb6-b967-4c056e3326a0/manager/0.log" Feb 02 10:16:59 crc kubenswrapper[4764]: I0202 10:16:59.761413 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-84f48565d4-qqnp6_8564e79b-9abd-41e1-a0a7-0d4b71011300/manager/0.log" Feb 02 10:16:59 crc kubenswrapper[4764]: I0202 10:16:59.839771 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mhq9r" event={"ID":"66dc0300-f006-4614-a263-812a27ce11f6","Type":"ContainerStarted","Data":"013aed9eae1e8e38b722f4836ff2c4fdf873aad33cff8bd25588dff59581b024"} Feb 02 10:16:59 crc kubenswrapper[4764]: I0202 10:16:59.844381 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7dd968899f-g2p6c_6d4fffad-a306-4f0b-ac40-7fa839f75aa3/manager/0.log" Feb 02 10:16:59 crc kubenswrapper[4764]: I0202 10:16:59.863275 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-mhq9r" podStartSLOduration=2.4456861659999998 podStartE2EDuration="5.863243051s" podCreationTimestamp="2026-02-02 10:16:54 +0000 UTC" firstStartedPulling="2026-02-02 10:16:55.79631366 +0000 UTC m=+4178.730037758" lastFinishedPulling="2026-02-02 10:16:59.213870555 +0000 UTC m=+4182.147594643" observedRunningTime="2026-02-02 10:16:59.862301136 +0000 UTC m=+4182.796025224" watchObservedRunningTime="2026-02-02 10:16:59.863243051 +0000 UTC m=+4182.796967139" Feb 02 10:17:00 crc kubenswrapper[4764]: I0202 10:17:00.040740 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-67bf948998-jhbzc_9e55cb1e-ed3b-4b7b-8b52-42436eb48223/manager/0.log" Feb 02 10:17:00 crc kubenswrapper[4764]: I0202 10:17:00.518156 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-585dbc889-7nrsg_f5cc9236-c70f-4449-a902-ef416aaf8781/manager/0.log" Feb 02 10:17:00 crc kubenswrapper[4764]: I0202 10:17:00.806773 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-55bff696bd-p7nmb_e98e75d4-963f-4a67-894e-302b7429eb01/manager/0.log" Feb 02 10:17:01 crc kubenswrapper[4764]: I0202 10:17:01.161659 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-6687f8d877-5ncb8_1844ef02-28fb-45ce-a697-54f1560c9087/manager/0.log" Feb 02 10:17:01 crc kubenswrapper[4764]: I0202 10:17:01.285708 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-5ff45dfdbfb5fxl_1b05c36c-dbb6-4ad4-8985-8d64551812ff/manager/0.log" Feb 02 10:17:01 crc kubenswrapper[4764]: I0202 10:17:01.515859 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-init-6bf6665fd-l67kz_ee6621b4-2d57-402e-bfda-df7f7f922dc8/operator/0.log" Feb 02 10:17:01 crc kubenswrapper[4764]: I0202 10:17:01.885495 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-8ctp4_50551dd4-71f5-4eb4-8736-ed04f0845c7e/registry-server/0.log" Feb 02 10:17:02 crc kubenswrapper[4764]: I0202 10:17:02.443371 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-5b964cf4cd-kskcp_1464055e-4745-45e3-be46-5b7700dce733/manager/0.log" Feb 02 10:17:02 crc kubenswrapper[4764]: I0202 10:17:02.451528 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-788c46999f-pk2hh_674c8820-f063-487b-9122-1ed46d1bc39f/manager/0.log" Feb 02 10:17:02 crc kubenswrapper[4764]: I0202 10:17:02.743176 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-qmjfr_36bffbf8-2209-4410-8005-d0164bb1eb34/operator/0.log" Feb 02 10:17:02 crc kubenswrapper[4764]: I0202 10:17:02.826069 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-646f757d77-g8bhc_a46546b9-e589-4eca-bca7-9150038f9ab8/manager/0.log" Feb 02 10:17:02 crc kubenswrapper[4764]: I0202 10:17:02.952918 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-68fc8c869-fb6zn_1a100291-702c-4908-8a63-acb05c4ef5e9/manager/0.log" Feb 02 10:17:03 crc kubenswrapper[4764]: I0202 10:17:03.075256 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-64b5b76f97-98xsk_7c983dd6-f499-41d5-a049-765a15b4f912/manager/0.log" Feb 02 10:17:03 crc kubenswrapper[4764]: I0202 10:17:03.138086 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-56f8bfcd9f-bd9cj_b8748b41-5e08-47bf-a853-ad8c4f2b0908/manager/0.log" Feb 02 10:17:03 crc kubenswrapper[4764]: I0202 10:17:03.277867 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-564965969-6kcc6_65c22016-b2f2-4a07-87be-480d18870dea/manager/0.log" Feb 02 10:17:04 crc kubenswrapper[4764]: I0202 10:17:04.704050 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-mhq9r" Feb 02 10:17:04 crc kubenswrapper[4764]: I0202 10:17:04.704439 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-mhq9r" Feb 02 10:17:05 crc kubenswrapper[4764]: I0202 10:17:05.758585 4764 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-mhq9r" podUID="66dc0300-f006-4614-a263-812a27ce11f6" containerName="registry-server" probeResult="failure" output=< Feb 02 10:17:05 crc kubenswrapper[4764]: timeout: failed to connect service ":50051" within 1s Feb 02 10:17:05 crc kubenswrapper[4764]: > Feb 02 10:17:14 crc kubenswrapper[4764]: I0202 10:17:14.758503 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-mhq9r" Feb 02 10:17:14 crc kubenswrapper[4764]: I0202 10:17:14.818500 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-mhq9r" Feb 02 10:17:14 crc kubenswrapper[4764]: I0202 10:17:14.998639 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-mhq9r"] Feb 02 10:17:15 crc kubenswrapper[4764]: I0202 10:17:15.971207 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-mhq9r" podUID="66dc0300-f006-4614-a263-812a27ce11f6" containerName="registry-server" containerID="cri-o://013aed9eae1e8e38b722f4836ff2c4fdf873aad33cff8bd25588dff59581b024" gracePeriod=2 Feb 02 10:17:16 crc kubenswrapper[4764]: I0202 10:17:16.402176 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mhq9r" Feb 02 10:17:16 crc kubenswrapper[4764]: I0202 10:17:16.514660 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qhcrm\" (UniqueName: \"kubernetes.io/projected/66dc0300-f006-4614-a263-812a27ce11f6-kube-api-access-qhcrm\") pod \"66dc0300-f006-4614-a263-812a27ce11f6\" (UID: \"66dc0300-f006-4614-a263-812a27ce11f6\") " Feb 02 10:17:16 crc kubenswrapper[4764]: I0202 10:17:16.514745 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/66dc0300-f006-4614-a263-812a27ce11f6-catalog-content\") pod \"66dc0300-f006-4614-a263-812a27ce11f6\" (UID: \"66dc0300-f006-4614-a263-812a27ce11f6\") " Feb 02 10:17:16 crc kubenswrapper[4764]: I0202 10:17:16.514809 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/66dc0300-f006-4614-a263-812a27ce11f6-utilities\") pod \"66dc0300-f006-4614-a263-812a27ce11f6\" (UID: \"66dc0300-f006-4614-a263-812a27ce11f6\") " Feb 02 10:17:16 crc kubenswrapper[4764]: I0202 10:17:16.515592 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/66dc0300-f006-4614-a263-812a27ce11f6-utilities" (OuterVolumeSpecName: "utilities") pod "66dc0300-f006-4614-a263-812a27ce11f6" (UID: "66dc0300-f006-4614-a263-812a27ce11f6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 10:17:16 crc kubenswrapper[4764]: I0202 10:17:16.525762 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/66dc0300-f006-4614-a263-812a27ce11f6-kube-api-access-qhcrm" (OuterVolumeSpecName: "kube-api-access-qhcrm") pod "66dc0300-f006-4614-a263-812a27ce11f6" (UID: "66dc0300-f006-4614-a263-812a27ce11f6"). InnerVolumeSpecName "kube-api-access-qhcrm". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 10:17:16 crc kubenswrapper[4764]: I0202 10:17:16.570377 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/66dc0300-f006-4614-a263-812a27ce11f6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "66dc0300-f006-4614-a263-812a27ce11f6" (UID: "66dc0300-f006-4614-a263-812a27ce11f6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 10:17:16 crc kubenswrapper[4764]: I0202 10:17:16.617275 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qhcrm\" (UniqueName: \"kubernetes.io/projected/66dc0300-f006-4614-a263-812a27ce11f6-kube-api-access-qhcrm\") on node \"crc\" DevicePath \"\"" Feb 02 10:17:16 crc kubenswrapper[4764]: I0202 10:17:16.617309 4764 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/66dc0300-f006-4614-a263-812a27ce11f6-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 10:17:16 crc kubenswrapper[4764]: I0202 10:17:16.617317 4764 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/66dc0300-f006-4614-a263-812a27ce11f6-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 10:17:16 crc kubenswrapper[4764]: I0202 10:17:16.980778 4764 generic.go:334] "Generic (PLEG): container finished" podID="66dc0300-f006-4614-a263-812a27ce11f6" containerID="013aed9eae1e8e38b722f4836ff2c4fdf873aad33cff8bd25588dff59581b024" exitCode=0 Feb 02 10:17:16 crc kubenswrapper[4764]: I0202 10:17:16.980820 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mhq9r" event={"ID":"66dc0300-f006-4614-a263-812a27ce11f6","Type":"ContainerDied","Data":"013aed9eae1e8e38b722f4836ff2c4fdf873aad33cff8bd25588dff59581b024"} Feb 02 10:17:16 crc kubenswrapper[4764]: I0202 10:17:16.980824 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mhq9r" Feb 02 10:17:16 crc kubenswrapper[4764]: I0202 10:17:16.980857 4764 scope.go:117] "RemoveContainer" containerID="013aed9eae1e8e38b722f4836ff2c4fdf873aad33cff8bd25588dff59581b024" Feb 02 10:17:16 crc kubenswrapper[4764]: I0202 10:17:16.980846 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mhq9r" event={"ID":"66dc0300-f006-4614-a263-812a27ce11f6","Type":"ContainerDied","Data":"8a529bca5d0592d3171e9ac6715fca6957648c5f39c8fd717f77051ac75d12d4"} Feb 02 10:17:17 crc kubenswrapper[4764]: I0202 10:17:17.006638 4764 scope.go:117] "RemoveContainer" containerID="4a73f68cadfc6fbcff4e3dfe3a65441e400a529b24316f62b1340de534ebf38a" Feb 02 10:17:17 crc kubenswrapper[4764]: I0202 10:17:17.021094 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-mhq9r"] Feb 02 10:17:17 crc kubenswrapper[4764]: I0202 10:17:17.031211 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-mhq9r"] Feb 02 10:17:17 crc kubenswrapper[4764]: I0202 10:17:17.042384 4764 scope.go:117] "RemoveContainer" containerID="2068ba96484fe8c8abc4b7349d8b5423cebb01eb4adec87129c1c40128962fd1" Feb 02 10:17:17 crc kubenswrapper[4764]: I0202 10:17:17.066673 4764 scope.go:117] "RemoveContainer" containerID="013aed9eae1e8e38b722f4836ff2c4fdf873aad33cff8bd25588dff59581b024" Feb 02 10:17:17 crc kubenswrapper[4764]: E0202 10:17:17.067093 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"013aed9eae1e8e38b722f4836ff2c4fdf873aad33cff8bd25588dff59581b024\": container with ID starting with 013aed9eae1e8e38b722f4836ff2c4fdf873aad33cff8bd25588dff59581b024 not found: ID does not exist" containerID="013aed9eae1e8e38b722f4836ff2c4fdf873aad33cff8bd25588dff59581b024" Feb 02 10:17:17 crc kubenswrapper[4764]: I0202 10:17:17.067130 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"013aed9eae1e8e38b722f4836ff2c4fdf873aad33cff8bd25588dff59581b024"} err="failed to get container status \"013aed9eae1e8e38b722f4836ff2c4fdf873aad33cff8bd25588dff59581b024\": rpc error: code = NotFound desc = could not find container \"013aed9eae1e8e38b722f4836ff2c4fdf873aad33cff8bd25588dff59581b024\": container with ID starting with 013aed9eae1e8e38b722f4836ff2c4fdf873aad33cff8bd25588dff59581b024 not found: ID does not exist" Feb 02 10:17:17 crc kubenswrapper[4764]: I0202 10:17:17.067164 4764 scope.go:117] "RemoveContainer" containerID="4a73f68cadfc6fbcff4e3dfe3a65441e400a529b24316f62b1340de534ebf38a" Feb 02 10:17:17 crc kubenswrapper[4764]: E0202 10:17:17.067421 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4a73f68cadfc6fbcff4e3dfe3a65441e400a529b24316f62b1340de534ebf38a\": container with ID starting with 4a73f68cadfc6fbcff4e3dfe3a65441e400a529b24316f62b1340de534ebf38a not found: ID does not exist" containerID="4a73f68cadfc6fbcff4e3dfe3a65441e400a529b24316f62b1340de534ebf38a" Feb 02 10:17:17 crc kubenswrapper[4764]: I0202 10:17:17.067452 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a73f68cadfc6fbcff4e3dfe3a65441e400a529b24316f62b1340de534ebf38a"} err="failed to get container status \"4a73f68cadfc6fbcff4e3dfe3a65441e400a529b24316f62b1340de534ebf38a\": rpc error: code = NotFound desc = could not find container \"4a73f68cadfc6fbcff4e3dfe3a65441e400a529b24316f62b1340de534ebf38a\": container with ID starting with 4a73f68cadfc6fbcff4e3dfe3a65441e400a529b24316f62b1340de534ebf38a not found: ID does not exist" Feb 02 10:17:17 crc kubenswrapper[4764]: I0202 10:17:17.067474 4764 scope.go:117] "RemoveContainer" containerID="2068ba96484fe8c8abc4b7349d8b5423cebb01eb4adec87129c1c40128962fd1" Feb 02 10:17:17 crc kubenswrapper[4764]: E0202 10:17:17.067713 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2068ba96484fe8c8abc4b7349d8b5423cebb01eb4adec87129c1c40128962fd1\": container with ID starting with 2068ba96484fe8c8abc4b7349d8b5423cebb01eb4adec87129c1c40128962fd1 not found: ID does not exist" containerID="2068ba96484fe8c8abc4b7349d8b5423cebb01eb4adec87129c1c40128962fd1" Feb 02 10:17:17 crc kubenswrapper[4764]: I0202 10:17:17.067739 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2068ba96484fe8c8abc4b7349d8b5423cebb01eb4adec87129c1c40128962fd1"} err="failed to get container status \"2068ba96484fe8c8abc4b7349d8b5423cebb01eb4adec87129c1c40128962fd1\": rpc error: code = NotFound desc = could not find container \"2068ba96484fe8c8abc4b7349d8b5423cebb01eb4adec87129c1c40128962fd1\": container with ID starting with 2068ba96484fe8c8abc4b7349d8b5423cebb01eb4adec87129c1c40128962fd1 not found: ID does not exist" Feb 02 10:17:17 crc kubenswrapper[4764]: I0202 10:17:17.836641 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="66dc0300-f006-4614-a263-812a27ce11f6" path="/var/lib/kubelet/pods/66dc0300-f006-4614-a263-812a27ce11f6/volumes" Feb 02 10:17:25 crc kubenswrapper[4764]: I0202 10:17:25.171262 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-qq96b_545e08e9-b2c4-493e-8fc7-8353fcab998d/control-plane-machine-set-operator/0.log" Feb 02 10:17:25 crc kubenswrapper[4764]: I0202 10:17:25.416150 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-98wxj_08bc3386-ee9a-4a3c-aa52-22d667971076/kube-rbac-proxy/0.log" Feb 02 10:17:25 crc kubenswrapper[4764]: I0202 10:17:25.487859 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-98wxj_08bc3386-ee9a-4a3c-aa52-22d667971076/machine-api-operator/0.log" Feb 02 10:17:39 crc kubenswrapper[4764]: I0202 10:17:39.915978 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-858654f9db-mkn7c_533ed531-c13b-4cbe-9355-e2f1096b905a/cert-manager-controller/0.log" Feb 02 10:17:40 crc kubenswrapper[4764]: I0202 10:17:40.100444 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-cf98fcc89-4wc9w_1d03ea6c-c159-4e5a-a38c-1bf190d3d454/cert-manager-cainjector/0.log" Feb 02 10:17:40 crc kubenswrapper[4764]: I0202 10:17:40.222644 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-687f57d79b-8gwjv_6cc98479-e4c6-45e2-912c-c3a479af0321/cert-manager-webhook/0.log" Feb 02 10:17:55 crc kubenswrapper[4764]: I0202 10:17:55.165725 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7754f76f8b-mxdpd_67f7c8f7-535f-4a07-a330-adf7c41da701/nmstate-console-plugin/0.log" Feb 02 10:17:55 crc kubenswrapper[4764]: I0202 10:17:55.338500 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-gswb4_7e5e0b71-cd10-4ccf-8100-f15a063cceb7/nmstate-handler/0.log" Feb 02 10:17:55 crc kubenswrapper[4764]: I0202 10:17:55.381011 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-54757c584b-jz8q9_dd14d932-c6a2-4d7e-a1bb-bf7cf4c30f00/kube-rbac-proxy/0.log" Feb 02 10:17:55 crc kubenswrapper[4764]: I0202 10:17:55.421242 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-54757c584b-jz8q9_dd14d932-c6a2-4d7e-a1bb-bf7cf4c30f00/nmstate-metrics/0.log" Feb 02 10:17:55 crc kubenswrapper[4764]: I0202 10:17:55.548826 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-646758c888-mv59b_2c1baf80-192e-47b6-852c-394b7e4ade4e/nmstate-operator/0.log" Feb 02 10:17:55 crc kubenswrapper[4764]: I0202 10:17:55.656403 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-8474b5b9d8-gz969_1a77ee37-2514-4a73-b2f1-565442966b16/nmstate-webhook/0.log" Feb 02 10:18:26 crc kubenswrapper[4764]: I0202 10:18:26.534585 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6968d8fdc4-qxqpn_aa337cba-937a-44e6-af3b-50f8f3438489/controller/0.log" Feb 02 10:18:26 crc kubenswrapper[4764]: I0202 10:18:26.560215 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6968d8fdc4-qxqpn_aa337cba-937a-44e6-af3b-50f8f3438489/kube-rbac-proxy/0.log" Feb 02 10:18:26 crc kubenswrapper[4764]: I0202 10:18:26.758804 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-rtp9n_9d0a4296-e773-4500-8a43-83796d835ad9/cp-frr-files/0.log" Feb 02 10:18:26 crc kubenswrapper[4764]: I0202 10:18:26.918709 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-rtp9n_9d0a4296-e773-4500-8a43-83796d835ad9/cp-frr-files/0.log" Feb 02 10:18:26 crc kubenswrapper[4764]: I0202 10:18:26.993853 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-rtp9n_9d0a4296-e773-4500-8a43-83796d835ad9/cp-reloader/0.log" Feb 02 10:18:27 crc kubenswrapper[4764]: I0202 10:18:27.009440 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-rtp9n_9d0a4296-e773-4500-8a43-83796d835ad9/cp-metrics/0.log" Feb 02 10:18:27 crc kubenswrapper[4764]: I0202 10:18:27.021824 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-rtp9n_9d0a4296-e773-4500-8a43-83796d835ad9/cp-reloader/0.log" Feb 02 10:18:27 crc kubenswrapper[4764]: I0202 10:18:27.240249 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-rtp9n_9d0a4296-e773-4500-8a43-83796d835ad9/cp-reloader/0.log" Feb 02 10:18:27 crc kubenswrapper[4764]: I0202 10:18:27.249927 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-rtp9n_9d0a4296-e773-4500-8a43-83796d835ad9/cp-frr-files/0.log" Feb 02 10:18:27 crc kubenswrapper[4764]: I0202 10:18:27.256474 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-rtp9n_9d0a4296-e773-4500-8a43-83796d835ad9/cp-metrics/0.log" Feb 02 10:18:27 crc kubenswrapper[4764]: I0202 10:18:27.338045 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-rtp9n_9d0a4296-e773-4500-8a43-83796d835ad9/cp-metrics/0.log" Feb 02 10:18:27 crc kubenswrapper[4764]: I0202 10:18:27.464714 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-rtp9n_9d0a4296-e773-4500-8a43-83796d835ad9/cp-reloader/0.log" Feb 02 10:18:27 crc kubenswrapper[4764]: I0202 10:18:27.467538 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-rtp9n_9d0a4296-e773-4500-8a43-83796d835ad9/cp-metrics/0.log" Feb 02 10:18:27 crc kubenswrapper[4764]: I0202 10:18:27.527283 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-rtp9n_9d0a4296-e773-4500-8a43-83796d835ad9/cp-frr-files/0.log" Feb 02 10:18:27 crc kubenswrapper[4764]: I0202 10:18:27.558083 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-rtp9n_9d0a4296-e773-4500-8a43-83796d835ad9/controller/0.log" Feb 02 10:18:27 crc kubenswrapper[4764]: I0202 10:18:27.650320 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-rtp9n_9d0a4296-e773-4500-8a43-83796d835ad9/frr-metrics/0.log" Feb 02 10:18:27 crc kubenswrapper[4764]: I0202 10:18:27.789400 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-rtp9n_9d0a4296-e773-4500-8a43-83796d835ad9/kube-rbac-proxy-frr/0.log" Feb 02 10:18:27 crc kubenswrapper[4764]: I0202 10:18:27.846215 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-rtp9n_9d0a4296-e773-4500-8a43-83796d835ad9/kube-rbac-proxy/0.log" Feb 02 10:18:28 crc kubenswrapper[4764]: I0202 10:18:28.021709 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-rtp9n_9d0a4296-e773-4500-8a43-83796d835ad9/reloader/0.log" Feb 02 10:18:28 crc kubenswrapper[4764]: I0202 10:18:28.123064 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7df86c4f6c-z9ptz_4757c3fb-ae2f-498d-a551-f6c2daee3a23/frr-k8s-webhook-server/0.log" Feb 02 10:18:28 crc kubenswrapper[4764]: I0202 10:18:28.359418 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-7848c48b6-2kvbb_f479cadf-cdb2-423f-b9c5-204bef98bf8a/manager/0.log" Feb 02 10:18:28 crc kubenswrapper[4764]: I0202 10:18:28.573580 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-5645b7d98c-hhmxc_1436d9d2-a684-4204-91d6-9da1a7df5fa2/webhook-server/0.log" Feb 02 10:18:28 crc kubenswrapper[4764]: I0202 10:18:28.815118 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-jw8j4_c8444bdb-052f-4c56-841a-b3c6718759d8/kube-rbac-proxy/0.log" Feb 02 10:18:28 crc kubenswrapper[4764]: I0202 10:18:28.888512 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-rtp9n_9d0a4296-e773-4500-8a43-83796d835ad9/frr/0.log" Feb 02 10:18:29 crc kubenswrapper[4764]: I0202 10:18:29.187023 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-jw8j4_c8444bdb-052f-4c56-841a-b3c6718759d8/speaker/0.log" Feb 02 10:18:43 crc kubenswrapper[4764]: I0202 10:18:43.522587 4764 patch_prober.go:28] interesting pod/machine-config-daemon-4ndm4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 10:18:43 crc kubenswrapper[4764]: I0202 10:18:43.523219 4764 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 10:18:45 crc kubenswrapper[4764]: I0202 10:18:45.085152 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcv6r4w_6021fc8f-ba41-4247-9b13-991cee9cb2a6/util/0.log" Feb 02 10:18:45 crc kubenswrapper[4764]: I0202 10:18:45.501432 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcv6r4w_6021fc8f-ba41-4247-9b13-991cee9cb2a6/util/0.log" Feb 02 10:18:45 crc kubenswrapper[4764]: I0202 10:18:45.589234 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcv6r4w_6021fc8f-ba41-4247-9b13-991cee9cb2a6/pull/0.log" Feb 02 10:18:45 crc kubenswrapper[4764]: I0202 10:18:45.594677 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcv6r4w_6021fc8f-ba41-4247-9b13-991cee9cb2a6/pull/0.log" Feb 02 10:18:45 crc kubenswrapper[4764]: I0202 10:18:45.752670 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcv6r4w_6021fc8f-ba41-4247-9b13-991cee9cb2a6/pull/0.log" Feb 02 10:18:45 crc kubenswrapper[4764]: I0202 10:18:45.778972 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcv6r4w_6021fc8f-ba41-4247-9b13-991cee9cb2a6/util/0.log" Feb 02 10:18:45 crc kubenswrapper[4764]: I0202 10:18:45.863995 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcv6r4w_6021fc8f-ba41-4247-9b13-991cee9cb2a6/extract/0.log" Feb 02 10:18:45 crc kubenswrapper[4764]: I0202 10:18:45.974104 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec71324cpz_08e7a6ba-02aa-4f59-8e6d-b967cb0f9715/util/0.log" Feb 02 10:18:46 crc kubenswrapper[4764]: I0202 10:18:46.155045 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec71324cpz_08e7a6ba-02aa-4f59-8e6d-b967cb0f9715/util/0.log" Feb 02 10:18:46 crc kubenswrapper[4764]: I0202 10:18:46.214745 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec71324cpz_08e7a6ba-02aa-4f59-8e6d-b967cb0f9715/pull/0.log" Feb 02 10:18:46 crc kubenswrapper[4764]: I0202 10:18:46.218015 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec71324cpz_08e7a6ba-02aa-4f59-8e6d-b967cb0f9715/pull/0.log" Feb 02 10:18:46 crc kubenswrapper[4764]: I0202 10:18:46.371532 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec71324cpz_08e7a6ba-02aa-4f59-8e6d-b967cb0f9715/util/0.log" Feb 02 10:18:46 crc kubenswrapper[4764]: I0202 10:18:46.433574 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec71324cpz_08e7a6ba-02aa-4f59-8e6d-b967cb0f9715/pull/0.log" Feb 02 10:18:46 crc kubenswrapper[4764]: I0202 10:18:46.478209 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec71324cpz_08e7a6ba-02aa-4f59-8e6d-b967cb0f9715/extract/0.log" Feb 02 10:18:46 crc kubenswrapper[4764]: I0202 10:18:46.608857 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5kwbp_4fe67111-6fe1-4886-b192-b75fea98586d/extract-utilities/0.log" Feb 02 10:18:46 crc kubenswrapper[4764]: I0202 10:18:46.812767 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5kwbp_4fe67111-6fe1-4886-b192-b75fea98586d/extract-utilities/0.log" Feb 02 10:18:46 crc kubenswrapper[4764]: I0202 10:18:46.819801 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5kwbp_4fe67111-6fe1-4886-b192-b75fea98586d/extract-content/0.log" Feb 02 10:18:46 crc kubenswrapper[4764]: I0202 10:18:46.902460 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5kwbp_4fe67111-6fe1-4886-b192-b75fea98586d/extract-content/0.log" Feb 02 10:18:47 crc kubenswrapper[4764]: I0202 10:18:47.098709 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5kwbp_4fe67111-6fe1-4886-b192-b75fea98586d/extract-utilities/0.log" Feb 02 10:18:47 crc kubenswrapper[4764]: I0202 10:18:47.162340 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5kwbp_4fe67111-6fe1-4886-b192-b75fea98586d/extract-content/0.log" Feb 02 10:18:47 crc kubenswrapper[4764]: I0202 10:18:47.460345 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-6kgcj_730067d3-5e02-4aa1-af49-9bb70546886f/extract-utilities/0.log" Feb 02 10:18:47 crc kubenswrapper[4764]: I0202 10:18:47.719377 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5kwbp_4fe67111-6fe1-4886-b192-b75fea98586d/registry-server/0.log" Feb 02 10:18:47 crc kubenswrapper[4764]: I0202 10:18:47.922515 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-6kgcj_730067d3-5e02-4aa1-af49-9bb70546886f/extract-utilities/0.log" Feb 02 10:18:48 crc kubenswrapper[4764]: I0202 10:18:48.071673 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-6kgcj_730067d3-5e02-4aa1-af49-9bb70546886f/extract-content/0.log" Feb 02 10:18:48 crc kubenswrapper[4764]: I0202 10:18:48.074330 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-6kgcj_730067d3-5e02-4aa1-af49-9bb70546886f/extract-content/0.log" Feb 02 10:18:48 crc kubenswrapper[4764]: I0202 10:18:48.263760 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-6kgcj_730067d3-5e02-4aa1-af49-9bb70546886f/extract-utilities/0.log" Feb 02 10:18:48 crc kubenswrapper[4764]: I0202 10:18:48.329209 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-6kgcj_730067d3-5e02-4aa1-af49-9bb70546886f/extract-content/0.log" Feb 02 10:18:48 crc kubenswrapper[4764]: I0202 10:18:48.560823 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-f4d5w_a03d1d75-d8a0-407c-ab02-d151d5ba00b6/marketplace-operator/0.log" Feb 02 10:18:48 crc kubenswrapper[4764]: I0202 10:18:48.664628 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-vp4vx_7e8b7ce6-dc1a-474e-83e0-659a275c7aaa/extract-utilities/0.log" Feb 02 10:18:49 crc kubenswrapper[4764]: I0202 10:18:49.061220 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-6kgcj_730067d3-5e02-4aa1-af49-9bb70546886f/registry-server/0.log" Feb 02 10:18:49 crc kubenswrapper[4764]: I0202 10:18:49.103258 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-vp4vx_7e8b7ce6-dc1a-474e-83e0-659a275c7aaa/extract-utilities/0.log" Feb 02 10:18:49 crc kubenswrapper[4764]: I0202 10:18:49.150319 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-vp4vx_7e8b7ce6-dc1a-474e-83e0-659a275c7aaa/extract-content/0.log" Feb 02 10:18:49 crc kubenswrapper[4764]: I0202 10:18:49.172517 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-vp4vx_7e8b7ce6-dc1a-474e-83e0-659a275c7aaa/extract-content/0.log" Feb 02 10:18:49 crc kubenswrapper[4764]: I0202 10:18:49.260508 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-vp4vx_7e8b7ce6-dc1a-474e-83e0-659a275c7aaa/extract-utilities/0.log" Feb 02 10:18:49 crc kubenswrapper[4764]: I0202 10:18:49.339328 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-vp4vx_7e8b7ce6-dc1a-474e-83e0-659a275c7aaa/extract-content/0.log" Feb 02 10:18:49 crc kubenswrapper[4764]: I0202 10:18:49.519221 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-vp4vx_7e8b7ce6-dc1a-474e-83e0-659a275c7aaa/registry-server/0.log" Feb 02 10:18:49 crc kubenswrapper[4764]: I0202 10:18:49.570797 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-rc4qs_9443ada1-5f42-40df-b8e2-2c072dd7f264/extract-utilities/0.log" Feb 02 10:18:49 crc kubenswrapper[4764]: I0202 10:18:49.776866 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-rc4qs_9443ada1-5f42-40df-b8e2-2c072dd7f264/extract-content/0.log" Feb 02 10:18:49 crc kubenswrapper[4764]: I0202 10:18:49.784809 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-rc4qs_9443ada1-5f42-40df-b8e2-2c072dd7f264/extract-utilities/0.log" Feb 02 10:18:49 crc kubenswrapper[4764]: I0202 10:18:49.868902 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-rc4qs_9443ada1-5f42-40df-b8e2-2c072dd7f264/extract-content/0.log" Feb 02 10:18:50 crc kubenswrapper[4764]: I0202 10:18:50.117538 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-rc4qs_9443ada1-5f42-40df-b8e2-2c072dd7f264/extract-content/0.log" Feb 02 10:18:50 crc kubenswrapper[4764]: I0202 10:18:50.125305 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-rc4qs_9443ada1-5f42-40df-b8e2-2c072dd7f264/extract-utilities/0.log" Feb 02 10:18:50 crc kubenswrapper[4764]: I0202 10:18:50.649518 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-rc4qs_9443ada1-5f42-40df-b8e2-2c072dd7f264/registry-server/0.log" Feb 02 10:19:13 crc kubenswrapper[4764]: I0202 10:19:13.522610 4764 patch_prober.go:28] interesting pod/machine-config-daemon-4ndm4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 10:19:13 crc kubenswrapper[4764]: I0202 10:19:13.523148 4764 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 10:19:43 crc kubenswrapper[4764]: I0202 10:19:43.522755 4764 patch_prober.go:28] interesting pod/machine-config-daemon-4ndm4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 10:19:43 crc kubenswrapper[4764]: I0202 10:19:43.523311 4764 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 10:19:43 crc kubenswrapper[4764]: I0202 10:19:43.523377 4764 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" Feb 02 10:19:43 crc kubenswrapper[4764]: I0202 10:19:43.524193 4764 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e7f0b09df430c74ab38568e7c16c9b37764975d9232ef539984f9cdd78407ca1"} pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 02 10:19:43 crc kubenswrapper[4764]: I0202 10:19:43.524276 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" containerName="machine-config-daemon" containerID="cri-o://e7f0b09df430c74ab38568e7c16c9b37764975d9232ef539984f9cdd78407ca1" gracePeriod=600 Feb 02 10:19:43 crc kubenswrapper[4764]: E0202 10:19:43.645813 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 10:19:44 crc kubenswrapper[4764]: I0202 10:19:44.251650 4764 generic.go:334] "Generic (PLEG): container finished" podID="d192f670-9f9d-4539-9641-e4bed73acdd4" containerID="e7f0b09df430c74ab38568e7c16c9b37764975d9232ef539984f9cdd78407ca1" exitCode=0 Feb 02 10:19:44 crc kubenswrapper[4764]: I0202 10:19:44.251698 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" event={"ID":"d192f670-9f9d-4539-9641-e4bed73acdd4","Type":"ContainerDied","Data":"e7f0b09df430c74ab38568e7c16c9b37764975d9232ef539984f9cdd78407ca1"} Feb 02 10:19:44 crc kubenswrapper[4764]: I0202 10:19:44.251735 4764 scope.go:117] "RemoveContainer" containerID="ee21cc1574aec6b1e483445c9c7de0962b1f3fa4c95e8c7525277156a9bf1c39" Feb 02 10:19:44 crc kubenswrapper[4764]: I0202 10:19:44.253396 4764 scope.go:117] "RemoveContainer" containerID="e7f0b09df430c74ab38568e7c16c9b37764975d9232ef539984f9cdd78407ca1" Feb 02 10:19:44 crc kubenswrapper[4764]: E0202 10:19:44.254324 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 10:19:57 crc kubenswrapper[4764]: I0202 10:19:57.831530 4764 scope.go:117] "RemoveContainer" containerID="e7f0b09df430c74ab38568e7c16c9b37764975d9232ef539984f9cdd78407ca1" Feb 02 10:19:57 crc kubenswrapper[4764]: E0202 10:19:57.832194 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 10:20:09 crc kubenswrapper[4764]: I0202 10:20:09.825627 4764 scope.go:117] "RemoveContainer" containerID="e7f0b09df430c74ab38568e7c16c9b37764975d9232ef539984f9cdd78407ca1" Feb 02 10:20:09 crc kubenswrapper[4764]: E0202 10:20:09.827009 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 10:20:23 crc kubenswrapper[4764]: I0202 10:20:23.825729 4764 scope.go:117] "RemoveContainer" containerID="e7f0b09df430c74ab38568e7c16c9b37764975d9232ef539984f9cdd78407ca1" Feb 02 10:20:23 crc kubenswrapper[4764]: E0202 10:20:23.827422 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 10:20:29 crc kubenswrapper[4764]: I0202 10:20:29.424408 4764 scope.go:117] "RemoveContainer" containerID="edf39fcc0244cd782d701fc4743f48b2701ea66017d3e26a58694c71fa146a07" Feb 02 10:20:38 crc kubenswrapper[4764]: I0202 10:20:38.825980 4764 scope.go:117] "RemoveContainer" containerID="e7f0b09df430c74ab38568e7c16c9b37764975d9232ef539984f9cdd78407ca1" Feb 02 10:20:38 crc kubenswrapper[4764]: E0202 10:20:38.826636 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 10:20:53 crc kubenswrapper[4764]: I0202 10:20:53.828554 4764 scope.go:117] "RemoveContainer" containerID="e7f0b09df430c74ab38568e7c16c9b37764975d9232ef539984f9cdd78407ca1" Feb 02 10:20:53 crc kubenswrapper[4764]: E0202 10:20:53.829575 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 10:21:07 crc kubenswrapper[4764]: I0202 10:21:07.831292 4764 scope.go:117] "RemoveContainer" containerID="e7f0b09df430c74ab38568e7c16c9b37764975d9232ef539984f9cdd78407ca1" Feb 02 10:21:07 crc kubenswrapper[4764]: E0202 10:21:07.832197 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 10:21:17 crc kubenswrapper[4764]: I0202 10:21:17.158435 4764 generic.go:334] "Generic (PLEG): container finished" podID="bc598811-a908-46d9-aadc-073b575a864d" containerID="f6d1ba99e6bbcb588b915ff9b5f81569c48a8feca4a9c9ace1f50a53385ea4f7" exitCode=0 Feb 02 10:21:17 crc kubenswrapper[4764]: I0202 10:21:17.158829 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-tvk74/must-gather-4qvng" event={"ID":"bc598811-a908-46d9-aadc-073b575a864d","Type":"ContainerDied","Data":"f6d1ba99e6bbcb588b915ff9b5f81569c48a8feca4a9c9ace1f50a53385ea4f7"} Feb 02 10:21:17 crc kubenswrapper[4764]: I0202 10:21:17.165914 4764 scope.go:117] "RemoveContainer" containerID="f6d1ba99e6bbcb588b915ff9b5f81569c48a8feca4a9c9ace1f50a53385ea4f7" Feb 02 10:21:17 crc kubenswrapper[4764]: I0202 10:21:17.269690 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-tvk74_must-gather-4qvng_bc598811-a908-46d9-aadc-073b575a864d/gather/0.log" Feb 02 10:21:22 crc kubenswrapper[4764]: I0202 10:21:22.825994 4764 scope.go:117] "RemoveContainer" containerID="e7f0b09df430c74ab38568e7c16c9b37764975d9232ef539984f9cdd78407ca1" Feb 02 10:21:22 crc kubenswrapper[4764]: E0202 10:21:22.826968 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 10:21:25 crc kubenswrapper[4764]: I0202 10:21:25.814344 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-tvk74/must-gather-4qvng"] Feb 02 10:21:25 crc kubenswrapper[4764]: I0202 10:21:25.817112 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-tvk74/must-gather-4qvng" podUID="bc598811-a908-46d9-aadc-073b575a864d" containerName="copy" containerID="cri-o://b3b304bdbe9119c2d0bf8545eb25701dcaa008bc8ef7b700a2e20300165bb7d6" gracePeriod=2 Feb 02 10:21:25 crc kubenswrapper[4764]: I0202 10:21:25.826829 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-tvk74/must-gather-4qvng"] Feb 02 10:21:26 crc kubenswrapper[4764]: I0202 10:21:26.283335 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-tvk74_must-gather-4qvng_bc598811-a908-46d9-aadc-073b575a864d/copy/0.log" Feb 02 10:21:26 crc kubenswrapper[4764]: I0202 10:21:26.285071 4764 generic.go:334] "Generic (PLEG): container finished" podID="bc598811-a908-46d9-aadc-073b575a864d" containerID="b3b304bdbe9119c2d0bf8545eb25701dcaa008bc8ef7b700a2e20300165bb7d6" exitCode=143 Feb 02 10:21:26 crc kubenswrapper[4764]: I0202 10:21:26.285277 4764 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ea07d975a552bdfd1c6a1d4aaf7b536c9866599a4aef53acf2d8757d2186e73e" Feb 02 10:21:26 crc kubenswrapper[4764]: I0202 10:21:26.326274 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-tvk74_must-gather-4qvng_bc598811-a908-46d9-aadc-073b575a864d/copy/0.log" Feb 02 10:21:26 crc kubenswrapper[4764]: I0202 10:21:26.327113 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-tvk74/must-gather-4qvng" Feb 02 10:21:26 crc kubenswrapper[4764]: I0202 10:21:26.505696 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-njs28\" (UniqueName: \"kubernetes.io/projected/bc598811-a908-46d9-aadc-073b575a864d-kube-api-access-njs28\") pod \"bc598811-a908-46d9-aadc-073b575a864d\" (UID: \"bc598811-a908-46d9-aadc-073b575a864d\") " Feb 02 10:21:26 crc kubenswrapper[4764]: I0202 10:21:26.506143 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/bc598811-a908-46d9-aadc-073b575a864d-must-gather-output\") pod \"bc598811-a908-46d9-aadc-073b575a864d\" (UID: \"bc598811-a908-46d9-aadc-073b575a864d\") " Feb 02 10:21:26 crc kubenswrapper[4764]: I0202 10:21:26.517008 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc598811-a908-46d9-aadc-073b575a864d-kube-api-access-njs28" (OuterVolumeSpecName: "kube-api-access-njs28") pod "bc598811-a908-46d9-aadc-073b575a864d" (UID: "bc598811-a908-46d9-aadc-073b575a864d"). InnerVolumeSpecName "kube-api-access-njs28". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 10:21:26 crc kubenswrapper[4764]: I0202 10:21:26.608949 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-njs28\" (UniqueName: \"kubernetes.io/projected/bc598811-a908-46d9-aadc-073b575a864d-kube-api-access-njs28\") on node \"crc\" DevicePath \"\"" Feb 02 10:21:26 crc kubenswrapper[4764]: I0202 10:21:26.707808 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc598811-a908-46d9-aadc-073b575a864d-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "bc598811-a908-46d9-aadc-073b575a864d" (UID: "bc598811-a908-46d9-aadc-073b575a864d"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 10:21:26 crc kubenswrapper[4764]: I0202 10:21:26.710923 4764 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/bc598811-a908-46d9-aadc-073b575a864d-must-gather-output\") on node \"crc\" DevicePath \"\"" Feb 02 10:21:27 crc kubenswrapper[4764]: I0202 10:21:27.296011 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-tvk74/must-gather-4qvng" Feb 02 10:21:27 crc kubenswrapper[4764]: I0202 10:21:27.835989 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc598811-a908-46d9-aadc-073b575a864d" path="/var/lib/kubelet/pods/bc598811-a908-46d9-aadc-073b575a864d/volumes" Feb 02 10:21:29 crc kubenswrapper[4764]: I0202 10:21:29.499079 4764 scope.go:117] "RemoveContainer" containerID="b3b304bdbe9119c2d0bf8545eb25701dcaa008bc8ef7b700a2e20300165bb7d6" Feb 02 10:21:29 crc kubenswrapper[4764]: I0202 10:21:29.542836 4764 scope.go:117] "RemoveContainer" containerID="10ef5001d0ddba0ef1876237d7bfd3920318548f77f9984bbe903f23353a6771" Feb 02 10:21:29 crc kubenswrapper[4764]: I0202 10:21:29.660734 4764 scope.go:117] "RemoveContainer" containerID="f6d1ba99e6bbcb588b915ff9b5f81569c48a8feca4a9c9ace1f50a53385ea4f7" Feb 02 10:21:34 crc kubenswrapper[4764]: I0202 10:21:34.825899 4764 scope.go:117] "RemoveContainer" containerID="e7f0b09df430c74ab38568e7c16c9b37764975d9232ef539984f9cdd78407ca1" Feb 02 10:21:34 crc kubenswrapper[4764]: E0202 10:21:34.826663 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 10:21:49 crc kubenswrapper[4764]: I0202 10:21:49.826242 4764 scope.go:117] "RemoveContainer" containerID="e7f0b09df430c74ab38568e7c16c9b37764975d9232ef539984f9cdd78407ca1" Feb 02 10:21:49 crc kubenswrapper[4764]: E0202 10:21:49.827352 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 10:22:04 crc kubenswrapper[4764]: I0202 10:22:04.825399 4764 scope.go:117] "RemoveContainer" containerID="e7f0b09df430c74ab38568e7c16c9b37764975d9232ef539984f9cdd78407ca1" Feb 02 10:22:04 crc kubenswrapper[4764]: E0202 10:22:04.826437 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 10:22:17 crc kubenswrapper[4764]: I0202 10:22:17.835511 4764 scope.go:117] "RemoveContainer" containerID="e7f0b09df430c74ab38568e7c16c9b37764975d9232ef539984f9cdd78407ca1" Feb 02 10:22:17 crc kubenswrapper[4764]: E0202 10:22:17.836438 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 10:22:31 crc kubenswrapper[4764]: I0202 10:22:31.826187 4764 scope.go:117] "RemoveContainer" containerID="e7f0b09df430c74ab38568e7c16c9b37764975d9232ef539984f9cdd78407ca1" Feb 02 10:22:31 crc kubenswrapper[4764]: E0202 10:22:31.827470 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 10:22:42 crc kubenswrapper[4764]: I0202 10:22:42.825790 4764 scope.go:117] "RemoveContainer" containerID="e7f0b09df430c74ab38568e7c16c9b37764975d9232ef539984f9cdd78407ca1" Feb 02 10:22:42 crc kubenswrapper[4764]: E0202 10:22:42.826505 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 10:22:56 crc kubenswrapper[4764]: I0202 10:22:56.826379 4764 scope.go:117] "RemoveContainer" containerID="e7f0b09df430c74ab38568e7c16c9b37764975d9232ef539984f9cdd78407ca1" Feb 02 10:22:56 crc kubenswrapper[4764]: E0202 10:22:56.828302 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 10:23:03 crc kubenswrapper[4764]: I0202 10:23:03.848102 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-m2gvg"] Feb 02 10:23:03 crc kubenswrapper[4764]: E0202 10:23:03.849231 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc598811-a908-46d9-aadc-073b575a864d" containerName="copy" Feb 02 10:23:03 crc kubenswrapper[4764]: I0202 10:23:03.849305 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc598811-a908-46d9-aadc-073b575a864d" containerName="copy" Feb 02 10:23:03 crc kubenswrapper[4764]: E0202 10:23:03.849361 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66dc0300-f006-4614-a263-812a27ce11f6" containerName="extract-utilities" Feb 02 10:23:03 crc kubenswrapper[4764]: I0202 10:23:03.849373 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="66dc0300-f006-4614-a263-812a27ce11f6" containerName="extract-utilities" Feb 02 10:23:03 crc kubenswrapper[4764]: E0202 10:23:03.849406 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66dc0300-f006-4614-a263-812a27ce11f6" containerName="registry-server" Feb 02 10:23:03 crc kubenswrapper[4764]: I0202 10:23:03.849415 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="66dc0300-f006-4614-a263-812a27ce11f6" containerName="registry-server" Feb 02 10:23:03 crc kubenswrapper[4764]: E0202 10:23:03.849440 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7582198-460d-45f8-a121-04b97151ac51" containerName="registry-server" Feb 02 10:23:03 crc kubenswrapper[4764]: I0202 10:23:03.849449 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7582198-460d-45f8-a121-04b97151ac51" containerName="registry-server" Feb 02 10:23:03 crc kubenswrapper[4764]: E0202 10:23:03.849466 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc598811-a908-46d9-aadc-073b575a864d" containerName="gather" Feb 02 10:23:03 crc kubenswrapper[4764]: I0202 10:23:03.849474 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc598811-a908-46d9-aadc-073b575a864d" containerName="gather" Feb 02 10:23:03 crc kubenswrapper[4764]: E0202 10:23:03.849489 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7582198-460d-45f8-a121-04b97151ac51" containerName="extract-utilities" Feb 02 10:23:03 crc kubenswrapper[4764]: I0202 10:23:03.849498 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7582198-460d-45f8-a121-04b97151ac51" containerName="extract-utilities" Feb 02 10:23:03 crc kubenswrapper[4764]: E0202 10:23:03.849521 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7582198-460d-45f8-a121-04b97151ac51" containerName="extract-content" Feb 02 10:23:03 crc kubenswrapper[4764]: I0202 10:23:03.849533 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7582198-460d-45f8-a121-04b97151ac51" containerName="extract-content" Feb 02 10:23:03 crc kubenswrapper[4764]: E0202 10:23:03.849592 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66dc0300-f006-4614-a263-812a27ce11f6" containerName="extract-content" Feb 02 10:23:03 crc kubenswrapper[4764]: I0202 10:23:03.849601 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="66dc0300-f006-4614-a263-812a27ce11f6" containerName="extract-content" Feb 02 10:23:03 crc kubenswrapper[4764]: I0202 10:23:03.849915 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc598811-a908-46d9-aadc-073b575a864d" containerName="gather" Feb 02 10:23:03 crc kubenswrapper[4764]: I0202 10:23:03.849952 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="66dc0300-f006-4614-a263-812a27ce11f6" containerName="registry-server" Feb 02 10:23:03 crc kubenswrapper[4764]: I0202 10:23:03.849973 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7582198-460d-45f8-a121-04b97151ac51" containerName="registry-server" Feb 02 10:23:03 crc kubenswrapper[4764]: I0202 10:23:03.849996 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc598811-a908-46d9-aadc-073b575a864d" containerName="copy" Feb 02 10:23:03 crc kubenswrapper[4764]: I0202 10:23:03.851698 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-m2gvg" Feb 02 10:23:03 crc kubenswrapper[4764]: I0202 10:23:03.863188 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-m2gvg"] Feb 02 10:23:03 crc kubenswrapper[4764]: I0202 10:23:03.954824 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ff70798-0df4-408a-9f49-06b462ce655a-catalog-content\") pod \"community-operators-m2gvg\" (UID: \"8ff70798-0df4-408a-9f49-06b462ce655a\") " pod="openshift-marketplace/community-operators-m2gvg" Feb 02 10:23:03 crc kubenswrapper[4764]: I0202 10:23:03.955392 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xfm9v\" (UniqueName: \"kubernetes.io/projected/8ff70798-0df4-408a-9f49-06b462ce655a-kube-api-access-xfm9v\") pod \"community-operators-m2gvg\" (UID: \"8ff70798-0df4-408a-9f49-06b462ce655a\") " pod="openshift-marketplace/community-operators-m2gvg" Feb 02 10:23:03 crc kubenswrapper[4764]: I0202 10:23:03.955431 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ff70798-0df4-408a-9f49-06b462ce655a-utilities\") pod \"community-operators-m2gvg\" (UID: \"8ff70798-0df4-408a-9f49-06b462ce655a\") " pod="openshift-marketplace/community-operators-m2gvg" Feb 02 10:23:04 crc kubenswrapper[4764]: I0202 10:23:04.057166 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xfm9v\" (UniqueName: \"kubernetes.io/projected/8ff70798-0df4-408a-9f49-06b462ce655a-kube-api-access-xfm9v\") pod \"community-operators-m2gvg\" (UID: \"8ff70798-0df4-408a-9f49-06b462ce655a\") " pod="openshift-marketplace/community-operators-m2gvg" Feb 02 10:23:04 crc kubenswrapper[4764]: I0202 10:23:04.057216 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ff70798-0df4-408a-9f49-06b462ce655a-utilities\") pod \"community-operators-m2gvg\" (UID: \"8ff70798-0df4-408a-9f49-06b462ce655a\") " pod="openshift-marketplace/community-operators-m2gvg" Feb 02 10:23:04 crc kubenswrapper[4764]: I0202 10:23:04.057329 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ff70798-0df4-408a-9f49-06b462ce655a-catalog-content\") pod \"community-operators-m2gvg\" (UID: \"8ff70798-0df4-408a-9f49-06b462ce655a\") " pod="openshift-marketplace/community-operators-m2gvg" Feb 02 10:23:04 crc kubenswrapper[4764]: I0202 10:23:04.057795 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ff70798-0df4-408a-9f49-06b462ce655a-utilities\") pod \"community-operators-m2gvg\" (UID: \"8ff70798-0df4-408a-9f49-06b462ce655a\") " pod="openshift-marketplace/community-operators-m2gvg" Feb 02 10:23:04 crc kubenswrapper[4764]: I0202 10:23:04.057856 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ff70798-0df4-408a-9f49-06b462ce655a-catalog-content\") pod \"community-operators-m2gvg\" (UID: \"8ff70798-0df4-408a-9f49-06b462ce655a\") " pod="openshift-marketplace/community-operators-m2gvg" Feb 02 10:23:04 crc kubenswrapper[4764]: I0202 10:23:04.252279 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xfm9v\" (UniqueName: \"kubernetes.io/projected/8ff70798-0df4-408a-9f49-06b462ce655a-kube-api-access-xfm9v\") pod \"community-operators-m2gvg\" (UID: \"8ff70798-0df4-408a-9f49-06b462ce655a\") " pod="openshift-marketplace/community-operators-m2gvg" Feb 02 10:23:04 crc kubenswrapper[4764]: I0202 10:23:04.484252 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-m2gvg" Feb 02 10:23:04 crc kubenswrapper[4764]: I0202 10:23:04.988975 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-m2gvg"] Feb 02 10:23:05 crc kubenswrapper[4764]: I0202 10:23:05.213768 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m2gvg" event={"ID":"8ff70798-0df4-408a-9f49-06b462ce655a","Type":"ContainerStarted","Data":"985c6262892792a9b4dbb3c042f4aa858897d2cda6cd1ca17d6b15d1c820482f"} Feb 02 10:23:06 crc kubenswrapper[4764]: I0202 10:23:06.226162 4764 generic.go:334] "Generic (PLEG): container finished" podID="8ff70798-0df4-408a-9f49-06b462ce655a" containerID="cd42bc1b1b1ee060ea130e1561aa76ff0183bbc3e2f14d5491111facf1cca89c" exitCode=0 Feb 02 10:23:06 crc kubenswrapper[4764]: I0202 10:23:06.226228 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m2gvg" event={"ID":"8ff70798-0df4-408a-9f49-06b462ce655a","Type":"ContainerDied","Data":"cd42bc1b1b1ee060ea130e1561aa76ff0183bbc3e2f14d5491111facf1cca89c"} Feb 02 10:23:06 crc kubenswrapper[4764]: I0202 10:23:06.229459 4764 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 02 10:23:08 crc kubenswrapper[4764]: I0202 10:23:08.249081 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m2gvg" event={"ID":"8ff70798-0df4-408a-9f49-06b462ce655a","Type":"ContainerStarted","Data":"f8721262663e42e4f78033a6661c86f3be275cd5ea5568d3cc1b4283b883c3d4"} Feb 02 10:23:09 crc kubenswrapper[4764]: I0202 10:23:09.260713 4764 generic.go:334] "Generic (PLEG): container finished" podID="8ff70798-0df4-408a-9f49-06b462ce655a" containerID="f8721262663e42e4f78033a6661c86f3be275cd5ea5568d3cc1b4283b883c3d4" exitCode=0 Feb 02 10:23:09 crc kubenswrapper[4764]: I0202 10:23:09.260753 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m2gvg" event={"ID":"8ff70798-0df4-408a-9f49-06b462ce655a","Type":"ContainerDied","Data":"f8721262663e42e4f78033a6661c86f3be275cd5ea5568d3cc1b4283b883c3d4"} Feb 02 10:23:10 crc kubenswrapper[4764]: I0202 10:23:10.278462 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m2gvg" event={"ID":"8ff70798-0df4-408a-9f49-06b462ce655a","Type":"ContainerStarted","Data":"49b99b134833bcab000e28b3d0022201c3fac219bfcb760d5599b2f73218dd6e"} Feb 02 10:23:10 crc kubenswrapper[4764]: I0202 10:23:10.307764 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-m2gvg" podStartSLOduration=3.8717604530000003 podStartE2EDuration="7.307745356s" podCreationTimestamp="2026-02-02 10:23:03 +0000 UTC" firstStartedPulling="2026-02-02 10:23:06.229183787 +0000 UTC m=+4549.162907875" lastFinishedPulling="2026-02-02 10:23:09.66516869 +0000 UTC m=+4552.598892778" observedRunningTime="2026-02-02 10:23:10.297169985 +0000 UTC m=+4553.230894073" watchObservedRunningTime="2026-02-02 10:23:10.307745356 +0000 UTC m=+4553.241469444" Feb 02 10:23:11 crc kubenswrapper[4764]: I0202 10:23:11.825745 4764 scope.go:117] "RemoveContainer" containerID="e7f0b09df430c74ab38568e7c16c9b37764975d9232ef539984f9cdd78407ca1" Feb 02 10:23:11 crc kubenswrapper[4764]: E0202 10:23:11.827282 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 10:23:14 crc kubenswrapper[4764]: I0202 10:23:14.485184 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-m2gvg" Feb 02 10:23:14 crc kubenswrapper[4764]: I0202 10:23:14.486178 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-m2gvg" Feb 02 10:23:14 crc kubenswrapper[4764]: I0202 10:23:14.547925 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-m2gvg" Feb 02 10:23:15 crc kubenswrapper[4764]: I0202 10:23:15.394509 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-m2gvg" Feb 02 10:23:15 crc kubenswrapper[4764]: I0202 10:23:15.441956 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-m2gvg"] Feb 02 10:23:17 crc kubenswrapper[4764]: I0202 10:23:17.352442 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-m2gvg" podUID="8ff70798-0df4-408a-9f49-06b462ce655a" containerName="registry-server" containerID="cri-o://49b99b134833bcab000e28b3d0022201c3fac219bfcb760d5599b2f73218dd6e" gracePeriod=2 Feb 02 10:23:17 crc kubenswrapper[4764]: I0202 10:23:17.846856 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-m2gvg" Feb 02 10:23:17 crc kubenswrapper[4764]: I0202 10:23:17.953098 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ff70798-0df4-408a-9f49-06b462ce655a-utilities\") pod \"8ff70798-0df4-408a-9f49-06b462ce655a\" (UID: \"8ff70798-0df4-408a-9f49-06b462ce655a\") " Feb 02 10:23:17 crc kubenswrapper[4764]: I0202 10:23:17.953375 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xfm9v\" (UniqueName: \"kubernetes.io/projected/8ff70798-0df4-408a-9f49-06b462ce655a-kube-api-access-xfm9v\") pod \"8ff70798-0df4-408a-9f49-06b462ce655a\" (UID: \"8ff70798-0df4-408a-9f49-06b462ce655a\") " Feb 02 10:23:17 crc kubenswrapper[4764]: I0202 10:23:17.953544 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ff70798-0df4-408a-9f49-06b462ce655a-catalog-content\") pod \"8ff70798-0df4-408a-9f49-06b462ce655a\" (UID: \"8ff70798-0df4-408a-9f49-06b462ce655a\") " Feb 02 10:23:17 crc kubenswrapper[4764]: I0202 10:23:17.955337 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8ff70798-0df4-408a-9f49-06b462ce655a-utilities" (OuterVolumeSpecName: "utilities") pod "8ff70798-0df4-408a-9f49-06b462ce655a" (UID: "8ff70798-0df4-408a-9f49-06b462ce655a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 10:23:17 crc kubenswrapper[4764]: I0202 10:23:17.959818 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8ff70798-0df4-408a-9f49-06b462ce655a-kube-api-access-xfm9v" (OuterVolumeSpecName: "kube-api-access-xfm9v") pod "8ff70798-0df4-408a-9f49-06b462ce655a" (UID: "8ff70798-0df4-408a-9f49-06b462ce655a"). InnerVolumeSpecName "kube-api-access-xfm9v". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 10:23:18 crc kubenswrapper[4764]: I0202 10:23:18.032755 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8ff70798-0df4-408a-9f49-06b462ce655a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8ff70798-0df4-408a-9f49-06b462ce655a" (UID: "8ff70798-0df4-408a-9f49-06b462ce655a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 10:23:18 crc kubenswrapper[4764]: I0202 10:23:18.055727 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xfm9v\" (UniqueName: \"kubernetes.io/projected/8ff70798-0df4-408a-9f49-06b462ce655a-kube-api-access-xfm9v\") on node \"crc\" DevicePath \"\"" Feb 02 10:23:18 crc kubenswrapper[4764]: I0202 10:23:18.055762 4764 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ff70798-0df4-408a-9f49-06b462ce655a-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 10:23:18 crc kubenswrapper[4764]: I0202 10:23:18.055772 4764 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ff70798-0df4-408a-9f49-06b462ce655a-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 10:23:18 crc kubenswrapper[4764]: I0202 10:23:18.367908 4764 generic.go:334] "Generic (PLEG): container finished" podID="8ff70798-0df4-408a-9f49-06b462ce655a" containerID="49b99b134833bcab000e28b3d0022201c3fac219bfcb760d5599b2f73218dd6e" exitCode=0 Feb 02 10:23:18 crc kubenswrapper[4764]: I0202 10:23:18.368074 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-m2gvg" Feb 02 10:23:18 crc kubenswrapper[4764]: I0202 10:23:18.368123 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m2gvg" event={"ID":"8ff70798-0df4-408a-9f49-06b462ce655a","Type":"ContainerDied","Data":"49b99b134833bcab000e28b3d0022201c3fac219bfcb760d5599b2f73218dd6e"} Feb 02 10:23:18 crc kubenswrapper[4764]: I0202 10:23:18.369277 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m2gvg" event={"ID":"8ff70798-0df4-408a-9f49-06b462ce655a","Type":"ContainerDied","Data":"985c6262892792a9b4dbb3c042f4aa858897d2cda6cd1ca17d6b15d1c820482f"} Feb 02 10:23:18 crc kubenswrapper[4764]: I0202 10:23:18.369310 4764 scope.go:117] "RemoveContainer" containerID="49b99b134833bcab000e28b3d0022201c3fac219bfcb760d5599b2f73218dd6e" Feb 02 10:23:18 crc kubenswrapper[4764]: I0202 10:23:18.393150 4764 scope.go:117] "RemoveContainer" containerID="f8721262663e42e4f78033a6661c86f3be275cd5ea5568d3cc1b4283b883c3d4" Feb 02 10:23:18 crc kubenswrapper[4764]: I0202 10:23:18.426902 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-m2gvg"] Feb 02 10:23:18 crc kubenswrapper[4764]: I0202 10:23:18.437670 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-m2gvg"] Feb 02 10:23:18 crc kubenswrapper[4764]: I0202 10:23:18.439239 4764 scope.go:117] "RemoveContainer" containerID="cd42bc1b1b1ee060ea130e1561aa76ff0183bbc3e2f14d5491111facf1cca89c" Feb 02 10:23:18 crc kubenswrapper[4764]: I0202 10:23:18.474161 4764 scope.go:117] "RemoveContainer" containerID="49b99b134833bcab000e28b3d0022201c3fac219bfcb760d5599b2f73218dd6e" Feb 02 10:23:18 crc kubenswrapper[4764]: E0202 10:23:18.474700 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"49b99b134833bcab000e28b3d0022201c3fac219bfcb760d5599b2f73218dd6e\": container with ID starting with 49b99b134833bcab000e28b3d0022201c3fac219bfcb760d5599b2f73218dd6e not found: ID does not exist" containerID="49b99b134833bcab000e28b3d0022201c3fac219bfcb760d5599b2f73218dd6e" Feb 02 10:23:18 crc kubenswrapper[4764]: I0202 10:23:18.474739 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"49b99b134833bcab000e28b3d0022201c3fac219bfcb760d5599b2f73218dd6e"} err="failed to get container status \"49b99b134833bcab000e28b3d0022201c3fac219bfcb760d5599b2f73218dd6e\": rpc error: code = NotFound desc = could not find container \"49b99b134833bcab000e28b3d0022201c3fac219bfcb760d5599b2f73218dd6e\": container with ID starting with 49b99b134833bcab000e28b3d0022201c3fac219bfcb760d5599b2f73218dd6e not found: ID does not exist" Feb 02 10:23:18 crc kubenswrapper[4764]: I0202 10:23:18.474767 4764 scope.go:117] "RemoveContainer" containerID="f8721262663e42e4f78033a6661c86f3be275cd5ea5568d3cc1b4283b883c3d4" Feb 02 10:23:18 crc kubenswrapper[4764]: E0202 10:23:18.475036 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f8721262663e42e4f78033a6661c86f3be275cd5ea5568d3cc1b4283b883c3d4\": container with ID starting with f8721262663e42e4f78033a6661c86f3be275cd5ea5568d3cc1b4283b883c3d4 not found: ID does not exist" containerID="f8721262663e42e4f78033a6661c86f3be275cd5ea5568d3cc1b4283b883c3d4" Feb 02 10:23:18 crc kubenswrapper[4764]: I0202 10:23:18.475079 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f8721262663e42e4f78033a6661c86f3be275cd5ea5568d3cc1b4283b883c3d4"} err="failed to get container status \"f8721262663e42e4f78033a6661c86f3be275cd5ea5568d3cc1b4283b883c3d4\": rpc error: code = NotFound desc = could not find container \"f8721262663e42e4f78033a6661c86f3be275cd5ea5568d3cc1b4283b883c3d4\": container with ID starting with f8721262663e42e4f78033a6661c86f3be275cd5ea5568d3cc1b4283b883c3d4 not found: ID does not exist" Feb 02 10:23:18 crc kubenswrapper[4764]: I0202 10:23:18.475097 4764 scope.go:117] "RemoveContainer" containerID="cd42bc1b1b1ee060ea130e1561aa76ff0183bbc3e2f14d5491111facf1cca89c" Feb 02 10:23:18 crc kubenswrapper[4764]: E0202 10:23:18.475311 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cd42bc1b1b1ee060ea130e1561aa76ff0183bbc3e2f14d5491111facf1cca89c\": container with ID starting with cd42bc1b1b1ee060ea130e1561aa76ff0183bbc3e2f14d5491111facf1cca89c not found: ID does not exist" containerID="cd42bc1b1b1ee060ea130e1561aa76ff0183bbc3e2f14d5491111facf1cca89c" Feb 02 10:23:18 crc kubenswrapper[4764]: I0202 10:23:18.475333 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cd42bc1b1b1ee060ea130e1561aa76ff0183bbc3e2f14d5491111facf1cca89c"} err="failed to get container status \"cd42bc1b1b1ee060ea130e1561aa76ff0183bbc3e2f14d5491111facf1cca89c\": rpc error: code = NotFound desc = could not find container \"cd42bc1b1b1ee060ea130e1561aa76ff0183bbc3e2f14d5491111facf1cca89c\": container with ID starting with cd42bc1b1b1ee060ea130e1561aa76ff0183bbc3e2f14d5491111facf1cca89c not found: ID does not exist" Feb 02 10:23:19 crc kubenswrapper[4764]: I0202 10:23:19.839545 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8ff70798-0df4-408a-9f49-06b462ce655a" path="/var/lib/kubelet/pods/8ff70798-0df4-408a-9f49-06b462ce655a/volumes" Feb 02 10:23:23 crc kubenswrapper[4764]: I0202 10:23:23.825923 4764 scope.go:117] "RemoveContainer" containerID="e7f0b09df430c74ab38568e7c16c9b37764975d9232ef539984f9cdd78407ca1" Feb 02 10:23:23 crc kubenswrapper[4764]: E0202 10:23:23.826595 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 10:23:34 crc kubenswrapper[4764]: I0202 10:23:34.826325 4764 scope.go:117] "RemoveContainer" containerID="e7f0b09df430c74ab38568e7c16c9b37764975d9232ef539984f9cdd78407ca1" Feb 02 10:23:34 crc kubenswrapper[4764]: E0202 10:23:34.827097 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 10:23:48 crc kubenswrapper[4764]: I0202 10:23:48.825750 4764 scope.go:117] "RemoveContainer" containerID="e7f0b09df430c74ab38568e7c16c9b37764975d9232ef539984f9cdd78407ca1" Feb 02 10:23:48 crc kubenswrapper[4764]: E0202 10:23:48.826562 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 10:24:01 crc kubenswrapper[4764]: I0202 10:24:01.826881 4764 scope.go:117] "RemoveContainer" containerID="e7f0b09df430c74ab38568e7c16c9b37764975d9232ef539984f9cdd78407ca1" Feb 02 10:24:01 crc kubenswrapper[4764]: E0202 10:24:01.828056 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 10:24:16 crc kubenswrapper[4764]: I0202 10:24:16.825429 4764 scope.go:117] "RemoveContainer" containerID="e7f0b09df430c74ab38568e7c16c9b37764975d9232ef539984f9cdd78407ca1" Feb 02 10:24:16 crc kubenswrapper[4764]: E0202 10:24:16.826258 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 10:24:21 crc kubenswrapper[4764]: I0202 10:24:21.074832 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-lgpbv/must-gather-b5bt7"] Feb 02 10:24:21 crc kubenswrapper[4764]: E0202 10:24:21.076616 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ff70798-0df4-408a-9f49-06b462ce655a" containerName="extract-content" Feb 02 10:24:21 crc kubenswrapper[4764]: I0202 10:24:21.076701 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ff70798-0df4-408a-9f49-06b462ce655a" containerName="extract-content" Feb 02 10:24:21 crc kubenswrapper[4764]: E0202 10:24:21.076964 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ff70798-0df4-408a-9f49-06b462ce655a" containerName="registry-server" Feb 02 10:24:21 crc kubenswrapper[4764]: I0202 10:24:21.077045 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ff70798-0df4-408a-9f49-06b462ce655a" containerName="registry-server" Feb 02 10:24:21 crc kubenswrapper[4764]: E0202 10:24:21.077125 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ff70798-0df4-408a-9f49-06b462ce655a" containerName="extract-utilities" Feb 02 10:24:21 crc kubenswrapper[4764]: I0202 10:24:21.077186 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ff70798-0df4-408a-9f49-06b462ce655a" containerName="extract-utilities" Feb 02 10:24:21 crc kubenswrapper[4764]: I0202 10:24:21.077421 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ff70798-0df4-408a-9f49-06b462ce655a" containerName="registry-server" Feb 02 10:24:21 crc kubenswrapper[4764]: I0202 10:24:21.078828 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-lgpbv/must-gather-b5bt7" Feb 02 10:24:21 crc kubenswrapper[4764]: I0202 10:24:21.082143 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-lgpbv"/"openshift-service-ca.crt" Feb 02 10:24:21 crc kubenswrapper[4764]: I0202 10:24:21.082380 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-lgpbv"/"kube-root-ca.crt" Feb 02 10:24:21 crc kubenswrapper[4764]: I0202 10:24:21.082508 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-lgpbv"/"default-dockercfg-6qtxf" Feb 02 10:24:21 crc kubenswrapper[4764]: I0202 10:24:21.092798 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-lgpbv/must-gather-b5bt7"] Feb 02 10:24:21 crc kubenswrapper[4764]: I0202 10:24:21.269680 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dq55k\" (UniqueName: \"kubernetes.io/projected/c9238785-9984-45c8-8eb8-f15bd97b7990-kube-api-access-dq55k\") pod \"must-gather-b5bt7\" (UID: \"c9238785-9984-45c8-8eb8-f15bd97b7990\") " pod="openshift-must-gather-lgpbv/must-gather-b5bt7" Feb 02 10:24:21 crc kubenswrapper[4764]: I0202 10:24:21.270130 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/c9238785-9984-45c8-8eb8-f15bd97b7990-must-gather-output\") pod \"must-gather-b5bt7\" (UID: \"c9238785-9984-45c8-8eb8-f15bd97b7990\") " pod="openshift-must-gather-lgpbv/must-gather-b5bt7" Feb 02 10:24:21 crc kubenswrapper[4764]: I0202 10:24:21.372280 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dq55k\" (UniqueName: \"kubernetes.io/projected/c9238785-9984-45c8-8eb8-f15bd97b7990-kube-api-access-dq55k\") pod \"must-gather-b5bt7\" (UID: \"c9238785-9984-45c8-8eb8-f15bd97b7990\") " pod="openshift-must-gather-lgpbv/must-gather-b5bt7" Feb 02 10:24:21 crc kubenswrapper[4764]: I0202 10:24:21.372418 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/c9238785-9984-45c8-8eb8-f15bd97b7990-must-gather-output\") pod \"must-gather-b5bt7\" (UID: \"c9238785-9984-45c8-8eb8-f15bd97b7990\") " pod="openshift-must-gather-lgpbv/must-gather-b5bt7" Feb 02 10:24:21 crc kubenswrapper[4764]: I0202 10:24:21.372853 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/c9238785-9984-45c8-8eb8-f15bd97b7990-must-gather-output\") pod \"must-gather-b5bt7\" (UID: \"c9238785-9984-45c8-8eb8-f15bd97b7990\") " pod="openshift-must-gather-lgpbv/must-gather-b5bt7" Feb 02 10:24:21 crc kubenswrapper[4764]: I0202 10:24:21.552921 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dq55k\" (UniqueName: \"kubernetes.io/projected/c9238785-9984-45c8-8eb8-f15bd97b7990-kube-api-access-dq55k\") pod \"must-gather-b5bt7\" (UID: \"c9238785-9984-45c8-8eb8-f15bd97b7990\") " pod="openshift-must-gather-lgpbv/must-gather-b5bt7" Feb 02 10:24:21 crc kubenswrapper[4764]: I0202 10:24:21.701028 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-lgpbv/must-gather-b5bt7" Feb 02 10:24:22 crc kubenswrapper[4764]: I0202 10:24:22.315674 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-lgpbv/must-gather-b5bt7"] Feb 02 10:24:22 crc kubenswrapper[4764]: I0202 10:24:22.935482 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-lgpbv/must-gather-b5bt7" event={"ID":"c9238785-9984-45c8-8eb8-f15bd97b7990","Type":"ContainerStarted","Data":"8b13540f8f2ed99b626fa43c5c43c456426eeabf0300ef98f9e7fe2aa7ddcdcc"} Feb 02 10:24:22 crc kubenswrapper[4764]: I0202 10:24:22.935808 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-lgpbv/must-gather-b5bt7" event={"ID":"c9238785-9984-45c8-8eb8-f15bd97b7990","Type":"ContainerStarted","Data":"d57970c433089112a43a97b7b0a86aa1fbb1612253f41f2a8f6b692ed649b7a3"} Feb 02 10:24:22 crc kubenswrapper[4764]: I0202 10:24:22.935822 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-lgpbv/must-gather-b5bt7" event={"ID":"c9238785-9984-45c8-8eb8-f15bd97b7990","Type":"ContainerStarted","Data":"3d14704e02cf0dc9b67c9a63d85e94b0d73d5d06d815bb03c10bacbf07f49c39"} Feb 02 10:24:22 crc kubenswrapper[4764]: I0202 10:24:22.958733 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-lgpbv/must-gather-b5bt7" podStartSLOduration=1.9587146560000002 podStartE2EDuration="1.958714656s" podCreationTimestamp="2026-02-02 10:24:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 10:24:22.951317919 +0000 UTC m=+4625.885042007" watchObservedRunningTime="2026-02-02 10:24:22.958714656 +0000 UTC m=+4625.892438734" Feb 02 10:24:27 crc kubenswrapper[4764]: I0202 10:24:27.473640 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-lgpbv/crc-debug-2mfzn"] Feb 02 10:24:27 crc kubenswrapper[4764]: I0202 10:24:27.475411 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-lgpbv/crc-debug-2mfzn" Feb 02 10:24:27 crc kubenswrapper[4764]: I0202 10:24:27.645361 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vtvzs\" (UniqueName: \"kubernetes.io/projected/e98aa151-d23e-4b9a-a169-149396d665e1-kube-api-access-vtvzs\") pod \"crc-debug-2mfzn\" (UID: \"e98aa151-d23e-4b9a-a169-149396d665e1\") " pod="openshift-must-gather-lgpbv/crc-debug-2mfzn" Feb 02 10:24:27 crc kubenswrapper[4764]: I0202 10:24:27.645433 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e98aa151-d23e-4b9a-a169-149396d665e1-host\") pod \"crc-debug-2mfzn\" (UID: \"e98aa151-d23e-4b9a-a169-149396d665e1\") " pod="openshift-must-gather-lgpbv/crc-debug-2mfzn" Feb 02 10:24:27 crc kubenswrapper[4764]: I0202 10:24:27.747378 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vtvzs\" (UniqueName: \"kubernetes.io/projected/e98aa151-d23e-4b9a-a169-149396d665e1-kube-api-access-vtvzs\") pod \"crc-debug-2mfzn\" (UID: \"e98aa151-d23e-4b9a-a169-149396d665e1\") " pod="openshift-must-gather-lgpbv/crc-debug-2mfzn" Feb 02 10:24:27 crc kubenswrapper[4764]: I0202 10:24:27.747707 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e98aa151-d23e-4b9a-a169-149396d665e1-host\") pod \"crc-debug-2mfzn\" (UID: \"e98aa151-d23e-4b9a-a169-149396d665e1\") " pod="openshift-must-gather-lgpbv/crc-debug-2mfzn" Feb 02 10:24:27 crc kubenswrapper[4764]: I0202 10:24:27.747809 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e98aa151-d23e-4b9a-a169-149396d665e1-host\") pod \"crc-debug-2mfzn\" (UID: \"e98aa151-d23e-4b9a-a169-149396d665e1\") " pod="openshift-must-gather-lgpbv/crc-debug-2mfzn" Feb 02 10:24:27 crc kubenswrapper[4764]: I0202 10:24:27.768090 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vtvzs\" (UniqueName: \"kubernetes.io/projected/e98aa151-d23e-4b9a-a169-149396d665e1-kube-api-access-vtvzs\") pod \"crc-debug-2mfzn\" (UID: \"e98aa151-d23e-4b9a-a169-149396d665e1\") " pod="openshift-must-gather-lgpbv/crc-debug-2mfzn" Feb 02 10:24:27 crc kubenswrapper[4764]: I0202 10:24:27.790515 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-lgpbv/crc-debug-2mfzn" Feb 02 10:24:27 crc kubenswrapper[4764]: I0202 10:24:27.981274 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-lgpbv/crc-debug-2mfzn" event={"ID":"e98aa151-d23e-4b9a-a169-149396d665e1","Type":"ContainerStarted","Data":"5e15962ae50f25b5ac14d26f89dc11079beecfd2ecbcd8aa537fd6eab4bb17cb"} Feb 02 10:24:28 crc kubenswrapper[4764]: I0202 10:24:28.989459 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-lgpbv/crc-debug-2mfzn" event={"ID":"e98aa151-d23e-4b9a-a169-149396d665e1","Type":"ContainerStarted","Data":"87d803d0f82aaabab517fc6bd41640a0d5b4f5438bb647422e6407ca98493ce8"} Feb 02 10:24:29 crc kubenswrapper[4764]: I0202 10:24:29.014659 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-lgpbv/crc-debug-2mfzn" podStartSLOduration=2.014635821 podStartE2EDuration="2.014635821s" podCreationTimestamp="2026-02-02 10:24:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 10:24:29.00821689 +0000 UTC m=+4631.941940988" watchObservedRunningTime="2026-02-02 10:24:29.014635821 +0000 UTC m=+4631.948359909" Feb 02 10:24:31 crc kubenswrapper[4764]: I0202 10:24:31.829336 4764 scope.go:117] "RemoveContainer" containerID="e7f0b09df430c74ab38568e7c16c9b37764975d9232ef539984f9cdd78407ca1" Feb 02 10:24:31 crc kubenswrapper[4764]: E0202 10:24:31.830139 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 10:24:45 crc kubenswrapper[4764]: I0202 10:24:45.828658 4764 scope.go:117] "RemoveContainer" containerID="e7f0b09df430c74ab38568e7c16c9b37764975d9232ef539984f9cdd78407ca1" Feb 02 10:24:47 crc kubenswrapper[4764]: I0202 10:24:47.158886 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" event={"ID":"d192f670-9f9d-4539-9641-e4bed73acdd4","Type":"ContainerStarted","Data":"c3ca98e86124ebc7a6731929c4f79572e15b8712d419753d62acb46e29b6778c"} Feb 02 10:25:00 crc kubenswrapper[4764]: I0202 10:25:00.429659 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-nb774"] Feb 02 10:25:00 crc kubenswrapper[4764]: I0202 10:25:00.443639 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nb774" Feb 02 10:25:00 crc kubenswrapper[4764]: I0202 10:25:00.443912 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-nb774"] Feb 02 10:25:00 crc kubenswrapper[4764]: I0202 10:25:00.522284 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/858615be-9087-4141-b5be-d6751cc64025-utilities\") pod \"redhat-operators-nb774\" (UID: \"858615be-9087-4141-b5be-d6751cc64025\") " pod="openshift-marketplace/redhat-operators-nb774" Feb 02 10:25:00 crc kubenswrapper[4764]: I0202 10:25:00.522635 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/858615be-9087-4141-b5be-d6751cc64025-catalog-content\") pod \"redhat-operators-nb774\" (UID: \"858615be-9087-4141-b5be-d6751cc64025\") " pod="openshift-marketplace/redhat-operators-nb774" Feb 02 10:25:00 crc kubenswrapper[4764]: I0202 10:25:00.522861 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bdpjj\" (UniqueName: \"kubernetes.io/projected/858615be-9087-4141-b5be-d6751cc64025-kube-api-access-bdpjj\") pod \"redhat-operators-nb774\" (UID: \"858615be-9087-4141-b5be-d6751cc64025\") " pod="openshift-marketplace/redhat-operators-nb774" Feb 02 10:25:00 crc kubenswrapper[4764]: I0202 10:25:00.624979 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/858615be-9087-4141-b5be-d6751cc64025-catalog-content\") pod \"redhat-operators-nb774\" (UID: \"858615be-9087-4141-b5be-d6751cc64025\") " pod="openshift-marketplace/redhat-operators-nb774" Feb 02 10:25:00 crc kubenswrapper[4764]: I0202 10:25:00.625264 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bdpjj\" (UniqueName: \"kubernetes.io/projected/858615be-9087-4141-b5be-d6751cc64025-kube-api-access-bdpjj\") pod \"redhat-operators-nb774\" (UID: \"858615be-9087-4141-b5be-d6751cc64025\") " pod="openshift-marketplace/redhat-operators-nb774" Feb 02 10:25:00 crc kubenswrapper[4764]: I0202 10:25:00.625419 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/858615be-9087-4141-b5be-d6751cc64025-catalog-content\") pod \"redhat-operators-nb774\" (UID: \"858615be-9087-4141-b5be-d6751cc64025\") " pod="openshift-marketplace/redhat-operators-nb774" Feb 02 10:25:00 crc kubenswrapper[4764]: I0202 10:25:00.625549 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/858615be-9087-4141-b5be-d6751cc64025-utilities\") pod \"redhat-operators-nb774\" (UID: \"858615be-9087-4141-b5be-d6751cc64025\") " pod="openshift-marketplace/redhat-operators-nb774" Feb 02 10:25:00 crc kubenswrapper[4764]: I0202 10:25:00.625763 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/858615be-9087-4141-b5be-d6751cc64025-utilities\") pod \"redhat-operators-nb774\" (UID: \"858615be-9087-4141-b5be-d6751cc64025\") " pod="openshift-marketplace/redhat-operators-nb774" Feb 02 10:25:00 crc kubenswrapper[4764]: I0202 10:25:00.649841 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bdpjj\" (UniqueName: \"kubernetes.io/projected/858615be-9087-4141-b5be-d6751cc64025-kube-api-access-bdpjj\") pod \"redhat-operators-nb774\" (UID: \"858615be-9087-4141-b5be-d6751cc64025\") " pod="openshift-marketplace/redhat-operators-nb774" Feb 02 10:25:00 crc kubenswrapper[4764]: I0202 10:25:00.768439 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nb774" Feb 02 10:25:01 crc kubenswrapper[4764]: I0202 10:25:01.365858 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-nb774"] Feb 02 10:25:02 crc kubenswrapper[4764]: I0202 10:25:02.288368 4764 generic.go:334] "Generic (PLEG): container finished" podID="858615be-9087-4141-b5be-d6751cc64025" containerID="4a39d8d588985a2e6bd8082130bc137a085433be68d8bf80cdb67f506f45be08" exitCode=0 Feb 02 10:25:02 crc kubenswrapper[4764]: I0202 10:25:02.288413 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nb774" event={"ID":"858615be-9087-4141-b5be-d6751cc64025","Type":"ContainerDied","Data":"4a39d8d588985a2e6bd8082130bc137a085433be68d8bf80cdb67f506f45be08"} Feb 02 10:25:02 crc kubenswrapper[4764]: I0202 10:25:02.288441 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nb774" event={"ID":"858615be-9087-4141-b5be-d6751cc64025","Type":"ContainerStarted","Data":"8e877b6c179a416f8e23558ab337453e118ee891dd335fd8ffca777d03bf02cb"} Feb 02 10:25:04 crc kubenswrapper[4764]: I0202 10:25:04.309522 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nb774" event={"ID":"858615be-9087-4141-b5be-d6751cc64025","Type":"ContainerStarted","Data":"6a0992d75a87ff27ab8cf98d46b8d463938bd60d5e33dc18bfacc4ec34ee4c0e"} Feb 02 10:25:06 crc kubenswrapper[4764]: I0202 10:25:06.333365 4764 generic.go:334] "Generic (PLEG): container finished" podID="e98aa151-d23e-4b9a-a169-149396d665e1" containerID="87d803d0f82aaabab517fc6bd41640a0d5b4f5438bb647422e6407ca98493ce8" exitCode=0 Feb 02 10:25:06 crc kubenswrapper[4764]: I0202 10:25:06.333415 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-lgpbv/crc-debug-2mfzn" event={"ID":"e98aa151-d23e-4b9a-a169-149396d665e1","Type":"ContainerDied","Data":"87d803d0f82aaabab517fc6bd41640a0d5b4f5438bb647422e6407ca98493ce8"} Feb 02 10:25:07 crc kubenswrapper[4764]: I0202 10:25:07.440865 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-lgpbv/crc-debug-2mfzn" Feb 02 10:25:07 crc kubenswrapper[4764]: I0202 10:25:07.475301 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-lgpbv/crc-debug-2mfzn"] Feb 02 10:25:07 crc kubenswrapper[4764]: I0202 10:25:07.483803 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-lgpbv/crc-debug-2mfzn"] Feb 02 10:25:07 crc kubenswrapper[4764]: I0202 10:25:07.556754 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vtvzs\" (UniqueName: \"kubernetes.io/projected/e98aa151-d23e-4b9a-a169-149396d665e1-kube-api-access-vtvzs\") pod \"e98aa151-d23e-4b9a-a169-149396d665e1\" (UID: \"e98aa151-d23e-4b9a-a169-149396d665e1\") " Feb 02 10:25:07 crc kubenswrapper[4764]: I0202 10:25:07.557308 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e98aa151-d23e-4b9a-a169-149396d665e1-host\") pod \"e98aa151-d23e-4b9a-a169-149396d665e1\" (UID: \"e98aa151-d23e-4b9a-a169-149396d665e1\") " Feb 02 10:25:07 crc kubenswrapper[4764]: I0202 10:25:07.557439 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e98aa151-d23e-4b9a-a169-149396d665e1-host" (OuterVolumeSpecName: "host") pod "e98aa151-d23e-4b9a-a169-149396d665e1" (UID: "e98aa151-d23e-4b9a-a169-149396d665e1"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 10:25:07 crc kubenswrapper[4764]: I0202 10:25:07.558232 4764 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e98aa151-d23e-4b9a-a169-149396d665e1-host\") on node \"crc\" DevicePath \"\"" Feb 02 10:25:07 crc kubenswrapper[4764]: I0202 10:25:07.569670 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e98aa151-d23e-4b9a-a169-149396d665e1-kube-api-access-vtvzs" (OuterVolumeSpecName: "kube-api-access-vtvzs") pod "e98aa151-d23e-4b9a-a169-149396d665e1" (UID: "e98aa151-d23e-4b9a-a169-149396d665e1"). InnerVolumeSpecName "kube-api-access-vtvzs". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 10:25:07 crc kubenswrapper[4764]: I0202 10:25:07.659658 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vtvzs\" (UniqueName: \"kubernetes.io/projected/e98aa151-d23e-4b9a-a169-149396d665e1-kube-api-access-vtvzs\") on node \"crc\" DevicePath \"\"" Feb 02 10:25:07 crc kubenswrapper[4764]: I0202 10:25:07.846952 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e98aa151-d23e-4b9a-a169-149396d665e1" path="/var/lib/kubelet/pods/e98aa151-d23e-4b9a-a169-149396d665e1/volumes" Feb 02 10:25:08 crc kubenswrapper[4764]: I0202 10:25:08.349911 4764 scope.go:117] "RemoveContainer" containerID="87d803d0f82aaabab517fc6bd41640a0d5b4f5438bb647422e6407ca98493ce8" Feb 02 10:25:08 crc kubenswrapper[4764]: I0202 10:25:08.349963 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-lgpbv/crc-debug-2mfzn" Feb 02 10:25:08 crc kubenswrapper[4764]: I0202 10:25:08.692032 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-lgpbv/crc-debug-wzcn2"] Feb 02 10:25:08 crc kubenswrapper[4764]: E0202 10:25:08.694257 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e98aa151-d23e-4b9a-a169-149396d665e1" containerName="container-00" Feb 02 10:25:08 crc kubenswrapper[4764]: I0202 10:25:08.694407 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="e98aa151-d23e-4b9a-a169-149396d665e1" containerName="container-00" Feb 02 10:25:08 crc kubenswrapper[4764]: I0202 10:25:08.694688 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="e98aa151-d23e-4b9a-a169-149396d665e1" containerName="container-00" Feb 02 10:25:08 crc kubenswrapper[4764]: I0202 10:25:08.695894 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-lgpbv/crc-debug-wzcn2" Feb 02 10:25:08 crc kubenswrapper[4764]: I0202 10:25:08.784255 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-67bmn\" (UniqueName: \"kubernetes.io/projected/619ef402-3c8b-4c08-a00c-c4dae3214fc4-kube-api-access-67bmn\") pod \"crc-debug-wzcn2\" (UID: \"619ef402-3c8b-4c08-a00c-c4dae3214fc4\") " pod="openshift-must-gather-lgpbv/crc-debug-wzcn2" Feb 02 10:25:08 crc kubenswrapper[4764]: I0202 10:25:08.784303 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/619ef402-3c8b-4c08-a00c-c4dae3214fc4-host\") pod \"crc-debug-wzcn2\" (UID: \"619ef402-3c8b-4c08-a00c-c4dae3214fc4\") " pod="openshift-must-gather-lgpbv/crc-debug-wzcn2" Feb 02 10:25:08 crc kubenswrapper[4764]: I0202 10:25:08.887151 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-67bmn\" (UniqueName: \"kubernetes.io/projected/619ef402-3c8b-4c08-a00c-c4dae3214fc4-kube-api-access-67bmn\") pod \"crc-debug-wzcn2\" (UID: \"619ef402-3c8b-4c08-a00c-c4dae3214fc4\") " pod="openshift-must-gather-lgpbv/crc-debug-wzcn2" Feb 02 10:25:08 crc kubenswrapper[4764]: I0202 10:25:08.887210 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/619ef402-3c8b-4c08-a00c-c4dae3214fc4-host\") pod \"crc-debug-wzcn2\" (UID: \"619ef402-3c8b-4c08-a00c-c4dae3214fc4\") " pod="openshift-must-gather-lgpbv/crc-debug-wzcn2" Feb 02 10:25:08 crc kubenswrapper[4764]: I0202 10:25:08.887311 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/619ef402-3c8b-4c08-a00c-c4dae3214fc4-host\") pod \"crc-debug-wzcn2\" (UID: \"619ef402-3c8b-4c08-a00c-c4dae3214fc4\") " pod="openshift-must-gather-lgpbv/crc-debug-wzcn2" Feb 02 10:25:08 crc kubenswrapper[4764]: I0202 10:25:08.910766 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-67bmn\" (UniqueName: \"kubernetes.io/projected/619ef402-3c8b-4c08-a00c-c4dae3214fc4-kube-api-access-67bmn\") pod \"crc-debug-wzcn2\" (UID: \"619ef402-3c8b-4c08-a00c-c4dae3214fc4\") " pod="openshift-must-gather-lgpbv/crc-debug-wzcn2" Feb 02 10:25:09 crc kubenswrapper[4764]: I0202 10:25:09.012239 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-lgpbv/crc-debug-wzcn2" Feb 02 10:25:09 crc kubenswrapper[4764]: I0202 10:25:09.363593 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-lgpbv/crc-debug-wzcn2" event={"ID":"619ef402-3c8b-4c08-a00c-c4dae3214fc4","Type":"ContainerStarted","Data":"a67e0013f05597c14b85afc64f819c7b0b9f2fd5e3a81276ec0b42cfeeeb60d9"} Feb 02 10:25:09 crc kubenswrapper[4764]: I0202 10:25:09.363647 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-lgpbv/crc-debug-wzcn2" event={"ID":"619ef402-3c8b-4c08-a00c-c4dae3214fc4","Type":"ContainerStarted","Data":"c7726d3b411dea6a79a8cd035913296f7875dcdc690cd7f7ab8dbcfda9f55116"} Feb 02 10:25:09 crc kubenswrapper[4764]: I0202 10:25:09.370163 4764 generic.go:334] "Generic (PLEG): container finished" podID="858615be-9087-4141-b5be-d6751cc64025" containerID="6a0992d75a87ff27ab8cf98d46b8d463938bd60d5e33dc18bfacc4ec34ee4c0e" exitCode=0 Feb 02 10:25:09 crc kubenswrapper[4764]: I0202 10:25:09.370214 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nb774" event={"ID":"858615be-9087-4141-b5be-d6751cc64025","Type":"ContainerDied","Data":"6a0992d75a87ff27ab8cf98d46b8d463938bd60d5e33dc18bfacc4ec34ee4c0e"} Feb 02 10:25:09 crc kubenswrapper[4764]: I0202 10:25:09.383035 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-lgpbv/crc-debug-wzcn2" podStartSLOduration=1.383018265 podStartE2EDuration="1.383018265s" podCreationTimestamp="2026-02-02 10:25:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 10:25:09.377099258 +0000 UTC m=+4672.310823346" watchObservedRunningTime="2026-02-02 10:25:09.383018265 +0000 UTC m=+4672.316742353" Feb 02 10:25:10 crc kubenswrapper[4764]: I0202 10:25:10.380668 4764 generic.go:334] "Generic (PLEG): container finished" podID="619ef402-3c8b-4c08-a00c-c4dae3214fc4" containerID="a67e0013f05597c14b85afc64f819c7b0b9f2fd5e3a81276ec0b42cfeeeb60d9" exitCode=0 Feb 02 10:25:10 crc kubenswrapper[4764]: I0202 10:25:10.380864 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-lgpbv/crc-debug-wzcn2" event={"ID":"619ef402-3c8b-4c08-a00c-c4dae3214fc4","Type":"ContainerDied","Data":"a67e0013f05597c14b85afc64f819c7b0b9f2fd5e3a81276ec0b42cfeeeb60d9"} Feb 02 10:25:10 crc kubenswrapper[4764]: I0202 10:25:10.384297 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nb774" event={"ID":"858615be-9087-4141-b5be-d6751cc64025","Type":"ContainerStarted","Data":"177e5ad50537de0d1eb09679f3614d0287637c76506ce0b49bee14636a966f71"} Feb 02 10:25:10 crc kubenswrapper[4764]: I0202 10:25:10.425843 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-nb774" podStartSLOduration=2.7746022200000002 podStartE2EDuration="10.425824101s" podCreationTimestamp="2026-02-02 10:25:00 +0000 UTC" firstStartedPulling="2026-02-02 10:25:02.290703166 +0000 UTC m=+4665.224427244" lastFinishedPulling="2026-02-02 10:25:09.941925037 +0000 UTC m=+4672.875649125" observedRunningTime="2026-02-02 10:25:10.419328717 +0000 UTC m=+4673.353052815" watchObservedRunningTime="2026-02-02 10:25:10.425824101 +0000 UTC m=+4673.359548199" Feb 02 10:25:10 crc kubenswrapper[4764]: I0202 10:25:10.769942 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-nb774" Feb 02 10:25:10 crc kubenswrapper[4764]: I0202 10:25:10.770004 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-nb774" Feb 02 10:25:11 crc kubenswrapper[4764]: I0202 10:25:11.490775 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-lgpbv/crc-debug-wzcn2" Feb 02 10:25:11 crc kubenswrapper[4764]: I0202 10:25:11.525865 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-lgpbv/crc-debug-wzcn2"] Feb 02 10:25:11 crc kubenswrapper[4764]: I0202 10:25:11.535787 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-lgpbv/crc-debug-wzcn2"] Feb 02 10:25:11 crc kubenswrapper[4764]: I0202 10:25:11.644133 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/619ef402-3c8b-4c08-a00c-c4dae3214fc4-host\") pod \"619ef402-3c8b-4c08-a00c-c4dae3214fc4\" (UID: \"619ef402-3c8b-4c08-a00c-c4dae3214fc4\") " Feb 02 10:25:11 crc kubenswrapper[4764]: I0202 10:25:11.644222 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-67bmn\" (UniqueName: \"kubernetes.io/projected/619ef402-3c8b-4c08-a00c-c4dae3214fc4-kube-api-access-67bmn\") pod \"619ef402-3c8b-4c08-a00c-c4dae3214fc4\" (UID: \"619ef402-3c8b-4c08-a00c-c4dae3214fc4\") " Feb 02 10:25:11 crc kubenswrapper[4764]: I0202 10:25:11.644308 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/619ef402-3c8b-4c08-a00c-c4dae3214fc4-host" (OuterVolumeSpecName: "host") pod "619ef402-3c8b-4c08-a00c-c4dae3214fc4" (UID: "619ef402-3c8b-4c08-a00c-c4dae3214fc4"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 10:25:11 crc kubenswrapper[4764]: I0202 10:25:11.644652 4764 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/619ef402-3c8b-4c08-a00c-c4dae3214fc4-host\") on node \"crc\" DevicePath \"\"" Feb 02 10:25:11 crc kubenswrapper[4764]: I0202 10:25:11.656563 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/619ef402-3c8b-4c08-a00c-c4dae3214fc4-kube-api-access-67bmn" (OuterVolumeSpecName: "kube-api-access-67bmn") pod "619ef402-3c8b-4c08-a00c-c4dae3214fc4" (UID: "619ef402-3c8b-4c08-a00c-c4dae3214fc4"). InnerVolumeSpecName "kube-api-access-67bmn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 10:25:11 crc kubenswrapper[4764]: I0202 10:25:11.746457 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-67bmn\" (UniqueName: \"kubernetes.io/projected/619ef402-3c8b-4c08-a00c-c4dae3214fc4-kube-api-access-67bmn\") on node \"crc\" DevicePath \"\"" Feb 02 10:25:11 crc kubenswrapper[4764]: I0202 10:25:11.824310 4764 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-nb774" podUID="858615be-9087-4141-b5be-d6751cc64025" containerName="registry-server" probeResult="failure" output=< Feb 02 10:25:11 crc kubenswrapper[4764]: timeout: failed to connect service ":50051" within 1s Feb 02 10:25:11 crc kubenswrapper[4764]: > Feb 02 10:25:11 crc kubenswrapper[4764]: I0202 10:25:11.835812 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="619ef402-3c8b-4c08-a00c-c4dae3214fc4" path="/var/lib/kubelet/pods/619ef402-3c8b-4c08-a00c-c4dae3214fc4/volumes" Feb 02 10:25:12 crc kubenswrapper[4764]: I0202 10:25:12.401627 4764 scope.go:117] "RemoveContainer" containerID="a67e0013f05597c14b85afc64f819c7b0b9f2fd5e3a81276ec0b42cfeeeb60d9" Feb 02 10:25:12 crc kubenswrapper[4764]: I0202 10:25:12.401659 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-lgpbv/crc-debug-wzcn2" Feb 02 10:25:12 crc kubenswrapper[4764]: I0202 10:25:12.766097 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-lgpbv/crc-debug-mx965"] Feb 02 10:25:12 crc kubenswrapper[4764]: E0202 10:25:12.766574 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="619ef402-3c8b-4c08-a00c-c4dae3214fc4" containerName="container-00" Feb 02 10:25:12 crc kubenswrapper[4764]: I0202 10:25:12.766590 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="619ef402-3c8b-4c08-a00c-c4dae3214fc4" containerName="container-00" Feb 02 10:25:12 crc kubenswrapper[4764]: I0202 10:25:12.766803 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="619ef402-3c8b-4c08-a00c-c4dae3214fc4" containerName="container-00" Feb 02 10:25:12 crc kubenswrapper[4764]: I0202 10:25:12.767596 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-lgpbv/crc-debug-mx965" Feb 02 10:25:12 crc kubenswrapper[4764]: I0202 10:25:12.868176 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2e318171-ef4e-4d2f-bcca-bfe22fa13637-host\") pod \"crc-debug-mx965\" (UID: \"2e318171-ef4e-4d2f-bcca-bfe22fa13637\") " pod="openshift-must-gather-lgpbv/crc-debug-mx965" Feb 02 10:25:12 crc kubenswrapper[4764]: I0202 10:25:12.868268 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bcd8t\" (UniqueName: \"kubernetes.io/projected/2e318171-ef4e-4d2f-bcca-bfe22fa13637-kube-api-access-bcd8t\") pod \"crc-debug-mx965\" (UID: \"2e318171-ef4e-4d2f-bcca-bfe22fa13637\") " pod="openshift-must-gather-lgpbv/crc-debug-mx965" Feb 02 10:25:12 crc kubenswrapper[4764]: I0202 10:25:12.970099 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2e318171-ef4e-4d2f-bcca-bfe22fa13637-host\") pod \"crc-debug-mx965\" (UID: \"2e318171-ef4e-4d2f-bcca-bfe22fa13637\") " pod="openshift-must-gather-lgpbv/crc-debug-mx965" Feb 02 10:25:12 crc kubenswrapper[4764]: I0202 10:25:12.971380 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bcd8t\" (UniqueName: \"kubernetes.io/projected/2e318171-ef4e-4d2f-bcca-bfe22fa13637-kube-api-access-bcd8t\") pod \"crc-debug-mx965\" (UID: \"2e318171-ef4e-4d2f-bcca-bfe22fa13637\") " pod="openshift-must-gather-lgpbv/crc-debug-mx965" Feb 02 10:25:12 crc kubenswrapper[4764]: I0202 10:25:12.971001 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2e318171-ef4e-4d2f-bcca-bfe22fa13637-host\") pod \"crc-debug-mx965\" (UID: \"2e318171-ef4e-4d2f-bcca-bfe22fa13637\") " pod="openshift-must-gather-lgpbv/crc-debug-mx965" Feb 02 10:25:12 crc kubenswrapper[4764]: I0202 10:25:12.997994 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bcd8t\" (UniqueName: \"kubernetes.io/projected/2e318171-ef4e-4d2f-bcca-bfe22fa13637-kube-api-access-bcd8t\") pod \"crc-debug-mx965\" (UID: \"2e318171-ef4e-4d2f-bcca-bfe22fa13637\") " pod="openshift-must-gather-lgpbv/crc-debug-mx965" Feb 02 10:25:13 crc kubenswrapper[4764]: I0202 10:25:13.085333 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-lgpbv/crc-debug-mx965" Feb 02 10:25:13 crc kubenswrapper[4764]: I0202 10:25:13.415416 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-lgpbv/crc-debug-mx965" event={"ID":"2e318171-ef4e-4d2f-bcca-bfe22fa13637","Type":"ContainerStarted","Data":"fcbfcd8f6c034d844c12eababe2a9586003b7e0beeb86b7426bbd37fbe25bbc3"} Feb 02 10:25:14 crc kubenswrapper[4764]: I0202 10:25:14.424700 4764 generic.go:334] "Generic (PLEG): container finished" podID="2e318171-ef4e-4d2f-bcca-bfe22fa13637" containerID="0d218f2ca37cd5c3b438940aac531d8fa576ab7f623680510f75d0fc3737fb5c" exitCode=0 Feb 02 10:25:14 crc kubenswrapper[4764]: I0202 10:25:14.424742 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-lgpbv/crc-debug-mx965" event={"ID":"2e318171-ef4e-4d2f-bcca-bfe22fa13637","Type":"ContainerDied","Data":"0d218f2ca37cd5c3b438940aac531d8fa576ab7f623680510f75d0fc3737fb5c"} Feb 02 10:25:14 crc kubenswrapper[4764]: I0202 10:25:14.465022 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-lgpbv/crc-debug-mx965"] Feb 02 10:25:14 crc kubenswrapper[4764]: I0202 10:25:14.477402 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-lgpbv/crc-debug-mx965"] Feb 02 10:25:15 crc kubenswrapper[4764]: I0202 10:25:15.532290 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-lgpbv/crc-debug-mx965" Feb 02 10:25:15 crc kubenswrapper[4764]: I0202 10:25:15.637608 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bcd8t\" (UniqueName: \"kubernetes.io/projected/2e318171-ef4e-4d2f-bcca-bfe22fa13637-kube-api-access-bcd8t\") pod \"2e318171-ef4e-4d2f-bcca-bfe22fa13637\" (UID: \"2e318171-ef4e-4d2f-bcca-bfe22fa13637\") " Feb 02 10:25:15 crc kubenswrapper[4764]: I0202 10:25:15.638028 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2e318171-ef4e-4d2f-bcca-bfe22fa13637-host\") pod \"2e318171-ef4e-4d2f-bcca-bfe22fa13637\" (UID: \"2e318171-ef4e-4d2f-bcca-bfe22fa13637\") " Feb 02 10:25:15 crc kubenswrapper[4764]: I0202 10:25:15.638162 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2e318171-ef4e-4d2f-bcca-bfe22fa13637-host" (OuterVolumeSpecName: "host") pod "2e318171-ef4e-4d2f-bcca-bfe22fa13637" (UID: "2e318171-ef4e-4d2f-bcca-bfe22fa13637"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 10:25:15 crc kubenswrapper[4764]: I0202 10:25:15.638843 4764 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2e318171-ef4e-4d2f-bcca-bfe22fa13637-host\") on node \"crc\" DevicePath \"\"" Feb 02 10:25:15 crc kubenswrapper[4764]: I0202 10:25:15.644152 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2e318171-ef4e-4d2f-bcca-bfe22fa13637-kube-api-access-bcd8t" (OuterVolumeSpecName: "kube-api-access-bcd8t") pod "2e318171-ef4e-4d2f-bcca-bfe22fa13637" (UID: "2e318171-ef4e-4d2f-bcca-bfe22fa13637"). InnerVolumeSpecName "kube-api-access-bcd8t". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 10:25:15 crc kubenswrapper[4764]: I0202 10:25:15.741182 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bcd8t\" (UniqueName: \"kubernetes.io/projected/2e318171-ef4e-4d2f-bcca-bfe22fa13637-kube-api-access-bcd8t\") on node \"crc\" DevicePath \"\"" Feb 02 10:25:15 crc kubenswrapper[4764]: I0202 10:25:15.838181 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2e318171-ef4e-4d2f-bcca-bfe22fa13637" path="/var/lib/kubelet/pods/2e318171-ef4e-4d2f-bcca-bfe22fa13637/volumes" Feb 02 10:25:16 crc kubenswrapper[4764]: I0202 10:25:16.441043 4764 scope.go:117] "RemoveContainer" containerID="0d218f2ca37cd5c3b438940aac531d8fa576ab7f623680510f75d0fc3737fb5c" Feb 02 10:25:16 crc kubenswrapper[4764]: I0202 10:25:16.441087 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-lgpbv/crc-debug-mx965" Feb 02 10:25:21 crc kubenswrapper[4764]: I0202 10:25:21.826432 4764 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-nb774" podUID="858615be-9087-4141-b5be-d6751cc64025" containerName="registry-server" probeResult="failure" output=< Feb 02 10:25:21 crc kubenswrapper[4764]: timeout: failed to connect service ":50051" within 1s Feb 02 10:25:21 crc kubenswrapper[4764]: > Feb 02 10:25:32 crc kubenswrapper[4764]: I0202 10:25:32.036140 4764 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-nb774" podUID="858615be-9087-4141-b5be-d6751cc64025" containerName="registry-server" probeResult="failure" output=< Feb 02 10:25:32 crc kubenswrapper[4764]: timeout: failed to connect service ":50051" within 1s Feb 02 10:25:32 crc kubenswrapper[4764]: > Feb 02 10:25:41 crc kubenswrapper[4764]: I0202 10:25:41.836399 4764 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-nb774" podUID="858615be-9087-4141-b5be-d6751cc64025" containerName="registry-server" probeResult="failure" output=< Feb 02 10:25:41 crc kubenswrapper[4764]: timeout: failed to connect service ":50051" within 1s Feb 02 10:25:41 crc kubenswrapper[4764]: > Feb 02 10:25:51 crc kubenswrapper[4764]: I0202 10:25:51.837091 4764 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-nb774" podUID="858615be-9087-4141-b5be-d6751cc64025" containerName="registry-server" probeResult="failure" output=< Feb 02 10:25:51 crc kubenswrapper[4764]: timeout: failed to connect service ":50051" within 1s Feb 02 10:25:51 crc kubenswrapper[4764]: > Feb 02 10:26:00 crc kubenswrapper[4764]: I0202 10:26:00.831199 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-nb774" Feb 02 10:26:00 crc kubenswrapper[4764]: I0202 10:26:00.906558 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-nb774" Feb 02 10:26:01 crc kubenswrapper[4764]: I0202 10:26:01.092438 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-nb774"] Feb 02 10:26:02 crc kubenswrapper[4764]: I0202 10:26:02.835952 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-nb774" podUID="858615be-9087-4141-b5be-d6751cc64025" containerName="registry-server" containerID="cri-o://177e5ad50537de0d1eb09679f3614d0287637c76506ce0b49bee14636a966f71" gracePeriod=2 Feb 02 10:26:03 crc kubenswrapper[4764]: I0202 10:26:03.864272 4764 generic.go:334] "Generic (PLEG): container finished" podID="858615be-9087-4141-b5be-d6751cc64025" containerID="177e5ad50537de0d1eb09679f3614d0287637c76506ce0b49bee14636a966f71" exitCode=0 Feb 02 10:26:03 crc kubenswrapper[4764]: I0202 10:26:03.864404 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nb774" event={"ID":"858615be-9087-4141-b5be-d6751cc64025","Type":"ContainerDied","Data":"177e5ad50537de0d1eb09679f3614d0287637c76506ce0b49bee14636a966f71"} Feb 02 10:26:04 crc kubenswrapper[4764]: I0202 10:26:04.074664 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nb774" Feb 02 10:26:04 crc kubenswrapper[4764]: I0202 10:26:04.164610 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/858615be-9087-4141-b5be-d6751cc64025-utilities\") pod \"858615be-9087-4141-b5be-d6751cc64025\" (UID: \"858615be-9087-4141-b5be-d6751cc64025\") " Feb 02 10:26:04 crc kubenswrapper[4764]: I0202 10:26:04.164821 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bdpjj\" (UniqueName: \"kubernetes.io/projected/858615be-9087-4141-b5be-d6751cc64025-kube-api-access-bdpjj\") pod \"858615be-9087-4141-b5be-d6751cc64025\" (UID: \"858615be-9087-4141-b5be-d6751cc64025\") " Feb 02 10:26:04 crc kubenswrapper[4764]: I0202 10:26:04.164888 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/858615be-9087-4141-b5be-d6751cc64025-catalog-content\") pod \"858615be-9087-4141-b5be-d6751cc64025\" (UID: \"858615be-9087-4141-b5be-d6751cc64025\") " Feb 02 10:26:04 crc kubenswrapper[4764]: I0202 10:26:04.165345 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/858615be-9087-4141-b5be-d6751cc64025-utilities" (OuterVolumeSpecName: "utilities") pod "858615be-9087-4141-b5be-d6751cc64025" (UID: "858615be-9087-4141-b5be-d6751cc64025"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 10:26:04 crc kubenswrapper[4764]: I0202 10:26:04.175224 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/858615be-9087-4141-b5be-d6751cc64025-kube-api-access-bdpjj" (OuterVolumeSpecName: "kube-api-access-bdpjj") pod "858615be-9087-4141-b5be-d6751cc64025" (UID: "858615be-9087-4141-b5be-d6751cc64025"). InnerVolumeSpecName "kube-api-access-bdpjj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 10:26:04 crc kubenswrapper[4764]: I0202 10:26:04.268230 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bdpjj\" (UniqueName: \"kubernetes.io/projected/858615be-9087-4141-b5be-d6751cc64025-kube-api-access-bdpjj\") on node \"crc\" DevicePath \"\"" Feb 02 10:26:04 crc kubenswrapper[4764]: I0202 10:26:04.268688 4764 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/858615be-9087-4141-b5be-d6751cc64025-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 10:26:04 crc kubenswrapper[4764]: I0202 10:26:04.302906 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/858615be-9087-4141-b5be-d6751cc64025-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "858615be-9087-4141-b5be-d6751cc64025" (UID: "858615be-9087-4141-b5be-d6751cc64025"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 10:26:04 crc kubenswrapper[4764]: I0202 10:26:04.371067 4764 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/858615be-9087-4141-b5be-d6751cc64025-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 10:26:04 crc kubenswrapper[4764]: I0202 10:26:04.881827 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nb774" event={"ID":"858615be-9087-4141-b5be-d6751cc64025","Type":"ContainerDied","Data":"8e877b6c179a416f8e23558ab337453e118ee891dd335fd8ffca777d03bf02cb"} Feb 02 10:26:04 crc kubenswrapper[4764]: I0202 10:26:04.881890 4764 scope.go:117] "RemoveContainer" containerID="177e5ad50537de0d1eb09679f3614d0287637c76506ce0b49bee14636a966f71" Feb 02 10:26:04 crc kubenswrapper[4764]: I0202 10:26:04.881917 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nb774" Feb 02 10:26:04 crc kubenswrapper[4764]: I0202 10:26:04.916453 4764 scope.go:117] "RemoveContainer" containerID="6a0992d75a87ff27ab8cf98d46b8d463938bd60d5e33dc18bfacc4ec34ee4c0e" Feb 02 10:26:04 crc kubenswrapper[4764]: I0202 10:26:04.972623 4764 scope.go:117] "RemoveContainer" containerID="4a39d8d588985a2e6bd8082130bc137a085433be68d8bf80cdb67f506f45be08" Feb 02 10:26:04 crc kubenswrapper[4764]: I0202 10:26:04.992499 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-nb774"] Feb 02 10:26:05 crc kubenswrapper[4764]: I0202 10:26:05.017343 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-nb774"] Feb 02 10:26:05 crc kubenswrapper[4764]: I0202 10:26:05.852871 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="858615be-9087-4141-b5be-d6751cc64025" path="/var/lib/kubelet/pods/858615be-9087-4141-b5be-d6751cc64025/volumes" Feb 02 10:26:30 crc kubenswrapper[4764]: I0202 10:26:30.849159 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-797597c45b-9mkpp_f06f976d-5e29-4e78-81f4-81de01ae8391/barbican-api/0.log" Feb 02 10:26:31 crc kubenswrapper[4764]: I0202 10:26:31.053994 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-797597c45b-9mkpp_f06f976d-5e29-4e78-81f4-81de01ae8391/barbican-api-log/0.log" Feb 02 10:26:31 crc kubenswrapper[4764]: I0202 10:26:31.075771 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-855c89b68b-ptbcl_b004a188-196b-4812-a9e6-e398062cf4d3/barbican-keystone-listener/0.log" Feb 02 10:26:31 crc kubenswrapper[4764]: I0202 10:26:31.163699 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-855c89b68b-ptbcl_b004a188-196b-4812-a9e6-e398062cf4d3/barbican-keystone-listener-log/0.log" Feb 02 10:26:31 crc kubenswrapper[4764]: I0202 10:26:31.252822 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-5686dc79cc-sb69r_4e3b3433-3d64-44ce-9a3a-14fcd0bdaa6d/barbican-worker/0.log" Feb 02 10:26:31 crc kubenswrapper[4764]: I0202 10:26:31.353400 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-5686dc79cc-sb69r_4e3b3433-3d64-44ce-9a3a-14fcd0bdaa6d/barbican-worker-log/0.log" Feb 02 10:26:31 crc kubenswrapper[4764]: I0202 10:26:31.516924 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-d9x8m_af835a1d-09cc-49ca-979f-41eaf146e926/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Feb 02 10:26:31 crc kubenswrapper[4764]: I0202 10:26:31.582045 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_242b61fa-f577-449a-86b2-4cafa5175bd6/ceilometer-central-agent/0.log" Feb 02 10:26:31 crc kubenswrapper[4764]: I0202 10:26:31.690557 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_242b61fa-f577-449a-86b2-4cafa5175bd6/ceilometer-notification-agent/0.log" Feb 02 10:26:31 crc kubenswrapper[4764]: I0202 10:26:31.740217 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_242b61fa-f577-449a-86b2-4cafa5175bd6/proxy-httpd/0.log" Feb 02 10:26:31 crc kubenswrapper[4764]: I0202 10:26:31.799843 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_242b61fa-f577-449a-86b2-4cafa5175bd6/sg-core/0.log" Feb 02 10:26:31 crc kubenswrapper[4764]: I0202 10:26:31.934020 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceph-client-edpm-deployment-openstack-edpm-ipam-bs4pz_3858fd68-f1b8-4e73-a0c9-5fb26dcbedd6/ceph-client-edpm-deployment-openstack-edpm-ipam/0.log" Feb 02 10:26:32 crc kubenswrapper[4764]: I0202 10:26:32.045264 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-d4lm7_329c5f09-ca39-4ced-a6b6-02db68c98510/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam/0.log" Feb 02 10:26:32 crc kubenswrapper[4764]: I0202 10:26:32.229977 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_4b8e0fd6-1949-4b4f-bbc5-b663c51f08c4/cinder-api/0.log" Feb 02 10:26:32 crc kubenswrapper[4764]: I0202 10:26:32.316834 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_4b8e0fd6-1949-4b4f-bbc5-b663c51f08c4/cinder-api-log/0.log" Feb 02 10:26:32 crc kubenswrapper[4764]: I0202 10:26:32.507371 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_371fb9ce-9ba4-429b-a68c-62d2e12d688c/probe/0.log" Feb 02 10:26:32 crc kubenswrapper[4764]: I0202 10:26:32.623690 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_371fb9ce-9ba4-429b-a68c-62d2e12d688c/cinder-backup/0.log" Feb 02 10:26:32 crc kubenswrapper[4764]: I0202 10:26:32.717796 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_10cfff03-5b66-41ef-948b-0bad4f6e9a91/cinder-scheduler/0.log" Feb 02 10:26:32 crc kubenswrapper[4764]: I0202 10:26:32.801055 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_10cfff03-5b66-41ef-948b-0bad4f6e9a91/probe/0.log" Feb 02 10:26:33 crc kubenswrapper[4764]: I0202 10:26:33.068960 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_1c67d8f7-21c3-456d-af99-c96d04090915/cinder-volume/0.log" Feb 02 10:26:33 crc kubenswrapper[4764]: I0202 10:26:33.450311 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_1c67d8f7-21c3-456d-af99-c96d04090915/probe/0.log" Feb 02 10:26:33 crc kubenswrapper[4764]: I0202 10:26:33.497974 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-rsd54_5a0ee3ef-d01f-4961-8362-8b6c4d3f37fc/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Feb 02 10:26:33 crc kubenswrapper[4764]: I0202 10:26:33.698372 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-c4mpn_c3b49d2a-ed47-45cb-9940-b5e87ff7a197/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Feb 02 10:26:33 crc kubenswrapper[4764]: I0202 10:26:33.877916 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6d94cddcb9-n4wc4_d61062ba-f497-4f61-9b51-27e62888797e/init/0.log" Feb 02 10:26:34 crc kubenswrapper[4764]: I0202 10:26:34.102575 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6d94cddcb9-n4wc4_d61062ba-f497-4f61-9b51-27e62888797e/init/0.log" Feb 02 10:26:34 crc kubenswrapper[4764]: I0202 10:26:34.173978 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_9f924031-1e49-4009-aeba-a4e5f3476373/glance-httpd/0.log" Feb 02 10:26:34 crc kubenswrapper[4764]: I0202 10:26:34.365245 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6d94cddcb9-n4wc4_d61062ba-f497-4f61-9b51-27e62888797e/dnsmasq-dns/0.log" Feb 02 10:26:34 crc kubenswrapper[4764]: I0202 10:26:34.426759 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_9f924031-1e49-4009-aeba-a4e5f3476373/glance-log/0.log" Feb 02 10:26:34 crc kubenswrapper[4764]: I0202 10:26:34.606652 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_687693a0-eb78-4922-991f-6738de00f468/glance-httpd/0.log" Feb 02 10:26:34 crc kubenswrapper[4764]: I0202 10:26:34.671055 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_687693a0-eb78-4922-991f-6738de00f468/glance-log/0.log" Feb 02 10:26:34 crc kubenswrapper[4764]: I0202 10:26:34.799395 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-698ff6d98d-6h7sk_174bcb22-4370-4c3f-9141-9c1cee0d2542/horizon/0.log" Feb 02 10:26:35 crc kubenswrapper[4764]: I0202 10:26:35.083292 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-9q5kq_7d388199-9f34-415c-b83c-d5f54d17d063/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Feb 02 10:26:35 crc kubenswrapper[4764]: I0202 10:26:35.088983 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-698ff6d98d-6h7sk_174bcb22-4370-4c3f-9141-9c1cee0d2542/horizon-log/0.log" Feb 02 10:26:35 crc kubenswrapper[4764]: I0202 10:26:35.220398 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-5469f_6b2831d8-cc8c-4ded-99e1-45ad94ff9776/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Feb 02 10:26:35 crc kubenswrapper[4764]: I0202 10:26:35.854980 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29500441-qpcvv_b77a0084-b391-4221-86d1-39629566a940/keystone-cron/0.log" Feb 02 10:26:35 crc kubenswrapper[4764]: I0202 10:26:35.887417 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-b47bfbff6-m265b_45891ac1-d70d-4c3d-af07-6a65855a9281/keystone-api/0.log" Feb 02 10:26:36 crc kubenswrapper[4764]: I0202 10:26:36.044545 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_2fce6f0b-cb36-4ff1-ab73-35c40d709502/kube-state-metrics/0.log" Feb 02 10:26:36 crc kubenswrapper[4764]: I0202 10:26:36.215672 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-ccmw8_f232691b-63e5-4e8a-9659-a6591c7dbe61/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Feb 02 10:26:36 crc kubenswrapper[4764]: I0202 10:26:36.357328 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_eed25ff8-c8ee-41d6-bb33-89e4c5f30724/manila-api/0.log" Feb 02 10:26:36 crc kubenswrapper[4764]: I0202 10:26:36.496867 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_eed25ff8-c8ee-41d6-bb33-89e4c5f30724/manila-api-log/0.log" Feb 02 10:26:36 crc kubenswrapper[4764]: I0202 10:26:36.825273 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_47fbbd57-48b1-4cbf-b857-8ef14b8c8529/probe/0.log" Feb 02 10:26:36 crc kubenswrapper[4764]: I0202 10:26:36.950349 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_b9fbc0ac-9686-4433-ac79-79b9e0cfb54b/manila-share/0.log" Feb 02 10:26:36 crc kubenswrapper[4764]: I0202 10:26:36.973731 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_b9fbc0ac-9686-4433-ac79-79b9e0cfb54b/probe/0.log" Feb 02 10:26:36 crc kubenswrapper[4764]: I0202 10:26:36.991449 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_47fbbd57-48b1-4cbf-b857-8ef14b8c8529/manila-scheduler/0.log" Feb 02 10:26:37 crc kubenswrapper[4764]: I0202 10:26:37.286991 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-6b9f774cdf-j8pdr_c7553cbf-5c61-4003-bd4a-aad61da77950/neutron-api/0.log" Feb 02 10:26:37 crc kubenswrapper[4764]: I0202 10:26:37.321661 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-6b9f774cdf-j8pdr_c7553cbf-5c61-4003-bd4a-aad61da77950/neutron-httpd/0.log" Feb 02 10:26:37 crc kubenswrapper[4764]: I0202 10:26:37.565356 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-6pxwj_a74b118e-ee59-4e87-967d-c8332a898a4c/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Feb 02 10:26:38 crc kubenswrapper[4764]: I0202 10:26:38.218525 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_db89c9c9-cad1-4e6b-aaf0-708e5b8b32b4/nova-api-log/0.log" Feb 02 10:26:38 crc kubenswrapper[4764]: I0202 10:26:38.268977 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_82f0b9ff-5694-4e86-acb1-6a8ef7b5b97a/nova-cell0-conductor-conductor/0.log" Feb 02 10:26:38 crc kubenswrapper[4764]: I0202 10:26:38.672677 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_b640955b-f755-48d2-9f0a-751b24424f35/nova-cell1-conductor-conductor/0.log" Feb 02 10:26:38 crc kubenswrapper[4764]: I0202 10:26:38.703620 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_db89c9c9-cad1-4e6b-aaf0-708e5b8b32b4/nova-api-api/0.log" Feb 02 10:26:38 crc kubenswrapper[4764]: I0202 10:26:38.731615 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_41b7a57b-b6a0-4af7-9304-e4c89c8e3812/nova-cell1-novncproxy-novncproxy/0.log" Feb 02 10:26:38 crc kubenswrapper[4764]: I0202 10:26:38.995476 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-h4h49_b5b2ecb9-0adf-489b-8a98-42cad0682d09/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam/0.log" Feb 02 10:26:39 crc kubenswrapper[4764]: I0202 10:26:39.085950 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_61eec87e-6dfc-41b8-a670-0914cec8b593/nova-metadata-log/0.log" Feb 02 10:26:39 crc kubenswrapper[4764]: I0202 10:26:39.480348 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_a9177511-b834-4e68-9f6f-8b1b4470b274/mysql-bootstrap/0.log" Feb 02 10:26:39 crc kubenswrapper[4764]: I0202 10:26:39.495161 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_dd11a817-6357-4405-983e-181451ce4daf/nova-scheduler-scheduler/0.log" Feb 02 10:26:39 crc kubenswrapper[4764]: I0202 10:26:39.697732 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_a9177511-b834-4e68-9f6f-8b1b4470b274/galera/0.log" Feb 02 10:26:39 crc kubenswrapper[4764]: I0202 10:26:39.721973 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_a9177511-b834-4e68-9f6f-8b1b4470b274/mysql-bootstrap/0.log" Feb 02 10:26:39 crc kubenswrapper[4764]: I0202 10:26:39.972499 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_1d6b6c31-819e-4098-b11b-6073cc751cad/mysql-bootstrap/0.log" Feb 02 10:26:40 crc kubenswrapper[4764]: I0202 10:26:40.150832 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_1d6b6c31-819e-4098-b11b-6073cc751cad/mysql-bootstrap/0.log" Feb 02 10:26:40 crc kubenswrapper[4764]: I0202 10:26:40.289863 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_1d6b6c31-819e-4098-b11b-6073cc751cad/galera/0.log" Feb 02 10:26:40 crc kubenswrapper[4764]: I0202 10:26:40.398073 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_54c115e4-eceb-49af-96ba-854f48802c73/openstackclient/0.log" Feb 02 10:26:40 crc kubenswrapper[4764]: I0202 10:26:40.676464 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-6v7f6_66662206-d732-404d-a81d-6303d9947d60/ovn-controller/0.log" Feb 02 10:26:40 crc kubenswrapper[4764]: I0202 10:26:40.821909 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-kgc8j_b3becd85-0167-4ffe-802d-67988c486908/openstack-network-exporter/0.log" Feb 02 10:26:41 crc kubenswrapper[4764]: I0202 10:26:41.007388 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_61eec87e-6dfc-41b8-a670-0914cec8b593/nova-metadata-metadata/0.log" Feb 02 10:26:41 crc kubenswrapper[4764]: I0202 10:26:41.047584 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-2p2bk_29a433e1-2651-43cd-99d1-95694063c980/ovsdb-server-init/0.log" Feb 02 10:26:41 crc kubenswrapper[4764]: I0202 10:26:41.209895 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-2p2bk_29a433e1-2651-43cd-99d1-95694063c980/ovsdb-server-init/0.log" Feb 02 10:26:41 crc kubenswrapper[4764]: I0202 10:26:41.306657 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-2p2bk_29a433e1-2651-43cd-99d1-95694063c980/ovsdb-server/0.log" Feb 02 10:26:41 crc kubenswrapper[4764]: I0202 10:26:41.324412 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-2p2bk_29a433e1-2651-43cd-99d1-95694063c980/ovs-vswitchd/0.log" Feb 02 10:26:41 crc kubenswrapper[4764]: I0202 10:26:41.567411 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-2qbnb_e6da450f-20b5-41ef-b9bb-a0df89b3d4f3/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Feb 02 10:26:41 crc kubenswrapper[4764]: I0202 10:26:41.631657 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_f56696ad-9508-4f7b-89fb-28b59af303a0/openstack-network-exporter/0.log" Feb 02 10:26:41 crc kubenswrapper[4764]: I0202 10:26:41.638397 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_f56696ad-9508-4f7b-89fb-28b59af303a0/ovn-northd/0.log" Feb 02 10:26:41 crc kubenswrapper[4764]: I0202 10:26:41.894369 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_cba91a53-7d3d-481b-bd55-3f0bbe7ccbb6/openstack-network-exporter/0.log" Feb 02 10:26:41 crc kubenswrapper[4764]: I0202 10:26:41.981658 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_cba91a53-7d3d-481b-bd55-3f0bbe7ccbb6/ovsdbserver-nb/0.log" Feb 02 10:26:42 crc kubenswrapper[4764]: I0202 10:26:42.122008 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_89c39976-ff91-4c6b-b369-5cba94e02db8/openstack-network-exporter/0.log" Feb 02 10:26:42 crc kubenswrapper[4764]: I0202 10:26:42.232509 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_89c39976-ff91-4c6b-b369-5cba94e02db8/ovsdbserver-sb/0.log" Feb 02 10:26:42 crc kubenswrapper[4764]: I0202 10:26:42.385678 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-68d4474c54-98j2g_f7a36a2c-9ade-4bb2-84d2-faa137207385/placement-api/0.log" Feb 02 10:26:42 crc kubenswrapper[4764]: I0202 10:26:42.509901 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_c60b198c-58ad-4363-afea-257f51f64e51/setup-container/0.log" Feb 02 10:26:42 crc kubenswrapper[4764]: I0202 10:26:42.557631 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-68d4474c54-98j2g_f7a36a2c-9ade-4bb2-84d2-faa137207385/placement-log/0.log" Feb 02 10:26:42 crc kubenswrapper[4764]: I0202 10:26:42.861376 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_c60b198c-58ad-4363-afea-257f51f64e51/setup-container/0.log" Feb 02 10:26:42 crc kubenswrapper[4764]: I0202 10:26:42.878049 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_c60b198c-58ad-4363-afea-257f51f64e51/rabbitmq/0.log" Feb 02 10:26:43 crc kubenswrapper[4764]: I0202 10:26:43.044765 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_ec28045f-e1d4-4b22-bb7d-3c6d2a1f1a38/setup-container/0.log" Feb 02 10:26:43 crc kubenswrapper[4764]: I0202 10:26:43.261677 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_ec28045f-e1d4-4b22-bb7d-3c6d2a1f1a38/setup-container/0.log" Feb 02 10:26:43 crc kubenswrapper[4764]: I0202 10:26:43.355440 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_ec28045f-e1d4-4b22-bb7d-3c6d2a1f1a38/rabbitmq/0.log" Feb 02 10:26:43 crc kubenswrapper[4764]: I0202 10:26:43.417821 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-hrmzd_a79016d5-22d2-4894-829c-7570224b6f69/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Feb 02 10:26:43 crc kubenswrapper[4764]: I0202 10:26:43.578157 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-8f7dl_feb92d24-e755-4837-8461-15e7e3a7b0e8/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Feb 02 10:26:43 crc kubenswrapper[4764]: I0202 10:26:43.855202 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-t5j4g_8649507c-417e-40fc-9a67-b6d6831f3ab9/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Feb 02 10:26:43 crc kubenswrapper[4764]: I0202 10:26:43.986767 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-97dd2_409c3e31-e6bf-409c-9bd8-7bcf433c0ce8/ssh-known-hosts-edpm-deployment/0.log" Feb 02 10:26:44 crc kubenswrapper[4764]: I0202 10:26:44.090434 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_09adc919-28cf-4d2d-b40d-183a2f51f06f/tempest-tests-tempest-tests-runner/0.log" Feb 02 10:26:44 crc kubenswrapper[4764]: I0202 10:26:44.287290 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_7576fa25-4c66-4f89-90a5-c2979b1c0eaa/test-operator-logs-container/0.log" Feb 02 10:26:44 crc kubenswrapper[4764]: I0202 10:26:44.402543 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-nwdxw_ddae0daa-153b-425b-92dd-d6c315ace4e5/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Feb 02 10:26:59 crc kubenswrapper[4764]: I0202 10:26:59.976699 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_cd76e53a-13fc-4293-b5ca-2324224850cc/memcached/0.log" Feb 02 10:27:13 crc kubenswrapper[4764]: I0202 10:27:13.543813 4764 patch_prober.go:28] interesting pod/machine-config-daemon-4ndm4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 10:27:13 crc kubenswrapper[4764]: I0202 10:27:13.544498 4764 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 10:27:18 crc kubenswrapper[4764]: I0202 10:27:18.911828 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_805de4b7130efd4fbe2290d0c51d78e6295b83d77efc3ff2d1b014ffe7z8r74_894b5d6a-cddd-43e6-a59e-6076f985e5a3/util/0.log" Feb 02 10:27:19 crc kubenswrapper[4764]: I0202 10:27:19.149374 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_805de4b7130efd4fbe2290d0c51d78e6295b83d77efc3ff2d1b014ffe7z8r74_894b5d6a-cddd-43e6-a59e-6076f985e5a3/util/0.log" Feb 02 10:27:19 crc kubenswrapper[4764]: I0202 10:27:19.150728 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_805de4b7130efd4fbe2290d0c51d78e6295b83d77efc3ff2d1b014ffe7z8r74_894b5d6a-cddd-43e6-a59e-6076f985e5a3/pull/0.log" Feb 02 10:27:19 crc kubenswrapper[4764]: I0202 10:27:19.158686 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_805de4b7130efd4fbe2290d0c51d78e6295b83d77efc3ff2d1b014ffe7z8r74_894b5d6a-cddd-43e6-a59e-6076f985e5a3/pull/0.log" Feb 02 10:27:19 crc kubenswrapper[4764]: I0202 10:27:19.378146 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_805de4b7130efd4fbe2290d0c51d78e6295b83d77efc3ff2d1b014ffe7z8r74_894b5d6a-cddd-43e6-a59e-6076f985e5a3/util/0.log" Feb 02 10:27:19 crc kubenswrapper[4764]: I0202 10:27:19.421751 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_805de4b7130efd4fbe2290d0c51d78e6295b83d77efc3ff2d1b014ffe7z8r74_894b5d6a-cddd-43e6-a59e-6076f985e5a3/pull/0.log" Feb 02 10:27:19 crc kubenswrapper[4764]: I0202 10:27:19.472881 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_805de4b7130efd4fbe2290d0c51d78e6295b83d77efc3ff2d1b014ffe7z8r74_894b5d6a-cddd-43e6-a59e-6076f985e5a3/extract/0.log" Feb 02 10:27:19 crc kubenswrapper[4764]: I0202 10:27:19.680810 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7b6c4d8c5f-wtkdr_9cbd726b-9dbb-46b3-9bcd-f3e9813b4f07/manager/0.log" Feb 02 10:27:19 crc kubenswrapper[4764]: I0202 10:27:19.772642 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-8d874c8fc-czpmn_7c8a2462-306c-43da-a538-9e34a11b5c21/manager/0.log" Feb 02 10:27:19 crc kubenswrapper[4764]: I0202 10:27:19.951050 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-6d9697b7f4-w4qf5_8b0e7cb4-257b-403b-8b92-4578ce77c388/manager/0.log" Feb 02 10:27:20 crc kubenswrapper[4764]: I0202 10:27:20.048722 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-8886f4c47-hvtx6_ee5cb41c-4d3d-4a9d-b1f7-1ede61e7b21c/manager/0.log" Feb 02 10:27:20 crc kubenswrapper[4764]: I0202 10:27:20.174916 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-69d6db494d-5f766_becd2ba3-ccad-4ea9-b3c7-6b39f6f8cb3d/manager/0.log" Feb 02 10:27:20 crc kubenswrapper[4764]: I0202 10:27:20.255604 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-5fb775575f-ntj47_82e16b73-9dfe-413b-b3dd-65da79b3d9dd/manager/0.log" Feb 02 10:27:20 crc kubenswrapper[4764]: I0202 10:27:20.580456 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-5f4b8bd54d-h4n52_ad8e7100-d5b8-4a56-bbda-056483f79113/manager/0.log" Feb 02 10:27:20 crc kubenswrapper[4764]: I0202 10:27:20.626646 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-79955696d6-mxvzm_b593c3f3-a225-4cb6-b967-4c056e3326a0/manager/0.log" Feb 02 10:27:20 crc kubenswrapper[4764]: I0202 10:27:20.774709 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-84f48565d4-qqnp6_8564e79b-9abd-41e1-a0a7-0d4b71011300/manager/0.log" Feb 02 10:27:20 crc kubenswrapper[4764]: I0202 10:27:20.873750 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7dd968899f-g2p6c_6d4fffad-a306-4f0b-ac40-7fa839f75aa3/manager/0.log" Feb 02 10:27:21 crc kubenswrapper[4764]: I0202 10:27:21.006198 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-67bf948998-jhbzc_9e55cb1e-ed3b-4b7b-8b52-42436eb48223/manager/0.log" Feb 02 10:27:21 crc kubenswrapper[4764]: I0202 10:27:21.108692 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-585dbc889-7nrsg_f5cc9236-c70f-4449-a902-ef416aaf8781/manager/0.log" Feb 02 10:27:21 crc kubenswrapper[4764]: I0202 10:27:21.337925 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-55bff696bd-p7nmb_e98e75d4-963f-4a67-894e-302b7429eb01/manager/0.log" Feb 02 10:27:21 crc kubenswrapper[4764]: I0202 10:27:21.401418 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-6687f8d877-5ncb8_1844ef02-28fb-45ce-a697-54f1560c9087/manager/0.log" Feb 02 10:27:21 crc kubenswrapper[4764]: I0202 10:27:21.567476 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-5ff45dfdbfb5fxl_1b05c36c-dbb6-4ad4-8985-8d64551812ff/manager/0.log" Feb 02 10:27:21 crc kubenswrapper[4764]: I0202 10:27:21.704082 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-init-6bf6665fd-l67kz_ee6621b4-2d57-402e-bfda-df7f7f922dc8/operator/0.log" Feb 02 10:27:22 crc kubenswrapper[4764]: I0202 10:27:22.087512 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-8ctp4_50551dd4-71f5-4eb4-8736-ed04f0845c7e/registry-server/0.log" Feb 02 10:27:22 crc kubenswrapper[4764]: I0202 10:27:22.166547 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-788c46999f-pk2hh_674c8820-f063-487b-9122-1ed46d1bc39f/manager/0.log" Feb 02 10:27:22 crc kubenswrapper[4764]: I0202 10:27:22.547279 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-5b964cf4cd-kskcp_1464055e-4745-45e3-be46-5b7700dce733/manager/0.log" Feb 02 10:27:22 crc kubenswrapper[4764]: I0202 10:27:22.874475 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-68fc8c869-fb6zn_1a100291-702c-4908-8a63-acb05c4ef5e9/manager/0.log" Feb 02 10:27:22 crc kubenswrapper[4764]: I0202 10:27:22.952489 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-qmjfr_36bffbf8-2209-4410-8005-d0164bb1eb34/operator/0.log" Feb 02 10:27:22 crc kubenswrapper[4764]: I0202 10:27:22.997506 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-646f757d77-g8bhc_a46546b9-e589-4eca-bca7-9150038f9ab8/manager/0.log" Feb 02 10:27:23 crc kubenswrapper[4764]: I0202 10:27:23.164365 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-64b5b76f97-98xsk_7c983dd6-f499-41d5-a049-765a15b4f912/manager/0.log" Feb 02 10:27:23 crc kubenswrapper[4764]: I0202 10:27:23.306081 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-56f8bfcd9f-bd9cj_b8748b41-5e08-47bf-a853-ad8c4f2b0908/manager/0.log" Feb 02 10:27:23 crc kubenswrapper[4764]: I0202 10:27:23.414017 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-564965969-6kcc6_65c22016-b2f2-4a07-87be-480d18870dea/manager/0.log" Feb 02 10:27:43 crc kubenswrapper[4764]: I0202 10:27:43.523070 4764 patch_prober.go:28] interesting pod/machine-config-daemon-4ndm4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 10:27:43 crc kubenswrapper[4764]: I0202 10:27:43.523654 4764 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 10:27:44 crc kubenswrapper[4764]: I0202 10:27:44.745668 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-qq96b_545e08e9-b2c4-493e-8fc7-8353fcab998d/control-plane-machine-set-operator/0.log" Feb 02 10:27:45 crc kubenswrapper[4764]: I0202 10:27:45.018896 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-98wxj_08bc3386-ee9a-4a3c-aa52-22d667971076/kube-rbac-proxy/0.log" Feb 02 10:27:45 crc kubenswrapper[4764]: I0202 10:27:45.053489 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-98wxj_08bc3386-ee9a-4a3c-aa52-22d667971076/machine-api-operator/0.log" Feb 02 10:27:59 crc kubenswrapper[4764]: I0202 10:27:59.842127 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-858654f9db-mkn7c_533ed531-c13b-4cbe-9355-e2f1096b905a/cert-manager-controller/0.log" Feb 02 10:27:59 crc kubenswrapper[4764]: I0202 10:27:59.971877 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-cf98fcc89-4wc9w_1d03ea6c-c159-4e5a-a38c-1bf190d3d454/cert-manager-cainjector/0.log" Feb 02 10:28:00 crc kubenswrapper[4764]: I0202 10:28:00.095395 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-687f57d79b-8gwjv_6cc98479-e4c6-45e2-912c-c3a479af0321/cert-manager-webhook/0.log" Feb 02 10:28:13 crc kubenswrapper[4764]: I0202 10:28:13.208100 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7754f76f8b-mxdpd_67f7c8f7-535f-4a07-a330-adf7c41da701/nmstate-console-plugin/0.log" Feb 02 10:28:13 crc kubenswrapper[4764]: I0202 10:28:13.522757 4764 patch_prober.go:28] interesting pod/machine-config-daemon-4ndm4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 10:28:13 crc kubenswrapper[4764]: I0202 10:28:13.523088 4764 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 10:28:13 crc kubenswrapper[4764]: I0202 10:28:13.523133 4764 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" Feb 02 10:28:13 crc kubenswrapper[4764]: I0202 10:28:13.523877 4764 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c3ca98e86124ebc7a6731929c4f79572e15b8712d419753d62acb46e29b6778c"} pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 02 10:28:13 crc kubenswrapper[4764]: I0202 10:28:13.523955 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" containerName="machine-config-daemon" containerID="cri-o://c3ca98e86124ebc7a6731929c4f79572e15b8712d419753d62acb46e29b6778c" gracePeriod=600 Feb 02 10:28:13 crc kubenswrapper[4764]: I0202 10:28:13.534428 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-gswb4_7e5e0b71-cd10-4ccf-8100-f15a063cceb7/nmstate-handler/0.log" Feb 02 10:28:13 crc kubenswrapper[4764]: I0202 10:28:13.629275 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-54757c584b-jz8q9_dd14d932-c6a2-4d7e-a1bb-bf7cf4c30f00/kube-rbac-proxy/0.log" Feb 02 10:28:14 crc kubenswrapper[4764]: I0202 10:28:14.314962 4764 generic.go:334] "Generic (PLEG): container finished" podID="d192f670-9f9d-4539-9641-e4bed73acdd4" containerID="c3ca98e86124ebc7a6731929c4f79572e15b8712d419753d62acb46e29b6778c" exitCode=0 Feb 02 10:28:14 crc kubenswrapper[4764]: I0202 10:28:14.315015 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" event={"ID":"d192f670-9f9d-4539-9641-e4bed73acdd4","Type":"ContainerDied","Data":"c3ca98e86124ebc7a6731929c4f79572e15b8712d419753d62acb46e29b6778c"} Feb 02 10:28:14 crc kubenswrapper[4764]: I0202 10:28:14.315475 4764 scope.go:117] "RemoveContainer" containerID="e7f0b09df430c74ab38568e7c16c9b37764975d9232ef539984f9cdd78407ca1" Feb 02 10:28:14 crc kubenswrapper[4764]: I0202 10:28:14.340899 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-54757c584b-jz8q9_dd14d932-c6a2-4d7e-a1bb-bf7cf4c30f00/nmstate-metrics/0.log" Feb 02 10:28:14 crc kubenswrapper[4764]: I0202 10:28:14.514781 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-646758c888-mv59b_2c1baf80-192e-47b6-852c-394b7e4ade4e/nmstate-operator/0.log" Feb 02 10:28:14 crc kubenswrapper[4764]: I0202 10:28:14.628526 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-8474b5b9d8-gz969_1a77ee37-2514-4a73-b2f1-565442966b16/nmstate-webhook/0.log" Feb 02 10:28:15 crc kubenswrapper[4764]: I0202 10:28:15.325157 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" event={"ID":"d192f670-9f9d-4539-9641-e4bed73acdd4","Type":"ContainerStarted","Data":"6b12505dc80e6ea3d9957c2c91cce9992a749373b0e1a31c7cffa2d5d19b9b76"} Feb 02 10:28:47 crc kubenswrapper[4764]: I0202 10:28:47.543344 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6968d8fdc4-qxqpn_aa337cba-937a-44e6-af3b-50f8f3438489/kube-rbac-proxy/0.log" Feb 02 10:28:47 crc kubenswrapper[4764]: I0202 10:28:47.550489 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6968d8fdc4-qxqpn_aa337cba-937a-44e6-af3b-50f8f3438489/controller/0.log" Feb 02 10:28:47 crc kubenswrapper[4764]: I0202 10:28:47.745463 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-rtp9n_9d0a4296-e773-4500-8a43-83796d835ad9/cp-frr-files/0.log" Feb 02 10:28:47 crc kubenswrapper[4764]: I0202 10:28:47.982781 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-rtp9n_9d0a4296-e773-4500-8a43-83796d835ad9/cp-reloader/0.log" Feb 02 10:28:47 crc kubenswrapper[4764]: I0202 10:28:47.986077 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-rtp9n_9d0a4296-e773-4500-8a43-83796d835ad9/cp-reloader/0.log" Feb 02 10:28:48 crc kubenswrapper[4764]: I0202 10:28:48.002900 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-rtp9n_9d0a4296-e773-4500-8a43-83796d835ad9/cp-frr-files/0.log" Feb 02 10:28:48 crc kubenswrapper[4764]: I0202 10:28:48.007556 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-rtp9n_9d0a4296-e773-4500-8a43-83796d835ad9/cp-metrics/0.log" Feb 02 10:28:48 crc kubenswrapper[4764]: I0202 10:28:48.185948 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-rtp9n_9d0a4296-e773-4500-8a43-83796d835ad9/cp-frr-files/0.log" Feb 02 10:28:48 crc kubenswrapper[4764]: I0202 10:28:48.195997 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-rtp9n_9d0a4296-e773-4500-8a43-83796d835ad9/cp-reloader/0.log" Feb 02 10:28:48 crc kubenswrapper[4764]: I0202 10:28:48.224431 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-rtp9n_9d0a4296-e773-4500-8a43-83796d835ad9/cp-metrics/0.log" Feb 02 10:28:48 crc kubenswrapper[4764]: I0202 10:28:48.267407 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-rtp9n_9d0a4296-e773-4500-8a43-83796d835ad9/cp-metrics/0.log" Feb 02 10:28:48 crc kubenswrapper[4764]: I0202 10:28:48.966521 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-rtp9n_9d0a4296-e773-4500-8a43-83796d835ad9/controller/0.log" Feb 02 10:28:49 crc kubenswrapper[4764]: I0202 10:28:49.012130 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-rtp9n_9d0a4296-e773-4500-8a43-83796d835ad9/cp-frr-files/0.log" Feb 02 10:28:49 crc kubenswrapper[4764]: I0202 10:28:49.030620 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-rtp9n_9d0a4296-e773-4500-8a43-83796d835ad9/cp-reloader/0.log" Feb 02 10:28:49 crc kubenswrapper[4764]: I0202 10:28:49.066857 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-rtp9n_9d0a4296-e773-4500-8a43-83796d835ad9/cp-metrics/0.log" Feb 02 10:28:49 crc kubenswrapper[4764]: I0202 10:28:49.252502 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-rtp9n_9d0a4296-e773-4500-8a43-83796d835ad9/frr-metrics/0.log" Feb 02 10:28:49 crc kubenswrapper[4764]: I0202 10:28:49.317378 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-rtp9n_9d0a4296-e773-4500-8a43-83796d835ad9/kube-rbac-proxy/0.log" Feb 02 10:28:49 crc kubenswrapper[4764]: I0202 10:28:49.366460 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-rtp9n_9d0a4296-e773-4500-8a43-83796d835ad9/kube-rbac-proxy-frr/0.log" Feb 02 10:28:49 crc kubenswrapper[4764]: I0202 10:28:49.569428 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-rtp9n_9d0a4296-e773-4500-8a43-83796d835ad9/reloader/0.log" Feb 02 10:28:49 crc kubenswrapper[4764]: I0202 10:28:49.645092 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7df86c4f6c-z9ptz_4757c3fb-ae2f-498d-a551-f6c2daee3a23/frr-k8s-webhook-server/0.log" Feb 02 10:28:50 crc kubenswrapper[4764]: I0202 10:28:50.016347 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-7848c48b6-2kvbb_f479cadf-cdb2-423f-b9c5-204bef98bf8a/manager/0.log" Feb 02 10:28:50 crc kubenswrapper[4764]: I0202 10:28:50.147901 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-5645b7d98c-hhmxc_1436d9d2-a684-4204-91d6-9da1a7df5fa2/webhook-server/0.log" Feb 02 10:28:50 crc kubenswrapper[4764]: I0202 10:28:50.428065 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-jw8j4_c8444bdb-052f-4c56-841a-b3c6718759d8/kube-rbac-proxy/0.log" Feb 02 10:28:50 crc kubenswrapper[4764]: I0202 10:28:50.521632 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-rtp9n_9d0a4296-e773-4500-8a43-83796d835ad9/frr/0.log" Feb 02 10:28:50 crc kubenswrapper[4764]: I0202 10:28:50.754691 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-jw8j4_c8444bdb-052f-4c56-841a-b3c6718759d8/speaker/0.log" Feb 02 10:29:05 crc kubenswrapper[4764]: I0202 10:29:05.872829 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcv6r4w_6021fc8f-ba41-4247-9b13-991cee9cb2a6/util/0.log" Feb 02 10:29:06 crc kubenswrapper[4764]: I0202 10:29:06.110736 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcv6r4w_6021fc8f-ba41-4247-9b13-991cee9cb2a6/util/0.log" Feb 02 10:29:06 crc kubenswrapper[4764]: I0202 10:29:06.193816 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcv6r4w_6021fc8f-ba41-4247-9b13-991cee9cb2a6/pull/0.log" Feb 02 10:29:06 crc kubenswrapper[4764]: I0202 10:29:06.236210 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcv6r4w_6021fc8f-ba41-4247-9b13-991cee9cb2a6/pull/0.log" Feb 02 10:29:06 crc kubenswrapper[4764]: I0202 10:29:06.366424 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcv6r4w_6021fc8f-ba41-4247-9b13-991cee9cb2a6/pull/0.log" Feb 02 10:29:06 crc kubenswrapper[4764]: I0202 10:29:06.372447 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcv6r4w_6021fc8f-ba41-4247-9b13-991cee9cb2a6/util/0.log" Feb 02 10:29:06 crc kubenswrapper[4764]: I0202 10:29:06.527856 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcv6r4w_6021fc8f-ba41-4247-9b13-991cee9cb2a6/extract/0.log" Feb 02 10:29:06 crc kubenswrapper[4764]: I0202 10:29:06.645696 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec71324cpz_08e7a6ba-02aa-4f59-8e6d-b967cb0f9715/util/0.log" Feb 02 10:29:06 crc kubenswrapper[4764]: I0202 10:29:06.843793 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec71324cpz_08e7a6ba-02aa-4f59-8e6d-b967cb0f9715/pull/0.log" Feb 02 10:29:06 crc kubenswrapper[4764]: I0202 10:29:06.864277 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec71324cpz_08e7a6ba-02aa-4f59-8e6d-b967cb0f9715/pull/0.log" Feb 02 10:29:06 crc kubenswrapper[4764]: I0202 10:29:06.902148 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec71324cpz_08e7a6ba-02aa-4f59-8e6d-b967cb0f9715/util/0.log" Feb 02 10:29:07 crc kubenswrapper[4764]: I0202 10:29:07.071011 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec71324cpz_08e7a6ba-02aa-4f59-8e6d-b967cb0f9715/pull/0.log" Feb 02 10:29:07 crc kubenswrapper[4764]: I0202 10:29:07.126883 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec71324cpz_08e7a6ba-02aa-4f59-8e6d-b967cb0f9715/extract/0.log" Feb 02 10:29:07 crc kubenswrapper[4764]: I0202 10:29:07.148559 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec71324cpz_08e7a6ba-02aa-4f59-8e6d-b967cb0f9715/util/0.log" Feb 02 10:29:07 crc kubenswrapper[4764]: I0202 10:29:07.735784 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5kwbp_4fe67111-6fe1-4886-b192-b75fea98586d/extract-utilities/0.log" Feb 02 10:29:07 crc kubenswrapper[4764]: I0202 10:29:07.957695 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5kwbp_4fe67111-6fe1-4886-b192-b75fea98586d/extract-content/0.log" Feb 02 10:29:07 crc kubenswrapper[4764]: I0202 10:29:07.957720 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5kwbp_4fe67111-6fe1-4886-b192-b75fea98586d/extract-utilities/0.log" Feb 02 10:29:07 crc kubenswrapper[4764]: I0202 10:29:07.983077 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5kwbp_4fe67111-6fe1-4886-b192-b75fea98586d/extract-content/0.log" Feb 02 10:29:08 crc kubenswrapper[4764]: I0202 10:29:08.224090 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5kwbp_4fe67111-6fe1-4886-b192-b75fea98586d/extract-utilities/0.log" Feb 02 10:29:08 crc kubenswrapper[4764]: I0202 10:29:08.244058 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5kwbp_4fe67111-6fe1-4886-b192-b75fea98586d/extract-content/0.log" Feb 02 10:29:08 crc kubenswrapper[4764]: I0202 10:29:08.485923 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-6kgcj_730067d3-5e02-4aa1-af49-9bb70546886f/extract-utilities/0.log" Feb 02 10:29:08 crc kubenswrapper[4764]: I0202 10:29:08.832535 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-6kgcj_730067d3-5e02-4aa1-af49-9bb70546886f/extract-utilities/0.log" Feb 02 10:29:08 crc kubenswrapper[4764]: I0202 10:29:08.835067 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-6kgcj_730067d3-5e02-4aa1-af49-9bb70546886f/extract-content/0.log" Feb 02 10:29:08 crc kubenswrapper[4764]: I0202 10:29:08.881687 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-6kgcj_730067d3-5e02-4aa1-af49-9bb70546886f/extract-content/0.log" Feb 02 10:29:08 crc kubenswrapper[4764]: I0202 10:29:08.942081 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5kwbp_4fe67111-6fe1-4886-b192-b75fea98586d/registry-server/0.log" Feb 02 10:29:09 crc kubenswrapper[4764]: I0202 10:29:09.063690 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-6kgcj_730067d3-5e02-4aa1-af49-9bb70546886f/extract-content/0.log" Feb 02 10:29:09 crc kubenswrapper[4764]: I0202 10:29:09.082820 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-6kgcj_730067d3-5e02-4aa1-af49-9bb70546886f/extract-utilities/0.log" Feb 02 10:29:09 crc kubenswrapper[4764]: I0202 10:29:09.364465 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-f4d5w_a03d1d75-d8a0-407c-ab02-d151d5ba00b6/marketplace-operator/0.log" Feb 02 10:29:09 crc kubenswrapper[4764]: I0202 10:29:09.473057 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-vp4vx_7e8b7ce6-dc1a-474e-83e0-659a275c7aaa/extract-utilities/0.log" Feb 02 10:29:09 crc kubenswrapper[4764]: I0202 10:29:09.885468 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-vp4vx_7e8b7ce6-dc1a-474e-83e0-659a275c7aaa/extract-content/0.log" Feb 02 10:29:09 crc kubenswrapper[4764]: I0202 10:29:09.957082 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-vp4vx_7e8b7ce6-dc1a-474e-83e0-659a275c7aaa/extract-utilities/0.log" Feb 02 10:29:09 crc kubenswrapper[4764]: I0202 10:29:09.957347 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-vp4vx_7e8b7ce6-dc1a-474e-83e0-659a275c7aaa/extract-content/0.log" Feb 02 10:29:09 crc kubenswrapper[4764]: I0202 10:29:09.996322 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-6kgcj_730067d3-5e02-4aa1-af49-9bb70546886f/registry-server/0.log" Feb 02 10:29:10 crc kubenswrapper[4764]: I0202 10:29:10.145808 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-vp4vx_7e8b7ce6-dc1a-474e-83e0-659a275c7aaa/extract-utilities/0.log" Feb 02 10:29:10 crc kubenswrapper[4764]: I0202 10:29:10.327148 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-vp4vx_7e8b7ce6-dc1a-474e-83e0-659a275c7aaa/extract-content/0.log" Feb 02 10:29:10 crc kubenswrapper[4764]: I0202 10:29:10.356123 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-vp4vx_7e8b7ce6-dc1a-474e-83e0-659a275c7aaa/registry-server/0.log" Feb 02 10:29:10 crc kubenswrapper[4764]: I0202 10:29:10.411539 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-rc4qs_9443ada1-5f42-40df-b8e2-2c072dd7f264/extract-utilities/0.log" Feb 02 10:29:10 crc kubenswrapper[4764]: I0202 10:29:10.753495 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-rc4qs_9443ada1-5f42-40df-b8e2-2c072dd7f264/extract-content/0.log" Feb 02 10:29:10 crc kubenswrapper[4764]: I0202 10:29:10.766746 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-rc4qs_9443ada1-5f42-40df-b8e2-2c072dd7f264/extract-utilities/0.log" Feb 02 10:29:10 crc kubenswrapper[4764]: I0202 10:29:10.770876 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-rc4qs_9443ada1-5f42-40df-b8e2-2c072dd7f264/extract-content/0.log" Feb 02 10:29:10 crc kubenswrapper[4764]: I0202 10:29:10.995371 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-rc4qs_9443ada1-5f42-40df-b8e2-2c072dd7f264/extract-content/0.log" Feb 02 10:29:11 crc kubenswrapper[4764]: I0202 10:29:11.052438 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-rc4qs_9443ada1-5f42-40df-b8e2-2c072dd7f264/extract-utilities/0.log" Feb 02 10:29:11 crc kubenswrapper[4764]: I0202 10:29:11.646853 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-rc4qs_9443ada1-5f42-40df-b8e2-2c072dd7f264/registry-server/0.log" Feb 02 10:29:30 crc kubenswrapper[4764]: E0202 10:29:30.323755 4764 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.217:51666->38.102.83.217:34003: write tcp 38.102.83.217:51666->38.102.83.217:34003: write: broken pipe Feb 02 10:30:00 crc kubenswrapper[4764]: I0202 10:30:00.157872 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29500470-6kczt"] Feb 02 10:30:00 crc kubenswrapper[4764]: E0202 10:30:00.158966 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="858615be-9087-4141-b5be-d6751cc64025" containerName="registry-server" Feb 02 10:30:00 crc kubenswrapper[4764]: I0202 10:30:00.158980 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="858615be-9087-4141-b5be-d6751cc64025" containerName="registry-server" Feb 02 10:30:00 crc kubenswrapper[4764]: E0202 10:30:00.158998 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="858615be-9087-4141-b5be-d6751cc64025" containerName="extract-content" Feb 02 10:30:00 crc kubenswrapper[4764]: I0202 10:30:00.159004 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="858615be-9087-4141-b5be-d6751cc64025" containerName="extract-content" Feb 02 10:30:00 crc kubenswrapper[4764]: E0202 10:30:00.159028 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="858615be-9087-4141-b5be-d6751cc64025" containerName="extract-utilities" Feb 02 10:30:00 crc kubenswrapper[4764]: I0202 10:30:00.159034 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="858615be-9087-4141-b5be-d6751cc64025" containerName="extract-utilities" Feb 02 10:30:00 crc kubenswrapper[4764]: E0202 10:30:00.159051 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e318171-ef4e-4d2f-bcca-bfe22fa13637" containerName="container-00" Feb 02 10:30:00 crc kubenswrapper[4764]: I0202 10:30:00.159057 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e318171-ef4e-4d2f-bcca-bfe22fa13637" containerName="container-00" Feb 02 10:30:00 crc kubenswrapper[4764]: I0202 10:30:00.159566 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e318171-ef4e-4d2f-bcca-bfe22fa13637" containerName="container-00" Feb 02 10:30:00 crc kubenswrapper[4764]: I0202 10:30:00.159585 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="858615be-9087-4141-b5be-d6751cc64025" containerName="registry-server" Feb 02 10:30:00 crc kubenswrapper[4764]: I0202 10:30:00.160438 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29500470-6kczt" Feb 02 10:30:00 crc kubenswrapper[4764]: I0202 10:30:00.172754 4764 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 02 10:30:00 crc kubenswrapper[4764]: I0202 10:30:00.173549 4764 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 02 10:30:00 crc kubenswrapper[4764]: I0202 10:30:00.179130 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h849n\" (UniqueName: \"kubernetes.io/projected/d4936582-46dd-4a92-8755-4c47d77eba56-kube-api-access-h849n\") pod \"collect-profiles-29500470-6kczt\" (UID: \"d4936582-46dd-4a92-8755-4c47d77eba56\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500470-6kczt" Feb 02 10:30:00 crc kubenswrapper[4764]: I0202 10:30:00.179211 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d4936582-46dd-4a92-8755-4c47d77eba56-config-volume\") pod \"collect-profiles-29500470-6kczt\" (UID: \"d4936582-46dd-4a92-8755-4c47d77eba56\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500470-6kczt" Feb 02 10:30:00 crc kubenswrapper[4764]: I0202 10:30:00.179251 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d4936582-46dd-4a92-8755-4c47d77eba56-secret-volume\") pod \"collect-profiles-29500470-6kczt\" (UID: \"d4936582-46dd-4a92-8755-4c47d77eba56\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500470-6kczt" Feb 02 10:30:00 crc kubenswrapper[4764]: I0202 10:30:00.180404 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29500470-6kczt"] Feb 02 10:30:00 crc kubenswrapper[4764]: I0202 10:30:00.281404 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h849n\" (UniqueName: \"kubernetes.io/projected/d4936582-46dd-4a92-8755-4c47d77eba56-kube-api-access-h849n\") pod \"collect-profiles-29500470-6kczt\" (UID: \"d4936582-46dd-4a92-8755-4c47d77eba56\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500470-6kczt" Feb 02 10:30:00 crc kubenswrapper[4764]: I0202 10:30:00.281500 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d4936582-46dd-4a92-8755-4c47d77eba56-config-volume\") pod \"collect-profiles-29500470-6kczt\" (UID: \"d4936582-46dd-4a92-8755-4c47d77eba56\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500470-6kczt" Feb 02 10:30:00 crc kubenswrapper[4764]: I0202 10:30:00.281558 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d4936582-46dd-4a92-8755-4c47d77eba56-secret-volume\") pod \"collect-profiles-29500470-6kczt\" (UID: \"d4936582-46dd-4a92-8755-4c47d77eba56\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500470-6kczt" Feb 02 10:30:00 crc kubenswrapper[4764]: I0202 10:30:00.282573 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d4936582-46dd-4a92-8755-4c47d77eba56-config-volume\") pod \"collect-profiles-29500470-6kczt\" (UID: \"d4936582-46dd-4a92-8755-4c47d77eba56\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500470-6kczt" Feb 02 10:30:00 crc kubenswrapper[4764]: I0202 10:30:00.291684 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d4936582-46dd-4a92-8755-4c47d77eba56-secret-volume\") pod \"collect-profiles-29500470-6kczt\" (UID: \"d4936582-46dd-4a92-8755-4c47d77eba56\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500470-6kczt" Feb 02 10:30:00 crc kubenswrapper[4764]: I0202 10:30:00.308859 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h849n\" (UniqueName: \"kubernetes.io/projected/d4936582-46dd-4a92-8755-4c47d77eba56-kube-api-access-h849n\") pod \"collect-profiles-29500470-6kczt\" (UID: \"d4936582-46dd-4a92-8755-4c47d77eba56\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500470-6kczt" Feb 02 10:30:00 crc kubenswrapper[4764]: I0202 10:30:00.505081 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29500470-6kczt" Feb 02 10:30:01 crc kubenswrapper[4764]: I0202 10:30:01.197012 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29500470-6kczt"] Feb 02 10:30:01 crc kubenswrapper[4764]: I0202 10:30:01.231454 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29500470-6kczt" event={"ID":"d4936582-46dd-4a92-8755-4c47d77eba56","Type":"ContainerStarted","Data":"b88b8188d01195b707191bcb0bdf17b242e13d6b008d1b096b76bb1dd588179e"} Feb 02 10:30:02 crc kubenswrapper[4764]: I0202 10:30:02.242224 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29500470-6kczt" event={"ID":"d4936582-46dd-4a92-8755-4c47d77eba56","Type":"ContainerStarted","Data":"6cd0506df8b192dd122165c1f9a71d3975358e33a9d24fae4367a9cfdf42e95c"} Feb 02 10:30:03 crc kubenswrapper[4764]: I0202 10:30:03.253222 4764 generic.go:334] "Generic (PLEG): container finished" podID="d4936582-46dd-4a92-8755-4c47d77eba56" containerID="6cd0506df8b192dd122165c1f9a71d3975358e33a9d24fae4367a9cfdf42e95c" exitCode=0 Feb 02 10:30:03 crc kubenswrapper[4764]: I0202 10:30:03.253437 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29500470-6kczt" event={"ID":"d4936582-46dd-4a92-8755-4c47d77eba56","Type":"ContainerDied","Data":"6cd0506df8b192dd122165c1f9a71d3975358e33a9d24fae4367a9cfdf42e95c"} Feb 02 10:30:04 crc kubenswrapper[4764]: I0202 10:30:04.650145 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29500470-6kczt" Feb 02 10:30:04 crc kubenswrapper[4764]: I0202 10:30:04.679184 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d4936582-46dd-4a92-8755-4c47d77eba56-config-volume\") pod \"d4936582-46dd-4a92-8755-4c47d77eba56\" (UID: \"d4936582-46dd-4a92-8755-4c47d77eba56\") " Feb 02 10:30:04 crc kubenswrapper[4764]: I0202 10:30:04.679442 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d4936582-46dd-4a92-8755-4c47d77eba56-secret-volume\") pod \"d4936582-46dd-4a92-8755-4c47d77eba56\" (UID: \"d4936582-46dd-4a92-8755-4c47d77eba56\") " Feb 02 10:30:04 crc kubenswrapper[4764]: I0202 10:30:04.679585 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h849n\" (UniqueName: \"kubernetes.io/projected/d4936582-46dd-4a92-8755-4c47d77eba56-kube-api-access-h849n\") pod \"d4936582-46dd-4a92-8755-4c47d77eba56\" (UID: \"d4936582-46dd-4a92-8755-4c47d77eba56\") " Feb 02 10:30:04 crc kubenswrapper[4764]: I0202 10:30:04.679921 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d4936582-46dd-4a92-8755-4c47d77eba56-config-volume" (OuterVolumeSpecName: "config-volume") pod "d4936582-46dd-4a92-8755-4c47d77eba56" (UID: "d4936582-46dd-4a92-8755-4c47d77eba56"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 10:30:04 crc kubenswrapper[4764]: I0202 10:30:04.680651 4764 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d4936582-46dd-4a92-8755-4c47d77eba56-config-volume\") on node \"crc\" DevicePath \"\"" Feb 02 10:30:04 crc kubenswrapper[4764]: I0202 10:30:04.688117 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4936582-46dd-4a92-8755-4c47d77eba56-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "d4936582-46dd-4a92-8755-4c47d77eba56" (UID: "d4936582-46dd-4a92-8755-4c47d77eba56"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 10:30:04 crc kubenswrapper[4764]: I0202 10:30:04.688178 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d4936582-46dd-4a92-8755-4c47d77eba56-kube-api-access-h849n" (OuterVolumeSpecName: "kube-api-access-h849n") pod "d4936582-46dd-4a92-8755-4c47d77eba56" (UID: "d4936582-46dd-4a92-8755-4c47d77eba56"). InnerVolumeSpecName "kube-api-access-h849n". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 10:30:04 crc kubenswrapper[4764]: I0202 10:30:04.783381 4764 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d4936582-46dd-4a92-8755-4c47d77eba56-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 02 10:30:04 crc kubenswrapper[4764]: I0202 10:30:04.783767 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h849n\" (UniqueName: \"kubernetes.io/projected/d4936582-46dd-4a92-8755-4c47d77eba56-kube-api-access-h849n\") on node \"crc\" DevicePath \"\"" Feb 02 10:30:05 crc kubenswrapper[4764]: I0202 10:30:05.291598 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29500470-6kczt" event={"ID":"d4936582-46dd-4a92-8755-4c47d77eba56","Type":"ContainerDied","Data":"b88b8188d01195b707191bcb0bdf17b242e13d6b008d1b096b76bb1dd588179e"} Feb 02 10:30:05 crc kubenswrapper[4764]: I0202 10:30:05.291657 4764 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b88b8188d01195b707191bcb0bdf17b242e13d6b008d1b096b76bb1dd588179e" Feb 02 10:30:05 crc kubenswrapper[4764]: I0202 10:30:05.291712 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29500470-6kczt" Feb 02 10:30:05 crc kubenswrapper[4764]: I0202 10:30:05.372547 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29500425-ppfcg"] Feb 02 10:30:05 crc kubenswrapper[4764]: I0202 10:30:05.383276 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29500425-ppfcg"] Feb 02 10:30:05 crc kubenswrapper[4764]: I0202 10:30:05.838686 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="237b5e52-e331-43b5-9d2d-1ec2c8e90854" path="/var/lib/kubelet/pods/237b5e52-e331-43b5-9d2d-1ec2c8e90854/volumes" Feb 02 10:30:30 crc kubenswrapper[4764]: I0202 10:30:30.293036 4764 scope.go:117] "RemoveContainer" containerID="395d3dca5cb98df00c029c52d20ebbda1fe995a4adede18c47a7d36365157315" Feb 02 10:30:43 crc kubenswrapper[4764]: I0202 10:30:43.523173 4764 patch_prober.go:28] interesting pod/machine-config-daemon-4ndm4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 10:30:43 crc kubenswrapper[4764]: I0202 10:30:43.523726 4764 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 10:31:09 crc kubenswrapper[4764]: I0202 10:31:09.286484 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-sh46q"] Feb 02 10:31:09 crc kubenswrapper[4764]: E0202 10:31:09.287553 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4936582-46dd-4a92-8755-4c47d77eba56" containerName="collect-profiles" Feb 02 10:31:09 crc kubenswrapper[4764]: I0202 10:31:09.287572 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4936582-46dd-4a92-8755-4c47d77eba56" containerName="collect-profiles" Feb 02 10:31:09 crc kubenswrapper[4764]: I0202 10:31:09.287815 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4936582-46dd-4a92-8755-4c47d77eba56" containerName="collect-profiles" Feb 02 10:31:09 crc kubenswrapper[4764]: I0202 10:31:09.289488 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-sh46q" Feb 02 10:31:09 crc kubenswrapper[4764]: I0202 10:31:09.296819 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-sh46q"] Feb 02 10:31:09 crc kubenswrapper[4764]: I0202 10:31:09.344409 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/85d746b2-5086-425c-b023-3d07f64186d8-utilities\") pod \"certified-operators-sh46q\" (UID: \"85d746b2-5086-425c-b023-3d07f64186d8\") " pod="openshift-marketplace/certified-operators-sh46q" Feb 02 10:31:09 crc kubenswrapper[4764]: I0202 10:31:09.344585 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/85d746b2-5086-425c-b023-3d07f64186d8-catalog-content\") pod \"certified-operators-sh46q\" (UID: \"85d746b2-5086-425c-b023-3d07f64186d8\") " pod="openshift-marketplace/certified-operators-sh46q" Feb 02 10:31:09 crc kubenswrapper[4764]: I0202 10:31:09.344648 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nrkvv\" (UniqueName: \"kubernetes.io/projected/85d746b2-5086-425c-b023-3d07f64186d8-kube-api-access-nrkvv\") pod \"certified-operators-sh46q\" (UID: \"85d746b2-5086-425c-b023-3d07f64186d8\") " pod="openshift-marketplace/certified-operators-sh46q" Feb 02 10:31:09 crc kubenswrapper[4764]: I0202 10:31:09.446361 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/85d746b2-5086-425c-b023-3d07f64186d8-utilities\") pod \"certified-operators-sh46q\" (UID: \"85d746b2-5086-425c-b023-3d07f64186d8\") " pod="openshift-marketplace/certified-operators-sh46q" Feb 02 10:31:09 crc kubenswrapper[4764]: I0202 10:31:09.446457 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/85d746b2-5086-425c-b023-3d07f64186d8-catalog-content\") pod \"certified-operators-sh46q\" (UID: \"85d746b2-5086-425c-b023-3d07f64186d8\") " pod="openshift-marketplace/certified-operators-sh46q" Feb 02 10:31:09 crc kubenswrapper[4764]: I0202 10:31:09.446497 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nrkvv\" (UniqueName: \"kubernetes.io/projected/85d746b2-5086-425c-b023-3d07f64186d8-kube-api-access-nrkvv\") pod \"certified-operators-sh46q\" (UID: \"85d746b2-5086-425c-b023-3d07f64186d8\") " pod="openshift-marketplace/certified-operators-sh46q" Feb 02 10:31:09 crc kubenswrapper[4764]: I0202 10:31:09.447304 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/85d746b2-5086-425c-b023-3d07f64186d8-utilities\") pod \"certified-operators-sh46q\" (UID: \"85d746b2-5086-425c-b023-3d07f64186d8\") " pod="openshift-marketplace/certified-operators-sh46q" Feb 02 10:31:09 crc kubenswrapper[4764]: I0202 10:31:09.447422 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/85d746b2-5086-425c-b023-3d07f64186d8-catalog-content\") pod \"certified-operators-sh46q\" (UID: \"85d746b2-5086-425c-b023-3d07f64186d8\") " pod="openshift-marketplace/certified-operators-sh46q" Feb 02 10:31:09 crc kubenswrapper[4764]: I0202 10:31:09.466806 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nrkvv\" (UniqueName: \"kubernetes.io/projected/85d746b2-5086-425c-b023-3d07f64186d8-kube-api-access-nrkvv\") pod \"certified-operators-sh46q\" (UID: \"85d746b2-5086-425c-b023-3d07f64186d8\") " pod="openshift-marketplace/certified-operators-sh46q" Feb 02 10:31:09 crc kubenswrapper[4764]: I0202 10:31:09.609456 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-sh46q" Feb 02 10:31:10 crc kubenswrapper[4764]: I0202 10:31:10.320110 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-sh46q"] Feb 02 10:31:10 crc kubenswrapper[4764]: I0202 10:31:10.965715 4764 generic.go:334] "Generic (PLEG): container finished" podID="85d746b2-5086-425c-b023-3d07f64186d8" containerID="5bcda5313d020098d7f9cef3234621b6aecfbf26d9137b938c56abea4d4ffe69" exitCode=0 Feb 02 10:31:10 crc kubenswrapper[4764]: I0202 10:31:10.965776 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sh46q" event={"ID":"85d746b2-5086-425c-b023-3d07f64186d8","Type":"ContainerDied","Data":"5bcda5313d020098d7f9cef3234621b6aecfbf26d9137b938c56abea4d4ffe69"} Feb 02 10:31:10 crc kubenswrapper[4764]: I0202 10:31:10.966013 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sh46q" event={"ID":"85d746b2-5086-425c-b023-3d07f64186d8","Type":"ContainerStarted","Data":"64a41663ec3440386e883b5362470b6ba7705da51f68d4789ea869067b02be5e"} Feb 02 10:31:10 crc kubenswrapper[4764]: I0202 10:31:10.968884 4764 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 02 10:31:11 crc kubenswrapper[4764]: I0202 10:31:11.081441 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-dc7vl"] Feb 02 10:31:11 crc kubenswrapper[4764]: I0202 10:31:11.084111 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dc7vl" Feb 02 10:31:11 crc kubenswrapper[4764]: I0202 10:31:11.094924 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dc7vl"] Feb 02 10:31:11 crc kubenswrapper[4764]: I0202 10:31:11.115615 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jwqsd\" (UniqueName: \"kubernetes.io/projected/5799d845-2481-4bc7-8e59-24931d92978c-kube-api-access-jwqsd\") pod \"redhat-marketplace-dc7vl\" (UID: \"5799d845-2481-4bc7-8e59-24931d92978c\") " pod="openshift-marketplace/redhat-marketplace-dc7vl" Feb 02 10:31:11 crc kubenswrapper[4764]: I0202 10:31:11.115747 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5799d845-2481-4bc7-8e59-24931d92978c-utilities\") pod \"redhat-marketplace-dc7vl\" (UID: \"5799d845-2481-4bc7-8e59-24931d92978c\") " pod="openshift-marketplace/redhat-marketplace-dc7vl" Feb 02 10:31:11 crc kubenswrapper[4764]: I0202 10:31:11.115830 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5799d845-2481-4bc7-8e59-24931d92978c-catalog-content\") pod \"redhat-marketplace-dc7vl\" (UID: \"5799d845-2481-4bc7-8e59-24931d92978c\") " pod="openshift-marketplace/redhat-marketplace-dc7vl" Feb 02 10:31:11 crc kubenswrapper[4764]: I0202 10:31:11.217163 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jwqsd\" (UniqueName: \"kubernetes.io/projected/5799d845-2481-4bc7-8e59-24931d92978c-kube-api-access-jwqsd\") pod \"redhat-marketplace-dc7vl\" (UID: \"5799d845-2481-4bc7-8e59-24931d92978c\") " pod="openshift-marketplace/redhat-marketplace-dc7vl" Feb 02 10:31:11 crc kubenswrapper[4764]: I0202 10:31:11.217568 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5799d845-2481-4bc7-8e59-24931d92978c-utilities\") pod \"redhat-marketplace-dc7vl\" (UID: \"5799d845-2481-4bc7-8e59-24931d92978c\") " pod="openshift-marketplace/redhat-marketplace-dc7vl" Feb 02 10:31:11 crc kubenswrapper[4764]: I0202 10:31:11.217719 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5799d845-2481-4bc7-8e59-24931d92978c-catalog-content\") pod \"redhat-marketplace-dc7vl\" (UID: \"5799d845-2481-4bc7-8e59-24931d92978c\") " pod="openshift-marketplace/redhat-marketplace-dc7vl" Feb 02 10:31:11 crc kubenswrapper[4764]: I0202 10:31:11.217969 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5799d845-2481-4bc7-8e59-24931d92978c-utilities\") pod \"redhat-marketplace-dc7vl\" (UID: \"5799d845-2481-4bc7-8e59-24931d92978c\") " pod="openshift-marketplace/redhat-marketplace-dc7vl" Feb 02 10:31:11 crc kubenswrapper[4764]: I0202 10:31:11.218162 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5799d845-2481-4bc7-8e59-24931d92978c-catalog-content\") pod \"redhat-marketplace-dc7vl\" (UID: \"5799d845-2481-4bc7-8e59-24931d92978c\") " pod="openshift-marketplace/redhat-marketplace-dc7vl" Feb 02 10:31:11 crc kubenswrapper[4764]: I0202 10:31:11.248560 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jwqsd\" (UniqueName: \"kubernetes.io/projected/5799d845-2481-4bc7-8e59-24931d92978c-kube-api-access-jwqsd\") pod \"redhat-marketplace-dc7vl\" (UID: \"5799d845-2481-4bc7-8e59-24931d92978c\") " pod="openshift-marketplace/redhat-marketplace-dc7vl" Feb 02 10:31:11 crc kubenswrapper[4764]: I0202 10:31:11.422892 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dc7vl" Feb 02 10:31:11 crc kubenswrapper[4764]: I0202 10:31:11.925324 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dc7vl"] Feb 02 10:31:12 crc kubenswrapper[4764]: W0202 10:31:12.059132 4764 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5799d845_2481_4bc7_8e59_24931d92978c.slice/crio-3c59c5045b73c12c5477c0081c8a115ab59da31cd917667ea09e580d336f625c WatchSource:0}: Error finding container 3c59c5045b73c12c5477c0081c8a115ab59da31cd917667ea09e580d336f625c: Status 404 returned error can't find the container with id 3c59c5045b73c12c5477c0081c8a115ab59da31cd917667ea09e580d336f625c Feb 02 10:31:12 crc kubenswrapper[4764]: I0202 10:31:12.988808 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sh46q" event={"ID":"85d746b2-5086-425c-b023-3d07f64186d8","Type":"ContainerStarted","Data":"9bf22edcc24d1ab8ad2adf280c503cdf6133782de216ffc5401d002da0bb84e4"} Feb 02 10:31:12 crc kubenswrapper[4764]: I0202 10:31:12.993832 4764 generic.go:334] "Generic (PLEG): container finished" podID="5799d845-2481-4bc7-8e59-24931d92978c" containerID="3bf1b0d93f6fe89ad94a36f7b05b596bee2d968ed5c9ac4b1bf48311b07241c4" exitCode=0 Feb 02 10:31:12 crc kubenswrapper[4764]: I0202 10:31:12.993889 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dc7vl" event={"ID":"5799d845-2481-4bc7-8e59-24931d92978c","Type":"ContainerDied","Data":"3bf1b0d93f6fe89ad94a36f7b05b596bee2d968ed5c9ac4b1bf48311b07241c4"} Feb 02 10:31:12 crc kubenswrapper[4764]: I0202 10:31:12.993919 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dc7vl" event={"ID":"5799d845-2481-4bc7-8e59-24931d92978c","Type":"ContainerStarted","Data":"3c59c5045b73c12c5477c0081c8a115ab59da31cd917667ea09e580d336f625c"} Feb 02 10:31:13 crc kubenswrapper[4764]: I0202 10:31:13.523330 4764 patch_prober.go:28] interesting pod/machine-config-daemon-4ndm4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 10:31:13 crc kubenswrapper[4764]: I0202 10:31:13.523657 4764 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 10:31:15 crc kubenswrapper[4764]: I0202 10:31:15.019118 4764 generic.go:334] "Generic (PLEG): container finished" podID="85d746b2-5086-425c-b023-3d07f64186d8" containerID="9bf22edcc24d1ab8ad2adf280c503cdf6133782de216ffc5401d002da0bb84e4" exitCode=0 Feb 02 10:31:15 crc kubenswrapper[4764]: I0202 10:31:15.019190 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sh46q" event={"ID":"85d746b2-5086-425c-b023-3d07f64186d8","Type":"ContainerDied","Data":"9bf22edcc24d1ab8ad2adf280c503cdf6133782de216ffc5401d002da0bb84e4"} Feb 02 10:31:15 crc kubenswrapper[4764]: I0202 10:31:15.029654 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dc7vl" event={"ID":"5799d845-2481-4bc7-8e59-24931d92978c","Type":"ContainerStarted","Data":"fbad3ae2aae0875222af37af2e59df84d1c7d25fa05340746a2f8694ef297acb"} Feb 02 10:31:16 crc kubenswrapper[4764]: I0202 10:31:16.040794 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sh46q" event={"ID":"85d746b2-5086-425c-b023-3d07f64186d8","Type":"ContainerStarted","Data":"eb597ec7970f45ac165b3f2cb402f206d8eef212b907ee4c1049a0fee666bf28"} Feb 02 10:31:16 crc kubenswrapper[4764]: I0202 10:31:16.043900 4764 generic.go:334] "Generic (PLEG): container finished" podID="5799d845-2481-4bc7-8e59-24931d92978c" containerID="fbad3ae2aae0875222af37af2e59df84d1c7d25fa05340746a2f8694ef297acb" exitCode=0 Feb 02 10:31:16 crc kubenswrapper[4764]: I0202 10:31:16.043978 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dc7vl" event={"ID":"5799d845-2481-4bc7-8e59-24931d92978c","Type":"ContainerDied","Data":"fbad3ae2aae0875222af37af2e59df84d1c7d25fa05340746a2f8694ef297acb"} Feb 02 10:31:16 crc kubenswrapper[4764]: I0202 10:31:16.069455 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-sh46q" podStartSLOduration=2.626044352 podStartE2EDuration="7.069429472s" podCreationTimestamp="2026-02-02 10:31:09 +0000 UTC" firstStartedPulling="2026-02-02 10:31:10.968542804 +0000 UTC m=+5033.902266892" lastFinishedPulling="2026-02-02 10:31:15.411927924 +0000 UTC m=+5038.345652012" observedRunningTime="2026-02-02 10:31:16.067627054 +0000 UTC m=+5039.001351142" watchObservedRunningTime="2026-02-02 10:31:16.069429472 +0000 UTC m=+5039.003153560" Feb 02 10:31:18 crc kubenswrapper[4764]: I0202 10:31:18.063381 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dc7vl" event={"ID":"5799d845-2481-4bc7-8e59-24931d92978c","Type":"ContainerStarted","Data":"c6cc10088e34d752568c43649592707b00fb8782bb5632adc5f416b9d9e82ecf"} Feb 02 10:31:18 crc kubenswrapper[4764]: I0202 10:31:18.092473 4764 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-dc7vl" podStartSLOduration=2.621434747 podStartE2EDuration="7.092457425s" podCreationTimestamp="2026-02-02 10:31:11 +0000 UTC" firstStartedPulling="2026-02-02 10:31:12.995667414 +0000 UTC m=+5035.929391502" lastFinishedPulling="2026-02-02 10:31:17.466690092 +0000 UTC m=+5040.400414180" observedRunningTime="2026-02-02 10:31:18.086311141 +0000 UTC m=+5041.020035229" watchObservedRunningTime="2026-02-02 10:31:18.092457425 +0000 UTC m=+5041.026181513" Feb 02 10:31:19 crc kubenswrapper[4764]: I0202 10:31:19.609975 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-sh46q" Feb 02 10:31:19 crc kubenswrapper[4764]: I0202 10:31:19.610277 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-sh46q" Feb 02 10:31:19 crc kubenswrapper[4764]: I0202 10:31:19.667664 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-sh46q" Feb 02 10:31:20 crc kubenswrapper[4764]: I0202 10:31:20.127623 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-sh46q" Feb 02 10:31:21 crc kubenswrapper[4764]: I0202 10:31:21.271702 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-sh46q"] Feb 02 10:31:21 crc kubenswrapper[4764]: I0202 10:31:21.423317 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-dc7vl" Feb 02 10:31:21 crc kubenswrapper[4764]: I0202 10:31:21.423384 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-dc7vl" Feb 02 10:31:21 crc kubenswrapper[4764]: I0202 10:31:21.466323 4764 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-dc7vl" Feb 02 10:31:22 crc kubenswrapper[4764]: I0202 10:31:22.105076 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-sh46q" podUID="85d746b2-5086-425c-b023-3d07f64186d8" containerName="registry-server" containerID="cri-o://eb597ec7970f45ac165b3f2cb402f206d8eef212b907ee4c1049a0fee666bf28" gracePeriod=2 Feb 02 10:31:22 crc kubenswrapper[4764]: I0202 10:31:22.188461 4764 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-dc7vl" Feb 02 10:31:22 crc kubenswrapper[4764]: I0202 10:31:22.556156 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-sh46q" Feb 02 10:31:22 crc kubenswrapper[4764]: I0202 10:31:22.682964 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/85d746b2-5086-425c-b023-3d07f64186d8-catalog-content\") pod \"85d746b2-5086-425c-b023-3d07f64186d8\" (UID: \"85d746b2-5086-425c-b023-3d07f64186d8\") " Feb 02 10:31:22 crc kubenswrapper[4764]: I0202 10:31:22.683182 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nrkvv\" (UniqueName: \"kubernetes.io/projected/85d746b2-5086-425c-b023-3d07f64186d8-kube-api-access-nrkvv\") pod \"85d746b2-5086-425c-b023-3d07f64186d8\" (UID: \"85d746b2-5086-425c-b023-3d07f64186d8\") " Feb 02 10:31:22 crc kubenswrapper[4764]: I0202 10:31:22.683330 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/85d746b2-5086-425c-b023-3d07f64186d8-utilities\") pod \"85d746b2-5086-425c-b023-3d07f64186d8\" (UID: \"85d746b2-5086-425c-b023-3d07f64186d8\") " Feb 02 10:31:22 crc kubenswrapper[4764]: I0202 10:31:22.685448 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/85d746b2-5086-425c-b023-3d07f64186d8-utilities" (OuterVolumeSpecName: "utilities") pod "85d746b2-5086-425c-b023-3d07f64186d8" (UID: "85d746b2-5086-425c-b023-3d07f64186d8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 10:31:22 crc kubenswrapper[4764]: I0202 10:31:22.705155 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/85d746b2-5086-425c-b023-3d07f64186d8-kube-api-access-nrkvv" (OuterVolumeSpecName: "kube-api-access-nrkvv") pod "85d746b2-5086-425c-b023-3d07f64186d8" (UID: "85d746b2-5086-425c-b023-3d07f64186d8"). InnerVolumeSpecName "kube-api-access-nrkvv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 10:31:22 crc kubenswrapper[4764]: I0202 10:31:22.740234 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/85d746b2-5086-425c-b023-3d07f64186d8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "85d746b2-5086-425c-b023-3d07f64186d8" (UID: "85d746b2-5086-425c-b023-3d07f64186d8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 10:31:22 crc kubenswrapper[4764]: I0202 10:31:22.787407 4764 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/85d746b2-5086-425c-b023-3d07f64186d8-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 10:31:22 crc kubenswrapper[4764]: I0202 10:31:22.787475 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nrkvv\" (UniqueName: \"kubernetes.io/projected/85d746b2-5086-425c-b023-3d07f64186d8-kube-api-access-nrkvv\") on node \"crc\" DevicePath \"\"" Feb 02 10:31:22 crc kubenswrapper[4764]: I0202 10:31:22.787497 4764 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/85d746b2-5086-425c-b023-3d07f64186d8-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 10:31:23 crc kubenswrapper[4764]: I0202 10:31:23.123155 4764 generic.go:334] "Generic (PLEG): container finished" podID="85d746b2-5086-425c-b023-3d07f64186d8" containerID="eb597ec7970f45ac165b3f2cb402f206d8eef212b907ee4c1049a0fee666bf28" exitCode=0 Feb 02 10:31:23 crc kubenswrapper[4764]: I0202 10:31:23.123897 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-sh46q" Feb 02 10:31:23 crc kubenswrapper[4764]: I0202 10:31:23.123957 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sh46q" event={"ID":"85d746b2-5086-425c-b023-3d07f64186d8","Type":"ContainerDied","Data":"eb597ec7970f45ac165b3f2cb402f206d8eef212b907ee4c1049a0fee666bf28"} Feb 02 10:31:23 crc kubenswrapper[4764]: I0202 10:31:23.124057 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sh46q" event={"ID":"85d746b2-5086-425c-b023-3d07f64186d8","Type":"ContainerDied","Data":"64a41663ec3440386e883b5362470b6ba7705da51f68d4789ea869067b02be5e"} Feb 02 10:31:23 crc kubenswrapper[4764]: I0202 10:31:23.124083 4764 scope.go:117] "RemoveContainer" containerID="eb597ec7970f45ac165b3f2cb402f206d8eef212b907ee4c1049a0fee666bf28" Feb 02 10:31:23 crc kubenswrapper[4764]: I0202 10:31:23.168066 4764 scope.go:117] "RemoveContainer" containerID="9bf22edcc24d1ab8ad2adf280c503cdf6133782de216ffc5401d002da0bb84e4" Feb 02 10:31:23 crc kubenswrapper[4764]: I0202 10:31:23.195437 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-sh46q"] Feb 02 10:31:23 crc kubenswrapper[4764]: I0202 10:31:23.203670 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-sh46q"] Feb 02 10:31:23 crc kubenswrapper[4764]: I0202 10:31:23.811114 4764 scope.go:117] "RemoveContainer" containerID="5bcda5313d020098d7f9cef3234621b6aecfbf26d9137b938c56abea4d4ffe69" Feb 02 10:31:23 crc kubenswrapper[4764]: I0202 10:31:23.845020 4764 scope.go:117] "RemoveContainer" containerID="eb597ec7970f45ac165b3f2cb402f206d8eef212b907ee4c1049a0fee666bf28" Feb 02 10:31:23 crc kubenswrapper[4764]: E0202 10:31:23.845485 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eb597ec7970f45ac165b3f2cb402f206d8eef212b907ee4c1049a0fee666bf28\": container with ID starting with eb597ec7970f45ac165b3f2cb402f206d8eef212b907ee4c1049a0fee666bf28 not found: ID does not exist" containerID="eb597ec7970f45ac165b3f2cb402f206d8eef212b907ee4c1049a0fee666bf28" Feb 02 10:31:23 crc kubenswrapper[4764]: I0202 10:31:23.845605 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb597ec7970f45ac165b3f2cb402f206d8eef212b907ee4c1049a0fee666bf28"} err="failed to get container status \"eb597ec7970f45ac165b3f2cb402f206d8eef212b907ee4c1049a0fee666bf28\": rpc error: code = NotFound desc = could not find container \"eb597ec7970f45ac165b3f2cb402f206d8eef212b907ee4c1049a0fee666bf28\": container with ID starting with eb597ec7970f45ac165b3f2cb402f206d8eef212b907ee4c1049a0fee666bf28 not found: ID does not exist" Feb 02 10:31:23 crc kubenswrapper[4764]: I0202 10:31:23.845708 4764 scope.go:117] "RemoveContainer" containerID="9bf22edcc24d1ab8ad2adf280c503cdf6133782de216ffc5401d002da0bb84e4" Feb 02 10:31:23 crc kubenswrapper[4764]: E0202 10:31:23.848580 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9bf22edcc24d1ab8ad2adf280c503cdf6133782de216ffc5401d002da0bb84e4\": container with ID starting with 9bf22edcc24d1ab8ad2adf280c503cdf6133782de216ffc5401d002da0bb84e4 not found: ID does not exist" containerID="9bf22edcc24d1ab8ad2adf280c503cdf6133782de216ffc5401d002da0bb84e4" Feb 02 10:31:23 crc kubenswrapper[4764]: I0202 10:31:23.848616 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9bf22edcc24d1ab8ad2adf280c503cdf6133782de216ffc5401d002da0bb84e4"} err="failed to get container status \"9bf22edcc24d1ab8ad2adf280c503cdf6133782de216ffc5401d002da0bb84e4\": rpc error: code = NotFound desc = could not find container \"9bf22edcc24d1ab8ad2adf280c503cdf6133782de216ffc5401d002da0bb84e4\": container with ID starting with 9bf22edcc24d1ab8ad2adf280c503cdf6133782de216ffc5401d002da0bb84e4 not found: ID does not exist" Feb 02 10:31:23 crc kubenswrapper[4764]: I0202 10:31:23.848640 4764 scope.go:117] "RemoveContainer" containerID="5bcda5313d020098d7f9cef3234621b6aecfbf26d9137b938c56abea4d4ffe69" Feb 02 10:31:23 crc kubenswrapper[4764]: E0202 10:31:23.850249 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5bcda5313d020098d7f9cef3234621b6aecfbf26d9137b938c56abea4d4ffe69\": container with ID starting with 5bcda5313d020098d7f9cef3234621b6aecfbf26d9137b938c56abea4d4ffe69 not found: ID does not exist" containerID="5bcda5313d020098d7f9cef3234621b6aecfbf26d9137b938c56abea4d4ffe69" Feb 02 10:31:23 crc kubenswrapper[4764]: I0202 10:31:23.850396 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5bcda5313d020098d7f9cef3234621b6aecfbf26d9137b938c56abea4d4ffe69"} err="failed to get container status \"5bcda5313d020098d7f9cef3234621b6aecfbf26d9137b938c56abea4d4ffe69\": rpc error: code = NotFound desc = could not find container \"5bcda5313d020098d7f9cef3234621b6aecfbf26d9137b938c56abea4d4ffe69\": container with ID starting with 5bcda5313d020098d7f9cef3234621b6aecfbf26d9137b938c56abea4d4ffe69 not found: ID does not exist" Feb 02 10:31:23 crc kubenswrapper[4764]: I0202 10:31:23.853874 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="85d746b2-5086-425c-b023-3d07f64186d8" path="/var/lib/kubelet/pods/85d746b2-5086-425c-b023-3d07f64186d8/volumes" Feb 02 10:31:23 crc kubenswrapper[4764]: I0202 10:31:23.867758 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-dc7vl"] Feb 02 10:31:24 crc kubenswrapper[4764]: I0202 10:31:24.139428 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-dc7vl" podUID="5799d845-2481-4bc7-8e59-24931d92978c" containerName="registry-server" containerID="cri-o://c6cc10088e34d752568c43649592707b00fb8782bb5632adc5f416b9d9e82ecf" gracePeriod=2 Feb 02 10:31:24 crc kubenswrapper[4764]: I0202 10:31:24.725716 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dc7vl" Feb 02 10:31:24 crc kubenswrapper[4764]: I0202 10:31:24.835004 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5799d845-2481-4bc7-8e59-24931d92978c-utilities\") pod \"5799d845-2481-4bc7-8e59-24931d92978c\" (UID: \"5799d845-2481-4bc7-8e59-24931d92978c\") " Feb 02 10:31:24 crc kubenswrapper[4764]: I0202 10:31:24.835093 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jwqsd\" (UniqueName: \"kubernetes.io/projected/5799d845-2481-4bc7-8e59-24931d92978c-kube-api-access-jwqsd\") pod \"5799d845-2481-4bc7-8e59-24931d92978c\" (UID: \"5799d845-2481-4bc7-8e59-24931d92978c\") " Feb 02 10:31:24 crc kubenswrapper[4764]: I0202 10:31:24.835288 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5799d845-2481-4bc7-8e59-24931d92978c-catalog-content\") pod \"5799d845-2481-4bc7-8e59-24931d92978c\" (UID: \"5799d845-2481-4bc7-8e59-24931d92978c\") " Feb 02 10:31:24 crc kubenswrapper[4764]: I0202 10:31:24.835737 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5799d845-2481-4bc7-8e59-24931d92978c-utilities" (OuterVolumeSpecName: "utilities") pod "5799d845-2481-4bc7-8e59-24931d92978c" (UID: "5799d845-2481-4bc7-8e59-24931d92978c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 10:31:24 crc kubenswrapper[4764]: I0202 10:31:24.847329 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5799d845-2481-4bc7-8e59-24931d92978c-kube-api-access-jwqsd" (OuterVolumeSpecName: "kube-api-access-jwqsd") pod "5799d845-2481-4bc7-8e59-24931d92978c" (UID: "5799d845-2481-4bc7-8e59-24931d92978c"). InnerVolumeSpecName "kube-api-access-jwqsd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 10:31:24 crc kubenswrapper[4764]: I0202 10:31:24.871426 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5799d845-2481-4bc7-8e59-24931d92978c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5799d845-2481-4bc7-8e59-24931d92978c" (UID: "5799d845-2481-4bc7-8e59-24931d92978c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 10:31:24 crc kubenswrapper[4764]: I0202 10:31:24.937574 4764 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5799d845-2481-4bc7-8e59-24931d92978c-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 10:31:24 crc kubenswrapper[4764]: I0202 10:31:24.939653 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jwqsd\" (UniqueName: \"kubernetes.io/projected/5799d845-2481-4bc7-8e59-24931d92978c-kube-api-access-jwqsd\") on node \"crc\" DevicePath \"\"" Feb 02 10:31:24 crc kubenswrapper[4764]: I0202 10:31:24.939740 4764 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5799d845-2481-4bc7-8e59-24931d92978c-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 10:31:25 crc kubenswrapper[4764]: I0202 10:31:25.151116 4764 generic.go:334] "Generic (PLEG): container finished" podID="5799d845-2481-4bc7-8e59-24931d92978c" containerID="c6cc10088e34d752568c43649592707b00fb8782bb5632adc5f416b9d9e82ecf" exitCode=0 Feb 02 10:31:25 crc kubenswrapper[4764]: I0202 10:31:25.151628 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dc7vl" event={"ID":"5799d845-2481-4bc7-8e59-24931d92978c","Type":"ContainerDied","Data":"c6cc10088e34d752568c43649592707b00fb8782bb5632adc5f416b9d9e82ecf"} Feb 02 10:31:25 crc kubenswrapper[4764]: I0202 10:31:25.151728 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dc7vl" event={"ID":"5799d845-2481-4bc7-8e59-24931d92978c","Type":"ContainerDied","Data":"3c59c5045b73c12c5477c0081c8a115ab59da31cd917667ea09e580d336f625c"} Feb 02 10:31:25 crc kubenswrapper[4764]: I0202 10:31:25.151747 4764 scope.go:117] "RemoveContainer" containerID="c6cc10088e34d752568c43649592707b00fb8782bb5632adc5f416b9d9e82ecf" Feb 02 10:31:25 crc kubenswrapper[4764]: I0202 10:31:25.151693 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dc7vl" Feb 02 10:31:25 crc kubenswrapper[4764]: I0202 10:31:25.192286 4764 scope.go:117] "RemoveContainer" containerID="fbad3ae2aae0875222af37af2e59df84d1c7d25fa05340746a2f8694ef297acb" Feb 02 10:31:25 crc kubenswrapper[4764]: I0202 10:31:25.198294 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-dc7vl"] Feb 02 10:31:25 crc kubenswrapper[4764]: I0202 10:31:25.207624 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-dc7vl"] Feb 02 10:31:25 crc kubenswrapper[4764]: I0202 10:31:25.669875 4764 scope.go:117] "RemoveContainer" containerID="3bf1b0d93f6fe89ad94a36f7b05b596bee2d968ed5c9ac4b1bf48311b07241c4" Feb 02 10:31:25 crc kubenswrapper[4764]: I0202 10:31:25.738090 4764 scope.go:117] "RemoveContainer" containerID="c6cc10088e34d752568c43649592707b00fb8782bb5632adc5f416b9d9e82ecf" Feb 02 10:31:25 crc kubenswrapper[4764]: E0202 10:31:25.738502 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c6cc10088e34d752568c43649592707b00fb8782bb5632adc5f416b9d9e82ecf\": container with ID starting with c6cc10088e34d752568c43649592707b00fb8782bb5632adc5f416b9d9e82ecf not found: ID does not exist" containerID="c6cc10088e34d752568c43649592707b00fb8782bb5632adc5f416b9d9e82ecf" Feb 02 10:31:25 crc kubenswrapper[4764]: I0202 10:31:25.738536 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c6cc10088e34d752568c43649592707b00fb8782bb5632adc5f416b9d9e82ecf"} err="failed to get container status \"c6cc10088e34d752568c43649592707b00fb8782bb5632adc5f416b9d9e82ecf\": rpc error: code = NotFound desc = could not find container \"c6cc10088e34d752568c43649592707b00fb8782bb5632adc5f416b9d9e82ecf\": container with ID starting with c6cc10088e34d752568c43649592707b00fb8782bb5632adc5f416b9d9e82ecf not found: ID does not exist" Feb 02 10:31:25 crc kubenswrapper[4764]: I0202 10:31:25.738564 4764 scope.go:117] "RemoveContainer" containerID="fbad3ae2aae0875222af37af2e59df84d1c7d25fa05340746a2f8694ef297acb" Feb 02 10:31:25 crc kubenswrapper[4764]: E0202 10:31:25.738847 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fbad3ae2aae0875222af37af2e59df84d1c7d25fa05340746a2f8694ef297acb\": container with ID starting with fbad3ae2aae0875222af37af2e59df84d1c7d25fa05340746a2f8694ef297acb not found: ID does not exist" containerID="fbad3ae2aae0875222af37af2e59df84d1c7d25fa05340746a2f8694ef297acb" Feb 02 10:31:25 crc kubenswrapper[4764]: I0202 10:31:25.738874 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fbad3ae2aae0875222af37af2e59df84d1c7d25fa05340746a2f8694ef297acb"} err="failed to get container status \"fbad3ae2aae0875222af37af2e59df84d1c7d25fa05340746a2f8694ef297acb\": rpc error: code = NotFound desc = could not find container \"fbad3ae2aae0875222af37af2e59df84d1c7d25fa05340746a2f8694ef297acb\": container with ID starting with fbad3ae2aae0875222af37af2e59df84d1c7d25fa05340746a2f8694ef297acb not found: ID does not exist" Feb 02 10:31:25 crc kubenswrapper[4764]: I0202 10:31:25.738894 4764 scope.go:117] "RemoveContainer" containerID="3bf1b0d93f6fe89ad94a36f7b05b596bee2d968ed5c9ac4b1bf48311b07241c4" Feb 02 10:31:25 crc kubenswrapper[4764]: E0202 10:31:25.739276 4764 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3bf1b0d93f6fe89ad94a36f7b05b596bee2d968ed5c9ac4b1bf48311b07241c4\": container with ID starting with 3bf1b0d93f6fe89ad94a36f7b05b596bee2d968ed5c9ac4b1bf48311b07241c4 not found: ID does not exist" containerID="3bf1b0d93f6fe89ad94a36f7b05b596bee2d968ed5c9ac4b1bf48311b07241c4" Feb 02 10:31:25 crc kubenswrapper[4764]: I0202 10:31:25.739302 4764 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3bf1b0d93f6fe89ad94a36f7b05b596bee2d968ed5c9ac4b1bf48311b07241c4"} err="failed to get container status \"3bf1b0d93f6fe89ad94a36f7b05b596bee2d968ed5c9ac4b1bf48311b07241c4\": rpc error: code = NotFound desc = could not find container \"3bf1b0d93f6fe89ad94a36f7b05b596bee2d968ed5c9ac4b1bf48311b07241c4\": container with ID starting with 3bf1b0d93f6fe89ad94a36f7b05b596bee2d968ed5c9ac4b1bf48311b07241c4 not found: ID does not exist" Feb 02 10:31:25 crc kubenswrapper[4764]: I0202 10:31:25.835724 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5799d845-2481-4bc7-8e59-24931d92978c" path="/var/lib/kubelet/pods/5799d845-2481-4bc7-8e59-24931d92978c/volumes" Feb 02 10:31:43 crc kubenswrapper[4764]: I0202 10:31:43.550966 4764 patch_prober.go:28] interesting pod/machine-config-daemon-4ndm4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 10:31:43 crc kubenswrapper[4764]: I0202 10:31:43.551391 4764 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 10:31:43 crc kubenswrapper[4764]: I0202 10:31:43.551427 4764 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" Feb 02 10:31:43 crc kubenswrapper[4764]: I0202 10:31:43.551840 4764 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6b12505dc80e6ea3d9957c2c91cce9992a749373b0e1a31c7cffa2d5d19b9b76"} pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 02 10:31:43 crc kubenswrapper[4764]: I0202 10:31:43.551882 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" containerName="machine-config-daemon" containerID="cri-o://6b12505dc80e6ea3d9957c2c91cce9992a749373b0e1a31c7cffa2d5d19b9b76" gracePeriod=600 Feb 02 10:31:44 crc kubenswrapper[4764]: E0202 10:31:44.403264 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 10:31:44 crc kubenswrapper[4764]: I0202 10:31:44.412213 4764 generic.go:334] "Generic (PLEG): container finished" podID="d192f670-9f9d-4539-9641-e4bed73acdd4" containerID="6b12505dc80e6ea3d9957c2c91cce9992a749373b0e1a31c7cffa2d5d19b9b76" exitCode=0 Feb 02 10:31:44 crc kubenswrapper[4764]: I0202 10:31:44.412269 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" event={"ID":"d192f670-9f9d-4539-9641-e4bed73acdd4","Type":"ContainerDied","Data":"6b12505dc80e6ea3d9957c2c91cce9992a749373b0e1a31c7cffa2d5d19b9b76"} Feb 02 10:31:44 crc kubenswrapper[4764]: I0202 10:31:44.412316 4764 scope.go:117] "RemoveContainer" containerID="c3ca98e86124ebc7a6731929c4f79572e15b8712d419753d62acb46e29b6778c" Feb 02 10:31:44 crc kubenswrapper[4764]: I0202 10:31:44.413285 4764 scope.go:117] "RemoveContainer" containerID="6b12505dc80e6ea3d9957c2c91cce9992a749373b0e1a31c7cffa2d5d19b9b76" Feb 02 10:31:44 crc kubenswrapper[4764]: E0202 10:31:44.413751 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 10:31:45 crc kubenswrapper[4764]: I0202 10:31:45.425521 4764 generic.go:334] "Generic (PLEG): container finished" podID="c9238785-9984-45c8-8eb8-f15bd97b7990" containerID="d57970c433089112a43a97b7b0a86aa1fbb1612253f41f2a8f6b692ed649b7a3" exitCode=0 Feb 02 10:31:45 crc kubenswrapper[4764]: I0202 10:31:45.425852 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-lgpbv/must-gather-b5bt7" event={"ID":"c9238785-9984-45c8-8eb8-f15bd97b7990","Type":"ContainerDied","Data":"d57970c433089112a43a97b7b0a86aa1fbb1612253f41f2a8f6b692ed649b7a3"} Feb 02 10:31:45 crc kubenswrapper[4764]: I0202 10:31:45.426888 4764 scope.go:117] "RemoveContainer" containerID="d57970c433089112a43a97b7b0a86aa1fbb1612253f41f2a8f6b692ed649b7a3" Feb 02 10:31:45 crc kubenswrapper[4764]: I0202 10:31:45.545880 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-lgpbv_must-gather-b5bt7_c9238785-9984-45c8-8eb8-f15bd97b7990/gather/0.log" Feb 02 10:31:58 crc kubenswrapper[4764]: I0202 10:31:58.826798 4764 scope.go:117] "RemoveContainer" containerID="6b12505dc80e6ea3d9957c2c91cce9992a749373b0e1a31c7cffa2d5d19b9b76" Feb 02 10:31:58 crc kubenswrapper[4764]: E0202 10:31:58.827924 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 10:32:00 crc kubenswrapper[4764]: I0202 10:32:00.255465 4764 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-lgpbv/must-gather-b5bt7"] Feb 02 10:32:00 crc kubenswrapper[4764]: I0202 10:32:00.259865 4764 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-lgpbv/must-gather-b5bt7" podUID="c9238785-9984-45c8-8eb8-f15bd97b7990" containerName="copy" containerID="cri-o://8b13540f8f2ed99b626fa43c5c43c456426eeabf0300ef98f9e7fe2aa7ddcdcc" gracePeriod=2 Feb 02 10:32:00 crc kubenswrapper[4764]: I0202 10:32:00.263830 4764 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-lgpbv/must-gather-b5bt7"] Feb 02 10:32:00 crc kubenswrapper[4764]: I0202 10:32:00.607824 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-lgpbv_must-gather-b5bt7_c9238785-9984-45c8-8eb8-f15bd97b7990/copy/0.log" Feb 02 10:32:00 crc kubenswrapper[4764]: I0202 10:32:00.608684 4764 generic.go:334] "Generic (PLEG): container finished" podID="c9238785-9984-45c8-8eb8-f15bd97b7990" containerID="8b13540f8f2ed99b626fa43c5c43c456426eeabf0300ef98f9e7fe2aa7ddcdcc" exitCode=143 Feb 02 10:32:00 crc kubenswrapper[4764]: I0202 10:32:00.985883 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-lgpbv_must-gather-b5bt7_c9238785-9984-45c8-8eb8-f15bd97b7990/copy/0.log" Feb 02 10:32:00 crc kubenswrapper[4764]: I0202 10:32:00.986315 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-lgpbv/must-gather-b5bt7" Feb 02 10:32:01 crc kubenswrapper[4764]: I0202 10:32:01.036386 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dq55k\" (UniqueName: \"kubernetes.io/projected/c9238785-9984-45c8-8eb8-f15bd97b7990-kube-api-access-dq55k\") pod \"c9238785-9984-45c8-8eb8-f15bd97b7990\" (UID: \"c9238785-9984-45c8-8eb8-f15bd97b7990\") " Feb 02 10:32:01 crc kubenswrapper[4764]: I0202 10:32:01.036472 4764 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/c9238785-9984-45c8-8eb8-f15bd97b7990-must-gather-output\") pod \"c9238785-9984-45c8-8eb8-f15bd97b7990\" (UID: \"c9238785-9984-45c8-8eb8-f15bd97b7990\") " Feb 02 10:32:01 crc kubenswrapper[4764]: I0202 10:32:01.042682 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c9238785-9984-45c8-8eb8-f15bd97b7990-kube-api-access-dq55k" (OuterVolumeSpecName: "kube-api-access-dq55k") pod "c9238785-9984-45c8-8eb8-f15bd97b7990" (UID: "c9238785-9984-45c8-8eb8-f15bd97b7990"). InnerVolumeSpecName "kube-api-access-dq55k". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 10:32:01 crc kubenswrapper[4764]: I0202 10:32:01.138729 4764 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dq55k\" (UniqueName: \"kubernetes.io/projected/c9238785-9984-45c8-8eb8-f15bd97b7990-kube-api-access-dq55k\") on node \"crc\" DevicePath \"\"" Feb 02 10:32:01 crc kubenswrapper[4764]: I0202 10:32:01.250738 4764 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c9238785-9984-45c8-8eb8-f15bd97b7990-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "c9238785-9984-45c8-8eb8-f15bd97b7990" (UID: "c9238785-9984-45c8-8eb8-f15bd97b7990"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 10:32:01 crc kubenswrapper[4764]: I0202 10:32:01.344994 4764 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/c9238785-9984-45c8-8eb8-f15bd97b7990-must-gather-output\") on node \"crc\" DevicePath \"\"" Feb 02 10:32:01 crc kubenswrapper[4764]: I0202 10:32:01.622738 4764 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-lgpbv_must-gather-b5bt7_c9238785-9984-45c8-8eb8-f15bd97b7990/copy/0.log" Feb 02 10:32:01 crc kubenswrapper[4764]: I0202 10:32:01.623356 4764 scope.go:117] "RemoveContainer" containerID="8b13540f8f2ed99b626fa43c5c43c456426eeabf0300ef98f9e7fe2aa7ddcdcc" Feb 02 10:32:01 crc kubenswrapper[4764]: I0202 10:32:01.623501 4764 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-lgpbv/must-gather-b5bt7" Feb 02 10:32:01 crc kubenswrapper[4764]: I0202 10:32:01.646063 4764 scope.go:117] "RemoveContainer" containerID="d57970c433089112a43a97b7b0a86aa1fbb1612253f41f2a8f6b692ed649b7a3" Feb 02 10:32:01 crc kubenswrapper[4764]: I0202 10:32:01.834921 4764 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c9238785-9984-45c8-8eb8-f15bd97b7990" path="/var/lib/kubelet/pods/c9238785-9984-45c8-8eb8-f15bd97b7990/volumes" Feb 02 10:32:10 crc kubenswrapper[4764]: I0202 10:32:10.825838 4764 scope.go:117] "RemoveContainer" containerID="6b12505dc80e6ea3d9957c2c91cce9992a749373b0e1a31c7cffa2d5d19b9b76" Feb 02 10:32:10 crc kubenswrapper[4764]: E0202 10:32:10.826643 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 10:32:23 crc kubenswrapper[4764]: I0202 10:32:23.826867 4764 scope.go:117] "RemoveContainer" containerID="6b12505dc80e6ea3d9957c2c91cce9992a749373b0e1a31c7cffa2d5d19b9b76" Feb 02 10:32:23 crc kubenswrapper[4764]: E0202 10:32:23.828581 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 10:32:38 crc kubenswrapper[4764]: I0202 10:32:38.825714 4764 scope.go:117] "RemoveContainer" containerID="6b12505dc80e6ea3d9957c2c91cce9992a749373b0e1a31c7cffa2d5d19b9b76" Feb 02 10:32:38 crc kubenswrapper[4764]: E0202 10:32:38.826411 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 10:32:49 crc kubenswrapper[4764]: I0202 10:32:49.825765 4764 scope.go:117] "RemoveContainer" containerID="6b12505dc80e6ea3d9957c2c91cce9992a749373b0e1a31c7cffa2d5d19b9b76" Feb 02 10:32:49 crc kubenswrapper[4764]: E0202 10:32:49.826359 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 10:33:00 crc kubenswrapper[4764]: I0202 10:33:00.826751 4764 scope.go:117] "RemoveContainer" containerID="6b12505dc80e6ea3d9957c2c91cce9992a749373b0e1a31c7cffa2d5d19b9b76" Feb 02 10:33:00 crc kubenswrapper[4764]: E0202 10:33:00.827611 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 10:33:15 crc kubenswrapper[4764]: I0202 10:33:15.826406 4764 scope.go:117] "RemoveContainer" containerID="6b12505dc80e6ea3d9957c2c91cce9992a749373b0e1a31c7cffa2d5d19b9b76" Feb 02 10:33:15 crc kubenswrapper[4764]: E0202 10:33:15.827756 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 10:33:26 crc kubenswrapper[4764]: I0202 10:33:26.826328 4764 scope.go:117] "RemoveContainer" containerID="6b12505dc80e6ea3d9957c2c91cce9992a749373b0e1a31c7cffa2d5d19b9b76" Feb 02 10:33:26 crc kubenswrapper[4764]: E0202 10:33:26.827113 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 10:33:37 crc kubenswrapper[4764]: I0202 10:33:37.831865 4764 scope.go:117] "RemoveContainer" containerID="6b12505dc80e6ea3d9957c2c91cce9992a749373b0e1a31c7cffa2d5d19b9b76" Feb 02 10:33:37 crc kubenswrapper[4764]: E0202 10:33:37.832734 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 10:33:48 crc kubenswrapper[4764]: I0202 10:33:48.828757 4764 scope.go:117] "RemoveContainer" containerID="6b12505dc80e6ea3d9957c2c91cce9992a749373b0e1a31c7cffa2d5d19b9b76" Feb 02 10:33:48 crc kubenswrapper[4764]: E0202 10:33:48.830289 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 10:34:00 crc kubenswrapper[4764]: I0202 10:34:00.826866 4764 scope.go:117] "RemoveContainer" containerID="6b12505dc80e6ea3d9957c2c91cce9992a749373b0e1a31c7cffa2d5d19b9b76" Feb 02 10:34:00 crc kubenswrapper[4764]: E0202 10:34:00.827884 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 10:34:13 crc kubenswrapper[4764]: I0202 10:34:13.825563 4764 scope.go:117] "RemoveContainer" containerID="6b12505dc80e6ea3d9957c2c91cce9992a749373b0e1a31c7cffa2d5d19b9b76" Feb 02 10:34:13 crc kubenswrapper[4764]: E0202 10:34:13.826392 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 10:34:28 crc kubenswrapper[4764]: I0202 10:34:28.826320 4764 scope.go:117] "RemoveContainer" containerID="6b12505dc80e6ea3d9957c2c91cce9992a749373b0e1a31c7cffa2d5d19b9b76" Feb 02 10:34:28 crc kubenswrapper[4764]: E0202 10:34:28.827801 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 10:34:41 crc kubenswrapper[4764]: I0202 10:34:41.826916 4764 scope.go:117] "RemoveContainer" containerID="6b12505dc80e6ea3d9957c2c91cce9992a749373b0e1a31c7cffa2d5d19b9b76" Feb 02 10:34:41 crc kubenswrapper[4764]: E0202 10:34:41.830113 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 10:34:55 crc kubenswrapper[4764]: I0202 10:34:55.826712 4764 scope.go:117] "RemoveContainer" containerID="6b12505dc80e6ea3d9957c2c91cce9992a749373b0e1a31c7cffa2d5d19b9b76" Feb 02 10:34:55 crc kubenswrapper[4764]: E0202 10:34:55.828285 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 10:35:06 crc kubenswrapper[4764]: I0202 10:35:06.825578 4764 scope.go:117] "RemoveContainer" containerID="6b12505dc80e6ea3d9957c2c91cce9992a749373b0e1a31c7cffa2d5d19b9b76" Feb 02 10:35:06 crc kubenswrapper[4764]: E0202 10:35:06.826360 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 10:35:17 crc kubenswrapper[4764]: I0202 10:35:17.136993 4764 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/openstack-operator-controller-manager-646f757d77-g8bhc" podUID="a46546b9-e589-4eca-bca7-9150038f9ab8" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.88:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 02 10:35:17 crc kubenswrapper[4764]: I0202 10:35:17.146317 4764 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/openstack-operator-controller-manager-646f757d77-g8bhc" podUID="a46546b9-e589-4eca-bca7-9150038f9ab8" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.88:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 02 10:35:21 crc kubenswrapper[4764]: I0202 10:35:21.826057 4764 scope.go:117] "RemoveContainer" containerID="6b12505dc80e6ea3d9957c2c91cce9992a749373b0e1a31c7cffa2d5d19b9b76" Feb 02 10:35:21 crc kubenswrapper[4764]: E0202 10:35:21.826797 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 10:35:34 crc kubenswrapper[4764]: I0202 10:35:34.824970 4764 scope.go:117] "RemoveContainer" containerID="6b12505dc80e6ea3d9957c2c91cce9992a749373b0e1a31c7cffa2d5d19b9b76" Feb 02 10:35:34 crc kubenswrapper[4764]: E0202 10:35:34.825684 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 10:35:48 crc kubenswrapper[4764]: I0202 10:35:48.826257 4764 scope.go:117] "RemoveContainer" containerID="6b12505dc80e6ea3d9957c2c91cce9992a749373b0e1a31c7cffa2d5d19b9b76" Feb 02 10:35:48 crc kubenswrapper[4764]: E0202 10:35:48.827334 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 10:35:53 crc kubenswrapper[4764]: I0202 10:35:53.505368 4764 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-ck2k9"] Feb 02 10:35:53 crc kubenswrapper[4764]: E0202 10:35:53.506442 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85d746b2-5086-425c-b023-3d07f64186d8" containerName="registry-server" Feb 02 10:35:53 crc kubenswrapper[4764]: I0202 10:35:53.506454 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="85d746b2-5086-425c-b023-3d07f64186d8" containerName="registry-server" Feb 02 10:35:53 crc kubenswrapper[4764]: E0202 10:35:53.506466 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9238785-9984-45c8-8eb8-f15bd97b7990" containerName="copy" Feb 02 10:35:53 crc kubenswrapper[4764]: I0202 10:35:53.506472 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9238785-9984-45c8-8eb8-f15bd97b7990" containerName="copy" Feb 02 10:35:53 crc kubenswrapper[4764]: E0202 10:35:53.506483 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5799d845-2481-4bc7-8e59-24931d92978c" containerName="registry-server" Feb 02 10:35:53 crc kubenswrapper[4764]: I0202 10:35:53.506490 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="5799d845-2481-4bc7-8e59-24931d92978c" containerName="registry-server" Feb 02 10:35:53 crc kubenswrapper[4764]: E0202 10:35:53.506507 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9238785-9984-45c8-8eb8-f15bd97b7990" containerName="gather" Feb 02 10:35:53 crc kubenswrapper[4764]: I0202 10:35:53.506512 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9238785-9984-45c8-8eb8-f15bd97b7990" containerName="gather" Feb 02 10:35:53 crc kubenswrapper[4764]: E0202 10:35:53.506528 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85d746b2-5086-425c-b023-3d07f64186d8" containerName="extract-content" Feb 02 10:35:53 crc kubenswrapper[4764]: I0202 10:35:53.506534 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="85d746b2-5086-425c-b023-3d07f64186d8" containerName="extract-content" Feb 02 10:35:53 crc kubenswrapper[4764]: E0202 10:35:53.506543 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85d746b2-5086-425c-b023-3d07f64186d8" containerName="extract-utilities" Feb 02 10:35:53 crc kubenswrapper[4764]: I0202 10:35:53.506551 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="85d746b2-5086-425c-b023-3d07f64186d8" containerName="extract-utilities" Feb 02 10:35:53 crc kubenswrapper[4764]: E0202 10:35:53.506567 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5799d845-2481-4bc7-8e59-24931d92978c" containerName="extract-content" Feb 02 10:35:53 crc kubenswrapper[4764]: I0202 10:35:53.506573 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="5799d845-2481-4bc7-8e59-24931d92978c" containerName="extract-content" Feb 02 10:35:53 crc kubenswrapper[4764]: E0202 10:35:53.506583 4764 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5799d845-2481-4bc7-8e59-24931d92978c" containerName="extract-utilities" Feb 02 10:35:53 crc kubenswrapper[4764]: I0202 10:35:53.506589 4764 state_mem.go:107] "Deleted CPUSet assignment" podUID="5799d845-2481-4bc7-8e59-24931d92978c" containerName="extract-utilities" Feb 02 10:35:53 crc kubenswrapper[4764]: I0202 10:35:53.506751 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="c9238785-9984-45c8-8eb8-f15bd97b7990" containerName="gather" Feb 02 10:35:53 crc kubenswrapper[4764]: I0202 10:35:53.506770 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="85d746b2-5086-425c-b023-3d07f64186d8" containerName="registry-server" Feb 02 10:35:53 crc kubenswrapper[4764]: I0202 10:35:53.506779 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="5799d845-2481-4bc7-8e59-24931d92978c" containerName="registry-server" Feb 02 10:35:53 crc kubenswrapper[4764]: I0202 10:35:53.506796 4764 memory_manager.go:354] "RemoveStaleState removing state" podUID="c9238785-9984-45c8-8eb8-f15bd97b7990" containerName="copy" Feb 02 10:35:53 crc kubenswrapper[4764]: I0202 10:35:53.509752 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ck2k9" Feb 02 10:35:53 crc kubenswrapper[4764]: I0202 10:35:53.538129 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-ck2k9"] Feb 02 10:35:53 crc kubenswrapper[4764]: I0202 10:35:53.549250 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6x4d8\" (UniqueName: \"kubernetes.io/projected/93ed86b9-8832-465e-acb4-27285492e32f-kube-api-access-6x4d8\") pod \"redhat-operators-ck2k9\" (UID: \"93ed86b9-8832-465e-acb4-27285492e32f\") " pod="openshift-marketplace/redhat-operators-ck2k9" Feb 02 10:35:53 crc kubenswrapper[4764]: I0202 10:35:53.549361 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/93ed86b9-8832-465e-acb4-27285492e32f-catalog-content\") pod \"redhat-operators-ck2k9\" (UID: \"93ed86b9-8832-465e-acb4-27285492e32f\") " pod="openshift-marketplace/redhat-operators-ck2k9" Feb 02 10:35:53 crc kubenswrapper[4764]: I0202 10:35:53.549483 4764 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/93ed86b9-8832-465e-acb4-27285492e32f-utilities\") pod \"redhat-operators-ck2k9\" (UID: \"93ed86b9-8832-465e-acb4-27285492e32f\") " pod="openshift-marketplace/redhat-operators-ck2k9" Feb 02 10:35:53 crc kubenswrapper[4764]: I0202 10:35:53.650584 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/93ed86b9-8832-465e-acb4-27285492e32f-catalog-content\") pod \"redhat-operators-ck2k9\" (UID: \"93ed86b9-8832-465e-acb4-27285492e32f\") " pod="openshift-marketplace/redhat-operators-ck2k9" Feb 02 10:35:53 crc kubenswrapper[4764]: I0202 10:35:53.650716 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/93ed86b9-8832-465e-acb4-27285492e32f-utilities\") pod \"redhat-operators-ck2k9\" (UID: \"93ed86b9-8832-465e-acb4-27285492e32f\") " pod="openshift-marketplace/redhat-operators-ck2k9" Feb 02 10:35:53 crc kubenswrapper[4764]: I0202 10:35:53.650764 4764 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6x4d8\" (UniqueName: \"kubernetes.io/projected/93ed86b9-8832-465e-acb4-27285492e32f-kube-api-access-6x4d8\") pod \"redhat-operators-ck2k9\" (UID: \"93ed86b9-8832-465e-acb4-27285492e32f\") " pod="openshift-marketplace/redhat-operators-ck2k9" Feb 02 10:35:53 crc kubenswrapper[4764]: I0202 10:35:53.651295 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/93ed86b9-8832-465e-acb4-27285492e32f-catalog-content\") pod \"redhat-operators-ck2k9\" (UID: \"93ed86b9-8832-465e-acb4-27285492e32f\") " pod="openshift-marketplace/redhat-operators-ck2k9" Feb 02 10:35:53 crc kubenswrapper[4764]: I0202 10:35:53.651339 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/93ed86b9-8832-465e-acb4-27285492e32f-utilities\") pod \"redhat-operators-ck2k9\" (UID: \"93ed86b9-8832-465e-acb4-27285492e32f\") " pod="openshift-marketplace/redhat-operators-ck2k9" Feb 02 10:35:53 crc kubenswrapper[4764]: I0202 10:35:53.672605 4764 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6x4d8\" (UniqueName: \"kubernetes.io/projected/93ed86b9-8832-465e-acb4-27285492e32f-kube-api-access-6x4d8\") pod \"redhat-operators-ck2k9\" (UID: \"93ed86b9-8832-465e-acb4-27285492e32f\") " pod="openshift-marketplace/redhat-operators-ck2k9" Feb 02 10:35:53 crc kubenswrapper[4764]: I0202 10:35:53.833337 4764 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ck2k9" Feb 02 10:35:54 crc kubenswrapper[4764]: I0202 10:35:54.410077 4764 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-ck2k9"] Feb 02 10:35:55 crc kubenswrapper[4764]: I0202 10:35:55.064001 4764 generic.go:334] "Generic (PLEG): container finished" podID="93ed86b9-8832-465e-acb4-27285492e32f" containerID="fedfaa70214bdf7e4c5eb6eefa7f540ebb201c04aa942f256a7adfa5b6b0105f" exitCode=0 Feb 02 10:35:55 crc kubenswrapper[4764]: I0202 10:35:55.064590 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ck2k9" event={"ID":"93ed86b9-8832-465e-acb4-27285492e32f","Type":"ContainerDied","Data":"fedfaa70214bdf7e4c5eb6eefa7f540ebb201c04aa942f256a7adfa5b6b0105f"} Feb 02 10:35:55 crc kubenswrapper[4764]: I0202 10:35:55.065775 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ck2k9" event={"ID":"93ed86b9-8832-465e-acb4-27285492e32f","Type":"ContainerStarted","Data":"248d410299702f80e9e2e607562891db8b68b8347b50b67543818ccf9b9c06c3"} Feb 02 10:35:57 crc kubenswrapper[4764]: I0202 10:35:57.085976 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ck2k9" event={"ID":"93ed86b9-8832-465e-acb4-27285492e32f","Type":"ContainerStarted","Data":"ace827a4a9af8246ef7c9e165fd4dfccc24a80a9b57fa2ba33294462dfed02b9"} Feb 02 10:36:00 crc kubenswrapper[4764]: I0202 10:36:00.825770 4764 scope.go:117] "RemoveContainer" containerID="6b12505dc80e6ea3d9957c2c91cce9992a749373b0e1a31c7cffa2d5d19b9b76" Feb 02 10:36:00 crc kubenswrapper[4764]: E0202 10:36:00.826524 4764 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4ndm4_openshift-machine-config-operator(d192f670-9f9d-4539-9641-e4bed73acdd4)\"" pod="openshift-machine-config-operator/machine-config-daemon-4ndm4" podUID="d192f670-9f9d-4539-9641-e4bed73acdd4" Feb 02 10:36:07 crc kubenswrapper[4764]: I0202 10:36:07.192554 4764 generic.go:334] "Generic (PLEG): container finished" podID="93ed86b9-8832-465e-acb4-27285492e32f" containerID="ace827a4a9af8246ef7c9e165fd4dfccc24a80a9b57fa2ba33294462dfed02b9" exitCode=0 Feb 02 10:36:07 crc kubenswrapper[4764]: I0202 10:36:07.193296 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ck2k9" event={"ID":"93ed86b9-8832-465e-acb4-27285492e32f","Type":"ContainerDied","Data":"ace827a4a9af8246ef7c9e165fd4dfccc24a80a9b57fa2ba33294462dfed02b9"} Feb 02 10:36:08 crc kubenswrapper[4764]: I0202 10:36:08.231728 4764 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ck2k9" event={"ID":"93ed86b9-8832-465e-acb4-27285492e32f","Type":"ContainerStarted","Data":"c47c1afa769cb616bded7e3f2bf8c0de28d446170e5d4564a47309688e992cc1"} var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515140077047024452 0ustar coreroot  Om77'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015140077047017367 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015140063762016511 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015140063762015461 5ustar corecore